
Introduction
- Integrate a large number of AI Models in your app through the groq api
- Includes a generous free plan with daily rate limit (without credit card)
- Includes 2b, 8b and 70b parameter models
Features of Groq Inference
- Lightning fast AI Models
- Supports 10+ Open Source AI Models- Gemma, LLama and Mixtral
- Best Free Plan that gives 500k tokens* daily for free to use in production
AI Models Supported (Google, Meta and Mixtral)
- gemma2-9b-it
- llama3-8b-8192
- llama3-70b-8192
- llama-guard-3-8b
- llama-3.1-8b-instant
- llama-3.1-70b-versatile
- llama-3.2-1b-preview*
- llama-3.2-3b-preview*
- llama-3.3-70b-versatile
- llama-3.3-70b-specdec
- mixtral-8x7b-32768
Groq AI Free Plan limits
The limit refreshes every 24 hrs
Model ID |
RPM |
RPD |
TPM |
TPD |
ASH |
ASD |
deepseek-r1-distill-llama-70b |
30 |
1,000 |
6,000 |
- |
- |
- |
deepseek-r1-distill-qwen-32b |
30 |
1,000 |
6,000 |
- |
- |
- |
gemma2-9b-it |
30 |
14,400 |
15,000 |
500,000 |
- |
- |
llama-3.1-8b-instant |
30 |
14,400 |
20,000 |
500,000 |
- |
- |
llama-3.1-70b-versatile |
30 |
14,400 |
6,000 |
200,000 |
- |
- |
llama-3.2-1b-preview |
30 |
7,000 |
7,000 |
500,000 |
- |
- |
llama-3.2-3b-preview |
30 |
7,000 |
7,000 |
500,000 |
- |
- |
llama-3.3-70b-specdec |
30 |
1,000 |
6,000 |
100,000 |
- |
- |
llama-3.3-70b-versatile |
30 |
1,000 |
6,000 |
100,000 |
- |
- |
llama-guard-3-8b |
30 |
14,400 |
15,000 |
500,000 |
- |
- |
llama3-8b-8192 |
30 |
14,400 |
30,000 |
500,000 |
- |
- |
llama3-70b-8192 |
30 |
14,400 |
6,000 |
500,000 |
- |
- |
mistral-saba-24b |
30 |
1,000 |
6,000 |
- |
- |
- |
mixtral-8x7b-32768 |
30 |
14,400 |
5,000 |
500,000 |
- |
- |
qwen-2.5-32b |
30 |
1,000 |
6,000 |
- |
- |
- |
qwen-2.5-coder-32b |
30 |
1,000 |
6,000 |
- |
- |
- |
qwen-qwq-32b |
30 |
1,000 |
6,000 |
- |
- |
- |
Note: Preview models are currently in testing. They shouldn't be used in production.
Blocks Documentation
Events:

GroqResponseReceived
Event triggered when AI response is received
Parameter |
Type |
statusCode |
number |
response |
text |

GroqRequestError
Event triggered when an error occurs in Groq API request
Parameter |
Type |
errorMessage |
text |

GroqMessageContentReceived
Event triggered when AI message content is extracted
Parameter |
Type |
messageContent |
text |
Methods:

SetAPIKey
Set the Groq API Key

SetModelName
Set the AI Model Name
Parameter |
Type |
model |
text |

SetAPIURL
Set the API Endpoint URL (Currently this block has no use but will be useful when Groq will introduce new API url)

ExtractJSONValue
Extract a specific value from the JSON response
use "choices[0].message.content" to get only the answer from the json
Parameter |
Type |
jsonString |
text |
fieldPath |
text |
Return Type: text

AskQuestion
Ask a question to the AI
Parameter |
Type |
userMessage |
text |
Generate API Key: GroqCloud
Sample Blocks
GroqTextMini (Free)
Features
- Use llama-8b-8192 model in your app
- Limit of maximum 500 tokens per response
- Check the difference between paid and free aix
GroqTextMini |
GroqText |
Free |
Paid(5.99$) |
Use llama-8b model |
Use 15+ AI Models (Lllama, Gemma, Mixtral, DeepSeek, Qwen, Distilled models) |
8b model |
1b, 2b, 3b, 8b, 32b, 70b, 80b models |
500 tokens |
Unlimited tokens depending on model capacity |
Download aix
com.sarthakdev.groqtextmini.aix (8.3 KB)
If you want to get GroqText extension you can get it from here for only 5.99$
Purchase Extension
You can purchase the extension instantly from the link below for just 12$= 6$ (launch price)
Subscribe to my free newsletter
https://newsletter.sarthakdev.in
4 Likes
Extract JSON value Function

Sample Json
{"id":"chatcmpl-7ceff50a-a5f1","object":"chat.completion","created":
1736604854,"model": "llama3-8b-8192","choices":
[{"index":0,"message":{"role":"assistant","content":"Hello! It's nice to meet you. Is there something I can help you with or would you like to chat?"},"logprobs":null,"finish_reason":"stop"}],"usage": {"queue_time":0.017792521000000002,"prompt_tokens":
11,"prompt_time":0.001729597,"completion_tokens": 25,"completion_time":0.020833333,"total_tokens":
36,"total_time":0.02256293},"system_fin
gerprint":"fp_a9","x_groq": {"id":"req_01jhavea"}}
- choices[0].message.content : Get content of message sent by assistant
- choices[0].finish_reason: Get the reason for the stop of the generation
- model : Get the model name
- usage.queue_time : Get the queue time
- usage.prompt_tokens : Get the prompt tokens
- usage.completion_rokens : Get the completion tokens
1 Like
Please contact me via dm for any questions or reply here. I will try to answer as soon as possible.
DeepSeek R1 is now supported
The extension now supports DeepSeek R1
Here's the ID: deepseek-r1-distill-llama-70b
1 Like
Qwen coder is now supported
The extension now supports Alibaba Qwen
Here's the ID: qwen-2.5-coder-32b
I don't think this blocks are correct. please make some corrections.
Move 'set label. text to ...' to the right place.
1 Like
Yes By mistake... will make the corrections
Introducing GroqTextMini (Free)
Features
- Use llama-8b-8192 model in your app
- Limit of maximum 500 tokens per response
- Check the difference between paid and free aix
GroqTextMini |
GroqText |
Free |
Paid(5.99$) |
Use llama-8b model |
Use 15+ AI Models (Lllama, Gemma, Mixtral, DeepSeek, Qwen, Distilled models) |
8b model |
1b, 2b, 3b, 8b, 32b, 70b, 80b models |
500 tokens |
Unlimited tokens depending on model capacity |
Download aix
com.sarthakdev.groqtextmini.aix (8.3 KB)
If you want to get GroqText extension you can get it from here for only 5.99$
1 Like
Qwen is now supported
The extension now supports Alibaba Qwen standard
Here's the ID: qwen-2.5-32b
Qwen is now supported
The extension now supports Alibaba Qwen QWQ
Here's the ID: qwen-qwq-32b
Mistral Saba is now supported
The extension now supports Mistral Saba
Here's the ID: mistral-saba-24b
Distilled models are now supported
The extension now supports distilled llama, qwen and deepseek models
Here's the ID: deepseek-r1-distill-qwen-32b, deepseek-r1-distill-llama-70b-specdec, deepseek-r1-distill-llama-70b
Updated Free Daily tokens and Rate Limits
Groq now supports 17 leading AI models from different providers all under a single API Key with daily free token usage
Model ID |
RPM |
RPD |
TPM |
TPD |
ASH |
ASD |
deepseek-r1-distill-llama-70b |
30 |
1,000 |
6,000 |
- |
- |
- |
deepseek-r1-distill-qwen-32b |
30 |
1,000 |
6,000 |
- |
- |
- |
gemma2-9b-it |
30 |
14,400 |
15,000 |
500,000 |
- |
- |
llama-3.1-8b-instant |
30 |
14,400 |
20,000 |
500,000 |
- |
- |
llama-3.1-70b-versatile |
30 |
14,400 |
6,000 |
200,000 |
- |
- |
llama-3.2-1b-preview |
30 |
7,000 |
7,000 |
500,000 |
- |
- |
llama-3.2-3b-preview |
30 |
7,000 |
7,000 |
500,000 |
- |
- |
llama-3.3-70b-specdec |
30 |
1,000 |
6,000 |
100,000 |
- |
- |
llama-3.3-70b-versatile |
30 |
1,000 |
6,000 |
100,000 |
- |
- |
llama-guard-3-8b |
30 |
14,400 |
15,000 |
500,000 |
- |
- |
llama3-8b-8192 |
30 |
14,400 |
30,000 |
500,000 |
- |
- |
llama3-70b-8192 |
30 |
14,400 |
6,000 |
500,000 |
- |
- |
mistral-saba-24b |
30 |
1,000 |
6,000 |
- |
- |
- |
mixtral-8x7b-32768 |
30 |
14,400 |
5,000 |
500,000 |
- |
- |
qwen-2.5-32b |
30 |
1,000 |
6,000 |
- |
- |
- |
qwen-2.5-coder-32b |
30 |
1,000 |
6,000 |
- |
- |
- |
qwen-qwq-32b |
30 |
1,000 |
6,000 |
- |
- |
- |
1 Like
Perfect..
What a work buddy!!
Just imagine!!
Nice work.
Congrats and all the best to you. Take my heart

1 Like
very good extension
you should use it for chatbots and if you have a budget then you can buy the full version, because its affordable
example app i made:
1 Like
after paid, when one model usage is reached, can switch to another model and keep using?
1 Like
$6 is one-off payment right?
1 Like
Yes updated versions will be free
Can your paid extension read image?