[Freemium] GroqText: DeepSeek, Llama, Gemma, Mixtral and Qwen in your app with a single extension (with free tokens daily)

ezgif-6-95a00689cb

Introduction

  • Integrate a large number of AI Models in your app through the groq api
  • Includes a generous free plan with daily rate limit (without credit card)
  • Includes 2b, 8b and 70b parameter models

Features of Groq Inference

  • Lightning fast AI Models
  • Supports 10+ Open Source AI Models- Gemma, LLama and Mixtral
  • Best Free Plan that gives 500k tokens* daily for free to use in production

AI Models Supported (Google, Meta and Mixtral)

  • gemma2-9b-it
  • llama3-8b-8192
  • llama3-70b-8192
  • llama-guard-3-8b
  • llama-3.1-8b-instant
  • llama-3.1-70b-versatile
  • llama-3.2-1b-preview*
  • llama-3.2-3b-preview*
  • llama-3.3-70b-versatile
  • llama-3.3-70b-specdec
  • mixtral-8x7b-32768

Groq AI Free Plan limits

The limit refreshes every 24 hrs

Model ID RPM RPD TPM TPD ASH ASD
deepseek-r1-distill-llama-70b 30 1,000 6,000 - - -
deepseek-r1-distill-qwen-32b 30 1,000 6,000 - - -
gemma2-9b-it 30 14,400 15,000 500,000 - -
llama-3.1-8b-instant 30 14,400 20,000 500,000 - -
llama-3.1-70b-versatile 30 14,400 6,000 200,000 - -
llama-3.2-1b-preview 30 7,000 7,000 500,000 - -
llama-3.2-3b-preview 30 7,000 7,000 500,000 - -
llama-3.3-70b-specdec 30 1,000 6,000 100,000 - -
llama-3.3-70b-versatile 30 1,000 6,000 100,000 - -
llama-guard-3-8b 30 14,400 15,000 500,000 - -
llama3-8b-8192 30 14,400 30,000 500,000 - -
llama3-70b-8192 30 14,400 6,000 500,000 - -
mistral-saba-24b 30 1,000 6,000 - - -
mixtral-8x7b-32768 30 14,400 5,000 500,000 - -
qwen-2.5-32b 30 1,000 6,000 - - -
qwen-2.5-coder-32b 30 1,000 6,000 - - -
qwen-qwq-32b 30 1,000 6,000 - - -

Note: Preview models are currently in testing. They shouldn't be used in production.

Blocks Documentation

Events:

GroqResponseReceivedBlock

GroqResponseReceived

Event triggered when AI response is received

Parameter Type
statusCode number
response text

GroqRequestErrorBlock

GroqRequestError

Event triggered when an error occurs in Groq API request

Parameter Type
errorMessage text

GroqMessageContentReceivedBlock

GroqMessageContentReceived

Event triggered when AI message content is extracted

Parameter Type
messageContent text

Methods:

SetAPIKeyBlock

SetAPIKey

Set the Groq API Key

Parameter Type
key text

SetModelNameBlock

SetModelName

Set the AI Model Name

Parameter Type
model text

SetAPIURLBlock

SetAPIURL

Set the API Endpoint URL (Currently this block has no use but will be useful when Groq will introduce new API url)

Parameter Type
url text

ExtractJSONValueBlock

ExtractJSONValue

Extract a specific value from the JSON response
use "choices[0].message.content" to get only the answer from the json

Parameter Type
jsonString text
fieldPath text

Return Type: text

AskQuestionBlock

AskQuestion

Ask a question to the AI

Parameter Type
userMessage text

Generate API Key: GroqCloud

Sample Blocks

GroqTextMini (Free)

Features

  1. Use llama-8b-8192 model in your app
  2. Limit of maximum 500 tokens per response
  3. Check the difference between paid and free aix
GroqTextMini GroqText
Free Paid(5.99$)
Use llama-8b model Use 15+ AI Models (Lllama, Gemma, Mixtral, DeepSeek, Qwen, Distilled models)
8b model 1b, 2b, 3b, 8b, 32b, 70b, 80b models
500 tokens Unlimited tokens depending on model capacity

Download aix

:robot: com.sarthakdev.groqtextmini.aix (8.3 KB)

If you want to get GroqText extension you can get it from here for only 5.99$

Purchase Extension

You can purchase the extension instantly from the link below for just 12$= 6$ (launch price)

Subscribe to my free newsletter

https://newsletter.sarthakdev.in

4 Likes

Sample Blocks to message hello to llm

Extract JSON value Function

ExtractJSONValueBlock

Sample Json
{"id":"chatcmpl-7ceff50a-a5f1","object":"chat.completion","created":
1736604854,"model": "llama3-8b-8192","choices":
[{"index":0,"message":{"role":"assistant","content":"Hello! It's nice to meet you. Is there something I can help you with or would you like to chat?"},"logprobs":null,"finish_reason":"stop"}],"usage": {"queue_time":0.017792521000000002,"prompt_tokens":
11,"prompt_time":0.001729597,"completion_tokens": 25,"completion_time":0.020833333,"total_tokens":
36,"total_time":0.02256293},"system_fin
gerprint":"fp_a9","x_groq": {"id":"req_01jhavea"}}
  • choices[0].message.content : Get content of message sent by assistant
  • choices[0].finish_reason: Get the reason for the stop of the generation
  • model : Get the model name
  • usage.queue_time : Get the queue time
  • usage.prompt_tokens : Get the prompt tokens
  • usage.completion_rokens : Get the completion tokens
1 Like

Please contact me via dm for any questions or reply here. I will try to answer as soon as possible.

:loudspeaker: DeepSeek R1 is now supported

The extension now supports DeepSeek R1
Here's the ID: deepseek-r1-distill-llama-70b

1 Like

:loudspeaker: Qwen coder is now supported

The extension now supports Alibaba Qwen
Here's the ID: qwen-2.5-coder-32b

I don't think this blocks are correct. please make some corrections.

Move 'set label. text to ...' to the right place.

1 Like

yes, exactly


Taifun

1 Like

Yes By mistake... will make the corrections

Introducing GroqTextMini (Free)

Features

  1. Use llama-8b-8192 model in your app
  2. Limit of maximum 500 tokens per response
  3. Check the difference between paid and free aix
GroqTextMini GroqText
Free Paid(5.99$)
Use llama-8b model Use 15+ AI Models (Lllama, Gemma, Mixtral, DeepSeek, Qwen, Distilled models)
8b model 1b, 2b, 3b, 8b, 32b, 70b, 80b models
500 tokens Unlimited tokens depending on model capacity

Download aix

:robot: com.sarthakdev.groqtextmini.aix (8.3 KB)

If you want to get GroqText extension you can get it from here for only 5.99$

1 Like

:loudspeaker: Qwen is now supported

The extension now supports Alibaba Qwen standard
Here's the ID: qwen-2.5-32b

:loudspeaker: Qwen is now supported

The extension now supports Alibaba Qwen QWQ
Here's the ID: qwen-qwq-32b

:loudspeaker: Mistral Saba is now supported

The extension now supports Mistral Saba
Here's the ID: mistral-saba-24b

:loudspeaker: Distilled models are now supported

The extension now supports distilled llama, qwen and deepseek models
Here's the ID: deepseek-r1-distill-qwen-32b, deepseek-r1-distill-llama-70b-specdec, deepseek-r1-distill-llama-70b

:rocket: Updated Free Daily tokens and Rate Limits

Groq now supports 17 leading AI models from different providers all under a single API Key with daily free token usage

Model ID RPM RPD TPM TPD ASH ASD
deepseek-r1-distill-llama-70b 30 1,000 6,000 - - -
deepseek-r1-distill-qwen-32b 30 1,000 6,000 - - -
gemma2-9b-it 30 14,400 15,000 500,000 - -
llama-3.1-8b-instant 30 14,400 20,000 500,000 - -
llama-3.1-70b-versatile 30 14,400 6,000 200,000 - -
llama-3.2-1b-preview 30 7,000 7,000 500,000 - -
llama-3.2-3b-preview 30 7,000 7,000 500,000 - -
llama-3.3-70b-specdec 30 1,000 6,000 100,000 - -
llama-3.3-70b-versatile 30 1,000 6,000 100,000 - -
llama-guard-3-8b 30 14,400 15,000 500,000 - -
llama3-8b-8192 30 14,400 30,000 500,000 - -
llama3-70b-8192 30 14,400 6,000 500,000 - -
mistral-saba-24b 30 1,000 6,000 - - -
mixtral-8x7b-32768 30 14,400 5,000 500,000 - -
qwen-2.5-32b 30 1,000 6,000 - - -
qwen-2.5-coder-32b 30 1,000 6,000 - - -
qwen-qwq-32b 30 1,000 6,000 - - -
1 Like

Perfect..

What a work buddy!!

Just imagine!!

Nice work.

Congrats and all the best to you. Take my heart

:heart: :heart: :heart:

1 Like

very good extension
you should use it for chatbots and if you have a budget then you can buy the full version, because its affordable

example app i made:

1 Like

after paid, when one model usage is reached, can switch to another model and keep using?

1 Like

Yes its possible

$6 is one-off payment right?

1 Like

Yes updated versions will be free

Can your paid extension read image?