Groq API

LLM

Fastest LLM inference — Llama, Mixtral, Gemma at unprecedented speed.

Groq provides the fastest LLM inference API. Run open-source models like Llama 3, Mixtral, and Gemma with industry-leading speed on custom LPU hardware.

Details

Auth Type
API Key (Bearer token)
Rate Limit
30 requests/minute (free)
Pricing
Free tier available
Full Docs
Step 1: Save your provider key

This is NOT your Callio key. Enter the API key from the provider's dashboard (e.g. The API key from the provider's dashboard).

API Key (Bearer token)

Get your API key at console.groq.com/keys.

Get API Credentials

Getting Started

1

Try It Instantly

Click "Try It" above to test the API in the playground

2

Add to Your Agent

Click "Add to Agent" to get your API key and integrate

Common Use Cases

Fast inference
Chat completion
Code generation
Real-time AI
Batch processing

💻 Code Examples

Get started quickly with these code examples in your favorite language

curl -X GET \
'https://callio.app/api/proxy/groq/forward?target=https%3A%2F%2Fapi.groq.com%2Fopenai%2Fv1%2Fendpoint' \
-H 'Authorization: Bearer YOUR_CALLIO_KEY' \
-H 'Content-Type: application/json'

💡 Tip: Replace YOUR_CALLIO_KEY with your actual Callio API key from the dashboard.

Ready to integrate Groq API?

Test endpoints live or generate your API key and start building in minutes

Browse More APIs