
Accelerate your AI with Groq's LPU Inference Engine. Power real-time applications and unleash the potential of large language models (LLMs) for faster, more efficient natural language processing.
32.2K
Similarweb

Pricing Model
Pricing Plans
Groq offers a pay-as-you-go pricing model for its AI services, charging based on the number of tokens processed. Their pricing varies depending on the specific AI model used, ranging from $0.18 per million output tokens for Llama 3.2 1b Vision to $5.00 per million output tokens for Deepseek Ri Distill Qwen 32B. They also offer various other models for ASR and vision tasks with different pricing structures. Batch API processing is available at a discounted rate for developers. For enterprises or on-premise deployments, custom solutions are available upon request. Key Points: Pay-as-you-go: Groq charges based on the number of tokens processed. Token Pricing: Varies by model, ranging from $0.18 to $5.00 per million output tokens. Model Variety: Offers models for LLMs, ASR, and vision tasks. Batch API Discount: 25% discount for developers using batch processing. Enterprise Solutions: Custom pricing available for enterprises and on-premise deployments.
---
---
Discover alternative AI tools similar to Groq that may better suit your needs.
Explore professional roles that benefit from using Groq.