Aller au contenu principal
Groq logo

Groq

Fast, low cost inference

Groq delivers ultra-fast AI inference powered by custom-built LPU (Language Processing Unit) silicon. The platform provides fast, low-cost inference with deterministic execution, supporting LLMs, speech-to-text, text-to-speech, and vision models through an OpenAI-compatible API.

Fonctionnalités

Custom LPU architecture
OpenAI-compatible API
LLM inference (Llama, Mixtral, Gemma)
Speech-to-text (Whisper)
Text-to-speech support
Image-to-text models
Prompt caching
Batch API with discounts
Global data center deployments
SOC 2, GDPR, HIPAA compliance

Avantages

  • + Extremely fast inference speeds
  • + OpenAI-compatible (easy migration)
  • + Competitive pricing
  • + Free tier for getting started
  • + Enterprise compliance certifications

Inconvénients

  • Limited model selection compared to OpenAI
  • Newer platform with less track record
  • Hardware availability can be constrained
  • No fine-tuning support yet