Groq

Ultra-fast LPU inference engine

https://groq.com📍 Mountain View, USAfreemiuminferencefastlpu

API サービス (1)

Chat CompletionsRESTbearer

Ultra-fast LLM inference — Llama, Mixtral, Gemma

https://api.groq.com/openai/v1

ドキュメント →
chatfastopen-source-models

AI モデル (2)

Llama 3.3 70B (on Groq)llama-3.3-70b-versatile
入力: $0.59/M出力: $0.79/MContext: 131K
ultra-fastversatilelpu
Llama 4 Scout (on Groq)meta-llama/llama-4-scout-17b-16e-instruct
入力: $0.11/M出力: $0.34/MContext: 131K
ultra-fastopen-sourcelpu