Groq provides fast and low-cost AI inference through its LPU architecture and GroqCloud platform.
Categories
Groq offers 5 LLM API models.
Speed benchmark average: 324 tok/s.
api.groq.com