Skip to main content
Vincony
GR
Groq
Text

Llama 4 Maverick (Groq)

groq/llama-4-maverick

2 credits / request
Compare with…Added 2026

Meta's Llama 4 Maverick served on Groq hardware for blazing-fast inference. Maverick is Meta's mixture-of-experts model offering strong multimodal and multilingual capabilities at high speed.

Key Features

Mixture-of-experts architecture for efficiency

Fast inference via Groq LPU

Strong multilingual support

Multimodal input support

Ideal Use Cases

1.

Real-time multilingual chatbots

2.

Fast content generation at scale

3.

Interactive applications requiring sub-second responses

Technical Specifications

Context Window128K tokens
ModalityText, Image → Text
ProviderGroq
CategoryText Generation
ArchitectureMixture-of-Experts
LatencyUltra-low (Groq LPU)

API Usage

1curl -X POST https://api.vincony.com/v1/chat/completions \
2 -H "Authorization: Bearer YOUR_API_KEY" \
3 -H "Content-Type: application/json" \
4 -d '{
5 "model": "groq/llama-4-maverick",
6 "messages": [
7 { "role": "user", "content": "Hello, Llama 4 Maverick (Groq)!" }
8 ]
9 }'

Replace YOUR_API_KEY with your Vincony API key. OpenAI-compatible endpoint — works with any OpenAI SDK.

Compare with Another Model

Or compare up to 3 models

Frequently Asked Questions

Try Llama 4 Maverick (Groq) now

Start using Llama 4 Maverick (Groq) instantly — 100 free credits, no credit card required. Access 343+ AI models through one platform.

Vincony — Access the World's Best AI Models