TO
Together
Text
Llama 4 Maverick (Together)
Meta's Llama 4 Maverick running on Together's GPU infrastructure, offering the mixture-of-experts model with competitive pricing and good throughput via OpenRouter.
Key Features
Mixture-of-experts architecture
Strong multilingual and multimodal capabilities
Competitive pricing via Together
Good throughput for batch processing
Ideal Use Cases
1.
Cost-effective text generation at scale
2.
Multilingual content creation
3.
Batch processing workloads
Technical Specifications
| Context Window | 128K tokens |
| Modality | Text, Image → Text |
| Provider | Together |
| Category | Text Generation |
| Architecture | Mixture-of-Experts |
API Usage
1 curl -X POST https://api.vincony.com/v1/chat/completions \ 2 -H "Authorization: Bearer YOUR_API_KEY" \ 3 -H "Content-Type: application/json" \ 4 -d '{ 5 "model": "together/llama-4-maverick", 6 "messages": [ 7 { "role": "user", "content": "Hello, Llama 4 Maverick (Together)!" } 8 ] 9 }'
Replace YOUR_API_KEY with your Vincony API key. OpenAI-compatible endpoint — works with any OpenAI SDK.
Compare with Another Model
Frequently Asked Questions
Try Llama 4 Maverick (Together) now
Start using Llama 4 Maverick (Together) instantly — 100 free credits, no credit card required. Access 343+ AI models through one platform.
More from Together
Use ← → to navigate between models · Esc to go back