Supported Models
ModelMax routes requests to the best available provider. All models are accessed through a single API.
Chat models
Text-in, text-out conversation models. Used with POST /v1/chat/completions.
Google Gemini
| Model ID | Input | Output | Notes |
|---|---|---|---|
gemini-3.1-pro-preview | text, image, audio, video | text | Latest Gemini Pro |
gemini-3-pro-preview | text, image, audio, video | text | Gemini 3.0 Pro |
gemini-3-flash-preview | text, image, audio, video | text | Fast, cost-effective |
gemini-3.1-flash-image-preview | text, image | text, image | Image generation capable |
gemini-3.1-flash-lite-preview | text, image, audio, video | text | Lightweight |
AWS Bedrock — DeepSeek
| Model ID | Input | Output | Notes |
|---|---|---|---|
deepseek-r1 | text | text | Reasoning model with chain-of-thought |
deepseek-v3.1 | text | text | General purpose |
deepseek-v3.2 | text | text | Latest general purpose |
AWS Bedrock — Qwen
| Model ID | Input | Output | Notes |
|---|---|---|---|
qwen3-coder-30b-a3b | text | text | Code-focused, lightweight |
qwen3-32b | text | text | General purpose |
qwen3-235b-a22b-2507 | text | text | Large, capable |
qwen3-coder-480b-a35b | text | text | Code-focused, high capacity |
qwen3-next-80b-a3b | text | text | Efficient MoE architecture |
qwen3-vl-235b-a22b | text, image | text | Vision-language model |
qwen3-coder-next | text | text | Latest coder model |
AWS Bedrock — MiniMax
| Model ID | Input | Output | Notes |
|---|---|---|---|
minimax-m2 | text | text | MiniMax M2 |
minimax-m2.1 | text | text | MiniMax M2.1 |
AWS Bedrock — Kimi (Moonshot)
| Model ID | Input | Output | Notes |
|---|---|---|---|
kimi-k2-thinking | text | text | With reasoning |
kimi-k2.5 | text | text | Latest Kimi |
Video models
Async video generation. Used with the Queue API (POST /v1/queue/{model}).
Google Gemini — Veo
| Model ID | Speed | Quality | Audio |
|---|---|---|---|
veo-3.1 | Standard | Highest | Yes |
veo-3.1-fast | Fast | High | Yes |
veo-3 | Standard | High | Yes |
veo-3-fast | Fast | Good | Yes |
veo-2 | Standard | Good | Yes |
All Veo models support:
- Text-to-video and image-to-video
- Up to 8 seconds duration
- 720p, 1080p, and 4K resolution
- Audio generation
Capability matrix
| Capability | Gemini Chat | Bedrock Chat | Qwen VL | Veo Video |
|---|---|---|---|---|
| Text input | Yes | Yes | Yes | Yes |
| Image input | Yes | — | Yes | Yes |
| Audio input | Yes | — | — | — |
| Video input | Yes | — | — | — |
| Text output | Yes | Yes | Yes | — |
| Image output | Partial | — | — | — |
| Video output | — | — | — | Yes |
| Streaming | Yes | Yes | Yes | — |
Choosing a model
For general chat: Start with gemini-3-flash-preview (fast, multimodal) or deepseek-v3.2 (strong text reasoning).
For code: Try qwen3-coder-480b-a35b or qwen3-coder-next.
For reasoning: Use deepseek-r1 (returns reasoning_content in the response).
For vision: Use gemini-3-flash-preview (images, audio, video input) or qwen3-vl-235b-a22b (images only).
For video generation: Use veo-3 for quality, veo-3-fast for speed.
