Model Lineup
MinT lists models by availability status.
Available Models
| Model Name | Training Type | Architecture | Size |
|---|---|---|---|
| Qwen/Qwen3-0.6B | Hybrid | Dense | Tiny |
| Qwen/Qwen3-4B-Instruct-2507 | Instruction | Dense | Compact |
| Qwen/Qwen3-30B-A3B-Instruct-2507 | Instruction | MoE | Medium |
| Qwen/Qwen3-235B-A22B-Instruct-2507 | Instruction | MoE | Large |
Coming Soon
| Model Name | Training Type | Architecture | Size |
|---|---|---|---|
| Qwen/Qwen3-30B-A3B | Hybrid | MoE | Medium |
| Qwen/Qwen3-30B-A3B-Base | Base | MoE | Medium |
| Qwen/Qwen3-8B | Hybrid | Dense | Small |
| Qwen/Qwen3-8B-Base | Base | Dense | Small |
| deepseek-ai/DeepSeek-V3.1 | Hybrid | MoE | Large |
| deepseek-ai/DeepSeek-V3.1-Base | Base | MoE | Large |
| moonshotai/Kimi-K2-Thinking* | Reasoning | MoE | Large |
| Qwen/Qwen3-VL-30B-A3B-Instruct | Vision | MoE | Medium |
| Qwen/Qwen3-VL-235B-A22B-Instruct | Vision | MoE | Large |
| π0 | Robotics | Dense | Small |
*Contact sales for Kimi-K2 access
Model Selection Recommendations
- Low-latency: Qwen3-0.6B or Qwen3-4B-Instruct-2507
- Balanced quality: Qwen3-30B-A3B-Instruct-2507
- Frontier scale: Qwen3-235B-A22B-Instruct-2507
Model Categories
By Training Type
- Hybrid - Mixed general + instruction behavior
- Instruction - Fine-tuned for instruction following
By Architecture
- Dense - Traditional transformer architecture
- MoE (Mixture of Experts) - Sparse activation for efficiency
By Size
- Tiny: under 1B parameters
- Compact: 1-4B parameters
- Medium: ~30B parameters
- Large: 200B+ parameters
Cost Efficiency
MoE models offer superior cost-efficiency: pricing scales with active parameters, not total model size. For example, a 235B MoE model with 22B active parameters costs the same as a 22B dense model.