AI Models / Compare
Phi-4 Mini
Tiny but capable 3.8B Phi model for on-device and low-latency inference.
- Creator
- Microsoft
- Lifecycle
- Active
- Context
- 128.0K
- Max output
- 16.4K
- Released
- Mar 1, 2025
- Status
- unknown
- Input
- $0.04 / 1M tokens
- Output
- $0.13 / 1M tokens
- Cached read
- — / 1M tokens
- Cached write
- — / 1M tokens
- Batch discount
- —%
- Source
- Phi-4 Mini pricing
- Verified
- Apr 5, 2026 (High)
Capabilities
- Modalities
- text→text
- Capabilities
- batchSupportfunctionCallingstructuredOutputs
- Strengths
- 128K context in 3.8B, On-device friendly
- Tradeoffs
- Lower reasoning ceiling than Phi-4
Migration Guidance
Best tiny model for on-device. Upgrade to Phi-4 for more demanding tasks.
Replacement models: phi-4
Other models from Microsoft
Phi-4