losclouds

AI Models / Compare

Phi-4 Mini

Tiny but capable 3.8B Phi model for on-device and low-latency inference.

Creator
Microsoft
Lifecycle
Active
Context
128.0K
Max output
16.4K
Released
Mar 1, 2025
Status
unknown
Input
$0.04 / 1M tokens
Output
$0.13 / 1M tokens
Cached read
/ 1M tokens
Cached write
/ 1M tokens
Batch discount
%
Source
Phi-4 Mini pricing
Verified
Apr 5, 2026 (High)

Capabilities

Modalities
texttext
Capabilities
batchSupportfunctionCallingstructuredOutputs
Strengths
128K context in 3.8B, On-device friendly
Tradeoffs
Lower reasoning ceiling than Phi-4
Official Links

Benchmark Coverage

BenchmarkVersionScoreDateSourceNotes

Release History

ReleaseAliasLifecycleRelease DateDeprecationShutdownSummary
Phi-4 Miniphi-4-miniActiveMar 1, 2025Current published model family snapshot.

Host Coverage

HostTypeContextPricing NoteDifferences
Azure AI Foundryfirst-party128.0K$0.04/$0.13 per MTok.
Migration Guidance

Best tiny model for on-device. Upgrade to Phi-4 for more demanding tasks.

Replacement models: phi-4

Change Events
DateTypeTitleDescriptionSource
Mar 1, 2025family_addedPhi-4 Mini publishedInitial public model family launch.Phi-4 Mini release notes

Other models from Microsoft

Phi-4