Phi 3
Phi-3 is a state of the art, lightweight model. It performs exceptionally well despite being half the size of Llama-3 8B. It is highly capable of being an on-device agentic LLM.
note
Model | Params | Context Length | GQA | Token Count | Knowledge Cutoff |
---|---|---|---|---|---|
Phi-3 Mini 128k Instruct | 3.8B | 128,000 | Yes | 3.3T | October 2023 |
Phi-3 Mini 128k Instruct
Model | Function Calling | MMLU | GPQA | GSM-8K | MATH | MT-bench | MT-bench Pairwise Comparison | |||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Win | Loss | Tie | Win Rate | Loss Rate | Adjusted Win Rate | |||||||
Phi-3 Mini 128k Instruct | - | 68.17 | 30.58 | 80.44 | 28.12 | 7.92 | 51 | 45 | 64 | 0.31875 | 0.28125 | 0.51875 |
Rubra Enhanced Phi-3 Mini 128k Instruct | 65.71% | 66.66 | 29.24 | 74.09 | 26.84 | 7.45 | 45 | 51 | 64 | 0.28125 | 0.31875 | 0.48125 |