Microsoft Open Source New Version of Phi-4: Inference Efficiency Rises 10 Times, Can Run on Laptops
Jin10 data reported on July 10, this morning, Microsoft open-sourced the latest version of the Phi-4 family, Phi-4-mini-flash-reasoning, on their official website. The mini-flash version continues the Phi-4 family's characteristic of small parameters and strong performance, specifically designed for scenarios limited by Computing Power, memory, and latency. It can run on a single GPU, making it suitable for edge devices such as laptops and tablets. Compared to the previous version, mini-flash utilizes Microsoft's self-developed innovative architecture SambaY, achieving a 10-fold increase in reasoning efficiency, with average latency reduced by 2 to 3 times, resulting in a significant overall improvement in reasoning performance.