📣 Creators, Exciting News!
Gate Square Certified Creator Application Is Now Live!
How to apply:
1️⃣ Open App → Tap [Square] at the bottom → Click your avatar in the top right
2️⃣ Tap [Get Certified] under your avatar
3️⃣ Once approved, you’ll get an exclusive verified badge that highlights your credibility and expertise!
Note: You need to update App to version 7.25.0 or above to apply.
The application channel is now open to KOLs, project teams, media, and business partners!
Super low threshold, just 500 followers + active posting to apply!
At Gate Square, everyone can be a community leader! �
Microsoft Open Source New Version of Phi-4: Inference Efficiency Rises 10 Times, Can Run on Laptops
Jin10 data reported on July 10, this morning, Microsoft open sourced the latest version of the Phi-4 family, Phi-4-mini-flash-reasoning, on its official website. The mini-flash version continues the Phi-4 family’s characteristics of small parameters and strong performance, specifically designed for scenarios limited by Computing Power, memory, and latency, capable of running on a single GPU, suitable for edge devices like laptops and tablets. Compared to the previous version, mini-flash utilizes Microsoft’s self-developed innovative architecture, SambaY, resulting in a big pump in inference efficiency by 10 times, with average latency reduced by 2-3 times, achieving a significant improvement in overall inference performance.