Year: 2025
-
Cost Cuts Come, Raptor Shines, Arrow Doubts Emerge in 1Q25 Intel Earnings Call
Newly installed Intel CEO Lip Bu Tan (LBT) held his first earnings call, covering the company’s 1Q25 finances and discussing its outlook. The company plans to cut costs by reorganizing and reducing capital expenditures (capex). Product plans are unchanged, but what was omitted from the 1Q25 review revealed more than what was discussed.
-
SambaNova Retreats from Training, Bets on Inference Services
SambaNova has reduced its workforce by 15% and refocused on AI inference services to navigate competition with GPU giant Nvidia. Employing CGRA and three memory tiers, its architecture on paper beats GPUs but real-world advantages are unclear.
-
Huawei’s Ascend 910C and 384-NPU CloudMatrix Fill China’s AI Void
As Huawei ramps up its AI chip efforts with the Ascend 910C NPU and CloudMatrix 384 system, it seeks to fill a void created by the banning of Nvidia GPUs. Having evolved over the past 5+ years, the Ascend 910 is competitive only in AI inference despite its original aim to tackle training.
-
MEXT Helps IT Leaders Find the Sweet Spot
MEXT software helps IT leaders find the configuration sweet spot that balances capability and cost. By preemptively loading pages from storage, it helps IT managers employ servers with less DRAM, reducing hardware costs without sacrificing performance.
-
Google Adds FP8 to Ironwood TPU; Can It Beat Blackwell?
Google’s TPU v7, Ironwood, offers 10x faster AI processing than v5p. How did Google achieve this speedup, and how does Ironwood compare with Nvidia’s Blackwell?
-
Nvidia Blackwell Shines, AMD MI325X Debuts in Latest MLPerf
MLCommons releases new MLPerf data-center inference results. Also known as The Nvidia Show, the semiannual benchmark report includes new tests in this edition. Nvidia extends its lead, AMD debuts MI325X. Explore scores, scaling, new benchmarks (Llama 3.1), and key AI hardware takeaways.