
SambaNova has reduced its workforce by 15% and refocused on AI inference services to navigate competition with GPU giant Nvidia. Employing CGRA and three memory tiers, its architecture on paper beats GPUs but real-world advantages are unclear. continue reading

As Huawei ramps up its AI chip efforts with the Ascend 910C NPU and CloudMatrix 384 system, it seeks to fill a void created by the banning of Nvidia GPUs. Having evolved over the past 5+ years, the Ascend 910 is competitive only in AI inference despite its original aim to tackle training. continue reading

MEXT software helps IT leaders find the configuration sweet spot that balances capability and cost. By preemptively loading pages from storage, it helps IT managers employ servers with less DRAM, reducing hardware costs without sacrificing performance. continue reading

Google’s TPU v7, Ironwood, offers 10x faster AI processing than v5p. How did Google achieve this speedup, and how does Ironwood compare with Nvidia’s Blackwell? continue reading

MLCommons releases new MLPerf data-center inference results. Also known as The Nvidia Show, the semiannual benchmark report includes new tests in this edition. Nvidia extends its lead, AMD debuts MI325X. Explore scores, scaling, new benchmarks (Llama 3.1), and key AI hardware takeaways. continue reading

MediaTek’s Arm-based Kompanio Ultra 910 rivals Intel’s Core Ultra 5 125U performance at lower power. What does this mean for Chromebook Plus, thin clients, and Windows on Arm? continue reading
Amazon AMD Arm auto Broadcom business Cerebras Ceva consumer CPU data center DPU DSP edge AI embedded FPGA Google GPU Imagination industrial Intel interconnect Marvell MCU MediaTek Meta Microsoft MLPerf NPU (AI accelerator) Nvidia NXP OpenAI PC process tech Qualcomm Renesas RISC-V SambaNova Semidynamics SiFive smartphone SoftBank software Tenstorrent Tesla