As Huawei ramps up its AI chip efforts with the Ascend 910C NPU and CloudMatrix 384 system, it seeks to fill a void created by the banning of Nvidia GPUs. Having evolved over the past 5+ years, the Ascend 910 is competitive only in AI inference despite its original aim to tackle training. continue reading
MEXT software helps IT leaders find the configuration sweet spot that balances capability and cost. By preemptively loading pages from storage, it helps IT managers employ servers with less DRAM, reducing hardware costs without sacrificing performance. continue reading
Google’s TPU v7, Ironwood, offers 10x faster AI processing than v5p. How did Google achieve this speedup, and how does Ironwood compare with Nvidia’s Blackwell? continue reading
MLCommons releases new MLPerf data-center inference results. Also known as The Nvidia Show, the semiannual benchmark report includes new tests in this edition. Nvidia extends its lead, AMD debuts MI325X. Explore scores, scaling, new benchmarks (Llama 3.1), and key AI hardware takeaways. continue reading
MediaTek’s Arm-based Kompanio Ultra 910 rivals Intel’s Core Ultra 5 125U performance at lower power. What does this mean for Chromebook Plus, thin clients, and Windows on Arm? continue reading
Researchers reverse-engineered an Nvidia Ampere GPU, revealing microarchitecture details such as compiler-controlled dependency management and caching. continue reading
Altera AMD Apple APU/IGP Arm auto Broadcom Cerebras Ceva Computex CPU data center DPU edge AI embedded Epyc FPGA Google GPU Imagination Immortalis Intel Mali Marvell MCU MediaTek memory Meta Microchip Microsoft MLPerf NPU (AI accelerator) Nvidia NXP PC process tech Qualcomm RISC-V SiFive smartphone SoftBank Syntiant Tenstorrent Tesla Zen 5