Startup InferX’s software quickly and securely loads and unloads AI models on Nvidia GPUs, reducing idle time versus dedicating a GPU to a model. continue reading
Newly installed Intel CEO Lip Bu Tan (LBT) held his first earnings call, covering the company’s 1Q25 finances and discussing its outlook. The company plans to cut costs by reorganizing and reducing capital expenditures (capex). Product plans are unchanged, but what was omitted from the 1Q25 review revealed more than what was discussed. continue reading
SambaNova has reduced its workforce by 15% and refocused on AI inference services to navigate competition with GPU giant Nvidia. Employing CGRA and three memory tiers, its architecture on paper beats GPUs but real-world advantages are unclear. continue reading
As Huawei ramps up its AI chip efforts with the Ascend 910C NPU and CloudMatrix 384 system, it seeks to fill a void created by the banning of Nvidia GPUs. Having evolved over the past 5+ years, the Ascend 910 is competitive only in AI inference despite its original aim to tackle training. continue reading
AMD Ampere Arm auto Broadcom Ceva Computex Condor CPU data center DPU edge AI embedded Epyc FPGA Google GPU Imagination Immortalis industrial Intel Mali Marvell MCU MediaTek memory Meta MEXT Microsoft MIPS MLPerf NPU (AI accelerator) Nvidia NXP PC policy process tech Qualcomm RISC-V smartphone SoftBank software Tenstorrent Tesla Zen 5