HPC News Bytes 20260202: Microsoft’s Inference Chip, H200 (Not H20) GPUs for China, Mega AI Data Center Deals

A happy month of February to you! The big players have dominated the HPC-AI news front of late, here’s a fast (8:44) recap of recent developments, including: Microsoft Maia 200 ….

HPC News Bytes 20260105: Denied H20s, China Now Granted H200s, Meta Acquires Manus, Sandia’s Photonics Quantum Chip

A happy and prosperous 2026 to you! The world of HPC-AI transitioned to the new year while generating its fair share of news, here’s a rapid (8:35) recap of recent developments, including ….

Arm and Meta Announce Extension of AI Partnership

Arm and Meta announced the extension of their partnership to scale AI across multiple layers of compute – spanning AI software and data center infrastructure. The multi-year partnership builds on the ongoing hardware and software co-design efforts between the two companies, combining Arm AI chips with Meta’s AI-driven products, infrastructure, and open technologies. “From milliwatt-scale […]

HPC News Bytes 20250630: The GPU-ASICs War – Chips from Microsoft, Meta, Google, and Why Engineering Projects Get Delayed

A happy Fourth of July week to you! There’s been much movement lately by the Big Tech companies developing their own AI chips, a.k.a., ASICs, this includes Meta, Microsoft and Google, to name three. In this episode ….

AI Inference: Meta Teams with Cerebras on Llama API

Meta has teamed with Cerebras on AI inference in Meta’s new Llama API, combining  Meta’s open-source Llama models with inference technology from Cerebras. Developers building on the Llama 4 Cerebras model in the API can expect speeds up to 18 times faster than traditional GPU-based solutions ….

AI Inference: Meta Collaborates with Cerebras on Llama API

Sunnyvale, CA — Meta has teamed with Cerebras on AI inference in Meta’s new Llama API, combining  Meta’s open-source Llama models with inference technology from Cerebras. Developers building on the Llama 4 Cerebras model in the API can expect speeds up to 18 times faster than traditional GPU-based solutions, according to Cerebras. “This acceleration unlocks […]

News Bytes Podcast 20250217: Arm Selling Its Own Chips to Meta?, Big xAI, Big Power, Big… Pollution?, TSMC in Intel Fab Takeover?, Europe’s Big AI Investment

A happy mid-winter day to you! The HPC-AI world produced its usual extraordinary onslaught of news this past week, here’s a quick (7:27) jaunt through recent developments: Arm to sell its own server CPU – with Meta as first customer?, Big AI, Big Power … Big Pollution?

HPC News Bytes 20250217: Arm Selling Its Own Chips to Meta?, Big xAI, Big Power, Big… Pollution?, TSMC in Intel Fab Takeover?, Europe’s Big AI Investment, Exascale in Industry

A happy mid-winter day to you! The HPC-AI world produced its usual extraordinary onslaught of news this past week, here’s …

HPC News Bytes 20250210: Big AI CAPEX Binge, More Data Center SMRs, Euro-Origin Quantum, Softbank Eyes Ampere

Good post-Super Bowl morning to you! The past week was somewhat calm relative to the riotous preceding period in the world of HPC-AI, but still, important developments emerged. Here’s a fast (5:52) overview ….

Lambda Launches Inference API

Dec. 13, 2024 — AI company Lambda today announced its Inference API, which the company said enables access to LLMs through a serverless AI for “a fraction of a cent.” The company said Lambda Inference API offers low-cost, scalable AI inference with such models as Meta’s recently released Llama 3.3 70B Instruct (FP8) at $0.20 […]