HPC News Bytes 20250929: ‘World Models’ Next Big AI Thing?, ‘Central Park-Sized’ Data Centers, Silicon Cooling Advancement, How to Tame a 6K+ Qubit Array

Good late-September morning to you! The HPC-AI world last week barreled ahead at breakneck ….

HPC News Bytes 20250818: US to Invest in Intel?; DeepSeek Training Trouble; RISC-V GPU Startup Seeks Funds; NSF-NVIDIA Partnerhip

A good August morning to you! It was a geopolitically-weighted week in the world of HPC-AI, here’s a quick (8:29) overview of recent developments, including: US government to invest in Intel … ?

Report: DeepSeek Model Delay Forces Use of NVIDIA GPUs

The Financial Times reported that DeepSeek was forced to delay the releas e of its new R2 large language model due to performance problems with Huawei Ascend processors, “highlighting the limits of Beijing’s push to replace US technology” …

Ceramic.ai Emerges from Stealth, Reports 2.5x Faster Model Training

Ceramic.ai emerged from stealth today with software for foundation model training infrastructure designed to enable enterprises to build and fine-tune generative AI models more efficiently. Founded by Anna Patterson, former Google VP of Engineering ….

Ceramic.ai Emerges from Stealth, Reports 2.5x Faster Model Training

SAN FRANCISCO — March 5, 2025 — Ceramic.ai emerged from stealth today with software for foundation model training infrastructure designed to enable enterprises to build and fine-tune generative AI models more efficiently. Founded by Anna Patterson, former Google VP of Engineering and Gradient Ventures founder, Ceramic.ai said it improves AI model training speed and cost-efficiency, […]

Lambda Launches Nvidia-Based Cloud Clusters for AI Model Training

SAN JOSE, July 24, 2024 — GPU cloud company Lambda has unveiled Lambda 1-Click Clusters, designed for AI engineers’ and researchers’ short-term access to multi-node GPU clusters in the cloud for large-scale AI model training. Lambda said the launch marks the first time such access to NVIDIA H100 Tensor Core GPUs on 2 to 64 […]

Cerebras Announces 2 AI Partnerships – for Large Language Model Training and ‘Generative AI’ Content Creation

Cerebras Systems, maker of the “dinner plate sized” AI processor, announced two alliances today, one with Cirrascale Cloud Services, provider of deep learning solutions for AVs, NLP and computer vision, and with Jasper, maker of an AI content platform for AI-based copywriting and content creation. Under the Cirrascale-Cerebras partnership, the two companies announced the availability […]

AWS Announces GA of EC2 Trn1 Instances for ML Model Training 

SEATTLE — Oct. 10, 2022 — Amazon Web Services today announced the general availability of Amazon Elastic Compute Cloud (Amazon EC2) Trn1 instances powered by AWS-designed Trainium chips. Trn1 instances are built for high-performance training of machine learning models in the cloud. AWS said the offering saves up to 50 percent cost-to-train savings over comparable […]

Azure Adopts AMD Instinct MI200 GPU for Large-Scale AI Training

SANTA CLARA, Calif. May 26, 2022 — Microsoft has announced the use of AMD Instinct MI200 GPU accelerators for large-scale AI training workloads. Microsoft also announced it is working with the PyTorch Core team and AMD data center software team to optimize the performance and developer experience for customers running PyTorch on Microsoft Azure. AMD […]

Google Cloud Says TPU-Powered Machine Learning Cluster Delivers 9 Exaflops Aggregate Power

As HPC enters the exascale era, watchers of the TOP500 list of the world’s most powerful supercomputers will look to see if the updated list, to be released this month at the ISC conference in Hamburg, will include systems that break the vaunted exaflops barrier. That said, Google Cloud yesterday unveiled what it called the […]