AWS Announces EC2 UltraCluster and GA of Trainium2 Instances

LAS VEGAS, Dec. 3, 2024 — At AWS re:Invent, Amazon Web Services today announced the general availability of AWS Trainium2-powered Amazon Elastic Compute Cloud (Amazon EC2) instances, introduced new Trn2 UltraServers, enabling customers to train ….

Oriole Networks Raises £10m for Faster LLM Training

London, 27 March 2024: Oriole Networks – a startup using light to train LLMs faster with less power – has raised £10 million in seed funding to improve AI performance and adoption, and solve AI’s energy problem. The round, which the company said is one of the UK’s largest seed raises in recent years, was co-led […]

Advancing HPC through oneAPI Heterogeneous Programming in Academia & Research

oneAPI offers an open industry effort supported by over 100 organizations. oneAPI is an open, unified, cross-architecture programming model for CPUs and accelerator architectures (GPUs, FPGAs, and others). Based on standards, the programming model simplifies software development and delivers uncompromised performance for accelerated compute without proprietary lock-in, while enabling the integration of existing code.

Los Alamos Claims Quantum Machine Learning Breakthrough: Training with Small Amounts of Data

Researchers at Los Alamos National Laboratory today announced a quantum machine learning “proof” they say shows that training a quantum neural network requires only a small amount of data, “(upending) previous assumptions stemming from classical computing’s huge appetite for data in machine learning, or artificial intelligence.” The lab said the theorem has direct applications, including […]

MLPerf: Latest Results Highlight ‘More Capable ML Training’

Open engineering consortium MLCommons has released new results from MLPerf Training v2.0, which measures how fast various platforms train machine learning models. The organizations said the latest MLPerf Training results “demonstrate broad industry participation and up to 1.8X greater performance ultimately paving the way for more capable intelligent systems….” As it has done with previous […]

Scalable Inferencing for Autonomous Trucking

In this sponsored post, Tim Miller, Vice President, Product Marketing, One Stop Systems, discusses autonomous trucking and that to achieve AI Level 4 (no driver) in the vehicles, powerful AI inference hardware supporting many different inferencing engines operating and coordinating simultaneously is required.

Azure Adopts AMD Instinct MI200 GPU for Large-Scale AI Training

SANTA CLARA, Calif. May 26, 2022 — Microsoft has announced the use of AMD Instinct MI200 GPU accelerators for large-scale AI training workloads. Microsoft also announced it is working with the PyTorch Core team and AMD data center software team to optimize the performance and developer experience for customers running PyTorch on Microsoft Azure. AMD […]

Rice Univ. Researchers Claim 15x AI Model Training Speed-up Using CPUs

Reports are circulating in AI circles that researchers from Rice University claim a breakthrough in AI model training acceleration – without using accelerators. Running AI software on commodity x86 CPUs, the Rice computer science team say  neural networks can be trained 15x faster than platforms utilizing GPUs. If valid, the new approach would be a double boon for organizations implementing AI strategies: faster model training using less costly microprocessors.