

Machine learning developers gained new abilities to develop and run their ML programs on the framework and hardware of their choice thanks to the OpenXLA Project, which today announced the availability of key open source components.
Data scientists and ML engineers often spend a lot of time optimizing their models to work on each hardware target. Whether they’re working in a framework like TensorFlow or PyTorch and targeting GPUs or TPUs, there was no way to avoid this manual effort, which consumed precious time and made it difficult to move applications at a later date.
This is the general problem targeted by the folks behind the OpenXLA Project, which was founded last fall and today includes Alibaba, Amazon Web Services, AMD, Apple, Arm, Cerebra Systems, Google, Graphcore, Hugging Face, Intel, Meta, and NVIDIA as its members.
By creating a unified machine learning compiler that works with a range of ML development frameworks and hardware platforms and runtimes, OpenXLA can accelerate the delivery of ML applications and provide greater code portability.
Today, the group announced the availability of three open source tools as part of the project. XLA is an ML compiler for CPUs, GPUs, and accelerators; StableHLO is an operation set for high-level operations (HLO) in ML that provides portability between frameworks and compilers; while IREE (Intermediate Representation Execution Environment) is an end-to-end MLIR (Multi-Level Intermediate Representation) compiler and runtime for mobile and edge deployments. All three are available for download from the OpenXLA GitHub site
Initial frameworks supported by OpenXLA including TensorFlow, PyTorch, and JAX, a new Google framework JAX is designed for transforming numerical functions, and is described as bringing together a modified version of autograd and TensorFlow while following the structure and workflow of NumPy. Initial hardware targets and optimizations include Intel CPU, Nvidia GPUs, Google TPUs, AMD GPU, Arm CPUs, AWS Trainium and Inferentia, Graphcore’s IPU, and Cerebras Wafer-Scale Engine (WSE). OpenXLA’s “target-independent optimizer” targets albebraic functions, op/kernel fusion, weight update sharding, full-graph layout propagation, scheduling, and SPMD for parallelism.
The OpenXLA compiler products can be used with a variety of ML use cases, including full-scale training of massive deep learning models, including large language models (LLMs) and even generative computer vision models like Stable Diffusion. It can also be used for inference; Waymo already uses OpenXLA for real-time inferencing on its self-driving cars, according to a post today on the Google open source blog.

The OpenXLA compiler ecosystem provides portability between ML development tools and hardware targets (Image source OpenXLA Project)
OpenXLA members touted some of their early successes with the new compiler. Alibaba, for instance, says it was able to train a GPT2 model on Nvidia GPUs 72% faster using OpenXLA, and saw an 88% speedup for a Swin Transformer training task on GPUs.
Hugging Face, meanwhile, said it saw about a 100% speedup when it paired XLA with its text generation model written in TensorFlow. “OpenXLA promises standardized building blocks upon which we can build much needed interoperability, and we can’t wait to follow and contribute!” said Morgan Funtowicz, head of machine learning optimization for the Brooklyn, New York, company.
Facebook was able to “achieve significant performance improvements on important projects,” including using XLA on PyTorch models running on Cloud TPUs, said Soumith Chintala, the lead maintainer for PyTorch.
The chip startups are pleased for XLA, which reduces the risks of adopting relatively new, unproven hardware for customers. “Our IPU compiler pipeline has used XLA since it was made public,” said David Norman, Graphcore’s director of software design. “Thanks to XLA’s platform independence and stability, it provides an ideal frontend for bringing up novel silicon.”
“OpenXLA helps extend our user reach and accelerated time to solution by providing the Cerebras Wafer-Scale Engine with a common interface to higher level ML frameworks,” says Andy Hock, a vice president and head of product at Cerebras. “We are tremendously excited to see the OpenXLA ecosystem available for even broader community engagement, contribution, and use on GitHub.”
AMD and Arm, which are battling bigger chipmakers for pieces of the ML training and serving pies, are also happy members of the OpenXLA Project.
“We value projects with open governance, flexible and broad applicability, cutting edge features and top-notch performance and are looking forward to the continued collaboration to expand open source ecosystem for ML developers,” Alan Lee, AMD’s corporate vice president of software development, said in the blog.
“The OpenXLA Project marks an important milestone on the path to simplifying ML software development,” said Peter Greenhalgh, vice president of technology and fellow at Arm. “We are fully supportive of the OpenXLA mission and look forward to leveraging the OpenXLA stability and standardization across the Arm Neoverse hardware and software roadmaps.”
Curiously absent are IBM, which continues to innovate on chips with its Power10 processor, and Microsoft, the world’s second largest provider behind AWS.
Related Items:
Google Announces Open Source ML Compiler Project, OpenXLA
AMD Joins New PyTorch Foundation as Founding Member
Inside Intel’s nGraph, a Universal Deep Learning Compiler
February 26, 2025
- Precisely Unveils AI-Powered Advancements and Elevated Data Governance Service in the Data Integrity Suite
- Redpanda Launches Snowflake Connector with Snowpipe Streaming for Faster Data Ingestion
- Qlik Study: 94% of Businesses Boost AI Investment, But Only 21% Have Fully Operationalized It
- Observo AI Launches Orion to Simplify Data Pipeline Management with AI
- Cribl Unveils Lakehouse for Scalable, Schema-Free Telemetry Data Management
- IBM Expands Granite LLM Family with New Multi-Modal and Reasoning AI Built for the Enterprise
- New Relic Enhances Observability Platform with AI-Driven Cloud Cost and Pipeline Insights
February 25, 2025
- Hydrolix Joins AWS ISV Accelerate Program
- New Relic Unveils 20+ AI Innovations to Enhance Intelligent Observability
- AMI Expands AI and GPU Management in Data Center Manager v6.0
- Gurobi Releases State of Mathematical Optimization in Data Science Report
- Red Hat Enhances Security and Virtualization Experience with Latest Version of Red Hat OpenShift
- Deutsche Telekom and Google Cloud Partner on Agentic AI for Autonomous Networks
- Wasabi Survey Finds 56% of Businesses Hindered by Cloud Storage Costs
- New Relic Enhances Observability with AI-Powered Video QoE and User Engagement Insights
- Dataminr and Esri Expand Partnership to Transform Location Intelligence with Real-Time Event Detection
- Cisco Expands Partnership with NVIDIA to Accelerate Enterprise AI Adoption
- IBM to Acquire DataStax to Enhance watsonx for GenAI and Unstructured Data
February 24, 2025
- OpenTelemetry Is Too Complicated, VictoriaMetrics Says
- What Are Reasoning Models and Why You Should Care
- PayPal Feeds the DL Beast with Huge Vault of Fraud Data
- Three Ways Data Products Empower Internal Users
- Keeping Data Private and Secure with Agentic AI
- Memgraph Bolsters AI Development with GraphRAG Support
- Three Data Challenges Leaders Need To Overcome to Successfully Implement AI
- What Leonardo DaVinci Teaches Us About Data Management
- Demystifying AI: What Every Business Leader Needs to Know
- Top-Down or Bottom-Up Data Model Design: Which is Best?
- More Features…
- Meet MATA, an AI Research Assistant for Scientific Data
- AI Agent Claims 80% Reduction in Time to Complete Data Tasks
- Snowflake Unleashes AI Agents to Unlock Enterprise Data
- IBM to Buy DataStax for Database, GenAI Capabilities
- DataRobot Expands AI Capabilities with Agnostiq Acquisition
- EDB Says It Tops Oracle, Other Databases in Benchmarks
- Databricks Unveils LakeFlow: A Unified and Intelligent Tool for Data Engineering
- Collibra Bolsters Position in Fast-Moving AI Governance Field
- Microsoft Open Sources Code Behind PostgreSQL-Based MongoDB Clone
- AI Making Data Analyst Job More Strategic, Alteryx Says
- More News In Brief…
- Informatica Reveals Surge in GenAI Investments as Nearly All Data Leaders Race Ahead
- Gartner Predicts 40% of Generative AI Solutions Will Be Multimodal By 2027
- PEAK:AIO Powers AI Data for University of Strathclyde’s MediForge Hub
- DataRobot Acquires Agnostiq to Accelerate Agentic AI Application Development
- Cloudera Welcomes Tom Brady as Keynote Speaker at ELEVATE26
- Starburst Closes Record FY25, Fueled by Rising AI Demand and Growing Enterprise Momentum
- TigerGraph Launches Savanna Cloud Platform to Scale Graph Analytics for AI
- Open Institute of Technology Expands Curricula and Enrollment to the United States
- General Assembly Launches Suite of Upskilling Programs to Prepare Businesses for an AI-Driven Economy
- EY and Microsoft Unveil AI Skills Passport to Bridge Workforce AI Training Gap
- More This Just In…