GitHunt
VA

A machine learning compiler for GPUs, CPUs, and ML accelerators

XLA

XLA (Accelerated Linear Algebra) is an open-source machine learning (ML)
compiler for GPUs, CPUs, and ML accelerators.

OpenXLA Ecosystem

The XLA compiler takes models from popular ML frameworks such as PyTorch,
TensorFlow, and JAX, and optimizes them for high-performance execution across
different hardware platforms including GPUs, CPUs, and ML accelerators.

openxla.org is the project's website.

Get started

If you want to use XLA to compile your ML project, refer to the corresponding
documentation for your ML framework:

If you're not contributing code to the XLA compiler, you don't need to clone and
build this repo. Everything here is intended for XLA contributors who want to
develop the compiler and XLA integrators who want to debug or add support for ML
frontends and hardware backends.

Contribute

If you'd like to contribute to XLA, review
How to Contribute and then see the
developer guide.

Contacts

  • For questions, contact the maintainers - maintainers at openxla.org

Resources

Code of Conduct

While under TensorFlow governance, all community spaces for SIG OpenXLA are
subject to the
TensorFlow Code of Conduct.

Languages

C++89.0%MLIR5.2%Starlark4.7%Python0.5%C0.2%Smarty0.2%Shell0.1%CMake0.0%LLVM0.0%Linker Script0.0%Batchfile0.0%
Apache License 2.0
Created December 13, 2023
Updated September 25, 2025
vam-google/xla | GitHunt