Getting Started with Building from Source

This document describes how to build the TT-XLA project on your local machine. You must build from source if you want to develop for TT-XLA. If you only want to run models, please choose one of the following sets of instructions instead:

The following topics are covered:

NOTE: If you encounter issues, please request assistance on the TT-XLA Issues page.

Configuring Hardware

Before setup can happen, you must configure your hardware. You can skip this section if you already completed the configuration steps. Otherwise, follow the instructions on the Getting Started page.

System Dependencies

TT-XLA has the following system dependencies:

  • Ubuntu 22.04
  • Python 3.10
  • python3.10-venv
  • Clang 17
  • GCC 11
  • Ninja
  • CMake 4.0.3

Installing Python

If your system already has Python installed, make sure it is Python 3.10:

python3 --version

If not, install Python:

sudo apt install python3.10

Installing CMake 4.0.3

To install CMake 4 or higher, do the following:

  1. Install CMake 4.0.3:
pip install cmake==4.0.3
  1. Check that the correct version of CMake is installed:
cmake --version

If you see cmake version 4.0.3 you are ready for the next section.

Installing Clang 17

To install Clang 17, do the following:

  1. Install Clang 17:
wget https://apt.llvm.org/llvm.sh
chmod u+x llvm.sh
sudo ./llvm.sh 17
sudo apt install -y libc++-17-dev libc++abi-17-dev
sudo ln -s /usr/bin/clang-17 /usr/bin/clang
sudo ln -s /usr/bin/clang++-17 /usr/bin/clang++
  1. Check that the selected GCC candidate using Clang 17 is using 11:
clang -v
  1. Look for the line that starts with: Selected GCC installation:. If it is something other than GCC 11, and you do not see GCC 11 listed as an option, please install GCC 11 using:
sudo apt-get install gcc-11 lib32stdc++-11-dev lib32gcc-11-dev
  1. If you see GCC 12 listed as installed and listed as the default choice, uninstall it with:
sudo rm -rf /usr/bin/../lib/gcc/x86_64-linux-gnu/12

Installing Ninja

To install Ninja, do the following:

sudo apt install ninja-build

Installing OpenMPI

To install OpenMPI, do the following:

sudo wget -q https://github.com/dmakoviichuk-tt/mpi-ulfm/releases/download/v5.0.7-ulfm/openmpi-ulfm_5.0.7-1_amd64.deb -O /tmp/openmpi-ulfm.deb && sudo apt install /tmp/openmpi-ulfm.deb

Installing Additional Dependencies

TT-XLA additionally requires the following libraries:

sudo apt install protobuf-compiler libprotobuf-dev
sudo apt install ccache
sudo apt install libnuma-dev
sudo apt install libhwloc-dev
sudo apt install libboost-all-dev

Build Process

TT-XLA integration with the TT-MLIR compiler is still in progress. Currently TT-XLA depends on the TT-MLIR toolchain to build from source. This build flow provides an easy way to experiment with TT-XLA, StableHLO, and the TT-MLIR infrastructure. The build process will be updated in the future to enhance the user experience.

Building the TT-MLIR Toolchain

Before compiling TT-XLA, the TT-MLIR toolchain needs to be built:

Building TT-XLA

Before running these commands to build TT-XLA, please ensure that the environment variable TTMLIR_TOOLCHAIN_DIR is set to point to the TT-MLIR toolchain directory created above as part of the TT-MLIR environment setup (for example export TTMLIR_TOOLCHAIN_DIR=/opt/ttmlir-toolchain/). You can also set export LOGGER_LEVEL=DEBUG in order to enable debug logs. To build TT-XLA do the following:

  1. Make sure you are not in the TT-MLIR build directory, and you are in the location where you want TT-XLA to install.

  2. Clone TT-XLA:

git clone https://github.com/tenstorrent/tt-xla.git
  1. Navigate into the TT-XLA folder:
cd tt-xla
  1. Run the following set of commands to build TT-XLA:
source venv/activate
cmake -G Ninja -B build # -DCMAKE_BUILD_TYPE=Debug in case you want debug build
cmake --build build

When the build completes, you are ready to set up the TT-XLA wheel.

Building and Installing a Wheel

To install and build a wheel do the following:

  1. Inside the tt-xla directory, navigate into the python_package directory and set up the wheel:
cd python_package
python setup.py bdist_wheel

The above command outputs a python_package/dist/pjrt_plugin_tt*.whl file which is self-contained.

  1. Install the pjrt_plugin_tt wheel:
pip install dist/pjrt_plugin_tt*.whl
  1. This step is not required, these are just example commands to test if the wheel is working. Open Python in the terminal and do the following:
python
import jax
tt_device = jax.devices("tt") # This will trigger plugin registration.
print(tt_device) # This prints the Tenstorrent device info if everything is OK.

The wheel has the following structure:

jax_plugins
`-- pjrt_plugin_tt
    |-- __init__.py
    |-- pjrt_plugin_tt.so   # Plugin itself.
    `-- tt-mlir             # Entire tt-mlir installation folder
        `-- install
            |-- include
            |   `-- ...
            |-- lib
            |   |-- libTTMLIRCompiler.so
            |   |-- libTTMLIRRuntime.so
            |   `-- ...
            `-- tt-metal    # We need to set TT_METAL_HOME to this dir when loading plugin
                |-- runtime
                |   `-- ...
                |-- tt_metal
                |   `-- ...
                `-- ttnn
                    `-- ...

It contains a custom Tenstorrent PJRT plugin (an .so file), __init__.py file which holds a Python function for registering the PJRT plugin with JAX and the tt-mlir installation dir. This is needed in order to dynamically link TT-MLIR libs in runtime and to resolve various tt-metal dependencies without which the plugin does not work.

Structuring wheel/folders this way allows JAX to automatically register the plugin upon usage (explained on OpenXLA's Develop a New JPRT Plugin page here).

Testing

The TT-XLA repo contains various tests in the tests directory. To run an individual test, pytest -svv is recommended in order to capture all potential error messages down the line. Multi-chip tests can be run only on specific Tenstorrent hardware, therefore these tests are structured in folders named by the Tenstorrent cards/systems they can be run on. For example, you can run pytest -v tests/jax/multi_chip/n300 only on a system with an n300 Tenstorrent card. Single-chip tests can be run on any system with the command pytest -v tests/jax/single_chip.

Common Build Errors

  • Building TT-XLA requires clang-17. Please make sure that clang-17 is installed on the system and clang/clang++ links to the correct version of the respective tools.
  • Please also see the TT-MLIR docs for common build errors.

Pre-commit

Pre-commit applies a git hook to the local repository such that linting is checked and applied on every git commit action. Install it from the root of the repository using:

source venv/activate
pre-commit install

If you have already committed something locally before installing the pre-commit hooks, you can run this command to check all files:

pre-commit run --all-files

For more information please visit pre-commit.