TT-Lang Specification

Specification Versions

Version

Date

Description of changes

0.1

12/15/2025

Initial version

0.2

01/20/2026

Remove ttl.Program

0.3

01/23/2026

Add specification for block operators and math functions

0.4

01/26/2026

Add ttl.math.broadcast

0.5

02/05/2026

Use dataflow buffer instead of circular buffer term

0.6

02/06/2026

Add rounding, mask, ttl.math.transpose, ttl.math.fill and ttl.math.where functions

0.7

02/09/2026

Move push and pop from ttl.DataflowBuffer to ttl.Block

0.8

02/09/2026

Formal block states

0.9

03/04/2026

Add ttl.GroupTransfer

0.9

03/06/2026

Add ttl.signpost

0.10

03/06/2026

Add debug printing

0.11

03/19/2026

Rename ttl.core to ttl.node

0.12

03/24/2026

Remove store(..., acc=True)

0.13

03/31/2026

Rename ttl.kernel to ttl.operation

0.14

04/02/2026

Add ttl.math.abs, ttl.math.neg and ttl.math.pow in addition to Python built-in operators.

Introduction

TT-Lang is a Python based domain specific language (DSL) designed to express low-level programs for Tenstorrent hardware. Here the low-level is used to contrast with the high-level programming model represented in TT-NN. Specifically, this means that computation and dataflow kernels are programmed separately and need to be explictly syncronized, while TT-NN hides this from the user in its built-in operations. Furthermore, instead of operating on entire tensors, like TT-NN, TT-Lang operates on chunks of tensors called blocks, for which the user has to pick the shape to achieve best performance. While based on Python the TT-Lang maintains a number of constraints to what parts of Python can be used in what context, hence the DSL nature of it. The TT-Lang is tightly integrated with TT-NN to provide seamless experience of mixing built-in TT-NN operations and user-written TT-Lang operations in a single program.

In addition to kernels, TT-Lang offers other abstractions familiar to TT-Metalium users such as dataflow buffers and semaphores. In contrast with TT-Metalium, TT-Lang also offers new, higher level abstractions. In addition to blocks these new abstractions include tensor slices and pipes that wrap the complexity of dealing with tensor memory layout and node-to-node communication correspondingly.

Operation function

Operation function is a Python function with ttl.operation decorator. This function takes input and output TT-NN tensors as arguments and returns None. An operation function contains definitions of kernel functions as well as objects shared by all kernel functions. A kernel function is a Python function with no arguments and returning None that is annotated by ttl.compute or ttl.datamovement decorators. The user can call TT-Lang operation function from a TT-NN program and is free to mix it with calling any of the built-in TT-NN operations.

Example

@ttl.operation()
def __foo(
    x: ttnn.Tensor, # input tensor
    y: ttnn.Tensor, # output tensor
) -> None:
    # ...

    @ttl.compute()
    def some_compute():
        # ...

    @ttl.datamovement()
    def some_dm0():
        # ...

    @ttl.datamovement()
    def some_dm1():
        # ...

# Simple wrapper to allow returning output tensor in TT-NN style
def foo(x: ttnn.Tensor) -> ttnn.Tensor:
    y = ttnn.zeros(x.shape, layout=ttnn.TILE_LAYOUT)
    __foo(x, y)
    return y

shape = ttnn.Shape([128, 128])

x = ttnn.rand(shape, layout=ttnn.TILE_LAYOUT)

ttnn.exp(y, foo(ttnn.abs(x)), fast_and_approximate_mode=True)

Grid

A grid defines a space of nodes to which the TT-Lang operation is submitted for execution. A node corresponds to a single Tensix Core and is a minimal unit capable of executing a TT-Lang program. In a single-chip case where node-to-node communication is conducted over Network-on-Chip (NoC), the grid is two dimensional. In a multi-chip case where chip-to-chip communication is conduced over TT-Fabric, the grid has additional mesh dimensions representing different levels of connectivity (same card, same host, same rack etc). There is also Single-Program-Multiple-Data (SPMD) mode in which the grid remains two dimensional while the TT-Lang operation is submitted for execution on multiple chips. In SPMD mode TT-Lang operation instances have the same behaviour on different chips while working on different partitions of data, which significantly simplifies reasoning about it.

Grid size function

The ttl.grid_size function returns the size of the grid. The function takes an argument that specifies how many dimensions to return. If requested dimensions are smaller than grid dimensions, the highest rank dimension is flattened. If requested dimensions are greater than grid dimensions, highest rank dimensions are padded with a value of one. The ttl.grid_size can be used inside an operation function as well as inside kernel functions.

Type alias/Function

Description

ttl.PositiveInt = Annotated[int, Gt(0)]

A positive integer. The metadata Gt(0), can be used by runtime type-checkers to enforce the integer constraints.

ttl.Size = ttl.PositiveInt

A size.

ttl.Shape = ttl.Size | Tuple[ttl.Size, ...]

A shape type. ttl.Size for 1D and tuple of ttl.Size otherwise.

ttl.grid_size(dims: ttl.Size) -> ttl.Shape

Return grid size in specified dimensionality. Returns ttl.Size for dims = 1 and a tuple of ttl.Size for other values of dims.

Example

# for (8, 8) single chip or SPMD grid gets x_size = 64
x_size = ttl.grid_size(dims = 1)

# for (8, 8, 8) multi-chip grid gets x_size = 8, y_size = 64
x_size, y_size = ttl.grid_size(dims = 2)

# for (8, 8) single-chip or SPMD grid gets x_size = 8, y_size = 8, z_size = 1
x_size, y_size, z_size = ttl.grid_size(dims = 3)

Node function

The ttl.node function returns node coordinates of the current node. Node coordinates are zero based and contiguous, which corresponds to a logical indexing scheme. The function takes an argument that specifies how many dimensions to return. If requested dimensions are smaller than grid dimensions, the highest rank dimension is flattened. If requested dimensions are greater than grid dimensions, highest rank dimensions are padded with a value of zero. The ttl.node can be used inside an operation function as well as inside kernel functions.

Type alias/Function

Description

ttl.NaturalInt = Annotated[int, Ge(0)]

Non-negative integer. The metadata Ge(0), can be used by runtime type-checkers to enforce the integer constraints.

ttl.Index = ttl.NaturalInt

An index, assumes non-negative indexes.

ttl.NodeCoord = ttl.Index | Tuple[ttl.Index, ...]

Node coordinates. ttl.Index for 1D and tuple of ttl.Index otherwise.

ttl.node(dims: ttl.Index) -> ttl.NodeCoord

Return node coordinates in specified dimensionality. Returns ttl.Index for dims = 1 and a tuple of ttl.Index for other values of dims.

Example

# for (8, 8) single chip or SPMD grid gets x = [0, 64)
x = ttl.node(dims = 1)

# for (8, 8, 8) multi-chip grid gets x = [0, 8), y = [0, 64)
x, y = ttl.node(dims = 2)

# for (8, 8) single-chip or SPMD grid gets x = [0, 8), y = [0, 8), z = 0
x, y, z = ttl.node(dims = 3)

Dataflow buffer

A dataflow buffer is a communication primitive for synchronizing the passing of data between kernel functions running on the same node. A dataflow buffer is created with the ttl.make_dataflow_buffer_like function by passing TT-NN tensor, shape and buffer factor.

The shape is expressed as a tuple with outermost dimension first and innermost dimension last. For ttl.math functions that take dimension indexes, the outermost dimension is indexed as 0, next to outermost as 1. It is possible to use negative dimension indexes to index from innermost dimension. This way the innermost dimension is indexed as -1, next to innermost as -2. The TT-NN tensor determines basic properties (likeness) such as data type and shape unit. The shape unit affects two innermost dimensions and is a whole tile (32 by 32 scalars) if the tensor has a tiled layout. For example, if a TT-NN tensor is of tiled layout and has shape of (2, 128, 32), the corresponding block that fits this entire tensor will have shape of (2, 4, 1). If tensor has a row-major layout the shape unit is an scalar. For the TT-NN tensor in the above example the corresponding block that fits this entire tensor will have shape of (2, 128, 32).

Shape determines the shape of a block returned by one of the acquisition functions. The size of a block in L1 memory is determined by shape, shape unit and data type. For example, for a block with shape (2, 4, 1), shape unit of a tile and BF16 data type, its size in L1 will be 2 * 4 * 32 * 1 * 32 * 2 = 16384 bytes. The buffer factor determines the total size of L1 memory allocated for a dataflow buffer. This size as a product of a block size and buffer factor. For the most common case buffer factor defaults to 2 to support double buffering. With double buffered dataflow buffer one kernel can write to a block while another is reading from a block thus enabling enabling the pipelining. For the example above, this means there will be a total of 32768 bytes of L1 memory allocated for the dataflow buffer.

There are two acquisition functions on a dataflow buffer object: wait and reserve. A dataflow buffer is constructed in the scope of an operation function but its object functions can only be used inside of kernel functions. Acquisition functions can be used with Python with statement, which will automatically release acquired blocks at the end of the with scope. Alternatively, if acquisition functions are used without the with the user must explicitly call a corresponding release function on the acquired block: pop for wait and push for reserve.

Example

x_dfb = ttl.make_dataflow_buffer_like(x,
    shape = (2, 2),
    buffer_factor = 2)

@ttl.datamovement()
def some_read():
    # acquire x_blk from x_dfb
    with x_dfb.reserve() as x_blk:

        # produce data into x_blk
        # ...

        # release x_blk implicitly by x_blk.push() at the end of the "with" scope

@ttl.compute()
def some_compute():
    # acquire x_blk from x_dfb
    x_blk = x_dfb.wait()

    # consume data in x_blk
    # ...

    x_blk.pop() # release x_blk explicitly

Type alias/Function

Description

ttl.make_dataflow_buffer_like(ttnn.Tensor: likeness_tensor, shape: ttl.Shape,   buffer_factor: ttl.Size) -> ttl.DataflowBuffer

Create a dataflow buffer by inheriting basic properties from likeness_tensor.

ttl.DataflowBuffer.reserve(self) -> ttl.Block

Reserve and return a block from a dataflow buffer. This function is blocking and will wait until a free block is available. A free block is typically used by a producer to write the data into.

ttl.Block.push(self)

Push a block to a dataflow buffer. This function is called by the producer to signal the consumer that a block filled with data is available. This function is non-blocking.

ttl.DataflowBuffer.wait(self) -> ttl.Block

Wait for and return a block from a dataflow buffer. This function is blocking and will wait until a block filled with data is available. A filled block is typically used by a consumer to read data from.

ttl.Block.pop(self)

Pop a block from a dataflow buffer. This function is called by the consumer to signal the producer that block is free and available. This function is non-blocking.

Block

A block represents memory acquired from a dataflow buffer. Block size is determined by the shape of a dataflow buffer and its memory is allocated when a dataflow buffer is created. Inside of a compute kernel a block can participate in a block expression with built-in Python operators and TT-Lang math functions as an operand. A block can also be a storage for the result of block expression by using store function. Inside of data movement kernels a block can participate in ttl.copy as a source or a destination.

Element-wise with broadcast example

# ---------------------
# Element-wise with broadcast with two outputs: Y = sqrt(A^2 + B^2), Z = sqrt(A^2 - B^2)
#
# Tensor   Torch shape  Shape in tiles
# A        N            NT
# B        1            1
# Y        N            NT
# Z        N            NT
#
# NT = N // TILE_SIZE

a_dfb = ttl.make_dataflow_buffer_like(A, shape = (1, ))
b_dfb = ttl.make_dataflow_buffer_like(B, shape = (1, ))
y_dfb = ttl.make_dataflow_buffer_like(Y, shape = (1, ))
z_dfb = ttl.make_dataflow_buffer_like(Z, shape = (1, ))

@ttl.datamovement()
def elwise_read():
    for nt in range(NT):

        # acquire a_blk and b_blk from a_dfb and b_dfb:

        with (
            a_dfb.reserve() as a_blk,
            b_dfb.reserve() as b_blk,
        ):
            # then copy:

            a_xf = ttl.copy(A[nt], a_blk)
            b_xf = ttl.copy(B[0], b_blk)

            a_xf.wait()
            b_xf.wait()

            # release a_blk and b_blk

@ttl.compute()
def elwise_compute():
    for _ in range(NT):

        # acquire a_blk, b_blk, y_blk and z_blk from a_dfb, b_dfb, y_dfb and z_dfb:

        with (
            a_dfb.wait() as a_blk,
            b_dfb.wait() as b_blk,
            y_dfb.reserve() as y_blk,
            z_dfb.reserve() as z_blk,
        ):
            # then compute y = sqrt(a^2 + b^2) and z = sqrt(a^2 - b^2):

            a_squared = a_blk ** 2
            b_squared = b_blk ** 2

            y = ttl.math.sqrt(a_squared + ttl.math.broadcast(b_squared, y_blk, dims=[0]))
            z = ttl.math.sqrt(a_squared - ttl.math.broadcast(b_squared, z_blk, dims=[0]))

            y_blk.store(y)
            z_blk.store(z)

            # release a_blk, b_blk and y_blk


@ttl.datamovement()
def elwise_write():
    for nt in range(NT):

        # acquire y_blk and z_blk from y_dfb and z_dfb:

        with (
            y_dfb.wait() as y_blk,
            z_dfb.wait() as z_blk,
        ):

            # then copy:

            y_xf = ttl.copy(y_blk, Y[nt])
            z_xf = ttl.copy(z_blk, Z[nt])
            y_xf.wait()
            z_xf.wait()

            # release y_blk and z_blk

Matmul example

# ---------------------
# Matmul with bias: Y = A @ B + C
#
# Tensor   Torch shape  Shape in tiles
# A        I, M, K      IT, MT, KT
# B        K, N         KT, NT
# C        M, N         MT, NT
# Y        I, M, N      IT, MT, NT
#
# IT = I // TILE_SIZE
# MT = M // TILE_SIZE
# NT = N // TILE_SIZE
# KT = K // TILE_SIZE

a_dfb = ttl.make_dataflow_buffer_like(A, shape = (1, 1, 1))
b_dfb = ttl.make_dataflow_buffer_like(B, shape = (1, 1))
c_dfb = ttl.make_dataflow_buffer_like(C, shape = (1, 1))
y_dfb = ttl.make_dataflow_buffer_like(Y, shape = (1, 1, 1))

@ttl.datamovement()
def matmul_read():
    for it in range(IT):
        for mt in range(MT):
            for nt in range(NT):

                # acquire c_blk from c_dfb:

                with c_dfb.reserve() as c_blk:

                    # then copy:

                    c_xf = ttl.copy(C[mt, nt], c_blk)
                    c_xf.wait()

                    # release c_blk

                for kt in range(KT):

                    # acquire a_blk and b_blk from a_dfb and b_dfb:

                    with (
                        a_dfb.reserve() as a_blk,
                        b_dfb.reserve() as b_blk,
                    ):
                        # then copy:

                        a_xf = ttl.copy(A[it, mt, kt], a_blk)
                        b_xf = ttl.copy(B[kt, nt], b_blk)

                        a_xf.wait()
                        b_xf.wait()

                        # release a_blk and b_blk

@ttl.compute()
def matmul_compute():
    for _ in range(IT):
        for _ in range(MT):
            for _ in range(NT):

                # acquire y_blk from y_dfb:

                with y_dfb.reserve() as y_blk:

                    # acquire c_blk from c_dfb:

                    y = ttl.math.fill(y_blk, 0)

                    for _ in range(KT):

                        # acquire a_blk and b_blk from a_dfb and b_dfb:

                        with (
                            a_dfb.wait() as a_blk,
                            b_dfb.wait() as b_blk,
                        ):

                            y += a_blk @ b_blk

                            # release a_blk and b_blk

                    with c_dfb.wait() as c_blk:

                        y = y + c_blk

                        # release c_blk

                    y_blk.store(y)

                    # release y_blk

@ttl.datamovement()
def matmul_write():
    for it in range(IT):
        for mt in range(MT):
            for nt in range(NT):

                # acquire y_blk from y_dfb:

                with y_dfb.wait() as y_blk:

                    # then copy:

                    y_xf = ttl.copy(y_blk, Y[it, mt, nt])
                    y_xf.wait()

                    # release y_blk

Function

Description

ttl.Block.store(self, expr: ttl.BlockExpr)

This function materializes the result of a block expression and stores it in the block. Block expression uses Python builtin math operators and ttl.math.xxx functions on block expression. This function is blocking so that block is safe to use immediately after the call.

ttl.BlockExpr.__pow__(self, exponent: ttl.NaturalInt) -> ttl.BlockExpr

Example of Python built-in operator. See full list in Appendix B. Block operators and math functions.

ttl.BlockExpr.__add__(self, other: ttl.BlockExpr) -> ttl.BlockExpr

ttl.BlockExpr.__iadd__(self, other: ttl.BlockExpr) -> ttl.BlockExpr

ttl.math.sqrt(expr: ttl.BlockExpr) -> ttl.BlockExpr

ttl.math.fill(value: float) -> ttl.BlockExpr

ttl.BlockExpr.__matmul__(self, other: ttl.BlockExpr) -> ttl.BlockExpr

ttl.Block diagram

Block states

Blocks have a life cycle that starts with acquisition by using dataflow buffer’s reserve or wait functions and ends with release by block’s push and pop functions correspondingly. During this life cycle there are restrictions on what operations and in what sequences a block can participate in. These restrictions are formalized by the table below, which summarizes the states, and the accompanying diagrams, which illustrate the legal transitions.

Block State

Description

MW

Must be Written: the block was reserved and contains garbage data and therefore must be written to.

MR

Must be Read: the block was waited on or written to and never read and therefore must be read from or pushed.

RW

Read-Write: the block was waited on or written to (MR) and then read from and therefore can be either read from more times or overwritten.

ROR(N)

Read Only while Reading: the block is being asynchronously read from by N ttl.copys.

NAW

No Access while Writing: the block is being asynchronously written to.

OS

Out of Scope: the block was pushed or popped.

Compute kernel reserve-push

Compute kernel wait-pop

Datamovement kernel reserve-push

Datamovement kernel wait-pop

Pipe

A pipe is a communication primitive for organizing the passing of data between data movement kernels on different nodes. A pipe is used as a source or a destination in the ttl.copy. The pipe is constructed with source node coordinate (src) and destination (dst), which is either a single node coordinate for unicast or node range for multicast. The node range uses a combination of dimension slices and values to describe a contiguous hypercube. The node range dimensions’ aspects will match the corresponding aspects returned by the grid_size function for the same number of dimensions.

Type alias/Function

Description

ttl.NodeRange = Tuple[ttl.Index | slice, ...]

A node range.

ttl.Pipe[DstT](src: ttl.NodeCoord, dst: DstT) -> ttl.Pipe[DstT]

Constructs pipe description to be used to construct pipe net. The dst argument is of DstT type, which can be either ttl.NodeCoord or ttl.NodeRange.

Pipe net

A pipe net is a communication primitive that groups pipes into a network. A pipe net is constructed from a list of pipes and encapsulates all necessary information to determine if a given node is source, destination or both and where and from which node or nodes the corresponding transfers will occur. Pipe net object has two functions: if_src and if_dst. Both functions have a single argument: condition body function.

Condition body function is invoked for each pipe in case of if_src if the current node is a source, and in case of if_dst if the current node is a destination. The condition body function has a single argument: a pipe identity that satisfies the condition. Condition body function can identify the source and the destination by its src and dst read-only properties correspondingly.

A pipe net is constructed in the scope of an operation function but can only be used with its if_src and if_dst functions inside of a data movement kernel function. The corresponding ttl.copy where a pipe is a source or a destination can be called only inside of a condition body function. Calls into if_src and if_dst can be nested within condition functions for different pipe nets.

Function

Description

ttl.PipeNet[DstT](pipes: List[ttl.Pipe[DstT]]) -> ttl.PipeNet[DstT]

Constructs pipe net.

ttl.PipeNet[DstT].if_src(self, cond_fun: Callable[[ttl.SrcPipeIdentity[DstT]], None])

Call condition function for each pipe in the pipe net that is a source.

ttl.PipeNet[DstT].if_dst(self, cond_fun: Callable[[ttl.DstPipeIdentity], None])

Call condition function for each pipe in the pipe net that is a destination.

@property ttl.SrcPipeIdentity[DstT].dst(self) -> DstT

Get destination node or node range for pipe in if_src.

@property ttl.DstPipeIdentity.src(self) -> ttl.NodeCoord

Get source node for pipe in if_dst.

ttl.PipeIdentity diagram

Gather example

# Grid:
#
# column
# x == 0
#   |
#   V
# (0, 0) (1, 0) (2, 0) (3, 0) <-- row y == 0
# (0, 1) (1, 1) (2, 1) (3, 1)
# (0, 2) (1, 2) (2, 2) (3, 2)
# (0, 3) (1, 3) (2, 3) (3, 3)

# ---------------------
# gather from row y to (0, y) with unicast

grid_x, grid_y = ttl.grid_size()

net = ttl.PipeNet([ttl.Pipe(
    src = (x, y),
    dst = (0, y)) for x in range(1, grid_x) for y in range(grid_y)])

# (1, 0) -> (0, 0) |             |
# (2, 0) -> (0, 0) | sequential  |
# (3, 0) -> (0, 0) |             |
# ...              |             | concurrent
#                                |
# (1, 1) -> (0, 1)               |
# ...                            |

@ttl.datamovement()
def dm():
    with dfb.reserve() as blk:

        def pipe_src(pipe):

            # write data into blk
            # ...

            # then copy blk to pipe:

            xf = ttl.copy(blk, pipe)
            xf.wait()

        def pipe_dst(pipe):

            # copy blk from pipe:

            xf = ttl.copy(pipe, blk)
            xf.wait()

            # then read data from blk
            # ...

        net.if_src(pipe_src)
        net.if_dst(pipe_dst)

Scatter example

# ---------------------
# scatter from (x, 0) to column x with multicast

grid_x, grid_y = ttl.grid_size()

net = ttl.PipeNet([ttl.Pipe(
    src = (x, 0),
    dst = (x, slice(1, grid_y))) for x in range(grid_x)])

# (0, 0) => (0, 1) (0, 2) (0, 3) ... |
# (1, 0) => (1, 1) (1, 2) (1, 3) ... | concurrent
# ...                                |

@ttl.datamovement()
def dm():
    with dfb.reserve() as blk:

        def pipe_src(pipe):

            # write data into blk
            # ...

            # then copy blk to pipe:

            xf = ttl.copy(blk, pipe)
            xf.wait()

        def pipe_dst(pipe):

            # copy blk from pipe:

            xf = ttl.copy(pipe, blk)
            xf.wait()

            # then read data from blk
            # ...

        net.if_src(pipe_src)
        net.if_dst(pipe_dst)

Scatter-gather example

# ---------------------
# scatter-gather column x with multicast/loopback

grid_x, grid_y = ttl.grid_size()

net = ttl.PipeNet([ttl.Pipe(
    src = (x, y),
    dst = (x, slice(0, grid_y))) for x in range(grid_x) for y in range(grid_y)])

# (0, 0) => (0, 0) (0, 1) (0, 2) ... |            |
# (0, 1) => (0, 0) (0, 1) (0, 2) ... | sequential |
# (0, 2) => (0, 0) (0, 1) (0, 2) ... |            |
# ...                                |            | concurrent
#                                                 |
# (1, 0) => (1, 0) (1, 1) (1, 2) ...              |
# ...                                             |

@ttl.datamovement()
def dm():
    with dfb.reserve() as blk:

        def pipe_src(pipe):

            # write data into blk
            # ...

            # then copy blk to pipe:

            xf = ttl.copy(blk, pipe)
            xf.wait()

        def pipe_dst(pipe):

            # copy blk from pipe:

            xf = ttl.copy(pipe, blk)
            xf.wait()

            # then read data from blk
            # ...

        net.if_src(pipe_src)
        net.if_dst(pipe_dst)

Forward to a +1 neighbor example

# ---------------------
# forward to a +1 neighbor in a column x

grid_x, grid_y = ttl.grid_size()

net = ttl.PipeNet([ttl.Pipe(
    src = (x, y),
    dst = (x, (y + 1) % grid_y)) for x in range(grid_x) for y in range(grid_y)])

# (0, 0) => (0, 1)  |
# (0, 1) => (0, 2)  |
# ...               |
# (0, 7)* => (0, 0) |
# ...               | concurrent
#                   |
# (1, 0) => (1, 1)  |
# ...               |
#
# * - assuming (8, 8) grid

@ttl.datamovement()
def dm():

    with (
        dfb_to_send.reserve() as blk_to_send,
        dfb_received.reserve() as blk_received,
    ):

        def pipe_src(pipe):

            # write data into blk_to_send
            # ...

            # then copy blk to blk_to_send:

            xf = ttl.copy(blk_to_send, pipe)
            xf.wait()

        def pipe_dst(pipe):

            # copy blk_received from pipe:

            xf = ttl.copy(pipe, blk_received)
            xf.wait()

            # then read data from blk_received
            # ...

        net.if_src(pipe_src)
        net.if_dst(pipe_dst)

Tensor slice

A tensor slice is a view into a TT-NN tensor defined in terms of a dimension slice or value for each of the tensor’s dimensions. A tensor slice can participate in ttl.copy as a source or a destination with the corresponding destination and source being a block. Tensor slice can only be used in the scope of a data movement kernel function.

Function

Description

ttnn.Tensor.__getitem__(self, *index: ttl.Index | slice) -> ttl.TensorSlice

Get a tensor slice from a TT-NN tensor.

Example

g = 2 # granularity
a_dfb = ttl.make_dataflow_buffer_like(A, shape = (g, 1))

row_tiles = A.shape[0] // ttl.TILE_SHAPE[0]
col_tiles = A.shape[1] // ttl.TILE_SHAPE[1]
cols_per_node = math.ceil(col_tiles / (grid_size(dims = 1)))

node_num = ttl.node(dims = 1)
start_ct = node_num * cols_per_node
end_ct = min(start_ct + cols_per_node, col_tiles)

@ttl.datamovement()
def dm():
    for ct in range(start_ct, end_ct):
        for rt in range(row_tiles // g):

            # acquire a_blk from a_dfb:

            with a_dfb.reserve() as a_blk:

                # then copy from a tensor slice of matching shape:

                row_slice = slice(rt * g, (rt + 1) * g) # explicit row slice
                a_xf = ttl.copy(
                    A[row_slice, ct:ct + 1], # in-line col slice
                    a_blk)
                a_xf.wait()

Copy

The ttl.copy function expresses a variety of data movements that always have two arguments: source and destination. ttl.copy returns a transfer handle object. A transfer handle has a wait function that serves as a barrier. When the wait returns the transfer is complete and data in the destination is safe to use. The ttl.copy can only be used inside of a data movement kernel function.

Group transfer

When ttl.copy function is called multiple times, instead of waiting on each transfer handle, it is possible to group handles and wait on all handles at once. This is done by instantiating ttl.GroupTransfer object and then adding handles with its add function. Once all handles are added wait_all function is called to wait for all transfers to complete.

Example

# ---------------------
# Nearest Neighbor Upsample
#
# Tensor              Torch shape
# input_images        N, HI, WI, C
# output_images       N, HO, WO, C
#
# HO = HI * scale_factor[0]
# WO = WI * scale_factor[1]

io_dfb = ttl.make_dataflow_buffer_like(
    input_images, shape=(C,), buffer_factor=2
)

@ttl.datamovement()
def reader():
    for n in range(N):
        for hi in range(HI):
            for wi in range(WI):
                with io_dfb.reserve() as io_blk:

                    # Copy input pixel channels

                    xf = ttl.copy(input_t[n, hi, wi, :], io_blk)

                    xf.wait()

@ttl.datamovement()
def writer():
    for n in range(N):
        for hi in range(HI):
            for wi in range(WI):
                with io_dfb.wait() as io_blk:
                    gxf = ttl.GroupTransfer()

                    for h_sf in range(scale_factor[0]):
                        for w_sf in range(scale_factor[1]):

                            # Copy output pixel channels

                            xf = ttl.copy(io_blk, output[n, hi * scale_factor[0] + h_sf, wi * scale_factor[1] + w_sf, :])

                            # Add transfer handle to a group

                            gxf.add(xf)

                    # Wait for all transfers to complete

                    gxf.wait_all()

Function

Description

ttl.copy(src: ttl.Block, dst: ttl.TensorSlice) -> ttl.Transfer

ttl.copy(src: ttl.TensorSlice, dst: ttl.Block) -> ttl.Transfer

ttl.copy(src: ttl.Block, dst: ttl.PipeIdentity) -> ttl.Transfer

ttl.copy(src: ttl.PipeIdentity, dst: ttl.Block) -> ttl.Transfer | Copy data between a block, a tensor slice, or a pipe. This function is non-blocking. The compiler statically checks if the shape of block and tensor slice are compatible and if the shape of block sent to a pipe is compatible with the shape of block received from the same pipe. When a pipe is used as a destination there must be a corresponding ttl.copy where the same pipe is used as source. Furthermore, ttl.copy with pipe must be guarded by pipe net’s if_src and is_dst where this pipe is destination and source correspondingly. | | ttl.Transfer.wait() | Wait for data transfer to complete. Transfer handle cannot be used after this function is called. This function is blocking. | | ttl.GroupTransfer.add(xf: ttl.Transfer) | Add transfer handle to a group. This function cannot be called after ttl.GroupTransfer.wait_all was called. | | ttl.GroupTransfer.wait_all() | Wait for all data transfers in group to complete. Group transfer cannot be used after this function is called. This function is blocking. |

Semaphore

A semaphore is a communication primitive for general synchronization between data movement kernels on different nodes. Each semaphore has an associated 32-bit unsigned integer semaphore value for each node. This value can be changed (set or incremented) by a data movement kernel on the local or a remote node. When changing semaphore value remotely a single node coordinate for unicast change or a node range for multicast change is specified. Only setting the semaphore value is supported as a multicast change. A data movement kernel can wait on a semaphore until its value satisfies a condition. It is possible to specify either a condition with exact value or a condition with minimum value. Only local data movement kernels can wait on a semaphore.

ttl.Semaphore class is constructed with its initial value that defaults to zero. A ttl.Semaphore instance can be constructed in a operation function scope. A ttl.Semaphore instance provides wait_eq, wait_ge and set functions for managing local semaphore value. To change a remote semaphore value an instance of ttl.UnicastRemoteSemaphore or ttl.MulticastRemoteSemaphore is obtained by calling get_remote and get_remote_multicast functions correspondingly. The ttl.UnicastRemoteSemaphore supports inc and set while ttl.MulticastRemoteSemaphore supports only set. Functions that change the value or wait on condition can be used only in the scope of a data movement kernel function. Functions that obtain remote semaphores can be used in scopes of both operation and data movement kernel functions.

One-to-many barrier example

node_num = ttl.node(dims = 1)
my_barrier = ttl.Semaphore()
all_barrier = my_barrier.get_remote_multicast()

@ttl.datamovement()
def dm():
    if node_num == 0:

        # do something on node 0 while non-0 nodes wait...

        all_barrier.set(1)
    else:
        my_barrier.wait_eq(1)

        # node 0 is done

Many-to-one barrier example

node_num = ttl.node(dims = 1)
my_barrier = ttl.Semaphore()
node_0_barrier = my_barrier.get_remote((0, 0))
non_0_node_count = grid_size(dims = 1) - 1

@ttl.datamovement()
def dm():
    if node_num != 0:

        # do something on non-0 nodes while node 0 waits...

        node_0_barrier.inc(1)
    else:
        my_barrier.wait_eq(non_0_node_count)

        # non-0 nodes are done

Function

Description

ttl.Count = ttl.NaturalInt

A type for semaphore value.

ttl.Semaphore.wait_eq(self, value: ttl.Count)

Wait until the local semaphore value is equal to specified value. This function is blocking. Can be used only in the scope of a data movement kernel function.

ttl.Semaphore.wait_ge(self, value: ttl.Count)

Wait until the local semaphore value is greater or equal to specified value. This function is blocking. Can be used only in the scope of a data movement kernel function.

ttl.Semaphore.set(self, value: ttl.Count)

Set the local semaphore value to specified value. This function is non-blocking. Can be used only in the scope of a data movement kernel function.

ttl.Semaphore.get_remote(self, ttl.NodeCoord: node) -> ttl.UnicastRemoteSemaphore

Get remote unicast semaphore for specified node coordinate. Returns an instance of ttl.UnicastRemoteSemaphore. Can be used in both operation and kernel function scopes.

ttl.Semaphore.get_remote_multicast(self, ttl.NodeRange: node_range) -> ttl.MulticastRemoteSemaphore

Get remote multicast semaphore for specified node range. When called with no arguments returns remote multicast semaphore for the entire grid. Returns an instance of ttl.MulticastRemoteSemaphore. Can be used in both operation and kernel function scopes.

ttl.UnicastRemoteSemaphore.set(self, value: ttl.Count)

Set remote unicast semaphore value to specified value. This function is non-blocking. Can be used only in the scope of a data movement kernel function.

ttl.UnicastRemoteSemaphore.inc(self, value: ttl.Count)

Increment remote unicast semaphore value by specified value. This function is non-blocking. Can be used only in the scope of a data movement kernel function.

ttl.MulticastRemoteSemaphore.set(self, value: ttl.Count)

Set remote multicast semaphore value to specified value. This function is non-blocking. Can be used only in the scope of a data movement kernel function.

Performance and debugging

TT-Lang provides a range for facilities to aid performance analisys and debugging. Generally, the description of these tools is outside of the scope of this specification with the exception of language extensions that are needed to support them.

Profiling signpost

Profiling signpost is a language construct that allows the user to specify a block of code that will be measured for performance during the program execution. This is achieved by using Python with statement in conjunction with ttl.signpost function. This function takes a string argument for a signpost name. This way the signpost will be identified in the profiling tool’s user interface.

Example

@ttl.datamovement()
def matmul_read():
    for it in range(IT):
        for mt in range(MT):
            for nt in range(NT):

                # Measure the entire iteration

                with ttl.signpost("i_m_n iteration"):

                    # Measure from reserve to push

                    with ttl.signpost("push c"):
                        with c_dfb.reserve() as c_blk:

                            # Measure only copy

                            with ttl.signpost("read c"):
                                c_xf = ttl.copy(C[mt, nt], c_blk)
                                c_xf.wait()

                    for kt in range(KT):
                        with ttl.signpost("push a and b"):

                            # Measure from reserve to push

                            with (
                                a_dfb.reserve() as a_blk,
                                b_dfb.reserve() as b_blk,
                            ):

                                # Measure only copy

                                with ttl.signpost("read a and b"):
                                    a_xf = ttl.copy(A[it, mt, kt], a_blk)
                                    b_xf = ttl.copy(B[kt, nt], b_blk)

                                    a_xf.wait()
                                    b_xf.wait()

Function

Description

ttl.signpost(str: name)

Declare as signpost. Can be used only with the with statement.

Debug printing

TT-Lang includes ability to print information to the standard output for debugging purpose. This is achieved by using the standard Python print function. In TT-Lang this function can be used with string constants, scalar variables, such as loop indexes or calculated slice bounds, as well as with TT-Lang specific objects, such as tensors and blocks. When print is used with TT-Lang objects there are additional attribute arguments, which enabling better control of the output content. Beacause of this, print is limited to only one TT-Lang object to be printed in conjunction any number of string and scalar variables.

Example

@ttl.datamovement()
def matmul_read():
    # Print first two pages of C

    print("C: ", C, num_pages=2)

    # Print first page of A and B

    print("A: ", A)
    print("B: ", B)

    for it in range(IT):
        for mt in range(MT):
            for nt in range(NT):
                with c_dfb.reserve() as c_blk:

                    # Print state of c_dfb dataflow buffer after reserve

                    print("c_dfb after reserve: ", c_dfb)

                    # Print iteration state and the content of c_blk block

                    print("it=", it, " mt=", mt, "nt=", nt, " c_blk: ", c_blk)

                    c_xf = ttl.copy(C[mt, nt], c_blk)
                    c_xf.wait()

                # Print state of c_dfb dataflow buffer after push

                print("c_dfb after push: ", c_dfb)

                for kt in range(KT):
                    with (
                        a_dfb.reserve() as a_blk,
                        b_dfb.reserve() as b_blk,
                    ):
                        # Print iteration state

                        print("kt=", kt)

                        # Print the content of a_blk block

                        print("a_blk:")
                        print(a_blk)

                        # Print the content of b_blk block

                        print("b_blk:")
                        print(b_blk)

                        a_xf = ttl.copy(A[it, mt, kt], a_blk)
                        b_xf = ttl.copy(B[kt, nt], b_blk)

                        a_xf.wait()
                        b_xf.wait()

Type

print function behavior

ttnn.Tensor

Print num_pages pages of a tensor. The num_pages attribute defaults to 1. For example, print(bias, num_pages=4).

ttl.Block

Print the content of a block. For example, print(bias_blk).

ttl.DataflowBuffer

Print the state of a dataflow buffer, which includes metadata such as size, page_size etc, as well as current value of its pointers: rd_ptr, wr_ptr and wr_tile_ptr. For example, print(bias_dfb).

Appendix A. Glossary

Term

Description

Domain specific language (DSL)

A language based on a constrained subset of the host language, Python in the case of TT-Lang.

Operation function

A Python function that encapsulates an operation written in TT-Lang, which can be used as TT-NN operation.

Kernel function

A Python function defined inside of the scope of an operation function that encapsulates a particular kernel behavior.

Data movement kernel function

A Python function that encapsulates data movement kernel behavior.

Compute kernel function

A Python function that encapsulates compute kernel behavior.

TT-NN tensor

Tensor representation in TT-NN environment. Encapsulates key meta information such as shape, data type, layout, storage and memory configuration.

Node

A minimal unit capable of executing a TT-Lang program.

Grid

A multidimensional space of nodes. A single chip is represented by a 2D grid. A multichip system is represented by 3D and higher dimensional grids.

Node coordinates

Coordinates of a given node within a grid. Each dimension is zero based and contiguous, which corresponds to logical indexing.

Dataflow buffer

A communication primitive for synchronizing the passing of data between kernels on the same node. Maintains memory space that is written by a producer and read by a consumer as well as synchronization mechanism necessary to communicate between producer and consumer to avoid data races.

Dataflow buffer’s shape

A shape of a block of memory acquired from a dataflow buffer to be either written by the producer or read by the consumer.

Dataflow buffer’s shape unit

A unit in which dataflow buffer shape is expressed. When a dataflow buffer is created in likeness of tiled TT-NN Tensor the unit is a tile. If it is created in likeness of row-major TT-NN the unit is a scalar.

Dataflow buffer’s buffer factor

A buffer factor determines how many block sized pages are allocated by the dataflow buffer. In the most case it is 2 pages to allow double buffering so that both consumer and producer can make progress by having one acquired block each to work with.

Dataflow buffer’s acquisition function

A blocking function that keeps a kernel waiting until a block becomes available in the dataflow buffer.

Dataflow buffer’s release function

A non-blocking function that releases a block back to the dataflow buffer to make it available to other kernels.

Block

A block of memory acquired from a dataflow buffer. In a compute kernel a block can participate in an expression as input, and also be used to store the expression’s result. In a data movement kernel a block can participate in copy operation as a source or destination.

Block expression

A block expression is a Python expression using built-in Python operators as well as TT-Lang math functions where operands are either blocks or block expressions.

Pipe

A pipe is a communication primitive for organizing the passing of data between data movement kernels on different nodes.

Pipe net

A pipe net is a communication primitive that groups pipes into a network. While a single pipe is capable of representing the passing of data from a single node, a network of pipes generalizes to a data passing pattern over the entire grid. A pipe net is constructed from the list of pipes, which is typically created by Python list comprehension over one or more aspects of a grid.

Pipe net’s condition body function

A Python function passed to be executed conditionally if the current node is a source, a destination, or both in the given pipe net. A condition function can be called multiple times sequentially if the current node participates in multiple pipes.

Tensor slice

A Python slice expression used with TT-NN tensor to specify a view to be used as a source or a destination in a copy operation.

Transfer handle

A handle to an asynchronous copy operation. A transfer handle is used as a barrier to ensure that operation is finished and the corresponding source or destination block is safe to use.

Semaphore

A communication primitive for general synchronization between data movement kernels on different nodes.

Semaphore value

A 32-bit unsigned integer value associated with a semaphore on each node. This value can be set or incremented by a data movement kernel on the local or a remote node.

Appendix B. Block operators and math functions

Binary operators and math functions

Function

Description

ttl.BlockExpr.__add__(self, other: ttl.BlockExpr) -> ttl.BlockExpr

Add two blocks element-wise. Example: a + b.

ttl.BlockExpr.__sub__(self, other: ttl.BlockExpr) -> ttl.BlockExpr

Two blocks subtracted second from first element-wise. Example: a - b.

ttl.BlockExpr.__mul__(self, other: ttl.BlockExpr) -> ttl.BlockExpr

Multiply two blocks element-wise. Example: a * b.

ttl.BlockExpr.__truediv__(self, other: ttl.BlockExpr) -> ttl.BlockExpr

Two blocks divided first by second element-wise. Example: a / b.

ttl.BlockExpr.__matmul__(self, other: ttl.BlockExpr) -> ttl.BlockExpr

Dot product of two blocks. If a has shape [M, K] and b has shape [K, N] then the result will have shape [M, N]. Example: a @ b.

ttl.math.max(a: ttl.BlockExpr, b: ttl.BlockExpr) -> ttl.BlockExpr

Element-wise maximum

ttl.math.min(a: ttl.BlockExpr, b: ttl.BlockExpr) -> ttl.BlockExpr

Element-wise minimum

In-place operators

| Function | Description | | ttl.BlockExpr.__iadd__(self, other: ttl.BlockExpr) -> ttl.BlockExpr | Add two blocks element-wise and replace first one with the result. Example: a += b. |

Basic unary math functions

Function

Description

ttl.BlockExpr.__abs__(self) -> ttl.BlockExpr

ttl.math.abs(expr: ttl.BlockExpr) -> ttl.BlockExpr

Absolute value. Example: abs(a), ttl.math.abs(a).

ttl.BlockExpr.__neg__(self) -> ttl.BlockExpr

ttl.math.neg(expr: ttl.BlockExpr) -> ttl.BlockExpr

Negation. Example: -a, ttl.math.neg(a).

ttl.BlockExpr.__pow__(self, exponent: ttl.NaturalInt) -> ttl.BlockExpr

ttl.math.pow(expr: ttl.BlockExpr, exponent: ttl.NaturalInt) -> ttl.BlockExpr

Power with scalar unsigned integer exponent. Example; a ** 2, ttl.math.pow(a, 2).

ttl.math.exp(expr: ttl.BlockExpr) -> ttl.BlockExpr

Natural base exponential (e^x)

ttl.math.exp2(expr: ttl.BlockExpr) -> ttl.BlockExpr

Base 2 exponential (2^x)

ttl.math.expm1(expr: ttl.BlockExpr) -> ttl.BlockExpr

Natural base exponential minus one (ttl.math.exp(x) - 1)

ttl.math.log(expr: ttl.BlockExpr) -> ttl.BlockExpr

Natural logarithm

ttl.math.logp1(expr: ttl.BlockExpr) -> ttl.BlockExpr

Natural logarithm of value plus 1 (ttl.math.log(x + 1))

ttl.math.sqrt(expr: ttl.BlockExpr) -> ttl.BlockExpr

Square root

ttl.math.square(expr: ttl.BlockExpr) -> ttl.BlockExpr

Square

ttl.math.rsqrt(expr: ttl.BlockExpr) -> ttl.BlockExpr

Reciprocal square root (1 / ttl.math.sqrt(x))

ttl.math.recip(expr: ttl.BlockExpr) -> ttl.BlockExpr

Reciprocal (1 / x)

ttl.math.rsub(a: ttl.BlockExpr, b: ttl.NaturalInt) -> ttl.BlockExpr

Subtract a from b where b is scalar unsigned integer (b - a)

Trigonometric unary math functions

Function

Description

ttl.math.tan(expr: ttl.BlockExpr) -> ttl.BlockExpr

Tangent

ttl.math.tanh(expr: ttl.BlockExpr) -> ttl.BlockExpr

Hyperbolic tangent

ttl.math.atan(expr: ttl.BlockExpr) -> ttl.BlockExpr

Arctangent

ttl.math.atanh(expr: ttl.BlockExpr) -> ttl.BlockExpr

Hyperbolic arctangent

ttl.math.sin(expr: ttl.BlockExpr) -> ttl.BlockExpr

Sine

ttl.math.asin(expr: ttl.BlockExpr) -> ttl.BlockExpr

Arcsine

ttl.math.asinh(expr: ttl.BlockExpr) -> ttl.BlockExpr

Hyperbolic arcsine

ttl.math.cos(expr: ttl.BlockExpr) -> ttl.BlockExpr

Cosine

ttl.math.acos(expr: ttl.BlockExpr) -> ttl.BlockExpr

Arccosine

ttl.math.acosh(expr: ttl.BlockExpr) -> ttl.BlockExpr

Hyperbolic arccosine

Activation functions

Function

Description

ttl.math.relu(expr: ttl.BlockExpr) -> ttl.BlockExpr

ReLU

ttl.math.relu_max(expr: ttl.BlockExpr, upper_limit: ttl.NaturalInt) -> ttl.BlockExpr

ReLU with upper limit (ttl.math.relu(ttl.math.min(x, upper_limit))))

ttl.math.relu_min(expr: ttl.BlockExpr, lower_limit: ttl.NaturalInt) -> ttl.BlockExpr

ReLU with lower limit (ttl.math.relu(ttl.math.max(x, lower_limit))))

ttl.math.leaky_relu(expr: ttl.BlockExpr, slope: ttl.NaturalInt) -> ttl.BlockExpr

Leaky ReLU

ttl.math.elu(expr: ttl.BlockExpr, slope: ttl.NaturalInt) -> ttl.BlockExpr

ELU

ttl.math.gelu(expr: ttl.BlockExpr) -> ttl.BlockExpr

GELU

ttl.math.sigmoid(expr: ttl.BlockExpr) -> ttl.BlockExpr

Sigmoid

ttl.math.celu(expr: ttl.BlockExpr, alpha: ttl.NaturalInt, alpha_recip: ttl.NaturalInt) -> ttl.BlockExpr

CELU

ttl.math.silu(expr: ttl.BlockExpr) -> ttl.BlockExpr

SiLU (Swish)

ttl.math.prelu(expr: ttl.BlockExpr, alpha: ttl.NaturalInt) -> ttl.BlockExpr

PReLU

ttl.math.softplus(expr: ttl.BlockExpr, beta: ttl.NaturalInt, beta_reciprocal: ttl.NaturalInt, threshold: ttl.NaturalInt) -> ttl.BlockExpr

Softplus

ttl.math.softsign(expr: ttl.BlockExpr) -> ttl.BlockExpr

Softsign

ttl.math.hardsigmoid(expr: ttl.BlockExpr) -> ttl.BlockExpr

Hardsigmoid

ttl.math.hardtanh(expr: ttl.BlockExpr, min: ttl.NaturalInt, max: ttl.NaturalInt) -> ttl.BlockExpr

Hardtanh

ttl.math.selu(expr: ttl.BlockExpr, scale: ttl.NaturalInt, alpha: ttl.NaturalInt) -> ttl.BlockExpr

SELU

Reduction, broadcast and transpose functions

Function

Description

ttl.math.reduce_sum(expr: ttl.BlockExpr, scaler: ttl.BlockExpr, dims: List[int]) -> ttl.BlockExpr

Scaled sum reduction over specified dimensions.

Example for reduction over dimension -1 (innermost): y.store(ttl.math.reduce_sum(a, s, dims=[-1])). Here if a has shape of (N, M) then y must have shape of (N, 1), and if a has shape of (I, N, M) then y must have shape of (I, N, 1), and so on.

Example for reduction over dimension 1 (next to outermost): y.store(ttl.math.reduce_max(a, s, dims=[1])). Here if a has shape of (N, M) then y must have shape of (N, 1), and if a has shape of (I, N, M) then y must have shape of (I, 1, M), and so on.

Example for reduction over two innermost dimensions: y.store(ttl.math.reduce_sum(a, s, dims=[-1, -2])). Here if a has shape of (N, M) then y must have shape of (1, 1), and if a has shape of (I, N, M) then y must have shape of (I, 1, 1), and so on. | | ttl.math.reduce_max(expr: ttl.BlockExpr, scaler: ttl.BlockExpr, dims: List[int]) -> ttl.BlockExpr | Scaled maximum reduction over specified dimensions. See examples for ttl.math.reduce_sum. | | ttl.math.broadcast(expr: ttl.BlockExpr, out_blk: ttl.Block, dims: List[int]) -> ttl.BlockExpr | Broadcast a block over specified dimensions. Produces block with shape expanded to be compatible with out_blk[1].

Example for broadcast over dimension -1 (innermost): y.store(ttl.math.broadcast(a, y, dims=[-1])). Here the store is the outer expression and therefore if y has shape of (N, M) then a must have shape of (N, 1), and if y has shape of (I, N, M) then a must have shape of (I, N, 1), and so on.

Example for broadcast over dimension 1 (next to outermost): y.store(b * ttl.math.broadcast(a, y, dims=[1])). Here the * is the outer expression and therefore if b has shape of (N, M) then a must have shape of (N, 1), and if b has shape of (I, N, M) then a must have shape of (I, 1, M), and so on.

Example for broadcast over two innermost dimensions: y.store(b + ttl.math.broadcast(a, y, dims=[-1, -2])). Here the + is the outer expression, but because the broadcast is on dims=[-1, -2] if b has shape of (N, M) then a must have shape of (1, 1), and if b has shape of (I, N, M) then a must have shape of (I, 1, 1), and so on. | | ttl.math.transpose(expr: ttl.BlockExpr) -> ttl.BlockExpr | Transpose a block. For argument block of shape (M, N) produces resulting block with shape (N, M). Supported only for 2-dimensional blocks. |

Rounding functions

Function

Description

ttl.math.frac(expr: ttl.BlockExpr) -> ttl.BlockExpr

Fractional portion

ttl.math.trunc(expr: ttl.BlockExpr) -> ttl.BlockExpr

Truncated integer portion

ttl.math.round(expr: ttl.BlockExpr, decimals: ttl.NaturalInt) -> ttl.BlockExpr

Rounds to the number of decimal places specified in decimals

ttl.math.floor(expr: ttl.BlockExpr) -> ttl.BlockExpr

Floor

ttl.math.ceil(expr: ttl.BlockExpr) -> ttl.BlockExpr

Ceil

ttl.math.clamp(expr: ttl.BlockExpr, min: ttl.NaturalInt, max: ttl.NaturalInt) -> ttl.BlockExpr

Clamp to specified min and max

ttl.math.threshold(expr: ttl.BlockExpr, threshold: ttl.NaturalInt, value: ttl.NaturalInt) -> ttl.BlockExpr

For all values greater than specified threshold replace with specified value

ttl.math.sign(expr: ttl.BlockExpr) -> ttl.BlockExpr

Replace positive element with 1, negative elements with -1 and leave zeroes as zero.

ttl.math.signbit(expr: ttl.BlockExpr) -> ttl.BlockExpr

Replace positive and positive zero elements with 1 and the rest with 0

Fill, mask and where functions

Function

Description

ttl.math.fill(out_blk: ttl.Block, value: float) -> ttl.BlockExpr

Fill a block with shape expanded to be compatible with out_blk[1] with specified value.

ttl.math.mask(expr: ttl.BlockExpr, mask: ttl.BlockExpr) -> ttl.BlockExpr

Mask a block with specified mask by replacing masked (corresponding mask element equals to 1) elements with 0.

ttl.math.mask_posinf(expr: ttl.BlockExpr, mask: ttl.BlockExpr) -> ttl.BlockExpr

Mask a block with specified mask by replacing masked (corresponding mask element equals to 1) elements with positive infinity.

ttl.math.where(condition: ttl.BlockExpr, true_value: ttl.BlockExpr, false_value: ttl.BlockExpr) -> ttl.BlockExpr

For each element in specified condition block return the corresponding element from true_value if true (condition element equals to 1) or the element from false_value if false (condition element equals to 0)

Appendix C. Naming guidelines

Object

Guideline

Tensor

Snake case. Example attention_mask, bias2.

Dataflow buffer

Snake case with dfb suffix. Example attention_mask_dfb, bias_dfb2

Block

Snake case with blk suffix. Example attention_mask_blk, bias_blk2

Transfer handle

Snake case with xf suffix. Example attention_mask_xf, bias_xf2

Pipe net

Snake case with net suffix. Example mcast_attention_mask_net, bias_net2

Appendix D. Functionality matrix

Functionality

Simulator

Compiler

2D grid ttl.grid_size and ttl.node with dims=2

0.1.7

0.1.7

2D grid ttl.grid_size and ttl.node with any dims

0.1.7

N/S

4D grid ttl.grid_size and ttl.node

N/S

N/S

SPMD

N/A

N/S

ttl.make_dataflow_buffer_like with 2D+ shape

0.1.7

0.1.7

ttl.make_dataflow_buffer_like with any shape

0.1.7

N/S

ttl.make_dataflow_buffer_like for tilized tensors

0.1.7

0.1.7

ttl.make_dataflow_buffer_like for row-major tensors

N/S

N/S

ttl.Block.store

0.1.7

0.1.7

Overwriting and accumulation through summation (+=) for block expressions

0.1.7

N/S

ttl.copy and ttl.Transfer

0.1.7

0.1.7

ttl.GroupTransfer

N/S

N/S

ttl.Semaphore on 2D grid

N/S

N/S

ttl.Semaphore on 4D grid

N/S

N/S

ttl.PipeNet and ttl.Pipe on 2D grid

0.1.7

N/S

ttl.PipeNet and ttl.Pipe on 4D grid

N/S

N/S

ttl.signpost (ignored in simulator)

0.1.7

0.1.7

Debug printing with print

0.1.7

0.1.7

Built-in unary math operators: -, abs

0.1.7

0.1.7

Built-in binary math operators: +, -, *, /

0.1.7

0.1.7

Built-in binary math operators: @, %, //, ^,

0.1.7

N/S

ttl.math unary math functions: exp, log, sqrt, rsqrt, tanh, sigmoid, relu, floor, recip, sin, cos, tan

0.1.7

0.1.7

ttl.math unary math functions: exp2, expm1, logp1, square, tan, atan, atanh, asin, asinh, acos, acosh, gelu, silu, softsign, hardsigmoid, selu, ceil, frac, trunc, sign, signbit, rsub, relu_max, relu_min, leaky_relu, elu, celu, prelu, softplus, hardtanh, round, clamp, threshold, fill

0.1.7

N/S

ttl.math binary math functions: min, max

0.1.7

0.1.7

ttl.math binary math functions: mask, mask_posinf

0.1.7

N/S

ttl.math.where

0.1.7

N/S

ttl.math.broadcast (compiler requires target block as argument)

0.1.7

0.1.7

ttl.math.reduce_max

0.1.7

N/S

ttl.math.reduce_sum

0.1.7

N/S

ttl.math.transpose

0.1.7

N/S

  • N/S - Not Supported

  • N/A - Not Applicable

Appendix E. Platform limitations

Description

Wormhole

Blackhole

Tile size in scalars

32, 32

32, 32

Maximum single chip grid size (unharvested)

8, 9

13, 10

Size of L1 memory (KB)

1464

1464

Maximum number of dataflow buffers

32

32