Kernel requirements

Kernels on the Hub must fulfill the requirements outlined on this page. By ensuring kernels are compliant, they can be used on a wide range of Linux systems and Torch builds.

You can use kernel-builder to build compliant kernels.

Directory layout

A kernel repository on the Hub must contain a build directory. This directory contains build variants of a kernel in the form of directories following the template <framework><version>-cxx<abiver>-<cu><cudaver>-<arch>-<os>. For example build/torch26-cxx98-cu118-x86_64-linux.

Each variant directory must contain a single directory with the same name as the repository (replacing - by _). For instance, kernels in the kernels-community/activation repository have a directories like build/<variant>/activation. This directory must be a Python package with an __init__.py file.

Build variants

A kernel can be compliant for a specific compute framework (e.g. CUDA) or architecture (e.g. x86_64). For compliance with a compute framework and architecture combination, all the variants from the build variant list must be available for that combination.

Versioning

Kernels are versioned on the Hub using Git tags. Version tags must be of the form v<major>.<minor>.<patch>. Versions are used by locking to resolve the version constraints.

Native Python module

Kernels will typically contain a native Python module with precompiled compute kernels and bindings. This module must fulfill the requirements outlined in this section. For all operating systems, a kernel must not have dynamic library dependencies outside:

Linux

These requirement can be checked with the ABI checker (see below).

macOS

The ABI3 requirement can be checked with the ABI checker (see below).

ABI checker

The manylinux_2_28 and Python ABI 3.9 version requirements can be checked with kernel-abi-check:


$ cargo install kernel-abi-check
$ kernel-abi-check result/relu/_relu_e87e0ca_dirty.abi3.so
🐍 Checking for compatibility with manylinux_2_28 and Python ABI version 3.9
✅ No compatibility issues found

Torch extension

Torch native extension functions must be registered in torch.ops.<namespace>. Since we allow loading of multiple versions of a module in the same Python process, namespace must be unique for each version of a kernel. Failing to do so will create clashes when different versions of the same kernel are loaded. Two suggested ways of doing this are:

Note: we recommend against appending a version number or git tag. Version numbers are typically not bumped on each commit, so users might use two different commits that happen to have the same version number. Git tags are not stable, so they do not provide a good way of guaranteeing uniqueness of the namespace.

Layers

A kernel can provide layers in addition to kernel functions. A layer from the Hub can replace the forward method of an existing layer for a certain device type. This makes it possible to provide more performant kernels for existing layers. See the layers documentation for more information on how to use layers.

Writing layers

To make the extension of layers safe, the layers must fulfill the following requirements:

There are two exceptions to the no class variables rule:

  1. The has_backward variable can be used to indicate whether the layer has a backward pass implemented (True when absent).
  2. The can_torch_compile variable can be used to indicate whether the layer supports torch.compile (False when absent).

This is an example of a pure layer:

class SiluAndMul(nn.Module):
    # This layer does not implement backward.
    has_backward: bool = False

    def forward(self, x: torch.Tensor):
        d = x.shape[-1] // 2
        output_shape = x.shape[:-1] + (d,)
        out = torch.empty(output_shape, dtype=x.dtype, device=x.device)
        ops.silu_and_mul(out, x)
        return out

For some layers, the forward method has to use state from the adopting class. In these cases, we recommend to use type annotations to indicate what member variables are expected. For instance:

class LlamaRMSNorm(nn.Module):
    weight: torch.Tensor
    variance_epsilon: float

    def forward(self, hidden_states: torch.Tensor) -> torch.Tensor:
        return rms_norm_fn(
            hidden_states,
            self.weight,
            bias=None,
            residual=None,
            eps=self.variance_epsilon,
            dropout_p=0.0,
            prenorm=False,
            residual_in_fp32=False,
        )

This layer expects the adopting layer to have weight and variance_epsilon member variables and uses them in the forward method.

Exporting layers

To accommodate portable loading, layers must be defined in the main __init__.py file. For example:

from . import layers

__all__ = [
  # ...
  "layers"
  # ...
]

Python requirements

< > Update on GitHub