Triton Plugins

Kernelize maintains and contributes to Triton and Triton Plugins. Below are example plugins we recommend as a starting point for new maintainers.

Triton CPU icon

Triton CPU

Triton backend that supports x86, ARM and RISC-V

Triton CPU is a Triton backend that generates optimized kernels for CPUs. Triton CPU serves as both as a Triton CPU backend and as a starting point for software developers to build their own backend.

Key Features:

  • Triton backend that generates optimal kernels for CPUs
  • Supportes existing Triton code
  • Leverages existing Triton knowledge and tools
  • Fork the open-source GitHub repo to save months on your Triton compiler development
Nexus icon

Nexus

Integrate Triton into your runtime for day-0 support of new models on the latest NPUs, CPUs and GPUs

Nexus integrates Triton kernels into an inference framework backend. It gathers hardware information and helps configure runtime frameworks to use the best kernels with no user configuration.

Key Features:

  • Extends existing inference platform runtimes
  • Optimizes layers on new target inference hardware
  • Works with vLLM and Ollama
  • Seamless integration with existing workflows

Platform Profiling Tools

TritonBench Dashboard icon

TritonBench Dashboard

Monitor and analyze performance metrics for Triton operators

TritonBench Dashboard displays performance information about key Triton operators on the target hardware.

Key Features:

  • Performance testing for key kernels
  • Nightly data for developers
  • Based on https://github.com/meta-pytorch/tritonbench

Hardware Tested Daily:

  • Nvidia H100
  • Nvidia B200
  • AMD MI350