Skip to content
Change the repository type filter

All

    Repositories list

    • server

      Public
      The Triton Inference Server provides an optimized cloud and edge inferencing solution.
      Python
      1.7k10k77487Updated Dec 8, 2025Dec 8, 2025
    • core

      Public
      The core library and APIs implementing the Triton Inference Server.
      C++
      121156020Updated Dec 8, 2025Dec 8, 2025
    • tensorrtllm_backend

      Public
      The Triton TensorRT-LLM Backend
      13291031723Updated Dec 8, 2025Dec 8, 2025
    • Triton backend that enables pre-process, post-processing and other logic to be implemented in Python.
      C++
      188661016Updated Dec 7, 2025Dec 7, 2025
    • model_analyzer

      Public
      Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Server models.
      Python
      8050001Updated Dec 6, 2025Dec 6, 2025
    • The Triton backend for the PyTorch TorchScript models.
      C++
      6216606Updated Dec 6, 2025Dec 6, 2025
    • openvino_backend

      Public
      OpenVINO backend for Triton.
      C++
      183461Updated Dec 5, 2025Dec 5, 2025
    • onnxruntime_backend

      Public
      The Triton backend for the ONNX Runtime.
      C++
      74169745Updated Dec 5, 2025Dec 5, 2025
    • fil_backend

      Public
      FIL backend for the Triton Inference Server
      Jupyter Notebook
      3883520Updated Dec 3, 2025Dec 3, 2025
    • common

      Public
      Common source, scripts and utilities shared across all Triton repositories.
      C++
      757807Updated Dec 3, 2025Dec 3, 2025
    • client

      Public
      Triton Python, C++ and Java client libraries, and GRPC-generated client examples for go, java and scala.
      Python
      2526665831Updated Nov 27, 2025Nov 27, 2025
    • Python
      3231909Updated Nov 26, 2025Nov 26, 2025
    • tensorflow_backend

      Public
      The Triton backend for TensorFlow.
      C++
      235502Updated Nov 22, 2025Nov 22, 2025
    • Triton CLI is an open source command line interface that enables users to create, deploy, and profile models served by the Triton Inference Server.
      Python
      57233Updated Nov 14, 2025Nov 14, 2025
    • tutorials

      Public
      This repository contains tutorials and examples for Triton Inference Server
      Python
      135812816Updated Nov 14, 2025Nov 14, 2025
    • third_party

      Public
      Third-party source packages that are modified for use in Triton.
      C
      64704Updated Nov 14, 2025Nov 14, 2025
    • The Triton backend for TensorRT.
      C++
      348001Updated Nov 14, 2025Nov 14, 2025
    • Simple Triton backend used for testing.
      C++
      5300Updated Nov 14, 2025Nov 14, 2025
    • An example Triton backend that demonstrates sending zero, one, or multiple responses for each request.
      C++
      8700Updated Nov 14, 2025Nov 14, 2025
    • TRITONCACHE implementation of a Redis cache
      C++
      41630Updated Nov 14, 2025Nov 14, 2025
    • Python
      391233318Updated Nov 14, 2025Nov 14, 2025
    • Implementation of a local in-memory cache for Triton Inference Server's TRITONCACHE API
      C++
      2610Updated Nov 14, 2025Nov 14, 2025
    • Example Triton backend that demonstrates most of the Triton Backend API.
      C++
      14700Updated Nov 14, 2025Nov 14, 2025
    • developer_tools

      Public
      C++
      92104Updated Nov 14, 2025Nov 14, 2025
    • The Triton repository agent that verifies model checksums.
      C++
      91100Updated Nov 14, 2025Nov 14, 2025
    • backend

      Public
      Common source, scripts and utilities for creating Triton backends.
      C++
      10136103Updated Nov 14, 2025Nov 14, 2025
    • dali_backend

      Public
      The Triton backend that allows running GPU-accelerated data pre-processing pipelines implemented in DALI's python API.
      C++
      33139236Updated Nov 13, 2025Nov 13, 2025
    • pytriton

      Public
      PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment in Python environments.
      Python
      55830110Updated Aug 13, 2025Aug 13, 2025
    • Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.
      Python
      2721530Updated Apr 22, 2025Apr 22, 2025
    • .github

      Public
      Community health files for NVIDIA Triton
      2200Updated Mar 27, 2025Mar 27, 2025