Top Repositories
ONNX-TensorRT: TensorRT backend for ONNX
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools
Repositories
13ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.
No description provided.
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools
ONNX Runtime: cross-platform, high performance scoring engine for ML models
Repository for ONNX working group artifacts
Tensors and Dynamic neural networks in Python with strong GPU acceleration
ONNX-TensorRT: TensorRT backend for ONNX
Protocol Buffers - Google's data interchange format
A conda-smithy repository for onnx.
Utils for building and publishing ONNX wheels
Open Neural Network Exchange