GitHub topics: triton-server
notAI-tech/fastDeploy
Deploy DL/ ML inference pipelines with minimal extra code.
Language: Python - Size: 15.7 MB - Last synced at: 1 day ago - Pushed at: 6 months ago - Stars: 98 - Forks: 17

Bobo-y/flexible-yolov5
More readable and flexible yolov5 with more backbone(gcn, resnet, shufflenet, moblienet, efficientnet, hrnet, swin-transformer, etc) and (cbam,dcn and so on), and tensorrt
Language: Python - Size: 12.4 MB - Last synced at: 24 days ago - Pushed at: 9 months ago - Stars: 673 - Forks: 118

xdilshod/yolov8-triton
Yolov8 onnx inference running on triton server
Language: Python - Size: 15.7 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

cnwangjie/triton-client-js
A Node.js client for the Triton Inference Server.
Language: JavaScript - Size: 171 KB - Last synced at: 3 days ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

levipereira/nvdsinfer_yolov7_efficient_nms
NvDsInferYolov7EfficientNMS for Gst-nvinferserver
Language: C++ - Size: 8.79 KB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 2 - Forks: 0

chiehpower/Setup-deeplearning-tools
Set up CI in DL/ cuda/ cudnn/ TensorRT/ onnx2trt/ onnxruntime/ onnxsim/ Pytorch/ Triton-Inference-Server/ Bazel/ Tesseract/ PaddleOCR/ NVIDIA-docker/ minIO/ Supervisord on AGX or PC from scratch.
Language: Python - Size: 4.7 MB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 43 - Forks: 6

kozistr/triton-grpc-proxy-rs
Proxy server for triton gRPC server that inferences embedding model in Rust
Language: Rust - Size: 108 KB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 16 - Forks: 2

levipereira/triton-server-yolo
This repository serves as an example of deploying the YOLO models on Triton Server for performance and testing purposes
Language: Shell - Size: 170 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 18 - Forks: 1

levipereira/triton-client-yolo
This repository utilizes the Triton Inference Server Client, which streamlines the complexity of model deployment.
Language: Python - Size: 889 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 2 - Forks: 0

levipereira/Docker-Yolov7-Nvidia-Kit
This repository provides an out-of-the-box deployment solution for creating an end-to-end procedure to train, deploy, and use Yolov7 models on Nvidia GPUs using Triton Server and Deepstream.
Language: Shell - Size: 23.4 KB - Last synced at: about 1 year ago - Pushed at: over 1 year ago - Stars: 2 - Forks: 0

levipereira/deepstream-yolo-triton-server-rtsp-out
The Purpose of this repository is to create a DeepStream/Triton-Server sample application that utilizes yolov7, yolov7-qat, yolov9 models to perform inference on video files or RTSP streams.
Language: Python - Size: 39.1 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

msclock/transformersplus 📦
Add Some plus extra features to transformers
Language: Python - Size: 258 KB - Last synced at: 24 days ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

Bobo-y/triton-backend-demo
This is a cutom triton-backend demo for process image (resize + norm)
Language: C++ - Size: 1.15 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 1 - Forks: 0

Bobo-y/django_ai_server_template
A template for delpoy AI server use django with tf_serving or triton_inference_serving
Language: Python - Size: 104 MB - Last synced at: 15 days ago - Pushed at: over 2 years ago - Stars: 6 - Forks: 1

omarabid59/yolov8-triton
Provides an ensemble model to deploy a YoloV8 ONNX model to Triton
Language: Python - Size: 11.7 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 11 - Forks: 4

Achiwilms/NVIDIA-Triton-Deployment-Quickstart
QuickStart Guide for Deploying a Basic ResNet Model on the Triton Inference Server
Language: Python - Size: 5.86 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

tamanna18/Triton-Inference-Server-Deployment-with-ONNX-Models
Triton Inference Server Deployment with ONNX Models
Size: 9.77 KB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 0 - Forks: 0
