An open API service providing repository metadata for many open source software ecosystems.

GitHub topics: triton-server

notAI-tech/fastDeploy

Deploy DL/ ML inference pipelines with minimal extra code.

Language: Python - Size: 15.7 MB - Last synced at: 1 day ago - Pushed at: 6 months ago - Stars: 98 - Forks: 17

Bobo-y/flexible-yolov5

More readable and flexible yolov5 with more backbone(gcn, resnet, shufflenet, moblienet, efficientnet, hrnet, swin-transformer, etc) and (cbam,dcn and so on), and tensorrt

Language: Python - Size: 12.4 MB - Last synced at: 24 days ago - Pushed at: 9 months ago - Stars: 673 - Forks: 118

xdilshod/yolov8-triton

Yolov8 onnx inference running on triton server

Language: Python - Size: 15.7 MB - Last synced at: 5 months ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

cnwangjie/triton-client-js

A Node.js client for the Triton Inference Server.

Language: JavaScript - Size: 171 KB - Last synced at: 3 days ago - Pushed at: 5 months ago - Stars: 0 - Forks: 0

levipereira/nvdsinfer_yolov7_efficient_nms

NvDsInferYolov7EfficientNMS for Gst-nvinferserver

Language: C++ - Size: 8.79 KB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 2 - Forks: 0

chiehpower/Setup-deeplearning-tools

Set up CI in DL/ cuda/ cudnn/ TensorRT/ onnx2trt/ onnxruntime/ onnxsim/ Pytorch/ Triton-Inference-Server/ Bazel/ Tesseract/ PaddleOCR/ NVIDIA-docker/ minIO/ Supervisord on AGX or PC from scratch.

Language: Python - Size: 4.7 MB - Last synced at: about 2 months ago - Pushed at: over 1 year ago - Stars: 43 - Forks: 6

kozistr/triton-grpc-proxy-rs

Proxy server for triton gRPC server that inferences embedding model in Rust

Language: Rust - Size: 108 KB - Last synced at: 9 months ago - Pushed at: 9 months ago - Stars: 16 - Forks: 2

levipereira/triton-server-yolo

This repository serves as an example of deploying the YOLO models on Triton Server for performance and testing purposes

Language: Shell - Size: 170 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 18 - Forks: 1

levipereira/triton-client-yolo

This repository utilizes the Triton Inference Server Client, which streamlines the complexity of model deployment.

Language: Python - Size: 889 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 2 - Forks: 0

levipereira/Docker-Yolov7-Nvidia-Kit

This repository provides an out-of-the-box deployment solution for creating an end-to-end procedure to train, deploy, and use Yolov7 models on Nvidia GPUs using Triton Server and Deepstream.

Language: Shell - Size: 23.4 KB - Last synced at: about 1 year ago - Pushed at: over 1 year ago - Stars: 2 - Forks: 0

levipereira/deepstream-yolo-triton-server-rtsp-out

The Purpose of this repository is to create a DeepStream/Triton-Server sample application that utilizes yolov7, yolov7-qat, yolov9 models to perform inference on video files or RTSP streams.

Language: Python - Size: 39.1 KB - Last synced at: about 1 year ago - Pushed at: about 1 year ago - Stars: 0 - Forks: 0

msclock/transformersplus 📦

Add Some plus extra features to transformers

Language: Python - Size: 258 KB - Last synced at: 24 days ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

Bobo-y/triton-backend-demo

This is a cutom triton-backend demo for process image (resize + norm)

Language: C++ - Size: 1.15 MB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 1 - Forks: 0

Bobo-y/django_ai_server_template

A template for delpoy AI server use django with tf_serving or triton_inference_serving

Language: Python - Size: 104 MB - Last synced at: 15 days ago - Pushed at: over 2 years ago - Stars: 6 - Forks: 1

omarabid59/yolov8-triton

Provides an ensemble model to deploy a YoloV8 ONNX model to Triton

Language: Python - Size: 11.7 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 11 - Forks: 4

Achiwilms/NVIDIA-Triton-Deployment-Quickstart

QuickStart Guide for Deploying a Basic ResNet Model on the Triton Inference Server

Language: Python - Size: 5.86 KB - Last synced at: over 1 year ago - Pushed at: over 1 year ago - Stars: 0 - Forks: 0

tamanna18/Triton-Inference-Server-Deployment-with-ONNX-Models

Triton Inference Server Deployment with ONNX Models

Size: 9.77 KB - Last synced at: almost 2 years ago - Pushed at: almost 2 years ago - Stars: 0 - Forks: 0