-
Notifications
You must be signed in to change notification settings - Fork 2
Open
Description
Stage 3: Optimizing TensorRT Inference
- TensorRT Loading YoloV8 Example
- Old Nvidia Guide for Inference
- New Nvidia Guide for Inference
- ONNX Example
- Experiment with FP16 Precision mode, compare and report results
- Experiment with INT8 Precision mode, (research process for calibrating as well), compare and report results
- Quantization Basics
- Quantization Blog
- Quantization Blog 2
- Quantization Blog 3
- Quantization Blog 4
- Knowledge Distillation Blog
- Knowledge Distillation Blog 2
- Batched NMS Blog
- Batched NMS Blog 2
- Batched NMS Blog 3
- Batched NMS Blog 4
- Batched NMS Blog 5
Metadata
Metadata
Assignees
Labels
No labels