-
Notifications
You must be signed in to change notification settings - Fork 62
Insights: NVIDIA/TensorRT-Model-Optimizer
Overview
-
0 Active pull requests
-
- 0 Merged pull requests
- 0 Open pull requests
- 1 Closed issue
- 2 New issues
There hasn’t been any commit activity on NVIDIA/TensorRT-Model-Optimizer in the last week.
Want to help out?
1 Issue closed by 1 person
-
Export fp8 ONNX
#162 closed
Mar 27, 2025
2 Issues opened by 2 people
-
Which instructions should I follow to quant my model from bf16 to nvfp4?
#164 opened
Mar 28, 2025 -
What is pre_quant_scale?
#163 opened
Mar 26, 2025
3 Unresolved conversations
Sometimes conversations happen on old items that aren’t yet closed. Here is a list of all the Issues and Pull Requests with unresolved conversations.
-
Large model offloaded to huggingface accelerate is not able to export the weights using unified export.
#157 commented on
Mar 24, 2025 • 0 new comments -
Unable to quantize pytorch model using huggingface export
#161 commented on
Mar 24, 2025 • 0 new comments -
PyTorch Quantization Failed to Quantize Scaled Dot Product
#149 commented on
Mar 27, 2025 • 0 new comments