- Notifications
You must be signed in to change notification settings - Fork202
Pull requests: NVIDIA/TensorRT-Model-Optimizer
Author
Uh oh!
There was an error while loading.Please reload this page.
Label
Uh oh!
There was an error while loading.Please reload this page.
Projects
Uh oh!
There was an error while loading.Please reload this page.
Milestones
Uh oh!
There was an error while loading.Please reload this page.
Reviews
Assignee
Assigned to nobodyLoading
Uh oh!
There was an error while loading.Please reload this page.
Sort
Pull requests list
Optimize calibrate_draft_vocab to read only required lines when calib…
#618 openedNov 27, 2025 byOfir408Loading…
Add all example e2e tests for github PR merge / nightly
#617 openedNov 27, 2025 bykevalmorabia97Loading…
Add build replacement library to the compress algorithm.
#616 openedNov 27, 2025 bydanielkorzekwaLoading…
[5680954,5620660@2][ONNX][Autocast] Update value info in converted graph
#611 openedNov 26, 2025 bygcunhaseLoading…
Add checkpoint save/load to ForwardHook + add IterativeChannelContributionHook
#610 openedNov 26, 2025 bydanielkorzekwaLoading…
Support attention quantization for diffusers >= 0.35.0
#608 openedNov 25, 2025 byshengliangxu • Draft
Convert compressed-tensor int4 format to GPTQ int4 format
#590 openedNov 20, 2025 byEdwardf0t1Loading…
Product Rename: TensorRT Model Optimizer to Model Optimizer
#583 openedNov 20, 2025 bykevalmorabia97Loading…
1 of 2 tasks
[OMNIML-2852] [2/n] Add Core Sparse Attention Infrastructure
#527 openedNov 7, 2025 bykaix-nvLoading…
[Draft] [5526696] Add kv cache quantization support for onnx quantization
#486 openedOct 31, 2025 byzhanghaocLoading…
Preserve original rope scaling type in export due to transformers library AutoConfig issue
#452 openedOct 17, 2025 byEdwardf0t1Loading…
ProTip! Updated in the last three days:updated:>2025-11-26.