Skip to content

Commit

Permalink
Add TensorRT-Model-Optimizer INT4 AWQ support in onnxruntime tools (#…
Browse files Browse the repository at this point in the history
…22390)

[TensorRT-Model-Optimizer](https://github.com/NVIDIA/TensorRT-Model-Optimizer)
have a implementation for INT4 AWQ. Adding the support in onnxruntime
tools to quantized the models with TensorRT-Model-Optimizer
  • Loading branch information
anujj authored Oct 11, 2024
1 parent cdebf37 commit 23d48ea
Showing 1 changed file with 351 additions and 2 deletions.
Loading

0 comments on commit 23d48ea

Please sign in to comment.