Skip to content

Commit

Permalink
Fix the logic how number of calibration batches are computed in PTQ t…
Browse files Browse the repository at this point in the history
…o match the documentation (#1918)

* Removing pytorch_quantization, replacing with lazy install

* Removing pytorch_quantization, replacing with lazy install

* Reorganize imports

* Reorganize imports

* Fix imports

* Fix imports

* Fix imports

* import_onnx_graphsurgeon_or_install

* import_onnx_graphsurgeon_or_install

* Remove TODOs

* Remove TODOs

* Fix imports

* Fix install command

* Merge master

* Add install

* update notebbok

* Update test

* use_fb_fake_quant

* Simplify export methods

* Simplify export methods

* Simplify export methods

* Fix use_fb_fake_quant

* Fix typo

* Make sure that we follow documentation on selecting number of calibration batches

* Remove extra call to quantize_module

* Set model to eval before PTQ

* Print quantization results to console

* Added tqdm descirption

* Set default calibration batches to 16

* Re-order arguments

* Re-run notebook

---------

Co-authored-by: Shay Aharon <[email protected]>
  • Loading branch information
BloodAxe and shaydeci authored Mar 18, 2024
1 parent b26eef1 commit 6473ed4
Show file tree
Hide file tree
Showing 6 changed files with 1,720 additions and 297 deletions.
Loading

0 comments on commit 6473ed4

Please sign in to comment.