Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for 8da4w quantization #884

Merged
merged 1 commit into from
Apr 30, 2024
Merged

Commits on Apr 26, 2024

  1. Add support for 8da4w quantization

    Summary: Add a new quantization for users to quantize their
    models using int8 per token dynamic activation + int4 per
    axis grouped weight quantization.
    
    Test Plan:
    tune run quantize --config quantization quantizer._component_=torchtune.utils.quantization.Int8DynActInt4WeightQuantizer quantizer.groupsize=256
    
    Reviewers: jerryzh168, kartikayk, ebsmothers
    
    Subscribers: jerryzh168, kartikayk, ebsmothers, supriyar
    andrewor14 committed Apr 26, 2024
    Configuration menu
    Copy the full SHA
    56fd119 View commit details
    Browse the repository at this point in the history