-
Notifications
You must be signed in to change notification settings - Fork 148
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
udpate llama7b_sparse_quantized example #2322
Conversation
|
||
output_dir = "output_llama7b_2:4_w4a16_channel" | ||
|
||
apply( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The number of arguments here is very confusing, especially since most of these are related to training...
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Talked to Ben and he is going to write up a README of just quantization without the training. This one is intended to be a more advanced readme showing how to do the full sparsity -> finetuning -> quantization flow
Co-authored-by: dbogunowicz <[email protected]>
* fix alias application with unit tests * style
No description provided.