Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

udpate llama7b_sparse_quantized example #2322

Merged
merged 6 commits into from
Jun 13, 2024
Merged

Conversation

bfineran
Copy link
Contributor

@bfineran bfineran commented Jun 7, 2024

No description provided.

@bfineran bfineran requested review from Satrat and markurtz June 7, 2024 16:49

output_dir = "output_llama7b_2:4_w4a16_channel"

apply(
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The number of arguments here is very confusing, especially since most of these are related to training...

Copy link

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Talked to Ben and he is going to write up a README of just quantization without the training. This one is intended to be a more advanced readme showing how to do the full sparsity -> finetuning -> quantization flow

dbogunowicz
dbogunowicz previously approved these changes Jun 10, 2024
examples/llama7b_one_shot_quantization.md Show resolved Hide resolved
examples/llama7b_sparse_quantized/README.md Outdated Show resolved Hide resolved
Sara Adkins added 2 commits June 12, 2024 11:19
@bfineran bfineran merged commit 5c1de1c into main Jun 13, 2024
18 checks passed
@bfineran bfineran deleted the compression-example-update branch June 13, 2024 20:04
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants