Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MPS CI runs #162

Merged
merged 17 commits into from
Apr 15, 2024
Merged

MPS CI runs #162

merged 17 commits into from
Apr 15, 2024

Conversation

mikekgfb
Copy link
Contributor

MPS quantization

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Meta Open Source bot. label Apr 13, 2024
@mikekgfb mikekgfb changed the title MPS quantization MPS CI runs Apr 13, 2024
@mikekgfb mikekgfb merged commit 9c8bf7b into main Apr 15, 2024
1 of 13 checks passed
@mikekgfb mikekgfb deleted the mps_quantization branch April 15, 2024 15:12
malfet pushed a commit that referenced this pull request Jul 17, 2024
* MPS quantization

* mps dtypes

* updates

* fix names

* typo

* no bfloat16 for older macOS

* fix typo

* remove failing embedding quantization from MPS runs

* bfloat -> current model precision

* typo

* missed bfloat16 to swotch to defaulkt precision

* remove int8 quantization on mps

* enable cpu fallback for mps on int4

* hack int4pack_mm for torch.float

* typo

* disable int4 because fp16 int4pack_mm not working for float16
malfet pushed a commit that referenced this pull request Jul 17, 2024
* MPS quantization

* mps dtypes

* updates

* fix names

* typo

* no bfloat16 for older macOS

* fix typo

* remove failing embedding quantization from MPS runs

* bfloat -> current model precision

* typo

* missed bfloat16 to swotch to defaulkt precision

* remove int8 quantization on mps

* enable cpu fallback for mps on int4

* hack int4pack_mm for torch.float

* typo

* disable int4 because fp16 int4pack_mm not working for float16
malfet pushed a commit that referenced this pull request Jul 17, 2024
* MPS quantization

* mps dtypes

* updates

* fix names

* typo

* no bfloat16 for older macOS

* fix typo

* remove failing embedding quantization from MPS runs

* bfloat -> current model precision

* typo

* missed bfloat16 to swotch to defaulkt precision

* remove int8 quantization on mps

* enable cpu fallback for mps on int4

* hack int4pack_mm for torch.float

* typo

* disable int4 because fp16 int4pack_mm not working for float16
malfet pushed a commit that referenced this pull request Jul 17, 2024
* MPS quantization

* mps dtypes

* updates

* fix names

* typo

* no bfloat16 for older macOS

* fix typo

* remove failing embedding quantization from MPS runs

* bfloat -> current model precision

* typo

* missed bfloat16 to swotch to defaulkt precision

* remove int8 quantization on mps

* enable cpu fallback for mps on int4

* hack int4pack_mm for torch.float

* typo

* disable int4 because fp16 int4pack_mm not working for float16
malfet pushed a commit that referenced this pull request Jul 17, 2024
* MPS quantization

* mps dtypes

* updates

* fix names

* typo

* no bfloat16 for older macOS

* fix typo

* remove failing embedding quantization from MPS runs

* bfloat -> current model precision

* typo

* missed bfloat16 to swotch to defaulkt precision

* remove int8 quantization on mps

* enable cpu fallback for mps on int4

* hack int4pack_mm for torch.float

* typo

* disable int4 because fp16 int4pack_mm not working for float16
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Meta Open Source bot.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants