Skip to content

Conversation

@vayuda
Copy link
Contributor

@vayuda vayuda commented Jun 12, 2024

Now allows the packing dimension to be -1
No longer need to declare the device to use in the function params as it takes it from the input data
Added rudimentary benchmark comparing with fp16 and using unpack in a weight only quantization model

  • faster for input vectors of size 4096 to 8192 but slower for sizes outside of that range (would like a review on the setup for another pr though)

@pytorch-bot
Copy link

pytorch-bot bot commented Jun 12, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/347

Note: Links to docs will display an error until the docs builds have been completed.

❌ 1 New Failure

As of commit bfd6c5f with merge base 950a893 (image):

NEW FAILURE - The following job has failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Jun 12, 2024
@vayuda vayuda requested a review from gau-nernst June 12, 2024 04:57
@msaroufim msaroufim self-requested a review June 13, 2024 02:37
@msaroufim msaroufim merged commit 6f44d25 into main Jun 13, 2024
@vayuda vayuda deleted the BitPack-v3 branch June 17, 2024 21:34
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants