Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bugfixes surrounding torch dtypes and QOL updates for torch #53

Closed
wants to merge 4 commits into from

Conversation

nreHieW
Copy link
Contributor

@nreHieW nreHieW commented Oct 9, 2024

This PR does the following:

  1. Bugfixes the conflicting torch dtypes in the attention computation by casting the returned key and values from the kv cache to the input types. This matches the JAX implementation where key, values are stored in fp32 and fp16 respectively post ROPE. (thanks @tensorqt for bringing this up) (should also close Torch doesn't work on mac. #50)

  2. qol fixes surrounding torch_main.py cleaning up imports and setting up prompt testing similar to main.py

  3. Fixes changes from Bugfix to varentropy calculation  #17 being overwritten

@nreHieW
Copy link
Contributor Author

nreHieW commented Oct 12, 2024

closing in favor of #72

@nreHieW nreHieW closed this Oct 12, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Torch doesn't work on mac.
1 participant