Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix problem with attention_size #144

Merged
merged 2 commits into from
May 25, 2024
Merged

fix problem with attention_size #144

merged 2 commits into from
May 25, 2024

Conversation

jettjaniak
Copy link
Contributor

@jettjaniak jettjaniak commented May 20, 2024

some sizes have intermediate_size = 2 * hidden_size, what's that about? (should be closer to 8/3)

@jettjaniak jettjaniak force-pushed the llama-conf-rope-fix branch 3 times, most recently from 0b8c8dc to 82c16ef Compare May 20, 2024 15:00
@jettjaniak
Copy link
Contributor Author

jettjaniak commented May 20, 2024

ValueError: hidden_size must be divisible by num_heads (got `hidden_size`: 54 and `num_heads`: 4)

@jettjaniak jettjaniak force-pushed the llama-conf-rope-fix branch 3 times, most recently from 1234a17 to aa4d664 Compare May 21, 2024 13:28
@jettjaniak jettjaniak force-pushed the llama-conf-rope-fix branch 2 times, most recently from 53044d0 to 033fc85 Compare May 25, 2024 12:39
@jettjaniak jettjaniak force-pushed the llama-conf-rope-fix branch from 033fc85 to 426c964 Compare May 25, 2024 12:59
@jettjaniak jettjaniak merged commit 8670bb8 into main May 25, 2024
1 check passed
@jettjaniak jettjaniak deleted the llama-conf-rope-fix branch May 25, 2024 13:03
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants