-
Notifications
You must be signed in to change notification settings - Fork 26
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix and rework GPT-TF.js #807
Conversation
1d88d35
to
03e5c7d
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
superb work, thanks for clearing the GPT's mud, every comments makes it more understandable!
yeah, sadly, as I forgot to merge the processing PR (#781) before you branched off, the whole processing pipeline changed a lot. sorry for the toes stepping (hopefully, it will simplify this PR).
btw, it seems that @xenova/transformer has been recently updated to @huggingface/transformer. did you try it out? maybe it'll help with the tokenizer usage (doesn't look much changed to me but you know best)
0a28b81
to
f7f96dc
Compare
maybe rename block size to context len, that would be more specific |
68d957b
to
8cbc96e
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
only cosmetic comments, thanks for the huge amount of work 🎉
8cbc96e
to
0781b7c
Compare
…odeling head and attention bias
… implement topk sampling
…ers following GPT2 convention, use LogLayer
… and language modeling head
…red task parameter for text tasks
…he model init config
0781b7c
to
30de4fb
Compare
well done! |
Closes #654