-
Notifications
You must be signed in to change notification settings - Fork 26.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Torch size missmatch in GPT-J model (Error) #13499
Comments
Hey @MantasLukauskas - could you please provide a reproducible code snippet? :-) |
Hey @patrickvonplaten, Fine-tuning was made in this one: You can use save steps 10 for faster saving :) After that, I try to load my model like in doc just use my model instead of EleutherAI/gpt-j-6B. Interesting that if I use pretrained Eleuther model everything works, but when I use my fine-tuned model with run_clm.py error occurs. Maybe that will help you to solve it or just have an idea of what is wrong there?
|
related #13581 |
Should be fixed by #13617 (comment) |
Environment info
transformers
version: 4.11.0.dev0Who can help
@patrickvonplaten, @LysandreJik, @patil-suraj
Information
I trained/fine-tuned the GPT-J model (model here: https://huggingface.co/EleutherAI/gpt-j-6B), fp16 like in this one suggestion (#13329) and now when I try to use the pipeline to load the model (as shown below) I got torch dimensions error (shown below as well). What solutions can be applied there?
Any ideas on how to solve it?
The text was updated successfully, but these errors were encountered: