Converting Hugging Face Safetensor to Checkpoint #53
Unanswered
Bojun-Feng
asked this question in
Q&A
Replies: 2 comments 6 replies
-
You probably want to use the PyTorch Lightning checkpoints here: https://huggingface.co/kaiyuy/leandojo-pl-ckpts |
Beta Was this translation helpful? Give feedback.
6 replies
-
Update: The new evaluation script takes Hugging Face (instead of PyTorch Ligthning) checkpoints as input. You can directly use models such as leandojo-lean4-tacgen-byt5-small. Please see the updated README for details. As a result, https://huggingface.co/kaiyuy/leandojo-pl-ckpts has been deleted. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello, I've been trying to reproduce the results of the paper through evaluating the hugging face model on the benchmark (without indexing).
However, I am having some trouble applying the hugging face safetensors to compatible torch lightning check points for evaluation. Currently I am using torch to change the safe tensor to a binary checkpoint:
Then, I run command according to the instructions:
However, it seems that dictionary keys are missing in the resulting checkpoint, I get a dictionary error for 'pytorch_lightning_version'.
I first manually added the pytorch lightning version, but am not sure what the parameter should be for other missing parameters such as 'max_inp_seq_len', and 'max_oup_seq_len'. Is there a conveinent way I can load the model from the safetensors or gain access to the original model checkpoints? I looked online but found no relevant information.
I am new to the repo and lean in general so this might be a stupid question. Would appreaciate any useful information or suggestions.
Beta Was this translation helpful? Give feedback.
All reactions