Converting Hugging Face Safetensor to Checkpoint #53
Unanswered
Bojun-Feng
asked this question in
Q&A
Replies: 2 comments 6 replies
-
You probably want to use the PyTorch Lightning checkpoints here: https://huggingface.co/kaiyuy/leandojo-pl-ckpts |
Beta Was this translation helpful? Give feedback.
6 replies
-
Update: The new evaluation script takes Hugging Face (instead of PyTorch Ligthning) checkpoints as input. You can directly use models such as leandojo-lean4-tacgen-byt5-small. Please see the updated README for details. As a result, https://huggingface.co/kaiyuy/leandojo-pl-ckpts has been deleted. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hello, I've been trying to reproduce the results of the paper through evaluating the hugging face model on the benchmark (without indexing).
However, I am having some trouble applying the hugging face safetensors to compatible torch lightning check points for evaluation. Currently I am using torch to change the safe tensor to a binary checkpoint:
Then, I run command according to the instructions:
However, it seems that dictionary keys are missing in the resulting checkpoint, I get a dictionary error for 'pytorch_lightning_version'.
I first manually added the pytorch lightning version, but am not sure what the parameter should be for other missing parameters such as 'max_inp_seq_len', and 'max_oup_seq_len'. Is there a conveinent way I can load the model from the safetensors or gain access to the original model checkpoints? I looked online but found no relevant information.
I am new to the repo and lean in general so this might be a stupid question. Would appreaciate any useful information or suggestions.
Beta Was this translation helpful? Give feedback.
All reactions