You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @raghavbj24 -- thank you for submitting this issue! Question for you: I see that your base_model is /home/ubuntu/llama-2-7b-hf_for_merge. Would the same "small size" phenomenon happen if you try to use meta-llama/Llama-2-7b-hf from https://huggingface.co/meta-llama/Llama-2-7b-hf? Please let me know. Thank you.
Hi @alexsherstinsky-- as per your suggestion I tried the base model as meta-llama/Llama-2-7b-hf from huggingface...but there is no difference and the size of the saved model is very small.
@raghavbj24 Could you please point me to the HuggingFace location where your model is saved and enable me to access it with "read" privileges? I am going to look into it thoroughly in the next few days. Thank you.
@raghavbj24 In parallel, if you do not mind: could you please rerun your experiment using this base model: alexsherstinsky/Mistral-7B-v0.1-sharded -- and let me know here what you see for the merged model size (and please also tell me the location where it will be saved). Thank you very much for your collaboration.
Hi,
I am trying to fine tune the llama-2 model with the help of the following config file->
The fine tuning is successful and I can see that merge and unload process was also completed as shown->
When I checked the disk size for this saved model it was 7.6MB only...indicating that the merge did not happen appropriately.
Environment:
Can someone help me in solving this
TIA
The text was updated successfully, but these errors were encountered: