Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Nanogpt implementation #17

Open
thistleknot opened this issue Aug 14, 2023 · 3 comments
Open

Nanogpt implementation #17

thistleknot opened this issue Aug 14, 2023 · 3 comments

Comments

@thistleknot
Copy link

Is it possible to provide an implementation of this using nanogpt and or litgpt?

@gaotianyu1350
Copy link
Member

Hi,

Thanks for your interest in MeZO! If they are in the transformers package, then it should be fairly easy to use -- just load these models and for prefix-tuning/lora, change the layer names accordingly.

@thistleknot
Copy link
Author

It's not. It's a complete rewrite

@gaotianyu1350
Copy link
Member

Then you can refer to the read me in the large_model folder. We highlighted the core part of the MeZO code which you can copy to your own codebase.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants