-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add more summarization models #2
Comments
I've updated in 0.2.3 to add Llama 3 8b instruct and removed some older transformer chat models that didn't perform well. It seems to do at least as well as gpt-3.5 did. Adding more transformer models should still be done, but it's a little more complicated as some of them seem to have different prompt formats. I still want to add Claude and other API providers. |
Ideally we use a similar HF hub search that was recently enabled for the embedding models. the only issue would be automatically sorting the chat templating for summarizing. I'm sure there is a library for this we could use |
Perhaps we use https://github.com/dottxt-ai/outlines |
I think we can search for "conversational" + "text-generation-interface" to get useful instruct models: |
#66 adds basic support for ollama, which provides a ton of models that can be run on a user's local machine (chat & embed): https://ollama.com/library |
this was accomplished with better huggingface support as well as ollama support |
We could easily add more models to the list of chat models used for summarization:
https://github.com/enjalot/latent-scope/blob/main/latentscope/models/chat_models.json
There are plenty of small open source LLMs on HuggingFace that could be added to the list as long as the transformers provider is compatible.
We should also add more Chat API models, currently only OpenAI and Mistral are implemented in providers but Cohere and Together should be easy to add.
Comments on this issue for which models to add would be a helpful start. Especially if there are fine-tuned small models focused on summarization (my experience with out-of-the box 1B and 7B is that their quality isn't great).
The text was updated successfully, but these errors were encountered: