Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Custom inference endpoints #306

Closed
himsgpt opened this issue Jul 31, 2023 · 6 comments
Closed

Custom inference endpoints #306

himsgpt opened this issue Jul 31, 2023 · 6 comments
Labels
duplicate This issue or pull request already exists enhancement New feature or request

Comments

@himsgpt
Copy link

himsgpt commented Jul 31, 2023

Problem

A lot of enterprises are building their own llm models, can we use them instead of chatgpt/hugging face, etc. Sagemaker is one option but I should be able to provide just an inference endpoint to use this for prompts in jupyter lab!

Proposed Solution

A lot of enterprises are building their own llm models, can we use them instead of chatgpt/hugging face, etc. Sagemaker is one option but I should be able to provide just an inference endpoint to use this for prompts in jupyter lab!

Additional context

@himsgpt himsgpt added the enhancement New feature or request label Jul 31, 2023
@welcome
Copy link

welcome bot commented Jul 31, 2023

Thank you for opening your first issue in this project! Engagement like this is essential for open source projects! 🤗

If you haven't done so already, check out Jupyter's Code of Conduct. Also, please try to follow the issue template as it helps other other community members to contribute more effectively.
welcome
You can meet the other Jovyans by joining our Discourse forum. There is also an intro thread there where you can stop by and say Hi! 👋

Welcome to the Jupyter community! 🎉

@JasonWeill
Copy link
Collaborator

Are you asking about locally hosted models? If so, this feature request is being discussed in #190.

@himsgpt
Copy link
Author

himsgpt commented Aug 1, 2023

Yes, enterprises have been building and deploying custom llm models in on-premises. We therefore want to use these custom endpoints in the jupyter lab instead of chatgpt, etc.

@JasonWeill
Copy link
Collaborator

@himsgpt Thanks for clarifying! Locally-hosted and on-prem models are high priorities for organizations that can't rely on third-party models for generative AI. Let's keep the conversation going in #190.

@JasonWeill JasonWeill added the duplicate This issue or pull request already exists label Aug 1, 2023
@c3-viral-lakhani
Copy link

@JasonWeill

We have our own inference endpoint, with model already deployed. Is it possible to configure the extension to point to the endpoint or any other way to use the endpoint directly?

Does the model have to come from GPT4All so that it can be used in jupyter-ai extension?

@JasonWeill
Copy link
Collaborator

@c3-viral-lakhani In #322, @dlqqq added support for OpenAI proxies. GPT4All is another way to use local models. If you have another locally deployed model, I recommend filing an issue and/or opening a pull request to add support for it to Jupyter AI, to ensure that the magic commands and chat UI work with your own endpoint. Thanks for your interest!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
duplicate This issue or pull request already exists enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

3 participants