Skip to content

Example Shiny for Python app which talks to the OpenAI API

License

Notifications You must be signed in to change notification settings

vscodesyncsettings/chatstream

 
 

Repository files navigation

chatstream for Shiny for Python

The chatstream package provides a Shiny for Python module for building AI chat applications. Please keep in mind that this is very much a work in progress, and the API is likely to change.

See this post on the Shiny blog to see some examples in action!

Chatstream currently supports the OpenAI API. To use this, you must have an OpenAI API key. You can get one from the OpenAI or from Azure's OpenAI Service. (Note that if you have use Azure, you will need to point the applications to the Azure endpoint instead of the default OpenAI endpoint.)

Installation

The chatstream package is not on PyPI, but can be installed with pip:

pip install chatstream@git+https://github.com/wch/chatstream.git

Alternatively, if you'd like to develop a local copy of the package, first clone the repository and then install it with pip:

cd chatstream
pip install -e .[dev]

Running examples

Before running any examples, you must set an environment variable named OPENAI_API_KEY with your OpenAI API key.

You can set the environment variable with the following command:

export OPENAI_API_KEY="<your_openai_api_key>"

Then run:

shiny run examples/basic/app.py --launch-browser

Some examples (like recipes) have a requirements.txt file. For those examples, first install the requirements, then run the application as normal:

pip install -r examples/recipes/requirements.txt
shiny run examples/recipes/app.py --launch-browser

FAQ

  • Does this work with Shinylive? It almost does. The openai package has dependencies which do not install on Pyodide, but chatstream currently has an openai_pyodide shim which uses the browser's fetch API. However, there is one more hurdle: the tiktoken package (which counts the number of tokens used by a piece of text) needs to be built to run on Pyodide.

  • Does this work with langchain? It currently does not. Note that most of the langchain interfaces do not support streaming responses, so instead of showing responses as each word comes in, there is a wait and then the entire response arrives at once.

About

Example Shiny for Python app which talks to the OpenAI API

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 88.0%
  • Makefile 6.5%
  • JavaScript 3.0%
  • CSS 2.5%