Welcome to the repository for the "Unparsed 2024 Talk" by Roger Kibbe. This repository contains all the materials and resources for the presentation delivered at the Unparsed 2024 conference.
This repository hosts the content for Roger Kibbe's talk at Unparsed 2024. The presentation explores the differences between prompt engineering, RAG and fine-tuning and explores the when, what and how of fine-tuning LLM
The presentation slides can be found in the presentation
directory.
Mistral 7B Fine Tune Juypter Notebook
To run the code examples locally, you will need to have Python installed. Follow the instructions below to set up your environment:
-
Clone the repository:
git clone https://github.com/rogerkibbe/unparsed-2024-talk.git cd unparsed-2024-talk
-
Create a virtual environment and activate it:
python3 -m venv venv source venv/bin/activate
-
Install the required dependencies:
pip install -r requirements.txt
-
Get a Mistral API key and create a .env file with the key:
MISTRAL_API_KEY=[Mistral API Key]
The fine-tuning code is located in the mistral_finetune.ipynb Juypter notebook. Open it to run the fine-tune.
The fine-tuning data is IT support Q&A for a Bastard Operator from Hell type chatbot. The data was created by Anthropic's Claude and hand edited.
gui_client.py is a python panel app. To run:
panel serve [file name]
cli_compare.py is a simple command line tool to compare the outputs of the BOFH fine-tuned model and the base Mistral 7B Model. To run:
python cli_compare.py