Skip to content

DCDmllm/IDEAL-Summary

Repository files navigation

IDEAL_Summary

IDEAL: Leveraging Infinite and Dynamic Characterizations of Large Language Models for Query-focused Summarization

Getting Started

1. Installation

conda create --name ideal_summary python=3.9
conda activate ideal_summary
pip install -r requirement.txt

Prepare pretrain model weights

Code prefixed with "LLama" are designed to work with the LLaMA 2 7B model weights. Similarly, code prefixed with "llama3" are compatible with the LLaMA 3 8B and LLaMA 3.1 8B model weights. You can obtain the official LLaMA consolidated format weights(Instruct version) by downloading them from the official Meta AI website or the Hugging Face model hub.

2. Data Processing

  1. Download datasets from their respective official repositories:

  2. Preprocess the datasets using the provided Jupyter notebook: data_process.ipynb.

3. Training, Inference, and Evaluation

To train, run inference, and evaluate the model, execute the following script:

bash exps/finetuning_*_generate_evaluate.sh

For multi-reference Rouge scores and Bert-score evaluations on the SQuALITY dataset, use the notebook multi_reference_evaluation_SQuAlITY.ipynb.

Results

The output directory includes the generated outputs of GPT-4o (2024-08-06) and our method on the test set reported in paper.

Acknowledgment

Our project is developed based on the following repositories:

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published