This is the repository for this paper. Including how we prompt and extract the results from ChatGPT via the OpenAI API.
The results are recorded here or you can go to official website to check the result.
- Use new openai api to force return json format
- Compare between select top-n snippets and summary the snippets
- Postprocessing when factoid type question have more than 5 entries
@inproceedings{hsueh2023bioasq,
title = {NCU-IISR: Prompt Engineering on GPT-4 to Stove Biological Problems in BioASQ 11b Phase B},
author = {Chun-Yu Hsueh and Yu Zhang and Yu-Wei Lu and Jen-Chieh Han and Wilailack Meesawad and Richard Tzong-Han Tsai},
year = 2023,
booktitle = {Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2023), Thessaloniki, Greece, September 18th to 21st, 2023},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
volume = 3497,
pages = {114--121},
url = {https://ceur-ws.org/Vol-3497/paper-009.pdf},
editor = {Mohammad Aliannejadi and Guglielmo Faggioli and Nicola Ferro and Michalis Vlachos}
}
- BioBERT1
- BioGPT (GPT2 based)2
- ChatGPT (GPT4)
- ChatGPT (GPT3.5)
- Fine-tune/Transfer GPT3,prompt turning
- OpenAI Evals
- BertScore
- ROUGE-SU4
- training 11b = training 10b + testing 10b
- BioLinkBERT
- OPT-175b
- Retriever method for snippets
- BM25 & TF-IDF
- OpenAI Embedding + Cos similarity
Footnotes
-
Inaccurate result on PapersWithCode, should follow PudMedQA homepage ↩