Skip to content

HuatuoGPT2, One-stage Training for Medical Adaption of LLMs. (An Open Medical GPT)

Notifications You must be signed in to change notification settings

eric-senyao/HuatuoGPT-II

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

35 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

HuatuoGPT2, One-stage Training for Medical Adaption of LLMs

HuatuoGPT-II

🖥️ Online Demo (7B) |⬇️ 7B Model |⬇️ 13B Model | ⬇️ 34B Model | 📃 Paper

✨ Latest News

  • [07/10/2024]: 🎉🎉🎉 Our paper is accepted for COLM 2024!
  • [06/24/2024] We have made all training data for HuatuoGPT2 publicly available. This includes the Pretraining dataset and the SFT dataset).
  • [01/10/2024] The HuatuoGPT2 model is now available on the Wisemodel platform.
  • [12/04/2023] We released the code and dataset for our evaluation.
  • [11/24/2023] We released the quantitative version of HuatuoGPT-II.
  • [11/21/2023] We released HuatuoGPT-II models. The HuatuoGPT-II will be available in 7B, 13B, and 34B versions.
  • [11/17/2023] We released the HuatuoGPT-II paper, achieving a new state-of-the-art in Chinese medical applications! Try our demo!

⚡ Introduction

Hello! Welcome to the repository for HuatuoGPT2.

HuatuoGPT2 employs an innovative domain adaptation method to significantly boost its medical knowledge and dialogue proficiency. It showcases state-of-the-art performance in several medical benchmarks, especially surpassing GPT-4 in expert evaluations and the fresh medical licensing exams.

The open-source release of HuatuoGPT-2 includes:

  • HuatuoGPT2 Model: Open-sourcing of 7B, 13B, and 34B versions.
  • Training Code: Training code for one-stage adaptation will be provided, enabling better model adaptation across various languages and domains.
  • HuatuoGPT2 Data: Release of partial pre-training and fine-tuning instructions.
  • Evaluation for Chinese Medical LLM: Comprehensive automatic evaluation methods for medical response capabilities of LLM and the fresh professional pharmacist exam assessment.

Note that we're still actively organizing our code and data. Please stay tuned for updates coming soon!

HuatuoGPT2

🌟 Performance

Compared with representative open-source models and closed-source models (including GPT-4), HuatuoGPT2 showed impressive performance on medical benchmarks. Here, we present two of the results.

  • Expert Evaluation: In assessments by medical professionals, HuatuoGPT-II's responses in Chinese medical contexts were favored over counterparts like GPT-4:
HuatuoGPT-II Win Rate Win Tie Fail
Single-round Medical Response
HuatuoGPT-II(7B) vs GPT-4 38 38 24
HuatuoGPT-II(7B) vs ChatGPT 52 33 15
HuatuoGPT-II(7B) vs Baichuan2-13B-Chat 63 19 18
HuatuoGPT-II(7B) vs HuatuoGPT 81 11 8
Multi-round Medical Dialogue
HuatuoGPT-II(7B) vs GPT-4 53 17 30
HuatuoGPT-II(7B) vs ChatGPT 56 11 33
HuatuoGPT-II(7B) vs Baichuan2-13B-Chat 63 19 18
HuatuoGPT-II(7B) vs HuatuoGPT 68 6 26
  • The Fresh Medical Exams: We collected the fresh 2023 Chinese National Pharmacist Licensure Examination, which started on October 21, 2023. This date is later than our data finalization. HuatuoGPT2 achieved the best results in this exam, as shown below.
HuatuoGPT2

👩‍⚕️ Model

Model Access

Our model is now available on Huggingface. You can Try our model in https://www.huatuogpt.cn/.

Model Backbone Checkpoint
HuatuoGPT2-7B Baichuan2-7B-Base HF Lnik
HuatuoGPT2-13B Baichuan2-13B-Base HF Lnik
HuatuoGPT2-34B Yi-34B HF Lnik

Quantization Model

A quantized version of HuatuoGPT2 is also provided, allowing users with constrained memory or computing resources to access our HuatuoGPT2.

Quantization Backbone Checkpoint
HuatuoGPT2-7B-4bits Baichuan2-7B-Base HF Lnik
HuatuoGPT2-7B-8bits Baichuan2-7B-Base HF Lnik
HuatuoGPT2-34B-4bits Yi-34B HF Lnik
HuatuoGPT2-34B-8bits Yi-34B HF Lnik

Model Inference

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
tokenizer = AutoTokenizer.from_pretrained("FreedomIntelligence/HuatuoGPT2-7B", use_fast=True, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("FreedomIntelligence/HuatuoGPT2-7B", device_map="auto", torch_dtype=torch.bfloat16, trust_remote_code=True)
messages = []
messages.append({"role": "user", "content": "肚子疼怎么办?"})
response = model.HuatuoChat(tokenizer, messages)
print(response)

Inference with Command Line

python cli_demo.py --model_name FreedomIntelligence/HuatuoGPT2-7B

📚 Data

We open source part of the training data.

Data Type # Training data Link
Medical Fine-tuning Instruction (GPT-4) 142,248 HF Link
Medical Pre-training Instruction 5,286,308 HF Link

🌈 One-stage adaption

Data Unification

HuatuoGPT2
  • HuatuoGPT2 transforms the pre-training corpus into (instruction, output) pairs using LLM. Utilize the script for Data Unification.
python adaption/data_unification/rewrite.py

One-stage training

HuatuoGPT2
  • We introduce a priority sampling approach, pre-processing data with this algorithm:
python adaption/one_stage_training/data_process.py
  • Then, training is conducted using one-stage training:
bash adaption/one_stage_training/train.sh

By adopting the One-stage Adaptation method, you will observe the following loss curve:

HuatuoGPT2

🧐 Evaluation

Automated Evaluation of Medical Response Quality

-- Evaluation code for the QA benchmarks.

accelerate launch evaluation/eval_qa.py --model_path=FreedomIntelligence/HuatuoGPT2-7B --data_path=./evaluation/data/eval_qa.json
  • Single-turn response evaluation using GPT-4:
python evaluation/eval_huatuo_inst.py
  • Multi-turn dialogue evaluation using GPT-4:
python evaluation/eval_huatuo_conv.py

The Fresh Medical Exams

Access our newest medical exam dataset via the link provided. The dataset includes complete exam questions, with exam dates noted to alert for potential leaks. We plan to release more updated exams in the future.

Examination #Question Exam Time Links
2023 Chinese National Pharmacist Licensure Examination (Pharmacy) 480 2023.10.22 huggingface
2023 Chinese National Pharmacist Licensure Examination (TCM) 480 2023.10.22 huggingface
Other Fresh Medical Examinations is in coming

🩺 HuatuoGPT Series

The HuatuoGPT series has so far launched two generations:

  • HuatuoGPT: A Doctor-like Medical Large Language Model
  • HuatuoGPT-II: An Domain-enhanced Medical Large Language Model

In the future, we will continue to release new versions of HuatuoGPT. Our goal is to enhance the capabilities of LLM in the Chinese medical field and to adhere to open-source principles (aligned with the ethos of FreedomIntelligence). We hope to work together with everyone to promote the development of medical LLM!

We are from the School of Data Science, the Chinese University of Hong Kong, Shenzhen (CUHKSZ) and the Shenzhen Research Institute of Big Data (SRIBD).

Citation

@misc{chen2023huatuogptii,
      title={HuatuoGPT-II, One-stage Training for Medical Adaption of LLMs}, 
      author={Junying Chen and Xidong Wang and Anningzhe Gao and Feng Jiang and Shunian Chen and Hongbo Zhang and Dingjie Song and Wenya Xie and Chuyi Kong and Jianquan Li and Xiang Wan and Haizhou Li and Benyou Wang},
      year={2023},
      eprint={2311.09774},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

@article{huatuogpt-2023,
  title={HuatuoGPT, Towards Taming Language Models To Be a Doctor},
  author={Hongbo Zhang and Junying Chen and Feng Jiang and Fei Yu and Zhihong Chen and Jianquan Li and Guiming Chen and Xiangbo Wu and Zhiyi Zhang and Qingying Xiao and Xiang Wan and Benyou Wang and Haizhou Li},
  journal={arXiv preprint arXiv:2305.15075},
  year={2023}
}

Star History

Star History Chart

About

HuatuoGPT2, One-stage Training for Medical Adaption of LLMs. (An Open Medical GPT)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 99.1%
  • Shell 0.9%