Skip to content

medal-contender/nbme-score-clinical-patient-notes

Repository files navigation

NBME - Score Clinical Patient Notes


Competition Overview

In this competition, you’ll identify specific clinical concepts in patient notes. Specifically, you'll develop an automated method to map clinical concepts from an exam rubric (e.g., “diminished appetite”) to various ways in which these concepts are expressed in clinical patient notes written by medical students (e.g., “eating less,” “clothes fit looser”). Great solutions will be both accurate and reliable.

If successful, you'll help tackle the biggest practical barriers in patient note scoring, making the approach more transparent, interpretable, and easing the development and administration of such assessments. As a result, medical practitioners will be able to explore the full potential of patient notes to reveal information relevant to clinical skills assessment.

Due Date

  • Team Merge Deadline - April 26, 2022
  • Submission Deadline - May 3, 2022

Members

- Jeongwon Kim ([email protected])
- Jaewoo Park ([email protected])
- Young Min Paik ([email protected])
- Hyeonhoon Lee ([email protected])

Competition Results


Submission notebook

  1. DeBERTa v3 Large DeepShare Model 10 fold + DeBERTa v3 Large Model 5 fold
  2. DeBERTa v3 Large DeepShare Model 10 fold + DeBERTa v3 Large Model 5 fold + RoBERTa Large

Task

This task is a PIPELINE to experiment with PLM(Pretrained Language Model), which is often used in NLP tasks, and the following models are tested
(In addition, we conducted experiments by modifying FC Layer of the each model.)

- RoBERTa
- DeBERTa
- Electra
- LUKE

Baseline 1

Baseline 1 is a task that predicts all tokens in the input and processes only tokens above the threshold, similar to the NER task in BERT.
It is a task in which the model recognizes and predicts 'text' and 'text_feature' by organizing Dataset in order of [‘text’, 'text_feature’].

Baseline 2, 3

Baseline 2, 3 is based on baseline 1, but similar to QA task, Dataset is organized in the order of ['text_feature’, 'text’].
baseline 2 is a Pytorch-based code and baseline 3 is Transformer-based code.

Main Issues

  • Dataset review

    1. typo problem
    • Many typos were found on a given dataset, and we solved these problems in the data preprocessing process.
    1. Related Issues
  • Data Embedding

    1. Tokenizer issue
    • In the process of embedding data, we found that each tokenizer calculates the position of the offset differently. To solve this problem, we have tried methods such as implementing token positioning functions, char-based embeddings, and we have been able to confirm good performance.
    1. Rlated Issues
  • Model performance

    1. Scheduler
    • We analyzed various studies and techniques related to schedulers in which the model can reach the global minimum, and applied the SOTA scheduler CosineAnnealingWarmupRestarts to improve model performance.
    1. Model tuning
    • To improve the performance of the model, we thought of a tuning method, and we improved the performance of the model by tuning the layer part.
    1. Ensemble Strategy
    • We discussed strategies for ensembling different models (Roberta, DeBERTa, etc.) because each model has a different structure and finally decided on three models to achieve the best performance.
    1. Related Issues

Program

  • Fetch Pretrained Models
$ sh ./download_pretrained_models.sh
  • Train
$ cd /nbme-score-clinical-patient-notes/nbme
$ python3 run_train.py \
          "--config-file", "nbme-score-clinical-patient-notes/configs/deberta.yaml", \
          "--train", \
          "--training-keyword", "deberta-test"

You can set launch.json for vscode as follow:

{
    "version": "0.2.0",
    "configurations": [
        {
            "name": "Python: Current File",
            "type": "python",
            "request": "launch",
            "program": "run_train.py",
            "console": "integratedTerminal",
            "cwd": "${workspaceFolder}/nbme",
            "args": ["--config-file", "../configs/deberta.yaml", "--train"]
        }
    ]
}

Requirements

Environment

  • Python 3.7 (To match with Kaggle environment)
  • Conda
  • git
  • git-lfs
  • CUDA 11.3 + PyTorch 1.10.1

Pytorch version may vary depanding on your hardware configurations.

Installation with virtual environment

git clone https://github.com/medal-contender/nbme-score-clinical-patient-notes.git
conda create -n nbme python=3.7
conda activate nbme
# PyTorch installation process may vary depending on your hardware
conda install pytorch torchvision torchaudio cudatoolkit=11.3 -c pytorch
cd nbme-score-clinical-patient-notes
pip install -r requirements.txt

Run this in WSL (or WSL2)

./download_pretrained_models.sh

To update the code

$ git pull

If you have local changes, and it causes to abort git pull, one way to get around this is the following:

# removing the local changes
$ git stash
# update
$ git pull
# put the local changes back on top of the recent update
$ git stash pop