Skip to content

Latest commit

 

History

History
27 lines (20 loc) · 1.52 KB

README.md

File metadata and controls

27 lines (20 loc) · 1.52 KB

OptiML Project - Text classification models optimization

Introduction

This project is a part of the course "Optimization for Machine Learning" at EPFL. The goal of this project is to classify texts and try to improve our models by using different optimization techniques on transformers models. In particular, we wish to compare different optimizers' generalization performance in text classification. For this we will fine-tune a RoBERTa model on a text classification dataset and compare the results of different optimization techniques.

Data

The data used in this project is the "text-classification-dataset-example" dataset from Hugging Face that can be found here and imported using the following code:

from datasets import load_dataset

dataset = load_dataset("cwchang/text-classification-dataset-example")

Models

We used the following models in our project:

  • RoBERTa (Robustly optimized BERT approach) : we fine-tuned the model on our dataset using our own training loop to see the effect of the optimization techniques on the model.

Code

The file main.ipynb contains the code to run the model on different optimizers according to a fixed learning rate. The file optml.ipynb runs the same code but with the cosine annealing with linear warm-up learning rate scheduler.

Authors