Finetuning various transformer architectures on a variety of standard NLP datasets using hugging face, pytorch, and other associated frameworks.
- Dataset: glue sst2
- Model: distillbert (uncased)
- Dataset: twitter airline sentiment
- Model: distillbert (cased)
- Dataset: glue rte
- Model: bert (cased)
- Dataset: conll2003
- Model: bert (cased)
- Dataset: brown corpus
- Model: bert (cased)
- Dataset: Helsinki-NLP kde4
- Model: Helsinki-NLP opus-mt-en-hi
P.S. This was done on a singular RTX2060Ti GPU, hence the selection of relatively lightweight models and datasets.