AmelieSchreiber's picture
Update README.md
8301fe6
metadata
license: mit
datasets:
  - AmelieSchreiber/cafa5_pickle_split
language:
  - en
metrics:
  - accuracy
  - f1
  - precision
  - recall
  - roc_auc
library_name: transformers
tags:
  - esm
  - esm2
  - protein language model
  - biology
  - cafa5

ESM-2 Pre-finetuned for CAFA-5 for Protein Function Prediction

This model is a pre-finetuned for CAFA-5 protein function prediction for four epochs. This model is meant to be finetuned in a second stage of training with a Low Rank Adaptation. The training script for both the pre-finetuning and second stage finetuning with LoRA is available here. This notebook allows you to pre-finetune the base model, and then use a LoRA for the second stage of training. Note, the second stage of training is a harder curriculum for the model as it uses class weights so that the model better captures the hierarchical (weighted) structure of the gene ontology (GO) terms that serve as the labels for the multilabel sequence classification task of predicting a protein's functions (GO terms).