monsoon-nlp commited on
Commit
83d48d0
1 Parent(s): 4272c6b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -15
README.md CHANGED
@@ -3,31 +3,29 @@ base_model: state-spaces/mamba-130m-hf
3
  tags:
4
  - generated_from_trainer
5
  model-index:
6
- - name: trainer
7
  results: []
 
 
8
  ---
9
 
10
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
- should probably proofread and complete it, then remove this comment. -->
12
 
13
- # trainer
 
14
 
15
- This model is a fine-tuned version of [state-spaces/mamba-130m-hf](https://huggingface.co/state-spaces/mamba-130m-hf) on an unknown dataset.
16
 
17
- ## Model description
18
 
19
- More information needed
20
 
21
- ## Intended uses & limitations
22
-
23
- More information needed
24
-
25
- ## Training and evaluation data
26
-
27
- More information needed
28
 
29
  ## Training procedure
30
 
 
 
31
  ### Training hyperparameters
32
 
33
  The following hyperparameters were used during training:
@@ -45,4 +43,4 @@ The following hyperparameters were used during training:
45
  - Transformers 4.40.0.dev0
46
  - Pytorch 2.2.1+cu121
47
  - Datasets 2.18.0
48
- - Tokenizers 0.15.2
 
3
  tags:
4
  - generated_from_trainer
5
  model-index:
6
+ - name: monsoon-nlp/mamba130-proteinpretrain-quinoa
7
  results: []
8
+ datasets:
9
+ - monsoon-nlp/greenbeing-proteins
10
  ---
11
 
12
+ # mamba130-proteinpretrain-quinoa
 
13
 
14
+ Full model finetuning of Mamba-130M-HF on the "research" split (quinoa
15
+ protein sequences) of GreenBeing-Proteins dataset.
16
 
17
+ Due to limits of V100 GPU, trained 510 steps x batches of 3, ~5% of the research split.
18
 
19
+ Requires GitHub main branch of Transformers (Mamba is not included in releases)
20
 
21
+ Considering training on natural language + proteins, or new "biotokens".
22
 
23
+ More details TBD
 
 
 
 
 
 
24
 
25
  ## Training procedure
26
 
27
+ Notebook: https://colab.research.google.com/drive/1W1rB6rRt8krHZSVYQ_TjbnD9OwzFQeGL
28
+
29
  ### Training hyperparameters
30
 
31
  The following hyperparameters were used during training:
 
43
  - Transformers 4.40.0.dev0
44
  - Pytorch 2.2.1+cu121
45
  - Datasets 2.18.0
46
+ - Tokenizers 0.15.2