rudyvdbrink commited on
Commit
3553205
·
verified ·
1 Parent(s): 93713ad

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -22
README.md CHANGED
@@ -1,25 +1,23 @@
1
- ---
2
- library_name: peft
3
- license: llama3.2
4
- base_model: meta-llama/Llama-3.2-1B
5
- tags:
6
- - generated_from_trainer
7
- metrics:
8
- - accuracy
9
- - f1
10
- - precision
11
- - recall
12
- model-index:
13
- - name: Llama-3.2-1B-binary-citation-classifier
14
- results: []
15
- ---
16
-
17
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
18
- should probably proofread and complete it, then remove this comment. -->
19
 
20
  # Llama-3.2-1B-binary-citation-classifier
21
 
22
- This model is a fine-tuned version of [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) on an unknown dataset.
 
23
  It achieves the following results on the evaluation set:
24
  - Loss: 0.5450
25
  - Accuracy: 0.746
@@ -29,18 +27,21 @@ It achieves the following results on the evaluation set:
29
 
30
  ## Model description
31
 
32
- More information needed
33
 
34
  ## Intended uses & limitations
35
 
36
- More information needed
37
 
38
  ## Training and evaluation data
39
 
40
- More information needed
41
 
42
  ## Training procedure
43
 
 
 
 
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
 
1
+ ---
2
+ library_name: peft
3
+ license: llama3.2
4
+ base_model: meta-llama/Llama-3.2-1B
5
+ tags:
6
+ - generated_from_trainer
7
+ metrics:
8
+ - accuracy
9
+ - f1
10
+ - precision
11
+ - recall
12
+ model-index:
13
+ - name: Llama-3.2-1B-binary-citation-classifier
14
+ results: []
15
+ ---
 
 
 
16
 
17
  # Llama-3.2-1B-binary-citation-classifier
18
 
19
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B) on a dataset of scientific abstracts and citation counts.
20
+ Its aim is to predict, based on an article abstract, if an article will be cited within five years or not.
21
  It achieves the following results on the evaluation set:
22
  - Loss: 0.5450
23
  - Accuracy: 0.746
 
27
 
28
  ## Model description
29
 
30
+ Llama-3.2-1B architecture, modified with a rank 8 LORA adapter.
31
 
32
  ## Intended uses & limitations
33
 
34
+ Intended use is binary classification. The training set consists of PubMed indexed neuroscience-related articles exclusively.
35
 
36
  ## Training and evaluation data
37
 
38
+ [Training and evalutation data](https://huggingface.co/datasets/rudyvdbrink/CitationDatabase)
39
 
40
  ## Training procedure
41
 
42
+ Pre-training following Meta's procedures.
43
+ LORA fine tuning with PEFT on 16k abstracts (8k cited, 8k uncited)
44
+
45
  ### Training hyperparameters
46
 
47
  The following hyperparameters were used during training: