DunnBC22 commited on
Commit
a9b8774
1 Parent(s): 348c6c0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -7
README.md CHANGED
@@ -7,11 +7,13 @@ metrics:
7
  model-index:
8
  - name: codet5-small-Generate_Docstrings_for_Python-Condensed
9
  results: []
 
 
 
 
 
10
  ---
11
 
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
-
15
  # codet5-small-Generate_Docstrings_for_Python-Condensed
16
 
17
  This model is a fine-tuned version of [Salesforce/codet5-small](https://huggingface.co/Salesforce/codet5-small) on the None dataset.
@@ -25,15 +27,19 @@ It achieves the following results on the evaluation set:
25
 
26
  ## Model description
27
 
28
- More information needed
 
 
 
 
29
 
30
  ## Intended uses & limitations
31
 
32
- More information needed
33
 
34
  ## Training and evaluation data
35
 
36
- More information needed
37
 
38
  ## Training procedure
39
 
@@ -63,4 +69,4 @@ The following hyperparameters were used during training:
63
  - Transformers 4.26.1
64
  - Pytorch 1.12.1
65
  - Datasets 2.9.0
66
- - Tokenizers 0.12.1
 
7
  model-index:
8
  - name: codet5-small-Generate_Docstrings_for_Python-Condensed
9
  results: []
10
+ datasets:
11
+ - calum/the-stack-smol-python-docstrings
12
+ language:
13
+ - en
14
+ pipeline_tag: text2text-generation
15
  ---
16
 
 
 
 
17
  # codet5-small-Generate_Docstrings_for_Python-Condensed
18
 
19
  This model is a fine-tuned version of [Salesforce/codet5-small](https://huggingface.co/Salesforce/codet5-small) on the None dataset.
 
27
 
28
  ## Model description
29
 
30
+ This model is trained to predict the docstring (the output) for a function (the input).
31
+
32
+ For more information on how it was created, check out the following link: https://github.com/DunnBC22/NLP_Projects/blob/main/Generate%20Docstrings/Smol%20Dataset/Code_T5_Project-Small%20Checkpoint.ipynb
33
+
34
+ For this model, I trimmed some of the longer samples to quicken the pace of training on consumer hardware.
35
 
36
  ## Intended uses & limitations
37
 
38
+ This model is intended to demonstrate my ability to solve a complex problem using technology.
39
 
40
  ## Training and evaluation data
41
 
42
+ Dataset Source: calum/the-stack-smol-python-docstrings (from HuggingFace Datasets; https://huggingface.co/datasets/calum/the-stack-smol-python-docstrings)
43
 
44
  ## Training procedure
45
 
 
69
  - Transformers 4.26.1
70
  - Pytorch 1.12.1
71
  - Datasets 2.9.0
72
+ - Tokenizers 0.12.1