ranamhamoud commited on
Commit
400901b
1 Parent(s): d1c461d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -15
README.md CHANGED
@@ -5,6 +5,9 @@ license: mit
5
  language:
6
  - en
7
  pipeline_tag: text-generation
 
 
 
8
  ---
9
 
10
  # Model Card for Educational Storytelling in Computer Science
@@ -15,13 +18,13 @@ This model, developed using Hugging Face’s transformer library, is designed fo
15
 
16
  ### Model Description
17
 
18
- This model is an innovative tool for teaching fundamental computer science concepts via educational storytelling. It generates interactive narratives tailored to specific CS topics requested by the user, such as loops, incorporating assessments to enhance learning and engagement.
19
 
20
- - **Developed by:** Rana M Khamoud
21
  - **Model type:** PEFT adapter model using LoRA from Meta's Llama2 7B
22
  - **Language(s) (NLP):** English
23
- - **License:** Specify License
24
- - **Finetuned from model [optional]:** Finetuned from Meta's Llama-2 7B model
25
 
26
  ### Model Sources
27
 
@@ -35,7 +38,7 @@ This model is an innovative tool for teaching fundamental computer science conce
35
 
36
  The model is designed to be used directly via an interactive interface where users can ask for stories about specific computer science topics. It's suitable for educational purposes, particularly in learning environments or as a supplementary learning tool.
37
 
38
- ### Downstream Use [optional]
39
 
40
  While primarily designed for educational storytelling, the model could potentially be adapted for other educational applications or interactive learning tools that require narrative generation.
41
 
@@ -61,12 +64,6 @@ Here's a general framework for initializing and running the model, detailed in t
61
 
62
  The model was trained on a custom dataset generated specifically for this project, aimed at creating educational content related to computer science topics. The data generation scripts and datasets are available at the linked GitHub repository.
63
 
64
- ### Training Procedure
65
-
66
- #### Preprocessing
67
-
68
- Specific preprocessing details were not provided but would typically include data cleaning and formatting to fit the model's input requirements.
69
-
70
  #### Training Hyperparameters
71
 
72
  The model was trained on an NVIDIA A100 machine using quantization techniques to optimize performance. Training involved configurations like LoRA adaptation and fine-tuning of Meta's Llama2 7B model under specified training arguments.
@@ -75,18 +72,17 @@ The model was trained on an NVIDIA A100 machine using quantization techniques to
75
 
76
  ### Testing Data, Factors & Metrics
77
 
78
- Further details on testing data and evaluation metrics are needed to provide insight into the model’s performance and accuracy.
79
 
80
  ### Results
81
 
82
- Results of the training and subsequent evaluations need to be provided to understand the effectiveness of the model in educational storytelling.
83
 
84
  ## Environmental Impact
85
 
86
  - **Hardware Type:** NVIDIA A100
87
- - **Hours used:** 5 hours
88
  - **Cloud Provider:** RunPod
89
- - **Compute Region:** Not specified (please provide if available)
90
  - **Carbon Emitted:** Estimates not provided
91
 
92
  [More Information Needed]
 
5
  language:
6
  - en
7
  pipeline_tag: text-generation
8
+ tags:
9
+ - educational
10
+ - storytelling
11
  ---
12
 
13
  # Model Card for Educational Storytelling in Computer Science
 
18
 
19
  ### Model Description
20
 
21
+ This model is an innovative tool for teaching fundamental computer science concepts via educational storytelling. It generates interactive stories tailored to specific CS topics requested by the user, such as algorithms, programming basics & more, incorporating assessments to enhance learning and engagement.
22
 
23
+ - **Developed by:** Ranam Hamoud & George Kanaan
24
  - **Model type:** PEFT adapter model using LoRA from Meta's Llama2 7B
25
  - **Language(s) (NLP):** English
26
+ - **License:** MIT License
27
+ - **Finetuned from model :** Finetuned from Meta's Llama-2 7B model
28
 
29
  ### Model Sources
30
 
 
38
 
39
  The model is designed to be used directly via an interactive interface where users can ask for stories about specific computer science topics. It's suitable for educational purposes, particularly in learning environments or as a supplementary learning tool.
40
 
41
+ ### Downstream Use
42
 
43
  While primarily designed for educational storytelling, the model could potentially be adapted for other educational applications or interactive learning tools that require narrative generation.
44
 
 
64
 
65
  The model was trained on a custom dataset generated specifically for this project, aimed at creating educational content related to computer science topics. The data generation scripts and datasets are available at the linked GitHub repository.
66
 
 
 
 
 
 
 
67
  #### Training Hyperparameters
68
 
69
  The model was trained on an NVIDIA A100 machine using quantization techniques to optimize performance. Training involved configurations like LoRA adaptation and fine-tuning of Meta's Llama2 7B model under specified training arguments.
 
72
 
73
  ### Testing Data, Factors & Metrics
74
 
75
+ Further details on testing data and evaluation metrics will be provided.
76
 
77
  ### Results
78
 
79
+ Results of the training and subsequent evaluations will be provided to understand the effectiveness of the model in educational storytelling.
80
 
81
  ## Environmental Impact
82
 
83
  - **Hardware Type:** NVIDIA A100
84
+ - **Hours used:** 8 hours
85
  - **Cloud Provider:** RunPod
 
86
  - **Carbon Emitted:** Estimates not provided
87
 
88
  [More Information Needed]