Locutusque commited on
Commit
777f7fe
1 Parent(s): 4e43f5c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -1
README.md CHANGED
@@ -11,6 +11,7 @@ metrics:
11
  - loss
12
  - reward
13
  - penalty
 
14
  ---
15
  # Model Card
16
  ## Model Details
@@ -31,4 +32,4 @@ The model architecture used in this model is GPT-2, a transformer-based language
31
  The model is evaluated based on several metrics, including loss, reward, penalty, BLEU score, and perplexity. The loss metric is calculated during training and reflects the difference between the predicted output and the actual output. The reward metric is based on the number of correct words generated by the model, while the penalty metric penalizes the model for repeating words consecutively. The BLEU score measures the similarity between the generated text and the ground truth text, while the perplexity metric measures how well the model is able to predict the next word in a sequence.
32
 
33
  ## Limitations and Bias
34
- Because I have a rather weak computer for machine learning, I was not able to train this model for too long. The model may output irrelevant answers, or even sometimes the responses can be nonsensical.
 
11
  - loss
12
  - reward
13
  - penalty
14
+ pipeline_tag: conversational
15
  ---
16
  # Model Card
17
  ## Model Details
 
32
  The model is evaluated based on several metrics, including loss, reward, penalty, BLEU score, and perplexity. The loss metric is calculated during training and reflects the difference between the predicted output and the actual output. The reward metric is based on the number of correct words generated by the model, while the penalty metric penalizes the model for repeating words consecutively. The BLEU score measures the similarity between the generated text and the ground truth text, while the perplexity metric measures how well the model is able to predict the next word in a sequence.
33
 
34
  ## Limitations and Bias
35
+ Because I have a rather weak computer for machine learning, I was not able to train this model for too long. The model may output irrelevant answers, or even sometimes the responses can be nonsensical.