Locutusque commited on
Commit
4e43f5c
1 Parent(s): a2c4ba8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -2
README.md CHANGED
@@ -4,10 +4,13 @@ datasets:
4
  - Locutusque/ColumnedChatCombined
5
  language:
6
  - en
 
7
  metrics:
8
  - bleu
9
  - perplexity
10
  - loss
 
 
11
  ---
12
  # Model Card
13
  ## Model Details
@@ -25,7 +28,7 @@ The model is trained on a large dataset of conversational data, consisting of in
25
  The model architecture used in this model is GPT-2, a transformer-based language model that is capable of generating high-quality text with a wide range of styles and tones. The GPT-2 architecture consists of a multi-layered transformer encoder-decoder, with self-attention mechanisms that allow the model to capture long-term dependencies and generate coherent text.
26
 
27
  ## Evaluation Metrics
28
- The model is evaluated based on several metrics, including loss, reward, BLEU score, and perplexity. The loss metric is calculated during training and reflects the difference between the predicted output and the actual output. The reward metric is based on the number of correct words generated by the model, while the penalty metric penalizes the model for repeating words consecutively. The BLEU score measures the similarity between the generated text and the ground truth text, while the perplexity metric measures how well the model is able to predict the next word in a sequence.
29
 
30
  ## Limitations and Bias
31
- One limitation of this model is that it may generate responses that are biased or inappropriate, depending on the nature of the training data. Care should be taken to ensure that the training data is diverse and representative of the target user population, in order to minimize the risk of bias or discrimination. Additionally, the model may struggle with generating responses that are highly domain-specific or technical in nature, as these may require specialized knowledge or context that is not present in the training data.
 
4
  - Locutusque/ColumnedChatCombined
5
  language:
6
  - en
7
+ - chi
8
  metrics:
9
  - bleu
10
  - perplexity
11
  - loss
12
+ - reward
13
+ - penalty
14
  ---
15
  # Model Card
16
  ## Model Details
 
28
  The model architecture used in this model is GPT-2, a transformer-based language model that is capable of generating high-quality text with a wide range of styles and tones. The GPT-2 architecture consists of a multi-layered transformer encoder-decoder, with self-attention mechanisms that allow the model to capture long-term dependencies and generate coherent text.
29
 
30
  ## Evaluation Metrics
31
+ The model is evaluated based on several metrics, including loss, reward, penalty, BLEU score, and perplexity. The loss metric is calculated during training and reflects the difference between the predicted output and the actual output. The reward metric is based on the number of correct words generated by the model, while the penalty metric penalizes the model for repeating words consecutively. The BLEU score measures the similarity between the generated text and the ground truth text, while the perplexity metric measures how well the model is able to predict the next word in a sequence.
32
 
33
  ## Limitations and Bias
34
+ Because I have a rather weak computer for machine learning, I was not able to train this model for too long. The model may output irrelevant answers, or even sometimes the responses can be nonsensical.