talalH commited on
Commit
30cbcdb
1 Parent(s): fff52f0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -7,7 +7,7 @@
7
  # Model
8
 
9
  <!-- Provide a quick summary of what the model is/does. -->
10
- This model card provides information about a fine-tuned T5 base model that has been specifically trained for generating summaries. The model utilizes transfer learning techniques and is based on a subset of the XSum and ChatGPT datasets. We have made some key modifications to the training process to optimize the model's performance and provide the best possible summaries, particularly supporting greater length outputs.
11
 
12
  **Dataset and Training:**
13
  The fine-tuned T5 base model is trained on a carefully curated subset of the XSum and ChatGPT datasets. These datasets contain a wide range of text samples, including news articles and conversational data. By utilizing this diverse data, the model gains a broader understanding of language and improves its ability to generate accurate and coherent summaries.
 
7
  # Model
8
 
9
  <!-- Provide a quick summary of what the model is/does. -->
10
+ This model card provides information about a fine-tuned T5 base model that has been specifically trained for generating summaries. We have made some key modifications to the training process to optimize the model's performance and provide the best possible summaries, particularly supporting greater length outputs. One notable difference between this model and other similar models is that it is trained on the target output length of 512. This means that the model is explicitly trained to generate summaries that are up to 512 tokens long. By focusing on this target output length, we aim to provide summaries that are more comprehensive and informative, while still maintaining a reasonable length for large text.
11
 
12
  **Dataset and Training:**
13
  The fine-tuned T5 base model is trained on a carefully curated subset of the XSum and ChatGPT datasets. These datasets contain a wide range of text samples, including news articles and conversational data. By utilizing this diverse data, the model gains a broader understanding of language and improves its ability to generate accurate and coherent summaries.