Update README.md
Browse files
README.md
CHANGED
|
@@ -27,7 +27,11 @@ inference: false
|
|
| 27 |
|
| 28 |
# bart-large-cnn-finetuned-samsum-lora
|
| 29 |
|
| 30 |
-
This model is a further fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the [samsum](https://huggingface.co/datasets/samsum) dataset.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
|
| 32 |
## Model description
|
| 33 |
|
|
@@ -62,7 +66,7 @@ Input texts longer than 4000 words can be converted into a list of sub-texts wit
|
|
| 62 |
Even though the model checkpoint is small, a huge input would crash the memory. Batching the inputs is advised.
|
| 63 |
Rule of thumb is that T4 can handle at a time a list of upto 14-15 elements, with each elements having 4000 words.
|
| 64 |
|
| 65 |
-
Note 'max_new_tokens=60' is used to limit the summary size. BART model has max generation length = 142 (default) and min generation length = 56.
|
| 66 |
|
| 67 |
```python
|
| 68 |
import torch
|
|
|
|
| 27 |
|
| 28 |
# bart-large-cnn-finetuned-samsum-lora
|
| 29 |
|
| 30 |
+
This model is a further fine-tuned version of [facebook/bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) on the [samsum](https://huggingface.co/datasets/samsum) dataset.
|
| 31 |
+
|
| 32 |
+
The base model [bart-large-cnn](https://huggingface.co/facebook/bart-large-cnn) is a fine-tuned verstion of BART model on the [CNN Daily Mail](https://huggingface.co/datasets/cnn_dailymail) dataset.
|
| 33 |
+
|
| 34 |
+
Check out [sooolee/flan-t5-base-cnn-samsum-lora](https://huggingface.co/sooolee/flan-t5-base-cnn-samsum-lora) the model fine-tuned for the same purpose.
|
| 35 |
|
| 36 |
## Model description
|
| 37 |
|
|
|
|
| 66 |
Even though the model checkpoint is small, a huge input would crash the memory. Batching the inputs is advised.
|
| 67 |
Rule of thumb is that T4 can handle at a time a list of upto 14-15 elements, with each elements having 4000 words.
|
| 68 |
|
| 69 |
+
Note 'max_new_tokens=60' is used in the example below to limit the summary size. BART model has max generation length = 142 (default) and min generation length = 56.
|
| 70 |
|
| 71 |
```python
|
| 72 |
import torch
|