pszemraj commited on
Commit
3df132e
1 Parent(s): 0ce67b2

update demo link

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -280,15 +280,14 @@ model-index:
280
 
281
  # Longformer Encoder-Decoder (LED) fine-tuned on Booksum
282
 
283
- demo:
284
 
285
- [![colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/gist/pszemraj/d9a0495861776168fd5cdcd7731bc4ee/example-long-t5-tglobal-base-16384-book-summary.ipynb)
286
 
287
  - A fine-tuned version of [allenai/led-large-16384](https://huggingface.co/allenai/led-large-16384) on the BookSum dataset.
288
- - Goal: a model that can generalize well and is useful in summarizing long text in academic and daily usage.
289
  - works well on lots of text and can handle 16384 tokens/batch (_if you have the GPU memory to handle that_)
290
 
291
- > Note: the API is set to generate a max of 64 tokens for runtime reasons, so the summaries may be truncated (depending on length of input text). For best results use python as below.
292
 
293
  ---
294
 
@@ -366,7 +365,7 @@ The following hyperparameters were used during training:
366
 
367
  #### In-between Epochs
368
 
369
- Unfortunately, don't have all records on-hand for middle epochs, the following should be representative:
370
 
371
  - learning_rate: 4e-05
372
  - train_batch_size: 2
 
280
 
281
  # Longformer Encoder-Decoder (LED) fine-tuned on Booksum
282
 
 
283
 
284
+ [![colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/gist/pszemraj/3eba944ddc9fc9a4a1bfb21e83b57620/summarization-token-batching.ipynb)
285
 
286
  - A fine-tuned version of [allenai/led-large-16384](https://huggingface.co/allenai/led-large-16384) on the BookSum dataset.
287
+ - Goal: a model that can generalize well and is useful in summarizing long text in academic and daily usage. See the demo linked above!
288
  - works well on lots of text and can handle 16384 tokens/batch (_if you have the GPU memory to handle that_)
289
 
290
+ > Note: the API is set to generate a max of 64 tokens for runtime reasons, so the summaries may be truncated (depending on the length of input text). For best results use python as below.
291
 
292
  ---
293
 
 
365
 
366
  #### In-between Epochs
367
 
368
+ Unfortunately, don't have all records on-hand for middle epochs; the following should be representative:
369
 
370
  - learning_rate: 4e-05
371
  - train_batch_size: 2