Text2Text Generation
Transformers
PyTorch
bart
feature-extraction
abertsch commited on
Commit
23e541d
1 Parent(s): 265086b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -4,9 +4,12 @@ datasets:
4
  - urialon/gov_report_validation
5
  - urialon/gov_report_test
6
  pipeline_tag: text2text-generation
 
7
  ---
8
  Model from the preprint [Unlimiformer: Long-Range Transformers with Unlimited Length Input](https://arxiv.org/abs/2305.01625)
9
 
10
  This is a BART-base model finetuned using Unlimiformer-aware early stopping, as described in section 3.1 of the paper. The model was finetuned on GovReport using the data processing pipeline from SLED; to load the validation or test set for use with these model, please use the datasets [urialon/gov_report_validation](https://huggingface.co/datasets/urialon/gov_report_validation) and [urialon/gov_report_test](https://huggingface.co/datasets/urialon/gov_report_test).
11
 
12
- This is generally a weaker model than the [alternating-training model](https://huggingface.co/abertsch/unlimiformer-bart-govreport-alternating) and a stronger model than the [baseline](https://huggingface.co/abertsch/bart-base-govreport).
 
 
4
  - urialon/gov_report_validation
5
  - urialon/gov_report_test
6
  pipeline_tag: text2text-generation
7
+ inference: false
8
  ---
9
  Model from the preprint [Unlimiformer: Long-Range Transformers with Unlimited Length Input](https://arxiv.org/abs/2305.01625)
10
 
11
  This is a BART-base model finetuned using Unlimiformer-aware early stopping, as described in section 3.1 of the paper. The model was finetuned on GovReport using the data processing pipeline from SLED; to load the validation or test set for use with these model, please use the datasets [urialon/gov_report_validation](https://huggingface.co/datasets/urialon/gov_report_validation) and [urialon/gov_report_test](https://huggingface.co/datasets/urialon/gov_report_test).
12
 
13
+ This is generally a weaker model than the [alternating-training model](https://huggingface.co/abertsch/unlimiformer-bart-govreport-alternating) and a stronger model than the [baseline](https://huggingface.co/abertsch/bart-base-govreport).
14
+
15
+ *The inference demo is disabled because you must add the Unlimiformer files to your repo before this model can handle unlimited length input!* See the [Unlimiformer GitHub](https://github.com/abertsch72/unlimiformer) for setup instructions.