andreiujica commited on
Commit
2684b92
1 Parent(s): de03f06

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -5
README.md CHANGED
@@ -60,15 +60,16 @@ It achieves the following results on the evaluation set:
60
 
61
  ## Model description
62
 
63
- More information needed
64
 
65
  ## Intended uses & limitations
66
 
67
- More information needed
 
 
68
 
69
- ## Training and evaluation data
70
-
71
- More information needed
72
 
73
  ## Training procedure
74
 
 
60
 
61
  ## Model description
62
 
63
+ This model is a long-form text generation model based on the Longformer Encoder-Decoder (LED) architecture. The LED architecture extends the Transformer model to handle long documents by incorporating sparse attention mechanisms. This makes it suitable for tasks such as summarization of lengthy patent documents, where traditional models might struggle with context length limitations. The model has been fine-tuned on the BigPatent dataset, a large collection of patent documents, to enhance its performance in generating concise and informative summaries.
64
 
65
  ## Intended uses & limitations
66
 
67
+ Intended uses
68
+ - Patent summarization: Generate concise summaries of patent documents.
69
+ - Long document summarization: Useful for summarizing other types of long-form documents beyond patents.
70
 
71
+ Limitations
72
+ - Context length: Although LED handles long documents better than standard Transformers, extremely lengthy documents might still present challenges.
 
73
 
74
  ## Training procedure
75