AgaMiko commited on
Commit
81049e6
1 Parent(s): 772dd5b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -0
README.md CHANGED
@@ -36,6 +36,12 @@ metrics:
36
 
37
  **Keywords generated with vlT5-base-keywords:** encoder-decoder architecture, keyword generation
38
 
 
 
 
 
 
 
39
  ## vlT5
40
 
41
  The biggest advantage is the transferability of the vlT5 model, as it works well on all domains and types of text. The downside is that the text length and the number of keywords are similar to the training data: the text piece of an abstract length generates approximately 3 to 5 keywords. It works both extractive and abstractively. Longer pieces of text must be split into smaller chunks, and then propagated to the model.
 
36
 
37
  **Keywords generated with vlT5-base-keywords:** encoder-decoder architecture, keyword generation
38
 
39
+ Results on demo model (different generation method, one model per language):
40
+
41
+ > Our vlT5 model is a keyword generation model based on encoder-decoder architecture using Transformer blocks presented by Google ([https://huggingface.co/t5-base](https://huggingface.co/t5-base)). The vlT5 was trained on scientific articles corpus to predict a given set of keyphrases based on the concatenation of the article’s abstract and title. It generates precise, yet not always complete keyphrases that describe the content of the article based only on the abstract.
42
+
43
+ **Keywords generated with vlT5-base-keywords:** encoder-decoder architecture, vlT5, keyword generation, scientific articles corpus
44
+
45
  ## vlT5
46
 
47
  The biggest advantage is the transferability of the vlT5 model, as it works well on all domains and types of text. The downside is that the text length and the number of keywords are similar to the training data: the text piece of an abstract length generates approximately 3 to 5 keywords. It works both extractive and abstractively. Longer pieces of text must be split into smaller chunks, and then propagated to the model.