Transformers
English
Inference Endpoints
howey commited on
Commit
300e95c
1 Parent(s): 8d82957

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -0
README.md ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ language:
5
+ - en
6
+ datasets:
7
+ - howey/unarXive
8
+ - howey/wiki_en
9
+ - howey/hupd
10
+ ---
11
+
12
+ ## Using HDT
13
+ To use the pre-trained model for masked language modeling, use the following snippet:
14
+ ```python
15
+ from transformers import AutoModelForMaskedLM, AutoTokenizer
16
+
17
+ # See the `MDLM` collection page on the hub for list of available models.
18
+ tokenizer = transformers.AutoTokenizer.from_pretrained('google-bert/bert-base-uncased')
19
+ model_name = 'howey/HDT-E'
20
+ model = AutoModelForMaskedLM.from_pretrained(model_name)
21
+ ```
22
+
23
+ For more details, please see our github repository: [HDT](https://github.com/autonomousvision/hdt)
24
+
25
+ ## Model Details
26
+ The model, which has a context length of `8192` and is similar in size to BERT with approximately `110M` parameters,
27
+ was trained on standard masked language modeling task with a Transformer-based architecture using our proposed hierarchical attention.
28
+ The training regimen comprised 24 hours on the ArXiv+Wikipedia+HUPD corpus, involving the processing of a total of `160 million` tokens.
29
+
30
+ For more details, please see our paper: [HDT: Hierarchical Document Transformer](https://arxiv.org/pdf/2407.08330).
31
+
32
+
33
+
34
+ ## Citation
35
+
36
+ <!-- If there is a paper or blog post introducing the model, the Bibtex information for that should go in this section. -->
37
+ Please cite our work using the bibtex below:
38
+
39
+ **BibTeX:**
40
+
41
+ ```
42
+ @inproceedings{He2024COLM,
43
+ title={HDT: Hierarchical Document Transformer},
44
+ author={Haoyu He and Markus Flicke and Jan Buchmann and Iryna Gurevych and Andreas Geiger},
45
+ year={2024},
46
+ booktitle={Conference on Language Modeling}
47
+ }
48
+ ```
49
+
50
+ ## Model Card Contact
51
+ Haoyu (haoyu.he@uni-tuebingen.de)