monologg commited on
Commit
ceacda4
1 Parent(s): 9419368

Move logo's position

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -9,12 +9,12 @@ widget:
9
 
10
  # KoBigBird
11
 
 
 
12
  Pretrained BigBird Model for Korean (**kobigbird-bert-base**)
13
 
14
  ## About
15
 
16
- <img style="padding-right: 20px" src="https://user-images.githubusercontent.com/28896432/140442206-e34b02d5-e279-47e5-9c2a-db1278b1c14d.png" width="100px" align="left" />
17
-
18
  BigBird, is a sparse-attention based transformer which extends Transformer based models, such as BERT to much longer sequences.
19
 
20
  BigBird relies on **block sparse attention** instead of normal attention (i.e. BERT's attention) and can handle sequences up to a length of 4096 at a much lower compute cost compared to BERT.
 
9
 
10
  # KoBigBird
11
 
12
+ <img src="https://user-images.githubusercontent.com/28896432/140442206-e34b02d5-e279-47e5-9c2a-db1278b1c14d.png" width="200"/>
13
+
14
  Pretrained BigBird Model for Korean (**kobigbird-bert-base**)
15
 
16
  ## About
17
 
 
 
18
  BigBird, is a sparse-attention based transformer which extends Transformer based models, such as BERT to much longer sequences.
19
 
20
  BigBird relies on **block sparse attention** instead of normal attention (i.e. BERT's attention) and can handle sequences up to a length of 4096 at a much lower compute cost compared to BERT.