Qubitium commited on
Commit
31ecda2
1 Parent(s): 8695694

add note about flash attn2 compat and quant

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -8,6 +8,11 @@ Special thanks to https://huggingface.co/fahadh4ilyas
8
  # use this to convert original dbrx models
9
  convert_v2.py
10
  ```
 
 
 
 
 
11
  ---
12
  inference: false
13
  license: other
 
8
  # use this to convert original dbrx models
9
  convert_v2.py
10
  ```
11
+
12
+ Known Issues:
13
+
14
+ 1. [TRAINING] padding in-compatibility with flash attention2. If you use padding, training may fail
15
+ 2. [QUANT GPTQ] testing...
16
  ---
17
  inference: false
18
  license: other