Update README.md
Browse files
README.md
CHANGED
@@ -69,7 +69,12 @@ This repo contains the following configurations under `./models/`:
|
|
69 |
+ Using shuffled batches (where each batch has the same durations) and a modified `rvq_levels_p` to help the NAR.
|
70 |
+ This model received LayerSkip-aware training, with layer dropout and early-exit loss to help try and bolster the model and enable self-speculation sampling.
|
71 |
+ I *need* to do heavy evaluation against the base model to ensure output quality does not drop before considering replacing the base model with this.
|
|
|
72 |
+ Goal is to utilize self-speculation sampling to enable speedups when possible.
|
|
|
|
|
|
|
|
|
73 |
|
74 |
Some additional configurations have been explored with, but experiments have not been fruitful:
|
75 |
* Exotic wrappers like `BitNet` seemed to yield little gains in inferencing, somehow. The memory savings is pretty much unneccessary as the models are already manageable at ~200M parameters.
|
|
|
69 |
+ Using shuffled batches (where each batch has the same durations) and a modified `rvq_levels_p` to help the NAR.
|
70 |
+ This model received LayerSkip-aware training, with layer dropout and early-exit loss to help try and bolster the model and enable self-speculation sampling.
|
71 |
+ I *need* to do heavy evaluation against the base model to ensure output quality does not drop before considering replacing the base model with this.
|
72 |
+
+ It currently does not seem to perform better even without early-exit...
|
73 |
+ Goal is to utilize self-speculation sampling to enable speedups when possible.
|
74 |
+
+ Current implementation will early-exit if the entropy/varentropy of the logits are low enough.
|
75 |
+
+ There doesn't seem to be any significant speedups...
|
76 |
+
+ Training is a pain, as float16 + AMP will fry the model fast, and training bfloat16 (with/without AMP) seems to harm the model overall.
|
77 |
+
+ I'd like to think more time training will help, but it doesn't seem to be worth it for a marginal speedup.
|
78 |
|
79 |
Some additional configurations have been explored with, but experiments have not been fruitful:
|
80 |
* Exotic wrappers like `BitNet` seemed to yield little gains in inferencing, somehow. The memory savings is pretty much unneccessary as the models are already manageable at ~200M parameters.
|