ecker commited on
Commit
6a78baa
·
verified ·
1 Parent(s): f26d377

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -0
README.md CHANGED
@@ -69,7 +69,12 @@ This repo contains the following configurations under `./models/`:
69
  + Using shuffled batches (where each batch has the same durations) and a modified `rvq_levels_p` to help the NAR.
70
  + This model received LayerSkip-aware training, with layer dropout and early-exit loss to help try and bolster the model and enable self-speculation sampling.
71
  + I *need* to do heavy evaluation against the base model to ensure output quality does not drop before considering replacing the base model with this.
 
72
  + Goal is to utilize self-speculation sampling to enable speedups when possible.
 
 
 
 
73
 
74
  Some additional configurations have been explored with, but experiments have not been fruitful:
75
  * Exotic wrappers like `BitNet` seemed to yield little gains in inferencing, somehow. The memory savings is pretty much unneccessary as the models are already manageable at ~200M parameters.
 
69
  + Using shuffled batches (where each batch has the same durations) and a modified `rvq_levels_p` to help the NAR.
70
  + This model received LayerSkip-aware training, with layer dropout and early-exit loss to help try and bolster the model and enable self-speculation sampling.
71
  + I *need* to do heavy evaluation against the base model to ensure output quality does not drop before considering replacing the base model with this.
72
+ + It currently does not seem to perform better even without early-exit...
73
  + Goal is to utilize self-speculation sampling to enable speedups when possible.
74
+ + Current implementation will early-exit if the entropy/varentropy of the logits are low enough.
75
+ + There doesn't seem to be any significant speedups...
76
+ + Training is a pain, as float16 + AMP will fry the model fast, and training bfloat16 (with/without AMP) seems to harm the model overall.
77
+ + I'd like to think more time training will help, but it doesn't seem to be worth it for a marginal speedup.
78
 
79
  Some additional configurations have been explored with, but experiments have not been fruitful:
80
  * Exotic wrappers like `BitNet` seemed to yield little gains in inferencing, somehow. The memory savings is pretty much unneccessary as the models are already manageable at ~200M parameters.