Nanobit commited on
Commit
6b50200
2 Parent(s): b9083a7 16f9e28

Merge pull request #178 from PocketDocLabs/main

Browse files

Update README.md to reflect current gradient checkpointing support

Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -387,7 +387,7 @@ train_on_inputs: false
387
  # don't use this, leads to wonky training (according to someone on the internet)
388
  group_by_length: false
389
 
390
- # does not work with current implementation of 4-bit LoRA
391
  gradient_checkpointing: false
392
 
393
  # stop training after this many evaluation losses have increased in a row
 
387
  # don't use this, leads to wonky training (according to someone on the internet)
388
  group_by_length: false
389
 
390
+ # Whether to use gradient checkpointing https://huggingface.co/docs/transformers/v4.18.0/en/performance#gradient-checkpointing
391
  gradient_checkpointing: false
392
 
393
  # stop training after this many evaluation losses have increased in a row