Spaces:
Runtime error
Runtime error
2.10.1
hf-hub:org/model_id
support for loading models w/ config and weights in Hugging Face Hub
2.10.0
- Added a ViT-bigG-14 model.
- Added an up-to-date example slurm script for large training jobs.
- Added a option to sync logs and checkpoints to S3 during training.
- New options for LR schedulers, constant and constant with cooldown
- Fix wandb autoresuming when resume is not set
- ConvNeXt
base
&base_w
pretrained models added timm-
model prefix removed from configstimm
augmentation + regularization (dropout / drop-path) supported
2.9.3
- Fix wandb collapsing multiple parallel runs into a single one
2.9.2
- Fix braceexpand memory explosion for complex webdataset urls
2.9.1
- Fix release
2.9.0
- Add training feature to auto-resume from the latest checkpoint on restart via
--resume latest
- Allow webp in webdataset
- Fix logging for number of samples when using gradient accumulation
- Add model configs for convnext xxlarge
2.8.2
- wrapped patchdropout in a torch.nn.Module
2.8.1
- relax protobuf dependency
- override the default patch dropout value in 'vision_cfg'
2.8.0
- better support for HF models
- add support for gradient accumulation
- CI fixes
- add support for patch dropout
- add convnext configs
2.7.0
- add multilingual H/14 xlm roberta large
2.6.1
- fix setup.py _read_reqs
2.6.0
- Make openclip training usable from pypi.
- Add xlm roberta large vit h 14 config.
2.5.0
- pretrained B/32 xlm roberta base: first multilingual clip trained on laion5B
- pretrained B/32 roberta base: first clip trained using an HF text encoder
2.4.1
- Add missing hf_tokenizer_name in CLIPTextCfg.
2.4.0
- Fix #211, missing RN50x64 config. Fix type of dropout param for ResNet models
- Bring back LayerNorm impl that casts to input for non bf16/fp16
- zero_shot.py: set correct tokenizer based on args
- training/params.py: remove hf params and get them from model config
2.3.1
- Implement grad checkpointing for hf model.
- custom_text: True if hf_model_name is set
- Disable hf tokenizer parallelism
2.3.0
- Generalizable Text Transformer with HuggingFace Models (@iejMac)
2.2.0
- Support for custom text tower
- Add checksum verification for pretrained model weights
2.1.0
- lot including sota models, bfloat16 option, better loading, better metrics
1.2.0
- ViT-B/32 trained on Laion2B-en
- add missing openai RN50x64 model
1.1.1
- ViT-B/16+
- Add grad checkpointing support
- more robust data loader