| | from typing import Callable |
| |
|
| | import dotenv |
| | import hydra |
| | from omegaconf import OmegaConf, DictConfig |
| |
|
| | |
| | |
| | dotenv.load_dotenv(override=True) |
| |
|
| | OmegaConf.register_new_resolver('eval', eval) |
| | OmegaConf.register_new_resolver('div_up', lambda x, y: (x + y - 1) // y) |
| | |
| | |
| | OmegaConf.register_new_resolver('datamodule', lambda attr: '${datamodule:' + str(attr) + '}') |
| |
|
| | |
| | import torch.backends |
| | torch.backends.cuda.matmul.allow_tf32 = True |
| | torch.backends.cudnn.allow_tf32 = True |
| |
|
| |
|
| | def dictconfig_filter_key(d: DictConfig, fn: Callable) -> DictConfig: |
| | """Only keep keys where fn(key) is True. Support nested DictConfig. |
| | """ |
| | |
| | |
| | return DictConfig({k: dictconfig_filter_key(v, fn) if isinstance(v, DictConfig) else v |
| | |
| | for k, v in d.items() if fn(k)}) |
| |
|
| |
|
| | @hydra.main(config_path="configs/", config_name="config.yaml") |
| | def main(config: DictConfig): |
| |
|
| | |
| | |
| | config = dictconfig_filter_key(config, lambda k: not k.startswith('__')) |
| |
|
| | |
| | |
| | from src.train import train |
| | from src.eval import evaluate |
| | from src.utils import utils |
| |
|
| | |
| | |
| | |
| | |
| | |
| | utils.extras(config) |
| |
|
| | |
| | if config.get("print_config"): |
| | utils.print_config(config, resolve=True) |
| |
|
| | |
| | mode = config.get('mode', 'train') |
| | if mode not in ['train', 'eval']: |
| | raise NotImplementedError(f'mode {mode} not supported') |
| | if mode == 'train': |
| | return train(config) |
| | elif mode == 'eval': |
| | return evaluate(config) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|