runtime error

: 'bert', 'use_custom_optimizer': 'adamw', 'output_learning_rate': 0.0001, 'arch_type': 2, 'add_label_name': True, 'normalize_embeddings': False, 'tie_weights': False, 'coil': True, 'colbert': False, 'token_dim': 16, 'label_frozen_layers': 2, 'do_train': True, 'do_eval': True, 'do_predict': False, 'per_device_train_batch_size': 1, 'gradient_accumulation_steps': 8, 'per_device_eval_batch_size': 1, 'learning_rate': 5e-05, 'num_train_epochs': 2, 'save_steps': 4900, 'evaluation_strategy': 'steps', 'eval_steps': 3000000, 'fp16': False, 'lr_scheduler_type': 'linear', 'dataloader_num_workers': 16, 'label_names': ['labels'], 'scenario': 'unseen_labels', 'ddp_find_unused_parameters': False, 'ignore_data_skip': True, 'seed': -1, 'output_dir': 'demo_tmp'} -------------------------------------------------------------------------------- Traceback (most recent call last): File "/home/user/app/app.py", line 12, in <module> model = DemoModel() File "/home/user/app/model.py", line 98, in __init__ self.model_args, self.data_args, self.training_args = parser.parse_dict(read_yaml_config(ARGS_FILE, output_dir = 'demo_tmp', extra_args = {})) File "/home/user/.local/lib/python3.10/site-packages/transformers/hf_*********.py", line 381, in parse_dict obj = dtype(**inputs) File "<string>", line 115, in __init__ File "/home/user/.local/lib/python3.10/site-packages/transformers/training_args.py", line 1340, in __post_init__ and (self.device.type != "cuda") File "/home/user/.local/lib/python3.10/site-packages/transformers/training_args.py", line 1764, in device return self._setup_devices File "/home/user/.local/lib/python3.10/site-packages/transformers/utils/generic.py", line 54, in __get__ cached = self.fget(obj) File "/home/user/.local/lib/python3.10/site-packages/transformers/training_args.py", line 1672, in _setup_devices raise ImportError( ImportError: Using the `Trainer` with `PyTorch` requires `accelerate>=0.20.1`: Please run `pip install transformers[torch]` or `pip install accelerate -U`

Container logs:

Fetching error logs...