| import torch | |
| if torch.cuda.is_available(): | |
| print('Using GPU') | |
| DEVICE = 'cuda' | |
| else: | |
| print('CUDA not available. Please connect to a GPU instance if possible.') | |
| DEVICE = 'cpu' | |
| XMEM_CONFIG = { | |
| 'top_k': 30, | |
| 'mem_every': 10, | |
| 'deep_update_every': -1, | |
| 'enable_long_term': True, | |
| 'enable_long_term_count_usage': True, | |
| 'num_prototypes': 128, | |
| 'min_mid_term_frames': 5, | |
| 'max_mid_term_frames': 10, | |
| 'max_long_term_elements': 5000, | |
| 'size': 480, | |
| 'device': DEVICE, | |
| } | |