File size: 1,329 Bytes
dd810eb 642d911 c1728bd 642d911 dd810eb 642d911 dd810eb 642d911 dd810eb 642d911 c1728bd dd810eb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 |
import torch
import argparse
import os
import wandb
import pytorch_lightning as pl
from pytorch_lightning.loggers import WandbLogger
from transformers import BartTokenizer
from idiomify.datamodules import IdiomifyDataModule
from idiomify.fetchers import fetch_config, fetch_idiomifier, fetch_tokenizer
from idiomify.paths import ROOT_DIR
def main():
parser = argparse.ArgumentParser()
parser.add_argument("--num_workers", type=int, default=os.cpu_count())
parser.add_argument("--fast_dev_run", action="store_true", default=False)
args = parser.parse_args()
config = fetch_config()['idiomifier']
config.update(vars(args))
# prepare the datamodule
with wandb.init(entity="eubinecto", project="idiomify", config=config) as run:
model = fetch_idiomifier(config['ver'], run) # fetch a pre-trained model
tokenizer = fetch_tokenizer(config['tokenizer_ver'], run)
datamodule = IdiomifyDataModule(config, tokenizer, run)
logger = WandbLogger(log_model=False)
trainer = pl.Trainer(fast_dev_run=config['fast_dev_run'],
gpus=torch.cuda.device_count(),
default_root_dir=str(ROOT_DIR),
logger=logger)
trainer.test(model, datamodule)
if __name__ == '__main__':
main()
|