{ "architectures": [ "GPT2LMHeadModel" ], "auto_map": { "AutoConfig": "GPT2Tokenizer", "AutoModel": "GPT2LMHeadModel" }, "batch_size": 12, "block_size": 100, "device": "cpu", "eval_interval": 250, "hidden_dropout_prob": 0.0, "n_embd": 300, "n_head": 6, "n_layer": 6, "learning_rate": 0.001, "max_iters": 6000, "torch_dtype": "float16", "transformers_version": "4.33.2", "vocab_size": 1000 }