hdlm-base-epsilon-0.0 / config.json
nimafathi's picture
Upload HDLM model with complete HF integration
30ca731 verified
{
"ngpus": 4,
"type": "aligned",
"gradient_accumulation_steps": 2,
"tokenizer": {
"tokens": 50257,
"model": "gpt2"
},
"training": {
"batch_size": 128,
"accum": 2,
"n_iters": 1250000,
"snapshot_freq": 10000,
"log_freq": 500,
"eval_freq": 10000,
"snapshot_freq_for_preemption": 3000,
"snapshot_sampling": true,
"ema": 0.9999,
"warmup_iter": -1,
"loss_type": "hybrid",
"epsilon": 0.0,
"lambda": 0.0
},
"data": {
"train": "openwebtext-train",
"valid": "wikitext103",
"cache_dir": "/home/toolkit/research-diffcodegen/data",
"debug": false
},
"graph": {
"type": "absorb",
"gamma": 1.0,
"file": "/home/toolkit/research-diffcodegen/data",
"report_all": false,
"expanded_sigma": true
},
"noise": {
"type": "loglinear",
"sigma_min": 0.0001,
"sigma_max": 2.0,
"ar_diffusion": false,
"expanded_sigma": true
},
"sampling": {
"predictor": "analytic",
"steps_per_level": 1,
"noise_removal": true,
"strategy": "direct",
"strategy_param": 0.9
},
"annealing": {
"type": "none",
"efficient": false,
"width": 1024,
"tau": 1024,
"eval_tau": 1024,
"steps_per_level": 1,
"sampling_method": "sdlm",
"diffusion_loss_weight": 1.0,
"ce_loss_weight": 1.0,
"sampling_eps": 0.0001,
"attention": {
"context_type": "block_causal",
"block_type": "full"
},
"match_inference": false
},
"eval": {
"batch_size": 16,
"perplexity": true,
"perplexity_batch_size": 8
},
"optim": {
"weight_decay": 0.1,
"optimizer": "AdamW",
"lr": 0.0002,
"beta1": 0.9,
"beta2": 0.95,
"eps": 1e-08,
"warmup": 10000,
"grad_clip": 1.0,
"scheduler": "cosine"
},
"experiment": {
"name": "MDLM",
"wandb_project": "Hybrid-SDLM-ALIGNED"
},
"model": {
"name": "HDLM",
"type": "ddit",
"hidden_size": 768,
"cond_dim": 128,
"length": 1024,
"n_blocks": 12,
"n_heads": 12,
"dropout": 0.1,
"scale_by_sigma": false,
"transformer_sigma_conditioning": false,
"hybrid_sigma_embedding": false,
"post_process_logits": false,
"use_timestep_embedding": false
},
"model_type": "epsilon_hybrid"
}