python train_cooccurence.py --input_file ~/data/wikipedia_proto/token.cooccur.pb.b64.bz2 --token_dictionary ~/data/wikipedia_proto/token.dict.pb.b64.bz2 --max_terms 20 --terms "news,apple,computer,physics,neural,democracy" --tensorboard_dir ~/data/wikipedia_training/logs --embedding_dim 64 --batch_size=65536 --steps_per_epoch=1000 --validation_steps=300 --checkpoint_dir ~/data/wikipedia_training/glove.{epoch:05d}-{val_loss:.5f}.hdf5 --num_epochs=10
