Text Generation
Transformers
English
codegen
Inference Endpoints
mhhmm commited on
Commit
0e1a117
1 Parent(s): f02a016

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +12 -0
config.json CHANGED
@@ -4,6 +4,18 @@
4
  "CodeGenForCausalLM"
5
  ],
6
  "tokenizer_class": "GPT2Tokenizer",
 
 
 
 
 
 
 
 
 
 
 
 
7
  "bias": "none",
8
  "fan_in_fan_out": false,
9
  "inference_mode": true,
 
4
  "CodeGenForCausalLM"
5
  ],
6
  "tokenizer_class": "GPT2Tokenizer",
7
+ "attention_probs_dropout_prob": 0.0,
8
+ "attn_pdrop": 0.0,
9
+ "bos_token_id": 0,
10
+ "embd_pdrop": 0.0,
11
+ "eos_token_id": 0,
12
+ "model_type": "codegen",
13
+ "pad_token_id": 1023,
14
+ "resid_pdrop": 0.0,
15
+ "rotary_dim": 4,
16
+ "scale_attn_weights": true,
17
+ "tie_word_embeddings": false,
18
+ "torch_dtype": "float32",
19
  "bias": "none",
20
  "fan_in_fan_out": false,
21
  "inference_mode": true,