Aung2024 commited on
Commit
0a13a5c
1 Parent(s): 59eab16

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +39 -32
config.json CHANGED
@@ -1,33 +1,40 @@
1
  {
2
- "alpha_pattern": {},
3
- "auto_mapping": null,
4
- "base_model_name_or_path": "EleutherAI/gpt-j-6B",
5
- "bias": "none",
6
- "fan_in_fan_out": false,
7
- "inference_mode": true,
8
- "init_lora_weights": true,
9
- "layer_replication": null,
10
- "layers_pattern": null,
11
- "layers_to_transform": null,
12
- "loftq_config": {},
13
- "lora_alpha": 32,
14
- "lora_dropout": 0.05,
15
- "megatron_config": null,
16
- "megatron_core": "megatron.core",
17
- "modules_to_save": [
18
- "qa_outputs"
19
- ],
20
- "peft_type": "LORA",
21
- "r": 16,
22
- "rank_pattern": {},
23
- "revision": null,
24
- "target_modules": [
25
- "k_proj",
26
- "out_proj",
27
- "v_proj",
28
- "q_proj"
29
- ],
30
- "task_type": "QUESTION_ANS",
31
- "use_dora": false,
32
- "use_rslora": false
33
- }
 
 
 
 
 
 
 
 
1
  {
2
+ "model_type": "gptj",
3
+ "vocab_size": 50257,
4
+ "n_positions": 2048,
5
+ "n_ctx": 2048,
6
+ "n_embd": 4096,
7
+ "n_layer": 28,
8
+ "n_head": 16,
9
+ "rotary_dim": 64,
10
+ "initializer_range": 0.02,
11
+ "layer_norm_epsilon": 1e-5,
12
+ "bos_token_id": 50256,
13
+ "eos_token_id": 50256,
14
+ "pad_token_id": 50256,
15
+ "task_specific_params": {
16
+ "text-generation": {
17
+ "do_sample": true,
18
+ "max_length": 50
19
+ }
20
+ },
21
+ "max_length": 20,
22
+ "min_length": 0,
23
+ "do_sample": false,
24
+ "early_stopping": false,
25
+ "num_beams": 1,
26
+ "temperature": 1.0,
27
+ "top_k": 50,
28
+ "top_p": 1.0,
29
+ "repetition_penalty": 1.0,
30
+ "length_penalty": 1.0,
31
+ "no_repeat_ngram_size": 0,
32
+ "encoder_no_repeat_ngram_size": 0,
33
+ "num_return_sequences": 1,
34
+ "chunk_size_feed_forward": 0,
35
+ "output_scores": false,
36
+ "return_dict_in_generate": false,
37
+ "forced_bos_token_id": null,
38
+ "forced_eos_token_id": null,
39
+ "remove_invalid_values": false
40
+ }