yzhang511 commited on
Commit
0cb69d7
1 Parent(s): 9dfe68c

Upload config.yaml

Browse files
Files changed (1) hide show
  1. config.yaml +97 -0
config.yaml ADDED
@@ -0,0 +1,97 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model_class: NDT1
2
+
3
+
4
+ encoder:
5
+
6
+ from_pt: null
7
+ stitching: false
8
+
9
+ masker:
10
+ force_active: true
11
+ mode: temporal
12
+ ratio: 0.3 # ratio of data to predict
13
+ zero_ratio: 1.0 # of the data to predict, ratio of zeroed out
14
+ random_ratio: 1.0 # of the not zeroed, ratio of randomly replaced
15
+ expand_prob: 0.0 # probability of expanding the mask in ``temporal`` mode
16
+ max_timespan: 1 # max span of mask if expanded
17
+ channels: null # neurons to mask in "co-smoothing" mode
18
+ timesteps: null # time steps to mask in ``forward-pred`` mode
19
+ mask_regions: ['all'] # brain regions to mask in ``inter-region`` mode
20
+ target_regions: ['all'] # brain regions to predict in ``intra-region`` mode
21
+ n_mask_regions: 1 # num of regions to choose from the list of mask_regions or target_regions
22
+
23
+ # context available for each timestep
24
+ context:
25
+ forward: -1
26
+ backward: -1
27
+
28
+ norm_and_noise:
29
+ active: false
30
+ smooth_sd: 2 # gaussian smoohing
31
+ norm: "zscore" # which normalization layer to use (null/layernorm/scalenorm/zscore)
32
+ eps: 1.e-7 # avoid dividing by zero when normalizing padded spikes
33
+ white_noise_sd: 1.0 # gaussian noise added to the inputs 1.0 originally
34
+ constant_offset_sd: 0.2 # gaussian noise added to the inputs but contsnat in the time dimension 0.2 originally
35
+
36
+ embedder:
37
+ n_channels: 668 # number of neurons recorded
38
+ n_blocks: 24 # number of blocks of experiments
39
+ n_dates: 24 # number of days of experiments
40
+ max_F: 100 # max feature len in timesteps
41
+
42
+ mode: linear # linear/embed/identity
43
+ mult: 2 # embedding multiplier. hiddden_sizd = n_channels * mult
44
+ adapt: false # adapt the embedding layer for each day
45
+ pos: true # embed position
46
+ act: softsign # activation for the embedding layers
47
+ scale: 1 # scale the embedding multiplying by this number
48
+ bias: true # use bias in the embedding layer
49
+ dropout: 0.2 # dropout in embedding layer
50
+
51
+ fixup_init: false # modify weight initialization
52
+ init_range: 0.1 # initialization range for embeddings
53
+ spike_log_init: false # special initialization
54
+ max_spikes: 0 # max number of spikes in a single time bin
55
+
56
+ tokenize_binary_mask: false
57
+ use_prompt: false
58
+ use_session: false
59
+
60
+ stack:
61
+ active: false # wether to stack consecutive timesteps
62
+ size: 32 # number of consecutive timesteps to stack
63
+ stride: 4 # stacking stride
64
+
65
+
66
+ transformer:
67
+ n_layers: 5 # number of transformer layers
68
+ hidden_size: 512 # hidden space of the transformer
69
+ use_scalenorm: false # use scalenorm instead of layernorm
70
+ use_rope: false # use rotary postional encoding
71
+ rope_theta: 10000.0 # rope angle of rotation
72
+
73
+
74
+ n_heads: 8 # number of attentiomn heads
75
+ attention_bias: true # learn bias in the attention layers
76
+
77
+ act: gelu # activiation function in mlp layers
78
+ inter_size: 1024 # intermediate dimension in the mlp layers
79
+ mlp_bias: true # learn bias in the mlp layers
80
+
81
+ dropout: 0.4 # dropout in transformer layers
82
+ fixup_init: true # modify weight initialization
83
+
84
+
85
+ factors:
86
+ active: false # project from hidden_size to factors
87
+ size: 8 # factors size
88
+ act: relu # activation function after projecting to factors
89
+ bias: true # use bias in projection to factors
90
+ dropout: 0.0 # dropout in projection to factors
91
+ fixup_init: false # modify weight initialization
92
+ init_range: 0.1 # initialization range for factors projetion
93
+
94
+ decoder:
95
+ from_pt: null
96
+
97
+