jiawenchim commited on
Commit
7ae6381
1 Parent(s): 9aa4f9c

Upload 7 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint/ckpt-0.data-00000-of-00001 filter=lfs diff=lfs merge=lfs -text
checkpoint/checkpoint ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ model_checkpoint_path: "ckpt-0"
2
+ all_model_checkpoint_paths: "ckpt-0"
3
+ all_model_checkpoint_timestamps: 1704519219.3977096
4
+ last_preserved_timestamp: 1704519218.1877174
checkpoint/ckpt-0.data-00000-of-00001 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43f10ec8766a8d3c7d866a52aa103caa8a6a61360c3dfa68b71e1425806018cd
3
+ size 43710540
checkpoint/ckpt-0.index ADDED
Binary file (22.7 kB). View file
 
pipeline.config ADDED
@@ -0,0 +1,188 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model {
2
+ ssd {
3
+ num_classes: 2
4
+ image_resizer {
5
+ fixed_shape_resizer {
6
+ height: 640
7
+ width: 640
8
+ }
9
+ }
10
+ feature_extractor {
11
+ type: "ssd_mobilenet_v1_fpn_keras"
12
+ depth_multiplier: 1.0
13
+ min_depth: 16
14
+ conv_hyperparams {
15
+ regularizer {
16
+ l2_regularizer {
17
+ weight: 4e-05
18
+ }
19
+ }
20
+ initializer {
21
+ random_normal_initializer {
22
+ mean: 0.0
23
+ stddev: 0.01
24
+ }
25
+ }
26
+ activation: RELU_6
27
+ batch_norm {
28
+ decay: 0.997
29
+ scale: true
30
+ epsilon: 0.001
31
+ }
32
+ }
33
+ override_base_feature_extractor_hyperparams: true
34
+ fpn {
35
+ min_level: 3
36
+ max_level: 7
37
+ }
38
+ }
39
+ box_coder {
40
+ faster_rcnn_box_coder {
41
+ y_scale: 10.0
42
+ x_scale: 10.0
43
+ height_scale: 5.0
44
+ width_scale: 5.0
45
+ }
46
+ }
47
+ matcher {
48
+ argmax_matcher {
49
+ matched_threshold: 0.5
50
+ unmatched_threshold: 0.5
51
+ ignore_thresholds: false
52
+ negatives_lower_than_unmatched: true
53
+ force_match_for_each_row: true
54
+ use_matmul_gather: true
55
+ }
56
+ }
57
+ similarity_calculator {
58
+ iou_similarity {
59
+ }
60
+ }
61
+ box_predictor {
62
+ weight_shared_convolutional_box_predictor {
63
+ conv_hyperparams {
64
+ regularizer {
65
+ l2_regularizer {
66
+ weight: 4e-05
67
+ }
68
+ }
69
+ initializer {
70
+ random_normal_initializer {
71
+ mean: 0.0
72
+ stddev: 0.01
73
+ }
74
+ }
75
+ activation: RELU_6
76
+ batch_norm {
77
+ decay: 0.997
78
+ scale: true
79
+ epsilon: 0.001
80
+ }
81
+ }
82
+ depth: 256
83
+ num_layers_before_predictor: 4
84
+ kernel_size: 3
85
+ class_prediction_bias_init: -4.6
86
+ }
87
+ }
88
+ anchor_generator {
89
+ multiscale_anchor_generator {
90
+ min_level: 3
91
+ max_level: 7
92
+ anchor_scale: 4.0
93
+ aspect_ratios: 1.0
94
+ aspect_ratios: 2.0
95
+ aspect_ratios: 0.5
96
+ scales_per_octave: 2
97
+ }
98
+ }
99
+ post_processing {
100
+ batch_non_max_suppression {
101
+ score_threshold: 1e-08
102
+ iou_threshold: 0.6
103
+ max_detections_per_class: 100
104
+ max_total_detections: 100
105
+ use_static_shapes: false
106
+ }
107
+ score_converter: SIGMOID
108
+ }
109
+ normalize_loss_by_num_matches: true
110
+ loss {
111
+ localization_loss {
112
+ weighted_smooth_l1 {
113
+ }
114
+ }
115
+ classification_loss {
116
+ weighted_sigmoid_focal {
117
+ gamma: 2.0
118
+ alpha: 0.25
119
+ }
120
+ }
121
+ classification_weight: 1.0
122
+ localization_weight: 1.0
123
+ }
124
+ encode_background_as_zeros: true
125
+ normalize_loc_loss_by_codesize: true
126
+ inplace_batchnorm_update: true
127
+ freeze_batchnorm: false
128
+ }
129
+ }
130
+ train_config {
131
+ batch_size: 10
132
+ data_augmentation_options {
133
+ random_horizontal_flip {
134
+ }
135
+ }
136
+ data_augmentation_options {
137
+ random_crop_image {
138
+ min_object_covered: 0.0
139
+ min_aspect_ratio: 0.75
140
+ max_aspect_ratio: 3.0
141
+ min_area: 0.75
142
+ max_area: 1.0
143
+ overlap_thresh: 0.0
144
+ }
145
+ }
146
+ sync_replicas: true
147
+ optimizer {
148
+ momentum_optimizer {
149
+ learning_rate {
150
+ cosine_decay_learning_rate {
151
+ learning_rate_base: 0.03
152
+ total_steps: 25000
153
+ warmup_learning_rate: 0.013333
154
+ warmup_steps: 2000
155
+ }
156
+ }
157
+ momentum_optimizer_value: 0.9
158
+ }
159
+ use_moving_average: false
160
+ }
161
+ fine_tune_checkpoint: "pretrained_models/ssd_mobilenet_v1_fpn_640x640_coco17_tpu-8/checkpoint/ckpt-0"
162
+ num_steps: 25000
163
+ startup_delay_steps: 0.0
164
+ replicas_to_aggregate: 8
165
+ max_number_of_boxes: 100
166
+ unpad_groundtruth_tensors: false
167
+ fine_tune_checkpoint_type: "detection"
168
+ fine_tune_checkpoint_version: V2
169
+ }
170
+ train_input_reader {
171
+ label_map_path: "/home/ubuntu/iti107_assignment/data/label_map.pbtxt"
172
+ tf_record_input_reader {
173
+ input_path: "/home/ubuntu/iti107_assignment/data/train.record"
174
+ }
175
+ }
176
+ eval_config {
177
+ metrics_set: "coco_detection_metrics"
178
+ use_moving_averages: false
179
+ batch_size: 1
180
+ }
181
+ eval_input_reader {
182
+ label_map_path: "/home/ubuntu/iti107_assignment/data/label_map.pbtxt"
183
+ shuffle: false
184
+ num_epochs: 1
185
+ tf_record_input_reader {
186
+ input_path: "/home/ubuntu/iti107_assignment/data/val.record"
187
+ }
188
+ }
saved_model/variables/variables.data-00000-of-00001 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b17e13201ad022089567aa1483d28f2634cb9f095799b5af8346891b2f1624ed
3
+ size 43765991
saved_model/variables/variables.index ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3a07da4407e12e8c3428cf46c48f6f1b036b06c4446a0e2c6a95ddccccadb30
3
+ size 19220