nanashi161382's picture
Upload v5 model
63d7664
version => v72
seq => 0
epoch => 16
test_unweighted_loss => 0.16244998574256897
test_weighted_loss => 0.7719346284866333
test_similarity => ['1.000', '0.588', '0.654', '0.659', '0.662', '0.690', '0.702', '0.714', '0.730', '0.737', '0.755', '0.755', '0.773', '0.784', '0.801', '0.814', '0.825', '0.827', '0.845', '0.852', '0.856', '0.873', '0.884', '0.887', '0.895', '0.897', '0.910', '0.917', '0.917', '0.925', '0.932', '0.934', '0.938', '0.932', '0.935', '0.944', '0.947', '0.950', '0.954', '0.956', '0.956', '0.956', '0.960', '0.963', '0.963', '0.964', '0.964', '0.961', '0.967', '0.972', '0.970', '0.971', '0.970', '0.974', '0.975', '0.977', '0.978', '0.978', '0.980', '0.984', '0.983', '0.984', '0.984', '0.983', '0.986', '0.987', '0.985', '0.987', '0.987', '0.986', '0.987', '0.986', '0.987', '0.986', '0.986', '0.986', '0.985']
test_reg => [-0.04472780600190163, 0.2619415521621704]
inference_test_weighted_loss => 1.4259785413742065
inference_test_similarity => ['1.000', '0.588', '0.530', '0.526', '0.521', '0.540', '0.536', '0.549', '0.556', '0.558', '0.558', '0.558', '0.570', '0.573', '0.573', '0.585', '0.584', '0.573', '0.581', '0.581', '0.580', '0.586', '0.583', '0.571', '0.580', '0.573', '0.571', '0.565', '0.562', '0.560', '0.560', '0.548', '0.542', '0.533', '0.531', '0.527', '0.522', '0.515', '0.512', '0.506', '0.499', '0.499', '0.494', '0.489', '0.483', '0.476', '0.473', '0.468', '0.466', '0.462', '0.455', '0.452', '0.449', '0.445', '0.441', '0.436', '0.433', '0.430', '0.425', '0.421', '0.417', '0.414', '0.409', '0.408', '0.404', '0.401', '0.398', '0.394', '0.391', '0.388', '0.385', '0.382', '0.380', '0.375', '0.374', '0.371', '0.398']
version => v73
seq => 1
epoch => 18
test_unweighted_loss => 0.1619832068681717
test_weighted_loss => 0.7709174752235413
test_similarity => ['1.000', '0.593', '0.649', '0.659', '0.657', '0.692', '0.703', '0.718', '0.732', '0.736', '0.757', '0.756', '0.773', '0.783', '0.803', '0.815', '0.827', '0.828', '0.847', '0.852', '0.858', '0.873', '0.884', '0.887', '0.897', '0.895', '0.909', '0.918', '0.919', '0.923', '0.934', '0.934', '0.937', '0.932', '0.938', '0.946', '0.947', '0.951', '0.954', '0.956', '0.958', '0.957', '0.961', '0.964', '0.963', '0.964', '0.964', '0.963', '0.967', '0.972', '0.970', '0.970', '0.970', '0.975', '0.976', '0.977', '0.977', '0.979', '0.980', '0.984', '0.983', '0.984', '0.984', '0.982', '0.986', '0.987', '0.985', '0.987', '0.987', '0.986', '0.986', '0.986', '0.987', '0.986', '0.986', '0.986', '0.985']
test_reg => [-0.04564117267727852, 0.2593025267124176]
inference_test_weighted_loss => 1.355571985244751
inference_test_similarity => ['1.000', '0.593', '0.528', '0.529', '0.523', '0.544', '0.542', '0.560', '0.568', '0.577', '0.577', '0.575', '0.589', '0.594', '0.592', '0.604', '0.606', '0.598', '0.603', '0.605', '0.604', '0.610', '0.607', '0.598', '0.605', '0.600', '0.599', '0.593', '0.590', '0.588', '0.587', '0.576', '0.571', '0.561', '0.562', '0.557', '0.550', '0.544', '0.540', '0.534', '0.528', '0.524', '0.519', '0.512', '0.507', '0.500', '0.496', '0.490', '0.487', '0.482', '0.476', '0.472', '0.467', '0.465', '0.461', '0.455', '0.453', '0.449', '0.444', '0.440', '0.437', '0.433', '0.428', '0.427', '0.423', '0.420', '0.417', '0.412', '0.409', '0.407', '0.403', '0.400', '0.397', '0.394', '0.391', '0.388', '0.427']
version => v75
seq => 2
epoch => 22
test_unweighted_loss => 0.16219504177570343
test_weighted_loss => 0.7735214829444885
test_similarity => ['1.000', '0.597', '0.647', '0.657', '0.660', '0.691', '0.701', '0.718', '0.733', '0.733', '0.756', '0.756', '0.771', '0.786', '0.799', '0.816', '0.825', '0.826', '0.843', '0.852', '0.855', '0.872', '0.882', '0.887', '0.898', '0.895', '0.908', '0.918', '0.916', '0.924', '0.935', '0.934', '0.939', '0.931', '0.937', '0.946', '0.948', '0.950', '0.954', '0.955', '0.957', '0.958', '0.961', '0.964', '0.964', '0.965', '0.964', '0.962', '0.967', '0.971', '0.969', '0.971', '0.970', '0.974', '0.976', '0.977', '0.977', '0.979', '0.980', '0.984', '0.983', '0.984', '0.984', '0.983', '0.986', '0.987', '0.985', '0.987', '0.987', '0.986', '0.987', '0.986', '0.987', '0.986', '0.986', '0.986', '0.985']
test_reg => [-0.046870097517967224, 0.2612517774105072]
inference_test_weighted_loss => 1.4126394987106323
inference_test_similarity => ['1.000', '0.597', '0.536', '0.525', '0.518', '0.535', '0.532', '0.551', '0.560', '0.565', '0.569', '0.570', '0.582', '0.584', '0.583', '0.593', '0.597', '0.585', '0.597', '0.600', '0.597', '0.604', '0.601', '0.591', '0.596', '0.592', '0.591', '0.586', '0.581', '0.578', '0.580', '0.568', '0.564', '0.552', '0.547', '0.546', '0.539', '0.535', '0.529', '0.523', '0.516', '0.515', '0.510', '0.504', '0.495', '0.489', '0.484', '0.477', '0.472', '0.467', '0.459', '0.454', '0.452', '0.446', '0.442', '0.436', '0.433', '0.428', '0.424', '0.420', '0.415', '0.410', '0.405', '0.403', '0.399', '0.396', '0.391', '0.387', '0.383', '0.380', '0.376', '0.373', '0.369', '0.366', '0.364', '0.361', '0.383']
version => v74
seq => 3
epoch => 20
test_unweighted_loss => 0.16184931993484497
test_weighted_loss => 0.7718237638473511
test_similarity => ['1.000', '0.591', '0.648', '0.657', '0.660', '0.689', '0.702', '0.716', '0.734', '0.737', '0.756', '0.754', '0.771', '0.783', '0.800', '0.818', '0.826', '0.827', '0.847', '0.850', '0.854', '0.871', '0.883', '0.888', '0.898', '0.898', '0.909', '0.918', '0.917', '0.925', '0.935', '0.934', '0.939', '0.933', '0.936', '0.945', '0.949', '0.950', '0.954', '0.956', '0.958', '0.957', '0.961', '0.963', '0.964', '0.965', '0.965', '0.962', '0.967', '0.972', '0.970', '0.971', '0.970', '0.973', '0.976', '0.977', '0.977', '0.980', '0.980', '0.984', '0.984', '0.984', '0.984', '0.983', '0.986', '0.987', '0.986', '0.987', '0.987', '0.986', '0.987', '0.986', '0.987', '0.986', '0.987', '0.986', '0.985']
test_reg => [-0.046316444873809814, 0.26389792561531067]
inference_test_weighted_loss => 1.3752267360687256
inference_test_similarity => ['1.000', '0.591', '0.532', '0.525', '0.520', '0.538', '0.536', '0.549', '0.560', '0.570', '0.567', '0.571', '0.583', '0.584', '0.584', '0.598', '0.600', '0.592', '0.605', '0.606', '0.603', '0.614', '0.613', '0.605', '0.611', '0.608', '0.608', '0.603', '0.599', '0.595', '0.596', '0.587', '0.582', '0.574', '0.573', '0.570', '0.562', '0.556', '0.550', '0.545', '0.538', '0.537', '0.532', '0.526', '0.519', '0.516', '0.510', '0.504', '0.498', '0.494', '0.487', '0.485', '0.481', '0.476', '0.470', '0.465', '0.463', '0.458', '0.453', '0.448', '0.444', '0.440', '0.436', '0.433', '0.429', '0.426', '0.422', '0.417', '0.414', '0.412', '0.408', '0.406', '0.402', '0.400', '0.397', '0.395', '0.446']