Unnamed: 0
int64
0
2.44k
repo
stringlengths
32
81
hash
stringlengths
40
40
diff
stringlengths
113
1.17k
old_path
stringlengths
5
84
rewrite
stringlengths
34
79
initial_state
stringlengths
75
980
final_state
stringlengths
76
980
0
https://:@github.com/emedvedev/attention-ocr.git
291042e7cb623c8a908e9badd132c1fa2360288c
@@ -465,7 +465,7 @@ class Model(object): mh = 32 mw = math.floor(1. * w / h * mh) img = img.resize( - (mw, h), + (mw, mh), Image.ANTIALIAS) img_data = np.asarray(img, dtype=np.uint8) for idx in xrange(len(output)):
aocr/model/model.py
ReplaceText(target='mh' @(468,25)->(468,26))
class Model(object): mh = 32 mw = math.floor(1. * w / h * mh) img = img.resize( (mw, h), Image.ANTIALIAS) img_data = np.asarray(img, dtype=np.uint8) for idx in xrange(len(output)):
class Model(object): mh = 32 mw = math.floor(1. * w / h * mh) img = img.resize( (mw, mh), Image.ANTIALIAS) img_data = np.asarray(img, dtype=np.uint8) for idx in xrange(len(output)):
1
https://:@github.com/emedvedev/attention-ocr.git
6e6593c27fe0e63118adadf11562b1c4699b14e3
@@ -477,7 +477,7 @@ class Model(object): attention_orig[i] = attention[int(i/4)-1] attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same') attention_orig = np.maximum(attention_orig, 0.3) - attention_out = np.zeros((h, mw)) + attention_out = np.zeros((mh, mw)) for i in xrange(mw): attention_out[:, i] = attention_orig[i] if len(img_data.shape) == 3:
aocr/model/model.py
ReplaceText(target='mh' @(480,42)->(480,43))
class Model(object): attention_orig[i] = attention[int(i/4)-1] attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same') attention_orig = np.maximum(attention_orig, 0.3) attention_out = np.zeros((h, mw)) for i in xrange(mw): attention_out[:, i] = attention_orig[i] if len(img_data.shape) == 3:
class Model(object): attention_orig[i] = attention[int(i/4)-1] attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same') attention_orig = np.maximum(attention_orig, 0.3) attention_out = np.zeros((mh, mw)) for i in xrange(mw): attention_out[:, i] = attention_orig[i] if len(img_data.shape) == 3:
2
https://:@github.com/emedvedev/attention-ocr.git
e741baf7170e72a974754908d323cacc0bd55247
@@ -133,7 +133,7 @@ class Model(object): self.target_weights = [] for i in xrange(self.decoder_size + 1): self.decoder_inputs.append( - tf.tile([0], [num_images]) + tf.tile([1], [num_images]) ) if i < self.decoder_size: self.target_weights.append(tf.tile([1.], [num_images]))
aocr/model/model.py
ReplaceText(target='1' @(136,29)->(136,30))
class Model(object): self.target_weights = [] for i in xrange(self.decoder_size + 1): self.decoder_inputs.append( tf.tile([0], [num_images]) ) if i < self.decoder_size: self.target_weights.append(tf.tile([1.], [num_images]))
class Model(object): self.target_weights = [] for i in xrange(self.decoder_size + 1): self.decoder_inputs.append( tf.tile([1], [num_images]) ) if i < self.decoder_size: self.target_weights.append(tf.tile([1.], [num_images]))
3
https://:@github.com/matthewdowney/TogglPy.git
d5b630aec58d29b85ccffa527d24766eef6f61f9
@@ -197,7 +197,7 @@ class Toggl(): day = datetime.now().day if not day else day hour = datetime.now().hour if not hour else hour - timestruct = datetime(year, month, day, hour - hourdiff).isoformat() + '.000Z' + timestruct = datetime(year, month, day, hour + hourdiff).isoformat() + '.000Z' data['time_entry']['start'] = timestruct data['time_entry']['duration'] = hourduration * 3600 data['time_entry']['pid'] = projectid
toggl/TogglPy.py
ReplaceText(target='+' @(200,53)->(200,54))
class Toggl(): day = datetime.now().day if not day else day hour = datetime.now().hour if not hour else hour timestruct = datetime(year, month, day, hour - hourdiff).isoformat() + '.000Z' data['time_entry']['start'] = timestruct data['time_entry']['duration'] = hourduration * 3600 data['time_entry']['pid'] = projectid
class Toggl(): day = datetime.now().day if not day else day hour = datetime.now().hour if not hour else hour timestruct = datetime(year, month, day, hour + hourdiff).isoformat() + '.000Z' data['time_entry']['start'] = timestruct data['time_entry']['duration'] = hourduration * 3600 data['time_entry']['pid'] = projectid
4
https://:@github.com/eEcoLiDAR/eEcoLiDAR.git
f0b2a0b7a5fdd41887ba40b7687c8161c0faba1e
@@ -28,6 +28,6 @@ class Test3FeatureExtractor(AbstractFeatureExtractor): return ['test3_a'] def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume): - t2a, t2c = utils.get_features(targetpc, targetindex, self.requires()) + t2a, t2c = utils.get_features(targetpc, self.requires(), targetindex) x, y, z = utils.get_point(targetpc, targetindex) return t2c - t2a - z # z
laserchicken/test_feature_extractor/feature_test23.py
ArgSwap(idxs=1<->2 @(31,19)->(31,37))
class Test3FeatureExtractor(AbstractFeatureExtractor): return ['test3_a'] def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume): t2a, t2c = utils.get_features(targetpc, targetindex, self.requires()) x, y, z = utils.get_point(targetpc, targetindex) return t2c - t2a - z # z
class Test3FeatureExtractor(AbstractFeatureExtractor): return ['test3_a'] def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume): t2a, t2c = utils.get_features(targetpc, self.requires(), targetindex) x, y, z = utils.get_point(targetpc, targetindex) return t2c - t2a - z # z
5
https://:@github.com/eEcoLiDAR/eEcoLiDAR.git
f0b2a0b7a5fdd41887ba40b7687c8161c0faba1e
@@ -31,7 +31,7 @@ class TestUtils(unittest.TestCase): pc[keys.point]["color"] = {"type": "double", "data": cols} pc[keys.point]["flavor"] = {"type": "double", "data": flavs} x, y, z = utils.get_point(pc, 2) - c, f = utils.get_features(pc, 2, ("color", "flavor")) + c, f = utils.get_features(pc, ("color", "flavor"), 2) self.assertEqual(c, 0.5 * (x + y)) self.assertEqual(f, 0.5 * (x - y))
laserchicken/test_utils.py
ArgSwap(idxs=1<->2 @(34,15)->(34,33))
class TestUtils(unittest.TestCase): pc[keys.point]["color"] = {"type": "double", "data": cols} pc[keys.point]["flavor"] = {"type": "double", "data": flavs} x, y, z = utils.get_point(pc, 2) c, f = utils.get_features(pc, 2, ("color", "flavor")) self.assertEqual(c, 0.5 * (x + y)) self.assertEqual(f, 0.5 * (x - y))
class TestUtils(unittest.TestCase): pc[keys.point]["color"] = {"type": "double", "data": cols} pc[keys.point]["flavor"] = {"type": "double", "data": flavs} x, y, z = utils.get_point(pc, 2) c, f = utils.get_features(pc, ("color", "flavor"), 2) self.assertEqual(c, 0.5 * (x + y)) self.assertEqual(f, 0.5 * (x - y))
6
https://:@github.com/eEcoLiDAR/eEcoLiDAR.git
502a365efda1393b130281803702d01f7e2d1dcd
@@ -29,7 +29,7 @@ class TestExtractEigenValues(unittest.TestCase): ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5)) self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor", - target_point_cloud[keys.provenance][0]["module"]) + target_point_cloud[keys.provenance][1]["module"]) @staticmethod def test_eigenvalues_of_too_few_points_results_in_0():
laserchicken/feature_extractor/test_eigenvals_feature_extractor.py
ReplaceText(target='1' @(32,61)->(32,62))
class TestExtractEigenValues(unittest.TestCase): ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5)) self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor", target_point_cloud[keys.provenance][0]["module"]) @staticmethod def test_eigenvalues_of_too_few_points_results_in_0():
class TestExtractEigenValues(unittest.TestCase): ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5)) self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor", target_point_cloud[keys.provenance][1]["module"]) @staticmethod def test_eigenvalues_of_too_few_points_results_in_0():
7
https://:@github.com/eEcoLiDAR/eEcoLiDAR.git
eb7b021147a60b57e5dec536bd6f118c213f0952
@@ -29,7 +29,7 @@ class TestExtractEigenValues(unittest.TestCase): ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5)) self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor", - target_point_cloud[keys.provenance][1]["module"]) + target_point_cloud[keys.provenance][-1]["module"]) @staticmethod def test_eigenvalues_of_too_few_points_results_in_0():
laserchicken/feature_extractor/test_eigenvals_feature_extractor.py
ReplaceText(target='-1' @(32,61)->(32,62))
class TestExtractEigenValues(unittest.TestCase): ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5)) self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor", target_point_cloud[keys.provenance][1]["module"]) @staticmethod def test_eigenvalues_of_too_few_points_results_in_0():
class TestExtractEigenValues(unittest.TestCase): ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5)) self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor", target_point_cloud[keys.provenance][-1]["module"]) @staticmethod def test_eigenvalues_of_too_few_points_results_in_0():
8
https://:@github.com/VinF/deer.git
66ea41db02c3361f18874dea7fd97720b1b06590
@@ -441,7 +441,7 @@ class CircularBuffer(object): if end == sys.maxsize: return self._data[self._lb+start:self._ub] - elif self._lb + end >= self._ub: + elif self._lb + end > self._ub: raise IndexError() else: return self._data[self._lb+start:self._lb+end]
General_Deep_Q_RL/agent.py
ReplaceText(target='>' @(444,28)->(444,30))
class CircularBuffer(object): if end == sys.maxsize: return self._data[self._lb+start:self._ub] elif self._lb + end >= self._ub: raise IndexError() else: return self._data[self._lb+start:self._lb+end]
class CircularBuffer(object): if end == sys.maxsize: return self._data[self._lb+start:self._ub] elif self._lb + end > self._ub: raise IndexError() else: return self._data[self._lb+start:self._lb+end]
9
https://:@github.com/VinF/deer.git
fd939e272d5441d48fd7d30bf24312c0f6bc8aaa
@@ -176,7 +176,7 @@ class MyEnv(Environment): # Lack of energy if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery): # If enough energy in the battery, use it - self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size*self.battery_eta + self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size/self.battery_eta else: # Otherwise: use what is left and then penalty reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh
General_Deep_Q_RL/environments/MG_two_storages_env.py
ReplaceText(target='/' @(179,126)->(179,127))
class MyEnv(Environment): # Lack of energy if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery): # If enough energy in the battery, use it self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size*self.battery_eta else: # Otherwise: use what is left and then penalty reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh
class MyEnv(Environment): # Lack of energy if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery): # If enough energy in the battery, use it self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size/self.battery_eta else: # Otherwise: use what is left and then penalty reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh
10
https://:@github.com/piccolbo/altair_recipes.git
4992dd864a317eaad641d0408f003c429ed24af6
@@ -6,7 +6,7 @@ from vega_datasets import data @viz_reg_test def test_boxplot_melted(): - return ar.boxplot(data.iris(), "species", "petalLength") + return ar.boxplot(data.iris(), "petalLength", "species") @viz_reg_test
tests/test_boxplot.py
ArgSwap(idxs=1<->2 @(9,11)->(9,21))
from vega_datasets import data @viz_reg_test def test_boxplot_melted(): return ar.boxplot(data.iris(), "species", "petalLength") @viz_reg_test
from vega_datasets import data @viz_reg_test def test_boxplot_melted(): return ar.boxplot(data.iris(), "petalLength", "species") @viz_reg_test
11
https://:@github.com/tailhook/zorro.git
dcbc37d47fe2a8de029f5a2f3ae13adf52e7aace
@@ -121,7 +121,7 @@ class RequestChannel(channel.PipelinedReqChannel): clen = int(headers.get('Content-Length', '0')) if clen < 0: raise EOFError("Wrong content length") - while pos[0] + clen < len(buf): + while pos[0] + clen > len(buf): readmore() return status, headers, buf[pos[0]:pos[0]+clen]
zorro/http.py
ReplaceText(target='>' @(124,32)->(124,33))
class RequestChannel(channel.PipelinedReqChannel): clen = int(headers.get('Content-Length', '0')) if clen < 0: raise EOFError("Wrong content length") while pos[0] + clen < len(buf): readmore() return status, headers, buf[pos[0]:pos[0]+clen]
class RequestChannel(channel.PipelinedReqChannel): clen = int(headers.get('Content-Length', '0')) if clen < 0: raise EOFError("Wrong content length") while pos[0] + clen > len(buf): readmore() return status, headers, buf[pos[0]:pos[0]+clen]
12
https://:@gitlab.com/eavise/brambox.git
f1faeed0b52d6f1c9c9ba6da818c1656f841622c
@@ -73,7 +73,7 @@ def test_multiclass(parser, df_anno_simple): parser = parser() with pytest.raises(ValueError) as errinfo: - bb.io.save(parser, df_anno_simple, 'path.txt') + bb.io.save(df_anno_simple, parser, 'path.txt') assert 'single-class problems' in str(errinfo.value)
test/io/parser/test_anno_cvc.py
ArgSwap(idxs=0<->1 @(76,8)->(76,18))
def test_multiclass(parser, df_anno_simple): parser = parser() with pytest.raises(ValueError) as errinfo: bb.io.save(parser, df_anno_simple, 'path.txt') assert 'single-class problems' in str(errinfo.value)
def test_multiclass(parser, df_anno_simple): parser = parser() with pytest.raises(ValueError) as errinfo: bb.io.save(df_anno_simple, parser, 'path.txt') assert 'single-class problems' in str(errinfo.value)
13
https://:@github.com/uber/h3-py.git
359924df907144c85ec323ae2804e2c0d173dfc5
@@ -631,7 +631,7 @@ def hex_ranges(h3_address_list, ring_size): (1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1 # hexRanges doesn't return distance array hex_range_list[ring_index].add( - h3_to_string(krings[i * num_hexagons + j])) + h3_to_string(krings[i * array_len + j])) return out
h3/h3.py
ReplaceText(target='array_len' @(634,40)->(634,52))
def hex_ranges(h3_address_list, ring_size): (1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1 # hexRanges doesn't return distance array hex_range_list[ring_index].add( h3_to_string(krings[i * num_hexagons + j])) return out
def hex_ranges(h3_address_list, ring_size): (1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1 # hexRanges doesn't return distance array hex_range_list[ring_index].add( h3_to_string(krings[i * array_len + j])) return out
14
https://:@github.com/polysquare/polysquare-generic-file-linter.git
e9dbb28ea30955ab59d1339c04f0710b24ba53aa
@@ -388,7 +388,7 @@ def _maybe_log_technical_terms(global_options, tool_options): terms = set(terms_file.read().splitlines()) # suppress(PYC70) terms_file.seek(0) # suppress(PYC70) terms_file.truncate(0) # suppress(PYC70) - tech_terms = freduce(lambda x, y: x + y, + tech_terms = freduce(lambda x, y: x | y, _drain(log_technical_terms_to_queue)) terms_file.write("\n".join(list(terms | # suppress(PYC70) set(tech_terms))))
polysquarelinter/linter.py
ReplaceText(target='|' @(391,48)->(391,49))
def _maybe_log_technical_terms(global_options, tool_options): terms = set(terms_file.read().splitlines()) # suppress(PYC70) terms_file.seek(0) # suppress(PYC70) terms_file.truncate(0) # suppress(PYC70) tech_terms = freduce(lambda x, y: x + y, _drain(log_technical_terms_to_queue)) terms_file.write("\n".join(list(terms | # suppress(PYC70) set(tech_terms))))
def _maybe_log_technical_terms(global_options, tool_options): terms = set(terms_file.read().splitlines()) # suppress(PYC70) terms_file.seek(0) # suppress(PYC70) terms_file.truncate(0) # suppress(PYC70) tech_terms = freduce(lambda x, y: x | y, _drain(log_technical_terms_to_queue)) terms_file.write("\n".join(list(terms | # suppress(PYC70) set(tech_terms))))
15
https://:@github.com/johntruckenbrodt/spatialist.git
c9d552e64cd47b30156b288e035d17debea48b45
@@ -300,7 +300,7 @@ def centerdist(obj1, obj2): def intersect(obj1, obj2): - if not (isinstance(obj1, Vector) or isinstance(obj2, Vector)): + if not (isinstance(obj1, Vector) and isinstance(obj2, Vector)): raise IOError('object must be of type Vector') obj1.reproject(obj2.srs)
pyroSAR/spatial/vector.py
ReplaceText(target='and' @(303,37)->(303,39))
def centerdist(obj1, obj2): def intersect(obj1, obj2): if not (isinstance(obj1, Vector) or isinstance(obj2, Vector)): raise IOError('object must be of type Vector') obj1.reproject(obj2.srs)
def centerdist(obj1, obj2): def intersect(obj1, obj2): if not (isinstance(obj1, Vector) and isinstance(obj2, Vector)): raise IOError('object must be of type Vector') obj1.reproject(obj2.srs)
16
https://:@github.com/Toblerity/Shapely.git
9f1b78e6fd5f4286f210b54827bdd26661f0ee7a
@@ -40,7 +40,7 @@ if __name__ == '__main__': ] if pattern: - tests = [f for f in docfiles if f.find(pattern) >= 0] + tests = [f for f in docfiles if f.find(pattern) == 0] else: tests = docfiles
tests/runalldoctests.py
ReplaceText(target='==' @(43,56)->(43,58))
if __name__ == '__main__': ] if pattern: tests = [f for f in docfiles if f.find(pattern) >= 0] else: tests = docfiles
if __name__ == '__main__': ] if pattern: tests = [f for f in docfiles if f.find(pattern) == 0] else: tests = docfiles
17
https://:@github.com/Toblerity/Shapely.git
d6fc8cc0e0d50b23ba0d7ca6195bc530b2f8d1b9
@@ -11,7 +11,7 @@ def halton(base): i = index while i > 0: result += f * (i % base) - i = i/base + i = i//base f = f/base return result i = 1
shapely/tests/test_unary_union.py
ReplaceText(target='//' @(14,17)->(14,18))
def halton(base): i = index while i > 0: result += f * (i % base) i = i/base f = f/base return result i = 1
def halton(base): i = index while i > 0: result += f * (i % base) i = i//base f = f/base return result i = 1
18
https://:@github.com/Toblerity/Shapely.git
5f0db7fdc052beeeef36aa1251f19175d0abeedb
@@ -36,7 +36,7 @@ if version is None: # Handle UTF-8 encoding of certain text files. open_kwds = {} -if sys.version_info > (3,): +if sys.version_info >= (3,): open_kwds['encoding'] = 'utf-8' with open('VERSION.txt', 'w', **open_kwds) as fp:
setup.py
ReplaceText(target='>=' @(39,20)->(39,21))
if version is None: # Handle UTF-8 encoding of certain text files. open_kwds = {} if sys.version_info > (3,): open_kwds['encoding'] = 'utf-8' with open('VERSION.txt', 'w', **open_kwds) as fp:
if version is None: # Handle UTF-8 encoding of certain text files. open_kwds = {} if sys.version_info >= (3,): open_kwds['encoding'] = 'utf-8' with open('VERSION.txt', 'w', **open_kwds) as fp:
19
https://:@github.com/svetlyak40wt/django-tagging-ng.git
0293b78ee0274d123eb70c1f8c5c01a5b36e2b40
@@ -163,7 +163,7 @@ class TaggedItemManager(models.Manager): associated with a given Tag or list of Tags. """ tags = get_tag_list(tags) - if len(tags) == 0: + if len(tags) == 1: tag = tags[0] # Optimisation for single tag else: return self.get_intersection_by_model(Model, tags)
models.py
ReplaceText(target='1' @(166,24)->(166,25))
class TaggedItemManager(models.Manager): associated with a given Tag or list of Tags. """ tags = get_tag_list(tags) if len(tags) == 0: tag = tags[0] # Optimisation for single tag else: return self.get_intersection_by_model(Model, tags)
class TaggedItemManager(models.Manager): associated with a given Tag or list of Tags. """ tags = get_tag_list(tags) if len(tags) == 1: tag = tags[0] # Optimisation for single tag else: return self.get_intersection_by_model(Model, tags)
20
https://:@github.com/svetlyak40wt/django-tagging-ng.git
3285d40e4c1de628886a7fa45a6d4cf6ed4cd7e7
@@ -163,7 +163,7 @@ class TaggedItemManager(models.Manager): associated with a given Tag or list of Tags. """ tags = get_tag_list(tags) - if len(tags) == 0: + if len(tags) == 1: tag = tags[0] # Optimisation for single tag else: return self.get_intersection_by_model(Model, tags)
models.py
ReplaceText(target='1' @(166,24)->(166,25))
class TaggedItemManager(models.Manager): associated with a given Tag or list of Tags. """ tags = get_tag_list(tags) if len(tags) == 0: tag = tags[0] # Optimisation for single tag else: return self.get_intersection_by_model(Model, tags)
class TaggedItemManager(models.Manager): associated with a given Tag or list of Tags. """ tags = get_tag_list(tags) if len(tags) == 1: tag = tags[0] # Optimisation for single tag else: return self.get_intersection_by_model(Model, tags)
21
https://:@github.com/zzzsochi/yadm.git
03efd06fe95c7d84264455c4fac5c8cbb17eb4dd
@@ -316,7 +316,7 @@ class QuerySet(BaseQuerySet): if data is None: if exc is not None: - raise exc(criteria) + raise exc(qs) else: return None
yadm/queryset.py
ReplaceText(target='qs' @(319,26)->(319,34))
class QuerySet(BaseQuerySet): if data is None: if exc is not None: raise exc(criteria) else: return None
class QuerySet(BaseQuerySet): if data is None: if exc is not None: raise exc(qs) else: return None
22
https://:@github.com/instacart/lore.git
1c1e0efdac6b27dc111eaa93bb99317c59aaffaf
@@ -196,7 +196,7 @@ class Base(object): def upload(self): self.fitting = 0 self.save() - lore.io.upload(self.remote_model_path(), self.model_path()) + lore.io.upload(self.model_path(), self.remote_model_path()) @classmethod def download(cls, fitting=0):
lore/models/base.py
ArgSwap(idxs=0<->1 @(199,8)->(199,22))
class Base(object): def upload(self): self.fitting = 0 self.save() lore.io.upload(self.remote_model_path(), self.model_path()) @classmethod def download(cls, fitting=0):
class Base(object): def upload(self): self.fitting = 0 self.save() lore.io.upload(self.model_path(), self.remote_model_path()) @classmethod def download(cls, fitting=0):
23
https://:@github.com/instacart/lore.git
1c1e0efdac6b27dc111eaa93bb99317c59aaffaf
@@ -78,7 +78,7 @@ class Base(lore.models.base.Base): def upload(self): super(Base, self).upload() - lore.io.upload(self.remote_weights_path(), self.weights_path()) + lore.io.upload(self.weights_path(), self.remote_weights_path()) @classmethod def download(cls, fitting=0):
lore/models/keras.py
ArgSwap(idxs=0<->1 @(81,8)->(81,22))
class Base(lore.models.base.Base): def upload(self): super(Base, self).upload() lore.io.upload(self.remote_weights_path(), self.weights_path()) @classmethod def download(cls, fitting=0):
class Base(lore.models.base.Base): def upload(self): super(Base, self).upload() lore.io.upload(self.weights_path(), self.remote_weights_path()) @classmethod def download(cls, fitting=0):
24
https://:@github.com/instacart/lore.git
f4ded2b3199d1c33ba6c9c79cd66b25d43c83c81
@@ -464,7 +464,7 @@ class Base(BaseEstimator): result = self.keras.predict(dataframe, batch_size=self.batch_size) if self.towers > 1: - result = numpy.mean(result, axis=0).squeeze(axis=0) + result = numpy.mean(result, axis=0).squeeze(axis=1) return result
lore/estimators/keras.py
ReplaceText(target='1' @(467,61)->(467,62))
class Base(BaseEstimator): result = self.keras.predict(dataframe, batch_size=self.batch_size) if self.towers > 1: result = numpy.mean(result, axis=0).squeeze(axis=0) return result
class Base(BaseEstimator): result = self.keras.predict(dataframe, batch_size=self.batch_size) if self.towers > 1: result = numpy.mean(result, axis=0).squeeze(axis=1) return result
25
https://:@github.com/baliga-lab/cmonkey2.git
3201a0e97688724450196da8fef96d283c855b3f
@@ -273,7 +273,7 @@ class ClusterMembership: #logging.warn("cluster %s already associated with %s", # str(cluster), str(column)) pass - if columns not in columns: + if column not in columns: columns.append(column) def remove_cluster_from_column(self, column, cluster):
cmonkey/membership.py
ReplaceText(target='column' @(276,11)->(276,18))
class ClusterMembership: #logging.warn("cluster %s already associated with %s", # str(cluster), str(column)) pass if columns not in columns: columns.append(column) def remove_cluster_from_column(self, column, cluster):
class ClusterMembership: #logging.warn("cluster %s already associated with %s", # str(cluster), str(column)) pass if column not in columns: columns.append(column) def remove_cluster_from_column(self, column, cluster):
26
https://:@github.com/baliga-lab/cmonkey2.git
48d14ac785b1013354a55a37239c66433fbf19eb
@@ -460,7 +460,7 @@ class ClusterMembership: max_score = sys.float_info.min for row in range(sm.num_rows()): if sm_values[row][0] > max_score: - max_score = sm[row][0] + max_score = sm_values[row][0] max_row = row return sm.row_names[max_row]
cmonkey/membership.py
ReplaceText(target='sm_values' @(463,32)->(463,34))
class ClusterMembership: max_score = sys.float_info.min for row in range(sm.num_rows()): if sm_values[row][0] > max_score: max_score = sm[row][0] max_row = row return sm.row_names[max_row]
class ClusterMembership: max_score = sys.float_info.min for row in range(sm.num_rows()): if sm_values[row][0] > max_score: max_score = sm_values[row][0] max_row = row return sm.row_names[max_row]
27
https://:@github.com/baliga-lab/cmonkey2.git
311548905a9def1cbdf63d2e0fd8a17346564742
@@ -17,7 +17,7 @@ class MicrobesOnlineTest(unittest.TestCase): # pylint: disable-msg=R0904 """test fixture""" if not os.path.exists('testcache'): os.mkdir('testcache') - self.service = mo.MicrobesOnline(mo.MICROBES_ONLINE_BASE_URL, 'testcache') + self.service = mo.MicrobesOnline('testcache', mo.MICROBES_ONLINE_BASE_URL) def tearDown(self): # pylint: disable-msg=C0103 """test cleanup"""
test/microbes_online_test.py
ArgSwap(idxs=0<->1 @(20,23)->(20,40))
class MicrobesOnlineTest(unittest.TestCase): # pylint: disable-msg=R0904 """test fixture""" if not os.path.exists('testcache'): os.mkdir('testcache') self.service = mo.MicrobesOnline(mo.MICROBES_ONLINE_BASE_URL, 'testcache') def tearDown(self): # pylint: disable-msg=C0103 """test cleanup"""
class MicrobesOnlineTest(unittest.TestCase): # pylint: disable-msg=R0904 """test fixture""" if not os.path.exists('testcache'): os.mkdir('testcache') self.service = mo.MicrobesOnline('testcache', mo.MICROBES_ONLINE_BASE_URL) def tearDown(self): # pylint: disable-msg=C0103 """test cleanup"""
28
https://:@github.com/muammar/ml4chem.git
dbb7de0379cb8881538d211899e4bec8794f16e3
@@ -344,7 +344,7 @@ def train(inputs, targets, model=None, data=None, optimizer=None, lr=None, logger.info('Training finished in {} hours {} minutes {:.2f} seconds.' .format(h, m, s)) logger.info('outputs') - logger.info(outputs) + logger.info(outputs_) logger.info('targets') logger.info(targets)
mlchem/models/neuralnetwork.py
ReplaceText(target='outputs_' @(347,16)->(347,23))
def train(inputs, targets, model=None, data=None, optimizer=None, lr=None, logger.info('Training finished in {} hours {} minutes {:.2f} seconds.' .format(h, m, s)) logger.info('outputs') logger.info(outputs) logger.info('targets') logger.info(targets)
def train(inputs, targets, model=None, data=None, optimizer=None, lr=None, logger.info('Training finished in {} hours {} minutes {:.2f} seconds.' .format(h, m, s)) logger.info('outputs') logger.info(outputs_) logger.info('targets') logger.info(targets)
29
https://:@github.com/chris7/pyquant.git
3730fbdb9789a59a65d38f5a2ae21c645086096f
@@ -624,7 +624,7 @@ def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si best_fit = np.array(best_fit) peak_func = bigauss_ndim if bigauss_fit else gauss_ndim # Get rid of peaks with low r^2 - if micro and r2_cutoff is not None: + if not micro and r2_cutoff is not None: final_fit = np.array([]) for peak_index in xrange(0, len(best_fit), step_size):
pyquant/peaks.py
ReplaceText(target='not ' @(627,7)->(627,7))
def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si best_fit = np.array(best_fit) peak_func = bigauss_ndim if bigauss_fit else gauss_ndim # Get rid of peaks with low r^2 if micro and r2_cutoff is not None: final_fit = np.array([]) for peak_index in xrange(0, len(best_fit), step_size):
def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si best_fit = np.array(best_fit) peak_func = bigauss_ndim if bigauss_fit else gauss_ndim # Get rid of peaks with low r^2 if not micro and r2_cutoff is not None: final_fit = np.array([]) for peak_index in xrange(0, len(best_fit), step_size):
30
https://:@github.com/chris7/pyquant.git
4a0755563e0a36fecf1f4393554cfaf4c1615c2c
@@ -615,7 +615,7 @@ def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe # By default, cross points returns the left side for i in xrange(len(cross_points)): index = cross_points[i] - if index < len(cross_points): + if index < len(ydata): if ydata[index] < ydata[index+1]: cross_points[i] = index+1
pyquant/utils.py
ReplaceText(target='ydata' @(618,23)->(618,35))
def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe # By default, cross points returns the left side for i in xrange(len(cross_points)): index = cross_points[i] if index < len(cross_points): if ydata[index] < ydata[index+1]: cross_points[i] = index+1
def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe # By default, cross points returns the left side for i in xrange(len(cross_points)): index = cross_points[i] if index < len(ydata): if ydata[index] < ydata[index+1]: cross_points[i] = index+1
31
https://:@github.com/chris7/pyquant.git
cd61286935d8ca64eb539851e39a98a0655ff400
@@ -611,7 +611,7 @@ def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe ydata = np.abs(ydata_peaks) if min_peak_width is None: - max_peak_width = int(len(ydata) / 2) + min_peak_width = int(len(ydata) / 2) if min_peak_width > 5: min_peak_width = 5
pyquant/utils.py
ReplaceText(target='min_peak_width' @(614,8)->(614,22))
def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe ydata = np.abs(ydata_peaks) if min_peak_width is None: max_peak_width = int(len(ydata) / 2) if min_peak_width > 5: min_peak_width = 5
def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe ydata = np.abs(ydata_peaks) if min_peak_width is None: min_peak_width = int(len(ydata) / 2) if min_peak_width > 5: min_peak_width = 5
32
https://:@github.com/ssec/sift.git
24ce052cd497c42c917de06f0c89a7c5be13ab50
@@ -599,7 +599,7 @@ class ProbeGraphDisplay (object) : x_point = self.workspace.get_content_point(x_uuid, point_xy) format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point) y_point = self.workspace.get_content_point(y_uuid, point_xy) - format_str, unit_str, y_point = self.document.convert_units(x_uuid, y_point) + format_str, unit_str, y_point = self.document.convert_units(y_uuid, y_point) else: x_point = None y_point = None
py/cspov/view/ProbeGraphs.py
ReplaceText(target='y_uuid' @(602,76)->(602,82))
class ProbeGraphDisplay (object) : x_point = self.workspace.get_content_point(x_uuid, point_xy) format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point) y_point = self.workspace.get_content_point(y_uuid, point_xy) format_str, unit_str, y_point = self.document.convert_units(x_uuid, y_point) else: x_point = None y_point = None
class ProbeGraphDisplay (object) : x_point = self.workspace.get_content_point(x_uuid, point_xy) format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point) y_point = self.workspace.get_content_point(y_uuid, point_xy) format_str, unit_str, y_point = self.document.convert_units(y_uuid, y_point) else: x_point = None y_point = None
33
https://:@github.com/threatwatch/twigs.git
7ee5d95178a459a8c2e8ff7855e3156e620c395c
@@ -85,7 +85,7 @@ def parse_inventory(email,data,params): asset_map = {} asset_map['owner'] = email asset_map['host'] = host - asset_map['id'] = host + asset_map['id'] = vmuuid asset_map['name'] = host asset_map['tags'] = [ ] asset_map['patch_tracker'] = { } # To help remove duplicate patches
twigs/azure.py
ReplaceText(target='vmuuid' @(88,30)->(88,34))
def parse_inventory(email,data,params): asset_map = {} asset_map['owner'] = email asset_map['host'] = host asset_map['id'] = host asset_map['name'] = host asset_map['tags'] = [ ] asset_map['patch_tracker'] = { } # To help remove duplicate patches
def parse_inventory(email,data,params): asset_map = {} asset_map['owner'] = email asset_map['host'] = host asset_map['id'] = vmuuid asset_map['name'] = host asset_map['tags'] = [ ] asset_map['patch_tracker'] = { } # To help remove duplicate patches
34
https://:@github.com/Keeper-Security/Commander.git
9bd55c8dd48ab62759bbbb8dfcd38ab364cec2dc
@@ -1550,7 +1550,7 @@ def prepare_record(params, record): else: if params.debug: print('Generated record key') unencrypted_key = os.urandom(32) - record_object['record_key'] = encrypt_aes(params.data_key, unencrypted_key) + record_object['record_key'] = encrypt_aes(unencrypted_key, params.data_key) record_object['revision'] = 0 data['title'] = record.title
keepercommander/api.py
ArgSwap(idxs=0<->1 @(1553,38)->(1553,49))
def prepare_record(params, record): else: if params.debug: print('Generated record key') unencrypted_key = os.urandom(32) record_object['record_key'] = encrypt_aes(params.data_key, unencrypted_key) record_object['revision'] = 0 data['title'] = record.title
def prepare_record(params, record): else: if params.debug: print('Generated record key') unencrypted_key = os.urandom(32) record_object['record_key'] = encrypt_aes(unencrypted_key, params.data_key) record_object['revision'] = 0 data['title'] = record.title
35
https://:@github.com/hpapaxen/rope.git
27d5085b30e89095e88339c96d9940e338482106
@@ -69,7 +69,7 @@ class JobSet(object): def get_percent_done(self): if self.count is not None and self.count > 0: - percent = self.done * 100 / self.count + percent = self.done * 100 // self.count return min(percent, 100) def get_name(self):
rope/base/taskhandle.py
ReplaceText(target='//' @(72,38)->(72,39))
class JobSet(object): def get_percent_done(self): if self.count is not None and self.count > 0: percent = self.done * 100 / self.count return min(percent, 100) def get_name(self):
class JobSet(object): def get_percent_done(self): if self.count is not None and self.count > 0: percent = self.done * 100 // self.count return min(percent, 100) def get_name(self):
36
https://:@github.com/hpapaxen/rope.git
27d5085b30e89095e88339c96d9940e338482106
@@ -524,7 +524,7 @@ class ProgressBar(object): self.text['text'] = text def _draw_shape(self): - width = int(self.canvas['width']) * self.percent / 100 + width = int(self.canvas['width']) * self.percent // 100 self.canvas.create_rectangle(0, 0, width, self.canvas['height'], fill=self.color) total_width = self.canvas['width']
rope/ui/uihelpers.py
ReplaceText(target='//' @(527,57)->(527,58))
class ProgressBar(object): self.text['text'] = text def _draw_shape(self): width = int(self.canvas['width']) * self.percent / 100 self.canvas.create_rectangle(0, 0, width, self.canvas['height'], fill=self.color) total_width = self.canvas['width']
class ProgressBar(object): self.text['text'] = text def _draw_shape(self): width = int(self.canvas['width']) * self.percent // 100 self.canvas.create_rectangle(0, 0, width, self.canvas['height'], fill=self.color) total_width = self.canvas['width']
37
https://:@github.com/hpapaxen/rope.git
2720419618aceab7fba51aaa4d66f7eae005b22d
@@ -129,7 +129,7 @@ class SimilarFinderTest(unittest.TestCase): source = 'x.a = 1\n' finder = similarfinder.SimilarFinder(source) result = list(finder.get_matches('${a} = 1')) - self.assertEquals(1, len(result)) + self.assertEquals(0, len(result)) def test_functions_not_matching_when_only_first_parameters(self): source = 'f(1, 2)\n'
ropetest/refactor/similarfindertest.py
ReplaceText(target='0' @(132,26)->(132,27))
class SimilarFinderTest(unittest.TestCase): source = 'x.a = 1\n' finder = similarfinder.SimilarFinder(source) result = list(finder.get_matches('${a} = 1')) self.assertEquals(1, len(result)) def test_functions_not_matching_when_only_first_parameters(self): source = 'f(1, 2)\n'
class SimilarFinderTest(unittest.TestCase): source = 'x.a = 1\n' finder = similarfinder.SimilarFinder(source) result = list(finder.get_matches('${a} = 1')) self.assertEquals(0, len(result)) def test_functions_not_matching_when_only_first_parameters(self): source = 'f(1, 2)\n'
38
https://:@github.com/hpapaxen/rope.git
0eb3cb58493cdaea83a4e24d47b5bd4dbd19f963
@@ -32,7 +32,7 @@ class BuiltinModule(pyobjects.AbstractModule): result.update(self.initial) for modname in self.submodules: name = modname.split('.')[-1] - result[name] = BuiltinModule(name, self.submodules) + result[name] = BuiltinModule(modname, self.submodules) return result @property
rope/base/builtins.py
ReplaceText(target='modname' @(35,41)->(35,45))
class BuiltinModule(pyobjects.AbstractModule): result.update(self.initial) for modname in self.submodules: name = modname.split('.')[-1] result[name] = BuiltinModule(name, self.submodules) return result @property
class BuiltinModule(pyobjects.AbstractModule): result.update(self.initial) for modname in self.submodules: name = modname.split('.')[-1] result[name] = BuiltinModule(modname, self.submodules) return result @property
39
https://:@github.com/hpapaxen/rope.git
528744bb4bc1b8076680f7c2c1bfac508ddca4f9
@@ -37,7 +37,7 @@ def relative(root, path): if os.path.samefile(root, path): return '/'.join(reversed(rel)) parent = os.path.dirname(path) - if not path or parent == path: + if not parent or parent == path: break rel.append(os.path.basename(path)) path = parent
rope/base/libutils.py
ReplaceText(target='parent' @(40,15)->(40,19))
def relative(root, path): if os.path.samefile(root, path): return '/'.join(reversed(rel)) parent = os.path.dirname(path) if not path or parent == path: break rel.append(os.path.basename(path)) path = parent
def relative(root, path): if os.path.samefile(root, path): return '/'.join(reversed(rel)) parent = os.path.dirname(path) if not parent or parent == path: break rel.append(os.path.basename(path)) path = parent
40
https://:@github.com/benjamincrom/baseball.git
6ef29729ad07458aebe709b4e42f56ecd3761ec4
@@ -111,7 +111,7 @@ def write_game_svg_and_html(game_id, game, output_path): html_filename = game_id + '.html' svg_text = game.get_svg_str() - html_text = HTML_WRAPPER.format(title=game_id, filename=html_filename) + html_text = HTML_WRAPPER.format(title=game_id, filename=svg_filename) output_svg_path = join(output_path, svg_filename) output_html_path = join(output_path, html_filename)
fetch_game.py
ReplaceText(target='svg_filename' @(114,60)->(114,73))
def write_game_svg_and_html(game_id, game, output_path): html_filename = game_id + '.html' svg_text = game.get_svg_str() html_text = HTML_WRAPPER.format(title=game_id, filename=html_filename) output_svg_path = join(output_path, svg_filename) output_html_path = join(output_path, html_filename)
def write_game_svg_and_html(game_id, game, output_path): html_filename = game_id + '.html' svg_text = game.get_svg_str() html_text = HTML_WRAPPER.format(title=game_id, filename=svg_filename) output_svg_path = join(output_path, svg_filename) output_html_path = join(output_path, html_filename)
41
https://:@github.com/sporestack/bitcash.git
dbc65e1b47426e0e4d286db5b27216ec36cb32cf
@@ -16,7 +16,7 @@ def test_set_fee_cache_time(): def test_get_fee(): - assert get_fee(fast=True) != get_fee(fast=False) + assert get_fee(fast=True) >= get_fee(fast=False) class TestFeeCache:
tests/network/test_fees.py
ReplaceText(target='>=' @(19,30)->(19,32))
def test_set_fee_cache_time(): def test_get_fee(): assert get_fee(fast=True) != get_fee(fast=False) class TestFeeCache:
def test_set_fee_cache_time(): def test_get_fee(): assert get_fee(fast=True) >= get_fee(fast=False) class TestFeeCache:
42
https://:@github.com/galaxy-genome-annotation/python-apollo.git
53e514b619844fa1f87179d738b9d29830027300
@@ -29,7 +29,7 @@ class ApolloTestCase(unittest.TestCase): """ org_info = wa.organisms.show_organism(org_id) - if 'directory' in org_info: + if 'directory' not in org_info: time.sleep(1) org_info = wa.organisms.show_organism(org_id)
test/__init__.py
ReplaceText(target=' not in ' @(32,22)->(32,26))
class ApolloTestCase(unittest.TestCase): """ org_info = wa.organisms.show_organism(org_id) if 'directory' in org_info: time.sleep(1) org_info = wa.organisms.show_organism(org_id)
class ApolloTestCase(unittest.TestCase): """ org_info = wa.organisms.show_organism(org_id) if 'directory' not in org_info: time.sleep(1) org_info = wa.organisms.show_organism(org_id)
43
https://:@github.com/jakubplichta/grafana-dashboard-builder.git
3228e6950d65b9bd347cacb56a9e85ec410b14ce
@@ -35,7 +35,7 @@ class Context(object): formatter = string.Formatter() (result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand) while result != to_expand: - (result, to_expand) = (formatter.vformat(to_expand, (), self._context), result) + (result, to_expand) = (formatter.vformat(result, (), self._context), result) return result elif isinstance(to_expand, list): return [self.expand_placeholders(value) for value in to_expand]
grafana_dashboards/context.py
ReplaceText(target='result' @(38,57)->(38,66))
class Context(object): formatter = string.Formatter() (result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand) while result != to_expand: (result, to_expand) = (formatter.vformat(to_expand, (), self._context), result) return result elif isinstance(to_expand, list): return [self.expand_placeholders(value) for value in to_expand]
class Context(object): formatter = string.Formatter() (result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand) while result != to_expand: (result, to_expand) = (formatter.vformat(result, (), self._context), result) return result elif isinstance(to_expand, list): return [self.expand_placeholders(value) for value in to_expand]
44
https://:@github.com/Phylliade/ikpy.git
815dbff3a521532a7b792c309902ffea82abac85
@@ -13,7 +13,7 @@ class TestFK(unittest.TestCase): one_move[5] = np.pi / 4 one_move[6] = -np.pi / 2 one_move[4] = -np.pi / 2 - self.test_pos = one_move + self.test_pos = all_zeros def test_fk_creature(self):
tests/test_fk.py
ReplaceText(target='all_zeros' @(16,24)->(16,32))
class TestFK(unittest.TestCase): one_move[5] = np.pi / 4 one_move[6] = -np.pi / 2 one_move[4] = -np.pi / 2 self.test_pos = one_move def test_fk_creature(self):
class TestFK(unittest.TestCase): one_move[5] = np.pi / 4 one_move[6] = -np.pi / 2 one_move[4] = -np.pi / 2 self.test_pos = all_zeros def test_fk_creature(self):
45
https://:@github.com/tingbot/tingbot-python.git
5374186675f6809faf9ce953fc35c81217348753
@@ -72,7 +72,7 @@ class RunLoop(object): while self.running: if len(self.timers) > 0: try: - self._wait(self.timers[0].next_fire_time) + self._wait(self.timers[-1].next_fire_time) except Exception as e: self._error(e) continue
tingbot/run_loop.py
ReplaceText(target='-1' @(75,43)->(75,44))
class RunLoop(object): while self.running: if len(self.timers) > 0: try: self._wait(self.timers[0].next_fire_time) except Exception as e: self._error(e) continue
class RunLoop(object): while self.running: if len(self.timers) > 0: try: self._wait(self.timers[-1].next_fire_time) except Exception as e: self._error(e) continue
46
https://:@github.com/nyoka-pmml/nyoka.git
8d5c0d31d0bf1e251abe686f06b614a16e5ffcfb
@@ -74,7 +74,7 @@ class TestMethods(unittest.TestCase): self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance") ##3 - self.assertEqual(pmml_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors) + self.assertEqual(pipeline_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors) def test_sklearn_03(self):
nyoka/tests/skl_to_pmml_UnitTest.py
ReplaceText(target='pipeline_obj' @(77,25)->(77,33))
class TestMethods(unittest.TestCase): self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance") ##3 self.assertEqual(pmml_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors) def test_sklearn_03(self):
class TestMethods(unittest.TestCase): self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance") ##3 self.assertEqual(pipeline_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors) def test_sklearn_03(self):
47
https://:@github.com/iris-edu/pyweed.git
77d919acd8d54d4879d1a34598e9e04f16fdf708
@@ -97,7 +97,7 @@ class WaveformEntry(AttribDict): self.error = None - self.start_time = self.distances.arrival + self.config.offsets[0] + self.start_time = self.distances.arrival - self.config.offsets[0] self.end_time = self.distances.arrival + self.config.offsets[1] self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_')
pyweed/waveforms_handler.py
ReplaceText(target='-' @(100,49)->(100,50))
class WaveformEntry(AttribDict): self.error = None self.start_time = self.distances.arrival + self.config.offsets[0] self.end_time = self.distances.arrival + self.config.offsets[1] self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_')
class WaveformEntry(AttribDict): self.error = None self.start_time = self.distances.arrival - self.config.offsets[0] self.end_time = self.distances.arrival + self.config.offsets[1] self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_')
48
https://:@github.com/anaxilaus/coindata.git
2e5067311c4eed50eed41c45f43ad63e8973e579
@@ -52,6 +52,6 @@ def dump_json(data, filepath): try: with open(filepath, 'w') as file: - json.dump(data, filepath) + json.dump(data, file) except TypeError as e: print("Data isn't JSON compatible.\n", e)
coindata/utils.py
ReplaceText(target='file' @(55,28)->(55,36))
def dump_json(data, filepath): try: with open(filepath, 'w') as file: json.dump(data, filepath) except TypeError as e: print("Data isn't JSON compatible.\n", e)
def dump_json(data, filepath): try: with open(filepath, 'w') as file: json.dump(data, file) except TypeError as e: print("Data isn't JSON compatible.\n", e)
49
https://:@github.com/Pixelapse/pyglass.git
a31e95cbc259ce61f5851d6f0d769792aaa182fe
@@ -20,7 +20,7 @@ def preview(src_path): preview_path = thumbnail_preview(src_path) if preview_path: - mimetype = magic.from_file(src_path, mime=True).lower() + mimetype = magic.from_file(preview_path, mime=True).lower() if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]: return preview_path
pyglass/quicklook/api.py
ReplaceText(target='preview_path' @(23,31)->(23,39))
def preview(src_path): preview_path = thumbnail_preview(src_path) if preview_path: mimetype = magic.from_file(src_path, mime=True).lower() if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]: return preview_path
def preview(src_path): preview_path = thumbnail_preview(src_path) if preview_path: mimetype = magic.from_file(preview_path, mime=True).lower() if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]: return preview_path
50
https://:@github.com/erezsh/plyplus.git
8cc69bebfcb2cb0480ac66d07fe1f4b8637bba11
@@ -784,7 +784,7 @@ class _Grammar(object): subtree.extend(child.tail) else: subtree.append(child) - p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) > 1 else subtree[0] + p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) != 1 else subtree[0] else: def p_rule(self, p): p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True)
plyplus/plyplus.py
ReplaceText(target='!=' @(787,98)->(787,99))
class _Grammar(object): subtree.extend(child.tail) else: subtree.append(child) p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) > 1 else subtree[0] else: def p_rule(self, p): p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True)
class _Grammar(object): subtree.extend(child.tail) else: subtree.append(child) p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) != 1 else subtree[0] else: def p_rule(self, p): p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True)
51
https://:@github.com/olls/graphics.git
877ef2670d4ad34c4fcf951dab0922419f081531
@@ -22,7 +22,7 @@ def colorStr(text, color=WHITE): return seq sys.stdout.write(seq + '\n') else: - return seq + return text sys.stdout.write(text + '\n') if __name__ == '__main__':
colors.py
ReplaceText(target='text' @(25,9)->(25,12))
def colorStr(text, color=WHITE): return seq sys.stdout.write(seq + '\n') else: return seq sys.stdout.write(text + '\n') if __name__ == '__main__':
def colorStr(text, color=WHITE): return seq sys.stdout.write(seq + '\n') else: return text sys.stdout.write(text + '\n') if __name__ == '__main__':
52
https://:@github.com/hkwi/twink.git
5ef359deb609fc55659596a8cb327abb9c7e4653
@@ -319,7 +319,7 @@ def ofp_action_set_field(message, offset): cursor = _cursor(offset) offset = cursor.offset - (type,len) = ofp_action_header(message, offset) + (type,len) = ofp_action_header(message, cursor) field = message[cursor.offset:offset+len] cursor.offset = offset+len return namedtuple("ofp_action_set_field",
twink/ofp4/parse.py
ReplaceText(target='cursor' @(322,41)->(322,47))
def ofp_action_set_field(message, offset): cursor = _cursor(offset) offset = cursor.offset (type,len) = ofp_action_header(message, offset) field = message[cursor.offset:offset+len] cursor.offset = offset+len return namedtuple("ofp_action_set_field",
def ofp_action_set_field(message, offset): cursor = _cursor(offset) offset = cursor.offset (type,len) = ofp_action_header(message, cursor) field = message[cursor.offset:offset+len] cursor.offset = offset+len return namedtuple("ofp_action_set_field",
53
https://:@github.com/hkwi/twink.git
fc6f5ad63cb12f9caf5455e3179ecfd9cd9de060
@@ -27,7 +27,7 @@ def _unpack(fmt, message, offset): return struct.unpack_from(fmt, message, offset) def _align(length): - return (length+7)/8*8 + return (length+7)//8*8 # 7.1 def ofp_header(version, type, length, xid):
twink/ofp4/build.py
ReplaceText(target='//' @(30,18)->(30,19))
def _unpack(fmt, message, offset): return struct.unpack_from(fmt, message, offset) def _align(length): return (length+7)/8*8 # 7.1 def ofp_header(version, type, length, xid):
def _unpack(fmt, message, offset): return struct.unpack_from(fmt, message, offset) def _align(length): return (length+7)//8*8 # 7.1 def ofp_header(version, type, length, xid):
54
https://:@github.com/biolab/orange3-datafusion.git
54f941a66a9b369a73190dfe2007e5b6dae1803a
@@ -33,7 +33,7 @@ def _find_completion(fuser, relation): for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type, relation.col_type): if fuser_relation._id == relation._id: - return fuser.complete(fuser_relation) + return fuser.complete(relation) return None
orangecontrib/datafusion/widgets/owcompletionscoring.py
ReplaceText(target='relation' @(36,34)->(36,48))
def _find_completion(fuser, relation): for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type, relation.col_type): if fuser_relation._id == relation._id: return fuser.complete(fuser_relation) return None
def _find_completion(fuser, relation): for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type, relation.col_type): if fuser_relation._id == relation._id: return fuser.complete(relation) return None
55
https://:@github.com/aiqm/torchani.git
5bb6691515e5e56fbe4994b140dd40b73043a33f
@@ -151,7 +151,7 @@ class PrepareInput(torch.nn.Module): new_tensors = [] for t in tensors: new_tensors.append(t.index_select(1, reverse)) - return (species, *tensors) + return (species, *new_tensors) def forward(self, species_coordinates): species, coordinates = species_coordinates
torchani/aev.py
ReplaceText(target='new_tensors' @(154,26)->(154,33))
class PrepareInput(torch.nn.Module): new_tensors = [] for t in tensors: new_tensors.append(t.index_select(1, reverse)) return (species, *tensors) def forward(self, species_coordinates): species, coordinates = species_coordinates
class PrepareInput(torch.nn.Module): new_tensors = [] for t in tensors: new_tensors.append(t.index_select(1, reverse)) return (species, *new_tensors) def forward(self, species_coordinates): species, coordinates = species_coordinates
56
https://:@github.com/aiqm/torchani.git
abc8f7f842ae4b273c6e867b392413dcadd9c921
@@ -593,7 +593,7 @@ def collate_fn(data, chunk_threshold, properties_info): if properties_info['padding_values'][i] is None: prop = torch.stack(prop) else: - prop = torch.nn.utils.rnn.pad_sequence(batch_species, + prop = torch.nn.utils.rnn.pad_sequence(prop, batch_first=True, padding_value=properties_info['padding_values'][i]) # sort with number of atoms
torchani/data/new.py
ReplaceText(target='prop' @(596,51)->(596,64))
def collate_fn(data, chunk_threshold, properties_info): if properties_info['padding_values'][i] is None: prop = torch.stack(prop) else: prop = torch.nn.utils.rnn.pad_sequence(batch_species, batch_first=True, padding_value=properties_info['padding_values'][i]) # sort with number of atoms
def collate_fn(data, chunk_threshold, properties_info): if properties_info['padding_values'][i] is None: prop = torch.stack(prop) else: prop = torch.nn.utils.rnn.pad_sequence(prop, batch_first=True, padding_value=properties_info['padding_values'][i]) # sort with number of atoms
57
https://:@github.com/aiqm/torchani.git
c18f4a5ea1f9732cc07c8816caa401981e43dc48
@@ -274,7 +274,7 @@ def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor, num_atoms = species.shape[1] num_species_pairs = angular_length // angular_sublength # PBC calculation is bypassed if there are no shifts - if shifts.numel() == 1: + if shifts.numel() == 0: atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr) else: atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr)
torchani/aev.py
ReplaceText(target='0' @(277,25)->(277,26))
def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor, num_atoms = species.shape[1] num_species_pairs = angular_length // angular_sublength # PBC calculation is bypassed if there are no shifts if shifts.numel() == 1: atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr) else: atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr)
def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor, num_atoms = species.shape[1] num_species_pairs = angular_length // angular_sublength # PBC calculation is bypassed if there are no shifts if shifts.numel() == 0: atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr) else: atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr)
58
https://:@github.com/cs207group4/cs207-FinalProject.git
e7f1cc613ace275a8d259de7455ee39ca063e029
@@ -120,7 +120,7 @@ class ChemSolver: r.set_initial_value(y0, 0) self._t = [0] self._y = [y0] - while r.successful() and r.t <= t1: + while r.successful() and r.t < t1: self._t.append(r.t + dt) self._y.append(r.integrate(r.t + dt)) self._t = np.array(self._t)
pychemkin/ChemSolver.py
ReplaceText(target='<' @(123,37)->(123,39))
class ChemSolver: r.set_initial_value(y0, 0) self._t = [0] self._y = [y0] while r.successful() and r.t <= t1: self._t.append(r.t + dt) self._y.append(r.integrate(r.t + dt)) self._t = np.array(self._t)
class ChemSolver: r.set_initial_value(y0, 0) self._t = [0] self._y = [y0] while r.successful() and r.t < t1: self._t.append(r.t + dt) self._y.append(r.integrate(r.t + dt)) self._t = np.array(self._t)
59
https://:@github.com/MrLeeh/pyads.git
d14fd2a7bb2d4b784a4f6a47b6981ba2a86b699c
@@ -97,7 +97,7 @@ def set_local_address(ams_netid): else: ams_netid_st = ams_netid - assert isinstance(ams_netid, SAmsNetId) + assert isinstance(ams_netid_st, SAmsNetId) if linux: return adsSetLocalAddress(ams_netid_st)
pyads/ads.py
ReplaceText(target='ams_netid_st' @(100,22)->(100,31))
def set_local_address(ams_netid): else: ams_netid_st = ams_netid assert isinstance(ams_netid, SAmsNetId) if linux: return adsSetLocalAddress(ams_netid_st)
def set_local_address(ams_netid): else: ams_netid_st = ams_netid assert isinstance(ams_netid_st, SAmsNetId) if linux: return adsSetLocalAddress(ams_netid_st)
60
https://:@github.com/pytorch/fairseq.git
0a836276129ef71fa6c44975dd02ab70bccc496d
@@ -58,7 +58,7 @@ class FConvEncoder(FairseqEncoder): self.projections = nn.ModuleList() self.convolutions = nn.ModuleList() for (out_channels, kernel_size) in convolutions: - pad = (kernel_size - 1) // 2 + pad = (kernel_size - 1) / 2 self.projections.append(Linear(in_channels, out_channels) if in_channels != out_channels else None) self.convolutions.append(
fairseq/models/fconv.py
ReplaceText(target='/' @(61,36)->(61,38))
class FConvEncoder(FairseqEncoder): self.projections = nn.ModuleList() self.convolutions = nn.ModuleList() for (out_channels, kernel_size) in convolutions: pad = (kernel_size - 1) // 2 self.projections.append(Linear(in_channels, out_channels) if in_channels != out_channels else None) self.convolutions.append(
class FConvEncoder(FairseqEncoder): self.projections = nn.ModuleList() self.convolutions = nn.ModuleList() for (out_channels, kernel_size) in convolutions: pad = (kernel_size - 1) / 2 self.projections.append(Linear(in_channels, out_channels) if in_channels != out_channels else None) self.convolutions.append(
61
https://:@github.com/pytorch/fairseq.git
f68a44359b6596997b931d2e662a899ffba9d407
@@ -62,7 +62,7 @@ class SinusoidalPositionalEmbedding(nn.Module): # recompute/expand embeddings if needed bsz, seq_len = input.size() max_pos = self.padding_idx + 1 + seq_len - if seq_len > self.weights.size(0): + if max_pos > self.weights.size(0): self.weights = SinusoidalPositionalEmbedding.get_embedding( max_pos, self.embedding_dim,
fairseq/modules/sinusoidal_positional_embedding.py
ReplaceText(target='max_pos' @(65,11)->(65,18))
class SinusoidalPositionalEmbedding(nn.Module): # recompute/expand embeddings if needed bsz, seq_len = input.size() max_pos = self.padding_idx + 1 + seq_len if seq_len > self.weights.size(0): self.weights = SinusoidalPositionalEmbedding.get_embedding( max_pos, self.embedding_dim,
class SinusoidalPositionalEmbedding(nn.Module): # recompute/expand embeddings if needed bsz, seq_len = input.size() max_pos = self.padding_idx + 1 + seq_len if max_pos > self.weights.size(0): self.weights = SinusoidalPositionalEmbedding.get_embedding( max_pos, self.embedding_dim,
62
https://:@github.com/pytorch/fairseq.git
762956a559e65e1e48df8f8b4df515d23b66fddb
@@ -82,7 +82,7 @@ def main(args): train_meter.start() valid_losses = [None] valid_subsets = args.valid_subset.split(',') - while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update: + while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update: # train for one epoch train(args, trainer, task, epoch_itr)
train.py
ReplaceText(target='<' @(85,47)->(85,49))
def main(args): train_meter.start() valid_losses = [None] valid_subsets = args.valid_subset.split(',') while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update: # train for one epoch train(args, trainer, task, epoch_itr)
def main(args): train_meter.start() valid_losses = [None] valid_subsets = args.valid_subset.split(',') while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update: # train for one epoch train(args, trainer, task, epoch_itr)
63
https://:@github.com/pytorch/fairseq.git
e9967cd334783f5da50deadc17cf8a4fc3380171
@@ -82,7 +82,7 @@ def main(args): train_meter.start() valid_losses = [None] valid_subsets = args.valid_subset.split(',') - while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update: + while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update: # train for one epoch train(args, trainer, task, epoch_itr)
train.py
ReplaceText(target='<' @(85,47)->(85,49))
def main(args): train_meter.start() valid_losses = [None] valid_subsets = args.valid_subset.split(',') while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update: # train for one epoch train(args, trainer, task, epoch_itr)
def main(args): train_meter.start() valid_losses = [None] valid_subsets = args.valid_subset.split(',') while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update: # train for one epoch train(args, trainer, task, epoch_itr)
64
https://:@github.com/pytorch/fairseq.git
7bcb487aad8504043d13c9b869d555aa565a46c7
@@ -49,7 +49,7 @@ class LabelSmoothedCrossEntropyCriterion(FairseqCriterion): sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens'] logging_output = { 'loss': utils.item(loss.data) if reduce else loss.data, - 'nll_loss': utils.item(nll_loss.data) if reduce else loss.data, + 'nll_loss': utils.item(nll_loss.data) if reduce else nll_loss.data, 'ntokens': sample['ntokens'], 'sample_size': sample_size, }
fairseq/criterions/label_smoothed_cross_entropy.py
ReplaceText(target='nll_loss' @(52,65)->(52,69))
class LabelSmoothedCrossEntropyCriterion(FairseqCriterion): sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens'] logging_output = { 'loss': utils.item(loss.data) if reduce else loss.data, 'nll_loss': utils.item(nll_loss.data) if reduce else loss.data, 'ntokens': sample['ntokens'], 'sample_size': sample_size, }
class LabelSmoothedCrossEntropyCriterion(FairseqCriterion): sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens'] logging_output = { 'loss': utils.item(loss.data) if reduce else loss.data, 'nll_loss': utils.item(nll_loss.data) if reduce else nll_loss.data, 'ntokens': sample['ntokens'], 'sample_size': sample_size, }
65
https://:@github.com/pytorch/fairseq.git
74efc21403477d103bd426ae64c37b7a30d8f4bf
@@ -154,7 +154,7 @@ class TestIncrementalDecoder(FairseqIncrementalDecoder): probs[:, i, self.dictionary.eos()] = 1.0 # random attention - attn = torch.rand(bbsz, src_len, tgt_len) + attn = torch.rand(bbsz, tgt_len, src_len) return Variable(probs), Variable(attn)
tests/utils.py
ArgSwap(idxs=1<->2 @(157,15)->(157,25))
class TestIncrementalDecoder(FairseqIncrementalDecoder): probs[:, i, self.dictionary.eos()] = 1.0 # random attention attn = torch.rand(bbsz, src_len, tgt_len) return Variable(probs), Variable(attn)
class TestIncrementalDecoder(FairseqIncrementalDecoder): probs[:, i, self.dictionary.eos()] = 1.0 # random attention attn = torch.rand(bbsz, tgt_len, src_len) return Variable(probs), Variable(attn)
66
https://:@github.com/pytorch/fairseq.git
dfd77717b91a6e233829735795ab49d6fd85c0b3
@@ -93,7 +93,7 @@ class CosineSchedule(FairseqLRScheduler): else: i = math.floor(curr_updates / self.period) t_i = self.period - t_curr = num_updates - (self.period * i) + t_curr = curr_updates - (self.period * i) lr_shrink = self.lr_shrink ** i min_lr = self.min_lr * lr_shrink
fairseq/optim/lr_scheduler/cosine_lr_scheduler.py
ReplaceText(target='curr_updates' @(96,25)->(96,36))
class CosineSchedule(FairseqLRScheduler): else: i = math.floor(curr_updates / self.period) t_i = self.period t_curr = num_updates - (self.period * i) lr_shrink = self.lr_shrink ** i min_lr = self.min_lr * lr_shrink
class CosineSchedule(FairseqLRScheduler): else: i = math.floor(curr_updates / self.period) t_i = self.period t_curr = curr_updates - (self.period * i) lr_shrink = self.lr_shrink ** i min_lr = self.min_lr * lr_shrink
67
https://:@github.com/pytorch/fairseq.git
0eea6923b9d7f408e667714709b070171ac7fe05
@@ -312,7 +312,7 @@ def make_positions(tensor, padding_idx, left_pad, onnx_trace=False): positions = range_buf.expand_as(tensor) if left_pad: positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1) - return positions * mask.long() + positions * (1 - mask.long()) + return positions * mask.long() + padding_idx * (1 - mask.long()) max_pos = padding_idx + 1 + tensor.size(1) if not hasattr(make_positions, 'range_buf'):
fairseq/utils.py
ReplaceText(target='padding_idx' @(315,41)->(315,50))
def make_positions(tensor, padding_idx, left_pad, onnx_trace=False): positions = range_buf.expand_as(tensor) if left_pad: positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1) return positions * mask.long() + positions * (1 - mask.long()) max_pos = padding_idx + 1 + tensor.size(1) if not hasattr(make_positions, 'range_buf'):
def make_positions(tensor, padding_idx, left_pad, onnx_trace=False): positions = range_buf.expand_as(tensor) if left_pad: positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1) return positions * mask.long() + padding_idx * (1 - mask.long()) max_pos = padding_idx + 1 + tensor.size(1) if not hasattr(make_positions, 'range_buf'):
68
https://:@github.com/pytorch/fairseq.git
4d3401b09f155995cd81fd394dfa50bf65ee8e5f
@@ -183,7 +183,7 @@ class Sampling(Search): lprobs = lprobs[:, ::beam_size, :].contiguous() # we exclude the first two vocab items, one of which is pad - assert self.pad == 1, 'sampling assumes the first two symbols can be ignored' + assert self.pad <= 1, 'sampling assumes the first two symbols can be ignored' lprobs_nopad = lprobs[:, :, 2:] # only sample from top-k candidates
fairseq/search.py
ReplaceText(target='<=' @(186,24)->(186,26))
class Sampling(Search): lprobs = lprobs[:, ::beam_size, :].contiguous() # we exclude the first two vocab items, one of which is pad assert self.pad == 1, 'sampling assumes the first two symbols can be ignored' lprobs_nopad = lprobs[:, :, 2:] # only sample from top-k candidates
class Sampling(Search): lprobs = lprobs[:, ::beam_size, :].contiguous() # we exclude the first two vocab items, one of which is pad assert self.pad <= 1, 'sampling assumes the first two symbols can be ignored' lprobs_nopad = lprobs[:, :, 2:] # only sample from top-k candidates
69
https://:@github.com/pytorch/fairseq.git
2340832fdd7acaaaf07626daa6a0cef6fda06cd1
@@ -160,7 +160,7 @@ def main(args): )) # update running id counter - start_id += len(results) + start_id += len(inputs) def cli_main():
interactive.py
ReplaceText(target='inputs' @(163,24)->(163,31))
def main(args): )) # update running id counter start_id += len(results) def cli_main():
def main(args): )) # update running id counter start_id += len(inputs) def cli_main():
70
https://:@github.com/pytorch/fairseq.git
39a60b844aad67aa59267d873edeb4948f6f0af9
@@ -351,7 +351,7 @@ class LSTMDecoder(FairseqIncrementalDecoder): self.additional_fc = Linear(hidden_size, out_embed_dim) if adaptive_softmax_cutoff is not None: # setting adaptive_softmax dropout to dropout_out for now but can be redefined - self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, embed_dim, adaptive_softmax_cutoff, + self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, hidden_size, adaptive_softmax_cutoff, dropout=dropout_out) elif not self.share_input_output_embed: self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out)
fairseq/models/lstm.py
ReplaceText(target='hidden_size' @(354,68)->(354,77))
class LSTMDecoder(FairseqIncrementalDecoder): self.additional_fc = Linear(hidden_size, out_embed_dim) if adaptive_softmax_cutoff is not None: # setting adaptive_softmax dropout to dropout_out for now but can be redefined self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, embed_dim, adaptive_softmax_cutoff, dropout=dropout_out) elif not self.share_input_output_embed: self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out)
class LSTMDecoder(FairseqIncrementalDecoder): self.additional_fc = Linear(hidden_size, out_embed_dim) if adaptive_softmax_cutoff is not None: # setting adaptive_softmax dropout to dropout_out for now but can be redefined self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, hidden_size, adaptive_softmax_cutoff, dropout=dropout_out) elif not self.share_input_output_embed: self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out)
71
https://:@github.com/pytorch/fairseq.git
49177c99c45f7d6e99a8f1500d16396e2d7b4519
@@ -498,7 +498,7 @@ class TransformerDecoder(FairseqIncrementalDecoder): del state_dict[k] version_key = '{}.version'.format(name) - if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) < 2: + if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) <= 2: # earlier checkpoints did not normalize after the stack of layers self.layer_norm = None self.normalize = False
fairseq/models/transformer.py
ReplaceText(target='<=' @(501,73)->(501,74))
class TransformerDecoder(FairseqIncrementalDecoder): del state_dict[k] version_key = '{}.version'.format(name) if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) < 2: # earlier checkpoints did not normalize after the stack of layers self.layer_norm = None self.normalize = False
class TransformerDecoder(FairseqIncrementalDecoder): del state_dict[k] version_key = '{}.version'.format(name) if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) <= 2: # earlier checkpoints did not normalize after the stack of layers self.layer_norm = None self.normalize = False
72
https://:@github.com/pytorch/fairseq.git
5d7a81099462e9f19715ce5fa37c03816a750e12
@@ -412,7 +412,7 @@ class LevenshteinTransformerModel(FairseqNATModel): max_lens = torch.zeros_like(output_tokens).fill_(255) else: if encoder_out.encoder_padding_mask is None: - max_src_len = encoder_out.encoder_out.size(1) + max_src_len = encoder_out.encoder_out.size(0) src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len) else: src_lens = (~encoder_out.encoder_padding_mask).sum(1)
fairseq/models/nat/levenshtein_transformer.py
ReplaceText(target='0' @(415,59)->(415,60))
class LevenshteinTransformerModel(FairseqNATModel): max_lens = torch.zeros_like(output_tokens).fill_(255) else: if encoder_out.encoder_padding_mask is None: max_src_len = encoder_out.encoder_out.size(1) src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len) else: src_lens = (~encoder_out.encoder_padding_mask).sum(1)
class LevenshteinTransformerModel(FairseqNATModel): max_lens = torch.zeros_like(output_tokens).fill_(255) else: if encoder_out.encoder_padding_mask is None: max_src_len = encoder_out.encoder_out.size(0) src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len) else: src_lens = (~encoder_out.encoder_padding_mask).sum(1)
73
https://:@github.com/pytorch/fairseq.git
431d604f696a15c06fceab56b4ace271bb85e74b
@@ -331,7 +331,7 @@ class SequenceGenerator(object): avg_attn_scores = avg_attn_scores[0] if avg_attn_scores is not None: if attn is None: - attn = scores.new(bsz * beam_size, src_tokens.size(1), max_len + 2) + attn = scores.new(bsz * beam_size, avg_attn_scores.size(1), max_len + 2) attn_buf = attn.clone() attn[:, :, step + 1].copy_(avg_attn_scores)
fairseq/sequence_generator.py
ReplaceText(target='avg_attn_scores' @(334,55)->(334,65))
class SequenceGenerator(object): avg_attn_scores = avg_attn_scores[0] if avg_attn_scores is not None: if attn is None: attn = scores.new(bsz * beam_size, src_tokens.size(1), max_len + 2) attn_buf = attn.clone() attn[:, :, step + 1].copy_(avg_attn_scores)
class SequenceGenerator(object): avg_attn_scores = avg_attn_scores[0] if avg_attn_scores is not None: if attn is None: attn = scores.new(bsz * beam_size, avg_attn_scores.size(1), max_len + 2) attn_buf = attn.clone() attn[:, :, step + 1].copy_(avg_attn_scores)
74
https://:@github.com/pytorch/fairseq.git
4f8b0643c80d6a41039ae29e94fca6b44de8791a
@@ -138,7 +138,7 @@ def should_stop_early(args, valid_loss): return False else: should_stop_early.num_runs += 1 - return should_stop_early.num_runs > args.patience + return should_stop_early.num_runs >= args.patience @metrics.aggregate('train')
fairseq_cli/train.py
ReplaceText(target='>=' @(141,42)->(141,43))
def should_stop_early(args, valid_loss): return False else: should_stop_early.num_runs += 1 return should_stop_early.num_runs > args.patience @metrics.aggregate('train')
def should_stop_early(args, valid_loss): return False else: should_stop_early.num_runs += 1 return should_stop_early.num_runs >= args.patience @metrics.aggregate('train')
75
https://:@github.com/pytorch/fairseq.git
9a718e29855713a51877237b2dcc25e39c234c82
@@ -110,5 +110,5 @@ class TranslationFromPretrainedBARTTask(TranslationTask): for s_t in src_tokens: s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)]) source_tokens.append(s_t) - dataset = LanguagePairDataset(src_tokens, src_lengths, self.source_dictionary) + dataset = LanguagePairDataset(source_tokens, src_lengths, self.source_dictionary) return dataset
fairseq/tasks/translation_from_pretrained_bart.py
ReplaceText(target='source_tokens' @(113,38)->(113,48))
class TranslationFromPretrainedBARTTask(TranslationTask): for s_t in src_tokens: s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)]) source_tokens.append(s_t) dataset = LanguagePairDataset(src_tokens, src_lengths, self.source_dictionary) return dataset
class TranslationFromPretrainedBARTTask(TranslationTask): for s_t in src_tokens: s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)]) source_tokens.append(s_t) dataset = LanguagePairDataset(source_tokens, src_lengths, self.source_dictionary) return dataset
76
https://:@github.com/pytorch/fairseq.git
b689b6ff3ab7b806217b8aa41821bb8fc85f7cd8
@@ -264,7 +264,7 @@ class LanguagePairDataset(FairseqDataset): tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]]) bos = self.src_dict.bos() - if self.src[index][-1] != bos: + if self.src[index][0] != bos: src_item = torch.cat([torch.LongTensor([bos]), self.src[index]]) if self.remove_eos_from_source:
fairseq/data/language_pair_dataset.py
ReplaceText(target='0' @(267,31)->(267,33))
class LanguagePairDataset(FairseqDataset): tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]]) bos = self.src_dict.bos() if self.src[index][-1] != bos: src_item = torch.cat([torch.LongTensor([bos]), self.src[index]]) if self.remove_eos_from_source:
class LanguagePairDataset(FairseqDataset): tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]]) bos = self.src_dict.bos() if self.src[index][0] != bos: src_item = torch.cat([torch.LongTensor([bos]), self.src[index]]) if self.remove_eos_from_source:
77
https://:@github.com/prprprus/PyMySQLPool.git
66b07cdf844554245cf209a72de89bd17133269c
@@ -169,7 +169,7 @@ class Pool(object): if self.ping_check: now = int(time()) timeout = now - if isinstance(int, self.ping_check): + if isinstance(self.ping_check, int): timeout = timeout - self.ping_check if not hasattr(c, '__ping_check_timestamp'): c.__ping_check_timestamp = now
pymysqlpool/pool.py
ArgSwap(idxs=0<->1 @(172,15)->(172,25))
class Pool(object): if self.ping_check: now = int(time()) timeout = now if isinstance(int, self.ping_check): timeout = timeout - self.ping_check if not hasattr(c, '__ping_check_timestamp'): c.__ping_check_timestamp = now
class Pool(object): if self.ping_check: now = int(time()) timeout = now if isinstance(self.ping_check, int): timeout = timeout - self.ping_check if not hasattr(c, '__ping_check_timestamp'): c.__ping_check_timestamp = now
78
https://:@github.com/dailymuse/oz.git
a15adf73c721d07b9dac886fcc27145e2449563c
@@ -173,7 +173,7 @@ class S3File(CDNFile): def copy(self, new_path, replace=False): """Uses boto to copy the file to the new path instead of uploading another file to the new key""" - if replace or get_file(new_path).exists(): + if replace or not get_file(new_path).exists(): self.key.copy(self.key.bucket, new_path) return True return False
oz/aws_cdn/__init__.py
ReplaceText(target='not ' @(176,22)->(176,22))
class S3File(CDNFile): def copy(self, new_path, replace=False): """Uses boto to copy the file to the new path instead of uploading another file to the new key""" if replace or get_file(new_path).exists(): self.key.copy(self.key.bucket, new_path) return True return False
class S3File(CDNFile): def copy(self, new_path, replace=False): """Uses boto to copy the file to the new path instead of uploading another file to the new key""" if replace or not get_file(new_path).exists(): self.key.copy(self.key.bucket, new_path) return True return False
79
https://:@github.com/juju/amulet.git
016bfab60aca89cbcb58e80f4103e371a77b06ba
@@ -77,7 +77,7 @@ class Deployment(object): pass # Copy the current parent directory to temp and deploy that elif self.charm_name: if charm_name == self.charm_name: - charm = os.getcwd() + charm_branch = os.getcwd() self.services[service] = {'branch': charm_branch} if units > 1:
amulet/deployer.py
ReplaceText(target='charm_branch' @(80,16)->(80,21))
class Deployment(object): pass # Copy the current parent directory to temp and deploy that elif self.charm_name: if charm_name == self.charm_name: charm = os.getcwd() self.services[service] = {'branch': charm_branch} if units > 1:
class Deployment(object): pass # Copy the current parent directory to temp and deploy that elif self.charm_name: if charm_name == self.charm_name: charm_branch = os.getcwd() self.services[service] = {'branch': charm_branch} if units > 1:
80
https://:@github.com/gitpython-developers/gitdb.git
ca829e0b341dd5c3ae1408b24702f2c75db6ec73
@@ -445,7 +445,7 @@ class DeltaApplyReader(LazyMixin): #{ Configuration - if not has_perf_mod: + if has_perf_mod: _set_cache_ = _set_cache_brute_ else: _set_cache_ = _set_cache_too_slow_without_c
stream.py
ReplaceText(target='' @(448,4)->(448,8))
class DeltaApplyReader(LazyMixin): #{ Configuration if not has_perf_mod: _set_cache_ = _set_cache_brute_ else: _set_cache_ = _set_cache_too_slow_without_c
class DeltaApplyReader(LazyMixin): #{ Configuration if has_perf_mod: _set_cache_ = _set_cache_brute_ else: _set_cache_ = _set_cache_too_slow_without_c
81
https://:@github.com/longld/peda.git
82fcb5a12c92c27fc5722772a84df47b996d3d03
@@ -4746,7 +4746,7 @@ class PEDACmd(object): step = peda.intsize() if not peda.is_address(address): # cannot determine address - msg("Invalid $SP address: 0x%x" % sp, "red") + msg("Invalid $SP address: 0x%x" % address, "red") return for i in range(count): if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)):
peda.py
ReplaceText(target='address' @(4749,46)->(4749,48))
class PEDACmd(object): step = peda.intsize() if not peda.is_address(address): # cannot determine address msg("Invalid $SP address: 0x%x" % sp, "red") return for i in range(count): if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)):
class PEDACmd(object): step = peda.intsize() if not peda.is_address(address): # cannot determine address msg("Invalid $SP address: 0x%x" % address, "red") return for i in range(count): if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)):
82
https://:@github.com/TheGhouls/oct.git
1f9ea29181962353fe0ea275cb4ba4ec9ae93142
@@ -18,7 +18,7 @@ class Report(object): self.set_statics() def set_statics(self): - if os.path.exists(self.results_dir): + if not os.path.exists(self.results_dir): return try: shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css'))
oct/results/reportwriter.py
ReplaceText(target='not ' @(21,11)->(21,11))
class Report(object): self.set_statics() def set_statics(self): if os.path.exists(self.results_dir): return try: shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css'))
class Report(object): self.set_statics() def set_statics(self): if not os.path.exists(self.results_dir): return try: shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css'))
83
https://:@github.com/ASPP/pelita.git
bee6872dbd95a1e526305ef39f42ac537fd2f708
@@ -105,7 +105,7 @@ def create_maze(layout_mesh): Mesh of lists of MazeComponents """ - maze_mesh = Mesh(layout_mesh.height, layout_mesh.width, + maze_mesh = Mesh(layout_mesh.width, layout_mesh.height, data=[[] for i in range(len(layout_mesh))]) for index in maze_mesh.iterkeys(): if layout_mesh[index] == CTFUniverse.wall:
pelita/universe.py
ArgSwap(idxs=0<->1 @(108,16)->(108,20))
def create_maze(layout_mesh): Mesh of lists of MazeComponents """ maze_mesh = Mesh(layout_mesh.height, layout_mesh.width, data=[[] for i in range(len(layout_mesh))]) for index in maze_mesh.iterkeys(): if layout_mesh[index] == CTFUniverse.wall:
def create_maze(layout_mesh): Mesh of lists of MazeComponents """ maze_mesh = Mesh(layout_mesh.width, layout_mesh.height, data=[[] for i in range(len(layout_mesh))]) for index in maze_mesh.iterkeys(): if layout_mesh[index] == CTFUniverse.wall:
84
https://:@github.com/ASPP/pelita.git
508cd180dce7b72ab248211c977c8525a9c023de
@@ -31,7 +31,7 @@ def __init__(self, index, initial_pos, team, homezone, @property def in_own_zone(self): - return self.homezone[0] <= self.current_pos[1] <= self.homezone[1] + return self.homezone[0] <= self.current_pos[0] <= self.homezone[1] def move(self, new_pos): self.current_pos = new_pos
pelita/universe.py
ReplaceText(target='0' @(34,52)->(34,53))
def __init__(self, index, initial_pos, team, homezone, @property def in_own_zone(self): return self.homezone[0] <= self.current_pos[1] <= self.homezone[1] def move(self, new_pos): self.current_pos = new_pos
def __init__(self, index, initial_pos, team, homezone, @property def in_own_zone(self): return self.homezone[0] <= self.current_pos[0] <= self.homezone[1] def move(self, new_pos): self.current_pos = new_pos
85
https://:@github.com/ASPP/pelita.git
6b76e416da2dc0d18224e47d7b176dad967e15b2
@@ -182,7 +182,7 @@ def a_star(self, initial, target): else: seen.append(current) for pos in self.adjacency[current]: - heapq.heappush(to_visit, (datamodel.manhattan_dist(current, pos), (pos))) + heapq.heappush(to_visit, (datamodel.manhattan_dist(target, pos), (pos))) # Now back-track using seen to determine how we got here. # Initialise the path with current node, i.e. position of food.
pelita/game_master.py
ReplaceText(target='target' @(185,71)->(185,78))
def a_star(self, initial, target): else: seen.append(current) for pos in self.adjacency[current]: heapq.heappush(to_visit, (datamodel.manhattan_dist(current, pos), (pos))) # Now back-track using seen to determine how we got here. # Initialise the path with current node, i.e. position of food.
def a_star(self, initial, target): else: seen.append(current) for pos in self.adjacency[current]: heapq.heappush(to_visit, (datamodel.manhattan_dist(target, pos), (pos))) # Now back-track using seen to determine how we got here. # Initialise the path with current node, i.e. position of food.
86
https://:@github.com/ASPP/pelita.git
fa2505d44ae3d3724f7fa979c0167f03bf7424f7
@@ -136,7 +136,7 @@ def play(self): if self.universe.teams[0].score < self.universe.teams[1].score: events.append(datamodel.TeamWins(1)) elif self.universe.teams[0].score > self.universe.teams[1].score: - events.append(datamodel.TeamWins(1)) + events.append(datamodel.TeamWins(0)) else: events.append(datamodel.GameDraw()) self.send_to_viewers(round_index, None, events)
pelita/game_master.py
ReplaceText(target='0' @(139,45)->(139,46))
def play(self): if self.universe.teams[0].score < self.universe.teams[1].score: events.append(datamodel.TeamWins(1)) elif self.universe.teams[0].score > self.universe.teams[1].score: events.append(datamodel.TeamWins(1)) else: events.append(datamodel.GameDraw()) self.send_to_viewers(round_index, None, events)
def play(self): if self.universe.teams[0].score < self.universe.teams[1].score: events.append(datamodel.TeamWins(1)) elif self.universe.teams[0].score > self.universe.teams[1].score: events.append(datamodel.TeamWins(0)) else: events.append(datamodel.GameDraw()) self.send_to_viewers(round_index, None, events)
87
https://:@github.com/ASPP/pelita.git
4044f845e54c2077d6896010c25fcc123fc10203
@@ -78,7 +78,7 @@ def test_equal_positions(self): layout = create_layout(layout_str) assert layout.bots == [(1, 1), (1, 1)] assert layout.enemy == [(1, 1), (1, 1)] - setup_test_game(layout=layout) + setup_test_game(layout=layout_str) def test_define_after(self): layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None)
test/test_team.py
ReplaceText(target='layout_str' @(81,31)->(81,37))
def test_equal_positions(self): layout = create_layout(layout_str) assert layout.bots == [(1, 1), (1, 1)] assert layout.enemy == [(1, 1), (1, 1)] setup_test_game(layout=layout) def test_define_after(self): layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None)
def test_equal_positions(self): layout = create_layout(layout_str) assert layout.bots == [(1, 1), (1, 1)] assert layout.enemy == [(1, 1), (1, 1)] setup_test_game(layout=layout_str) def test_define_after(self): layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None)
88
https://:@github.com/ASPP/pelita.git
6fd0a9d2af44c491c1cc6774c3a169e97e2040be
@@ -398,7 +398,7 @@ def _team(self): @property def turn(self): """ The turn of our bot. """ - return self.bot_index // 2 + return self.bot_index % 2 @property def other(self):
pelita/player/team.py
ReplaceText(target='%' @(401,30)->(401,32))
def _team(self): @property def turn(self): """ The turn of our bot. """ return self.bot_index // 2 @property def other(self):
def _team(self): @property def turn(self): """ The turn of our bot. """ return self.bot_index % 2 @property def other(self):
89
https://:@github.com/iotaledger/ccurl.interface.py.git
eb7f9190d24995d3f8d03a8350382ab6045a6e67
@@ -44,7 +44,7 @@ gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your mwm = 14 # target is mainnet -bundle = entangled_interface.local_attach_to_tangle(pb, gta['trunkTransaction'], gta['branchTransaction'], mwm) +bundle = entangled_interface.local_attach_to_tangle(pb, gta['branchTransaction'],gta['trunkTransaction'], mwm) bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
examples/with_entangled.py
ArgSwap(idxs=1<->2 @(47,9)->(47,51))
gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your mwm = 14 # target is mainnet bundle = entangled_interface.local_attach_to_tangle(pb, gta['trunkTransaction'], gta['branchTransaction'], mwm) bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your mwm = 14 # target is mainnet bundle = entangled_interface.local_attach_to_tangle(pb, gta['branchTransaction'],gta['trunkTransaction'], mwm) bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
90
https://:@github.com/softlayer/softlayer-python.git
53731de7e51d31475cc224aceb0f3ff7217cdafd
@@ -153,7 +153,7 @@ class NetworkManager(object): ('privateResidenceFlag', private_residence), ('state', state), ('postalCode', postal_code)]: - if key is not None: + if value is not None: update[key] = value # If there's anything to update, update it
SoftLayer/managers/network.py
ReplaceText(target='value' @(156,15)->(156,18))
class NetworkManager(object): ('privateResidenceFlag', private_residence), ('state', state), ('postalCode', postal_code)]: if key is not None: update[key] = value # If there's anything to update, update it
class NetworkManager(object): ('privateResidenceFlag', private_residence), ('state', state), ('postalCode', postal_code)]: if value is not None: update[key] = value # If there's anything to update, update it
91
https://:@github.com/softlayer/softlayer-python.git
dcf66e15711e47c594f20ffac7605bfc6d1a8746
@@ -15,7 +15,7 @@ import click type=click.Choice(['vs', 'vlan', 'server']), help='Firewall type', required=True) -@click.option('--high-availability', '--ha', +@click.option('--ha', '--high-availability', is_flag=True, help='High available firewall option') @environment.pass_env
SoftLayer/CLI/firewall/add.py
ArgSwap(idxs=0<->1 @(18,1)->(18,13))
import click type=click.Choice(['vs', 'vlan', 'server']), help='Firewall type', required=True) @click.option('--high-availability', '--ha', is_flag=True, help='High available firewall option') @environment.pass_env
import click type=click.Choice(['vs', 'vlan', 'server']), help='Firewall type', required=True) @click.option('--ha', '--high-availability', is_flag=True, help='High available firewall option') @environment.pass_env
92
https://:@github.com/softlayer/softlayer-python.git
f0840e302d486d6002a14419bbde85c1deedaf6a
@@ -271,7 +271,7 @@ class BlockStorageManager(utils.IdentifierMixin, object): package, 'performance_storage_iscsi' ), - storage_utils.find_performance_space_price(package, iops), + storage_utils.find_performance_space_price(package, size), storage_utils.find_performance_iops_price(package, size, iops), ] elif storage_type == 'storage_service_enterprise':
SoftLayer/managers/block.py
ReplaceText(target='size' @(274,68)->(274,72))
class BlockStorageManager(utils.IdentifierMixin, object): package, 'performance_storage_iscsi' ), storage_utils.find_performance_space_price(package, iops), storage_utils.find_performance_iops_price(package, size, iops), ] elif storage_type == 'storage_service_enterprise':
class BlockStorageManager(utils.IdentifierMixin, object): package, 'performance_storage_iscsi' ), storage_utils.find_performance_space_price(package, size), storage_utils.find_performance_iops_price(package, size, iops), ] elif storage_type == 'storage_service_enterprise':
93
https://:@github.com/softlayer/softlayer-python.git
4418057fc0e3632aba2d89b6e42494c79cadd16a
@@ -367,7 +367,7 @@ class VSManager(utils.IdentifierMixin, object): if datacenter: data["datacenter"] = {"name": datacenter} - if private_vlan and public_vlan: + if private_vlan or public_vlan: network_components = self._create_network_components(public_vlan, private_vlan, private_subnet, public_subnet) data.update(network_components)
SoftLayer/managers/vs.py
ReplaceText(target='or' @(370,24)->(370,27))
class VSManager(utils.IdentifierMixin, object): if datacenter: data["datacenter"] = {"name": datacenter} if private_vlan and public_vlan: network_components = self._create_network_components(public_vlan, private_vlan, private_subnet, public_subnet) data.update(network_components)
class VSManager(utils.IdentifierMixin, object): if datacenter: data["datacenter"] = {"name": datacenter} if private_vlan or public_vlan: network_components = self._create_network_components(public_vlan, private_vlan, private_subnet, public_subnet) data.update(network_components)
94
https://:@github.com/softlayer/softlayer-python.git
58b27c6bf5400a717acd00b7866964ef11f36e59
@@ -87,6 +87,6 @@ def cli(env, identifier): for guest in guests: real_guest = guest.get('virtualGuest') member_table.add_row([ - guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate')) + real_guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate')) ]) env.fout(member_table)
SoftLayer/CLI/autoscale/detail.py
ReplaceText(target='real_guest' @(90,12)->(90,17))
def cli(env, identifier): for guest in guests: real_guest = guest.get('virtualGuest') member_table.add_row([ guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate')) ]) env.fout(member_table)
def cli(env, identifier): for guest in guests: real_guest = guest.get('virtualGuest') member_table.add_row([ real_guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate')) ]) env.fout(member_table)
95
https://:@github.com/data-8/datascience.git
fd9aceb598290fb89a0f3131c3fb39dde18ef543
@@ -446,7 +446,7 @@ class Table(collections.abc.Mapping): count | points 9 | 10 """ - percentiles = [percentile(self[column_name], p) for column_name in self] + percentiles = [percentile(p, self[column_name]) for column_name in self] return Table(percentiles, self.column_labels) ################## # Export/Display #
datascience/tables.py
ArgSwap(idxs=0<->1 @(449,23)->(449,33))
class Table(collections.abc.Mapping): count | points 9 | 10 """ percentiles = [percentile(self[column_name], p) for column_name in self] return Table(percentiles, self.column_labels) ################## # Export/Display #
class Table(collections.abc.Mapping): count | points 9 | 10 """ percentiles = [percentile(p, self[column_name]) for column_name in self] return Table(percentiles, self.column_labels) ################## # Export/Display #
96
https://:@github.com/data-8/datascience.git
084450f127ecc490b887cad82fa43cda5f9b32fe
@@ -2255,7 +2255,7 @@ class Table(collections.abc.MutableMapping): space_count[labels[i]] += 1 return updated_labels return labels - yticks = make_unique_labels(labels) + yticks = make_unique_labels(yticks) print("yticks: " + str(yticks)) print("ylabel: " + str(ylabel))
datascience/tables.py
ReplaceText(target='yticks' @(2258,36)->(2258,42))
class Table(collections.abc.MutableMapping): space_count[labels[i]] += 1 return updated_labels return labels yticks = make_unique_labels(labels) print("yticks: " + str(yticks)) print("ylabel: " + str(ylabel))
class Table(collections.abc.MutableMapping): space_count[labels[i]] += 1 return updated_labels return labels yticks = make_unique_labels(yticks) print("yticks: " + str(yticks)) print("ylabel: " + str(ylabel))
97
https://:@github.com/dnaeon/py-vpoller.git
81769f6f8d9cb0dfc8cbc39a44027afa7d459636
@@ -51,6 +51,6 @@ def task(name, required=None): result = {'success': 1, 'msg': e.message} finally: return result - registry.register(name=name, fn=fn, required=required) + registry.register(name=name, fn=wrapper, required=required) return wrapper return decorator
src/vpoller/decorators.py
ReplaceText(target='wrapper' @(54,40)->(54,42))
def task(name, required=None): result = {'success': 1, 'msg': e.message} finally: return result registry.register(name=name, fn=fn, required=required) return wrapper return decorator
def task(name, required=None): result = {'success': 1, 'msg': e.message} finally: return result registry.register(name=name, fn=wrapper, required=required) return wrapper return decorator
98
https://:@github.com/enthought/qt_binder.git
68381b406035f2ce9666cb8ef1ab2e8e57cf8bf8
@@ -58,4 +58,4 @@ else: loader = RecordingUiLoader() ui = loader.load(path) - return ui, ui.to_be_bound() + return ui, loader.to_be_bound()
qt_binder/qt/ui_loader.py
ReplaceText(target='loader' @(61,19)->(61,21))
else: loader = RecordingUiLoader() ui = loader.load(path) return ui, ui.to_be_bound()
else: loader = RecordingUiLoader() ui = loader.load(path) return ui, loader.to_be_bound()
99
https://:@github.com/ggozad/behaving.git
1bc546aa03f9d42ff78a0a79e0894e488edc9add
@@ -40,7 +40,7 @@ def should_receive_email(context, address): def click_link_in_email(context, address): mails = context.mail.user_messages(address) assert mails, u'message not found' - mail = email.message_from_string(mails[-1]) + mail = email.message_from_string(mails[0]) links = URL_RE.findall(str(mail).replace('=\n', '')) assert links, u'link not found' url = links[0]
src/behaving/mail/steps.py
ReplaceText(target='0' @(43,43)->(43,45))
def should_receive_email(context, address): def click_link_in_email(context, address): mails = context.mail.user_messages(address) assert mails, u'message not found' mail = email.message_from_string(mails[-1]) links = URL_RE.findall(str(mail).replace('=\n', '')) assert links, u'link not found' url = links[0]
def should_receive_email(context, address): def click_link_in_email(context, address): mails = context.mail.user_messages(address) assert mails, u'message not found' mail = email.message_from_string(mails[0]) links = URL_RE.findall(str(mail).replace('=\n', '')) assert links, u'link not found' url = links[0]

See Allamanis et al., 2021 (NeurIPS 2021) for more information.

Downloads last month
2
Edit dataset card