Dataset Viewer
	| Unnamed: 0
				 int64 0 2.44k | repo
				 stringlengths 32 81 | hash
				 stringlengths 40 40 | diff
				 stringlengths 113 1.17k | old_path
				 stringlengths 5 84 | rewrite
				 stringlengths 34 79 | initial_state
				 stringlengths 75 980 | final_state
				 stringlengths 76 980 | 
|---|---|---|---|---|---|---|---|
| 0 | 
	https://:@github.com/emedvedev/attention-ocr.git | 
	291042e7cb623c8a908e9badd132c1fa2360288c | 
	@@ -465,7 +465,7 @@ class Model(object):
             mh = 32
             mw = math.floor(1. * w / h * mh)
             img = img.resize(
-                    (mw, h),
+                    (mw, mh),
                     Image.ANTIALIAS)
             img_data = np.asarray(img, dtype=np.uint8)
             for idx in xrange(len(output)):
 | 
	aocr/model/model.py | 
	ReplaceText(target='mh' @(468,25)->(468,26)) | 
	class Model(object):
             mh = 32
             mw = math.floor(1. * w / h * mh)
             img = img.resize(
                     (mw, h),
                     Image.ANTIALIAS)
             img_data = np.asarray(img, dtype=np.uint8)
             for idx in xrange(len(output)): | 
	class Model(object):
             mh = 32
             mw = math.floor(1. * w / h * mh)
             img = img.resize(
                     (mw, mh),
                     Image.ANTIALIAS)
             img_data = np.asarray(img, dtype=np.uint8)
             for idx in xrange(len(output)): | 
| 1 | 
	https://:@github.com/emedvedev/attention-ocr.git | 
	6e6593c27fe0e63118adadf11562b1c4699b14e3 | 
	@@ -477,7 +477,7 @@ class Model(object):
                         attention_orig[i] = attention[int(i/4)-1]
                 attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same')
                 attention_orig = np.maximum(attention_orig, 0.3)
-                attention_out = np.zeros((h, mw))
+                attention_out = np.zeros((mh, mw))
                 for i in xrange(mw):
                     attention_out[:, i] = attention_orig[i]
                 if len(img_data.shape) == 3:
 | 
	aocr/model/model.py | 
	ReplaceText(target='mh' @(480,42)->(480,43)) | 
	class Model(object):
                         attention_orig[i] = attention[int(i/4)-1]
                 attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same')
                 attention_orig = np.maximum(attention_orig, 0.3)
                 attention_out = np.zeros((h, mw))
                 for i in xrange(mw):
                     attention_out[:, i] = attention_orig[i]
                 if len(img_data.shape) == 3: | 
	class Model(object):
                         attention_orig[i] = attention[int(i/4)-1]
                 attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same')
                 attention_orig = np.maximum(attention_orig, 0.3)
                 attention_out = np.zeros((mh, mw))
                 for i in xrange(mw):
                     attention_out[:, i] = attention_orig[i]
                 if len(img_data.shape) == 3: | 
| 2 | 
	https://:@github.com/emedvedev/attention-ocr.git | 
	e741baf7170e72a974754908d323cacc0bd55247 | 
	@@ -133,7 +133,7 @@ class Model(object):
             self.target_weights = []
             for i in xrange(self.decoder_size + 1):
                 self.decoder_inputs.append(
-                    tf.tile([0], [num_images])
+                    tf.tile([1], [num_images])
                 )
                 if i < self.decoder_size:
                     self.target_weights.append(tf.tile([1.], [num_images]))
 | 
	aocr/model/model.py | 
	ReplaceText(target='1' @(136,29)->(136,30)) | 
	class Model(object):
             self.target_weights = []
             for i in xrange(self.decoder_size + 1):
                 self.decoder_inputs.append(
                     tf.tile([0], [num_images])
                 )
                 if i < self.decoder_size:
                     self.target_weights.append(tf.tile([1.], [num_images])) | 
	class Model(object):
             self.target_weights = []
             for i in xrange(self.decoder_size + 1):
                 self.decoder_inputs.append(
                     tf.tile([1], [num_images])
                 )
                 if i < self.decoder_size:
                     self.target_weights.append(tf.tile([1.], [num_images])) | 
| 3 | 
	https://:@github.com/matthewdowney/TogglPy.git | 
	d5b630aec58d29b85ccffa527d24766eef6f61f9 | 
	@@ -197,7 +197,7 @@ class Toggl():
         day = datetime.now().day if not day else day
         hour = datetime.now().hour if not hour else hour
 
-        timestruct = datetime(year, month, day, hour - hourdiff).isoformat() + '.000Z'
+        timestruct = datetime(year, month, day, hour + hourdiff).isoformat() + '.000Z'
         data['time_entry']['start'] = timestruct
         data['time_entry']['duration'] = hourduration * 3600
         data['time_entry']['pid'] = projectid
 | 
	toggl/TogglPy.py | 
	ReplaceText(target='+' @(200,53)->(200,54)) | 
	class Toggl():
         day = datetime.now().day if not day else day
         hour = datetime.now().hour if not hour else hour
 
         timestruct = datetime(year, month, day, hour - hourdiff).isoformat() + '.000Z'
         data['time_entry']['start'] = timestruct
         data['time_entry']['duration'] = hourduration * 3600
         data['time_entry']['pid'] = projectid | 
	class Toggl():
         day = datetime.now().day if not day else day
         hour = datetime.now().hour if not hour else hour
 
         timestruct = datetime(year, month, day, hour + hourdiff).isoformat() + '.000Z'
         data['time_entry']['start'] = timestruct
         data['time_entry']['duration'] = hourduration * 3600
         data['time_entry']['pid'] = projectid | 
| 4 | 
	https://:@github.com/eEcoLiDAR/eEcoLiDAR.git | 
	f0b2a0b7a5fdd41887ba40b7687c8161c0faba1e | 
	@@ -28,6 +28,6 @@ class Test3FeatureExtractor(AbstractFeatureExtractor):
         return ['test3_a']
 
     def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume):
-        t2a, t2c = utils.get_features(targetpc, targetindex, self.requires())
+        t2a, t2c = utils.get_features(targetpc, self.requires(), targetindex)
         x, y, z = utils.get_point(targetpc, targetindex)
         return t2c - t2a - z  # z
 | 
	laserchicken/test_feature_extractor/feature_test23.py | 
	ArgSwap(idxs=1<->2 @(31,19)->(31,37)) | 
	class Test3FeatureExtractor(AbstractFeatureExtractor):
         return ['test3_a']
 
     def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume):
         t2a, t2c = utils.get_features(targetpc, targetindex, self.requires())
         x, y, z = utils.get_point(targetpc, targetindex)
         return t2c - t2a - z  # z | 
	class Test3FeatureExtractor(AbstractFeatureExtractor):
         return ['test3_a']
 
     def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume):
         t2a, t2c = utils.get_features(targetpc, self.requires(), targetindex)
         x, y, z = utils.get_point(targetpc, targetindex)
         return t2c - t2a - z  # z | 
| 5 | 
	https://:@github.com/eEcoLiDAR/eEcoLiDAR.git | 
	f0b2a0b7a5fdd41887ba40b7687c8161c0faba1e | 
	@@ -31,7 +31,7 @@ class TestUtils(unittest.TestCase):
         pc[keys.point]["color"] = {"type": "double", "data": cols}
         pc[keys.point]["flavor"] = {"type": "double", "data": flavs}
         x, y, z = utils.get_point(pc, 2)
-        c, f = utils.get_features(pc, 2, ("color", "flavor"))
+        c, f = utils.get_features(pc, ("color", "flavor"), 2)
         self.assertEqual(c, 0.5 * (x + y))
         self.assertEqual(f, 0.5 * (x - y))
 
 | 
	laserchicken/test_utils.py | 
	ArgSwap(idxs=1<->2 @(34,15)->(34,33)) | 
	class TestUtils(unittest.TestCase):
         pc[keys.point]["color"] = {"type": "double", "data": cols}
         pc[keys.point]["flavor"] = {"type": "double", "data": flavs}
         x, y, z = utils.get_point(pc, 2)
         c, f = utils.get_features(pc, 2, ("color", "flavor"))
         self.assertEqual(c, 0.5 * (x + y))
         self.assertEqual(f, 0.5 * (x - y))
  | 
	class TestUtils(unittest.TestCase):
         pc[keys.point]["color"] = {"type": "double", "data": cols}
         pc[keys.point]["flavor"] = {"type": "double", "data": flavs}
         x, y, z = utils.get_point(pc, 2)
         c, f = utils.get_features(pc, ("color", "flavor"), 2)
         self.assertEqual(c, 0.5 * (x + y))
         self.assertEqual(f, 0.5 * (x - y))
  | 
| 6 | 
	https://:@github.com/eEcoLiDAR/eEcoLiDAR.git | 
	502a365efda1393b130281803702d01f7e2d1dcd | 
	@@ -29,7 +29,7 @@ class TestExtractEigenValues(unittest.TestCase):
                          ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
 
         self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
-                         target_point_cloud[keys.provenance][0]["module"])
+                         target_point_cloud[keys.provenance][1]["module"])
 
     @staticmethod
     def test_eigenvalues_of_too_few_points_results_in_0():
 | 
	laserchicken/feature_extractor/test_eigenvals_feature_extractor.py | 
	ReplaceText(target='1' @(32,61)->(32,62)) | 
	class TestExtractEigenValues(unittest.TestCase):
                          ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
 
         self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
                          target_point_cloud[keys.provenance][0]["module"])
 
     @staticmethod
     def test_eigenvalues_of_too_few_points_results_in_0(): | 
	class TestExtractEigenValues(unittest.TestCase):
                          ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
 
         self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
                          target_point_cloud[keys.provenance][1]["module"])
 
     @staticmethod
     def test_eigenvalues_of_too_few_points_results_in_0(): | 
| 7 | 
	https://:@github.com/eEcoLiDAR/eEcoLiDAR.git | 
	eb7b021147a60b57e5dec536bd6f118c213f0952 | 
	@@ -29,7 +29,7 @@ class TestExtractEigenValues(unittest.TestCase):
                          ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
 
         self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
-                         target_point_cloud[keys.provenance][1]["module"])
+                         target_point_cloud[keys.provenance][-1]["module"])
 
     @staticmethod
     def test_eigenvalues_of_too_few_points_results_in_0():
 | 
	laserchicken/feature_extractor/test_eigenvals_feature_extractor.py | 
	ReplaceText(target='-1' @(32,61)->(32,62)) | 
	class TestExtractEigenValues(unittest.TestCase):
                          ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
 
         self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
                          target_point_cloud[keys.provenance][1]["module"])
 
     @staticmethod
     def test_eigenvalues_of_too_few_points_results_in_0(): | 
	class TestExtractEigenValues(unittest.TestCase):
                          ["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
 
         self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
                          target_point_cloud[keys.provenance][-1]["module"])
 
     @staticmethod
     def test_eigenvalues_of_too_few_points_results_in_0(): | 
| 8 | 
	https://:@github.com/VinF/deer.git | 
	66ea41db02c3361f18874dea7fd97720b1b06590 | 
	@@ -441,7 +441,7 @@ class CircularBuffer(object):
         
         if end == sys.maxsize:
             return self._data[self._lb+start:self._ub]
-        elif self._lb + end >= self._ub:
+        elif self._lb + end > self._ub:
                 raise IndexError()
         else:
             return self._data[self._lb+start:self._lb+end]
 | 
	General_Deep_Q_RL/agent.py | 
	ReplaceText(target='>' @(444,28)->(444,30)) | 
	class CircularBuffer(object):
         
         if end == sys.maxsize:
             return self._data[self._lb+start:self._ub]
         elif self._lb + end >= self._ub:
                 raise IndexError()
         else:
             return self._data[self._lb+start:self._lb+end] | 
	class CircularBuffer(object):
         
         if end == sys.maxsize:
             return self._data[self._lb+start:self._ub]
         elif self._lb + end > self._ub:
                 raise IndexError()
         else:
             return self._data[self._lb+start:self._lb+end] | 
| 9 | 
	https://:@github.com/VinF/deer.git | 
	fd939e272d5441d48fd7d30bf24312c0f6bc8aaa | 
	@@ -176,7 +176,7 @@ class MyEnv(Environment):
         # Lack of energy
             if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery):
             # If enough energy in the battery, use it
-                self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size*self.battery_eta
+                self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size/self.battery_eta
             else:
             # Otherwise: use what is left and then penalty                
                 reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh
 | 
	General_Deep_Q_RL/environments/MG_two_storages_env.py | 
	ReplaceText(target='/' @(179,126)->(179,127)) | 
	class MyEnv(Environment):
         # Lack of energy
             if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery):
             # If enough energy in the battery, use it
                 self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size*self.battery_eta
             else:
             # Otherwise: use what is left and then penalty                
                 reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh | 
	class MyEnv(Environment):
         # Lack of energy
             if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery):
             # If enough energy in the battery, use it
                 self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size/self.battery_eta
             else:
             # Otherwise: use what is left and then penalty                
                 reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh | 
| 10 | 
	https://:@github.com/piccolbo/altair_recipes.git | 
	4992dd864a317eaad641d0408f003c429ed24af6 | 
	@@ -6,7 +6,7 @@ from vega_datasets import data
 
 @viz_reg_test
 def test_boxplot_melted():
-    return ar.boxplot(data.iris(), "species", "petalLength")
+    return ar.boxplot(data.iris(), "petalLength", "species")
 
 
 @viz_reg_test
 | 
	tests/test_boxplot.py | 
	ArgSwap(idxs=1<->2 @(9,11)->(9,21)) | 
	from vega_datasets import data
 
 @viz_reg_test
 def test_boxplot_melted():
     return ar.boxplot(data.iris(), "species", "petalLength")
 
 
 @viz_reg_test | 
	from vega_datasets import data
 
 @viz_reg_test
 def test_boxplot_melted():
     return ar.boxplot(data.iris(), "petalLength", "species")
 
 
 @viz_reg_test | 
| 11 | 
	https://:@github.com/tailhook/zorro.git | 
	dcbc37d47fe2a8de029f5a2f3ae13adf52e7aace | 
	@@ -121,7 +121,7 @@ class RequestChannel(channel.PipelinedReqChannel):
             clen = int(headers.get('Content-Length', '0'))
             if clen < 0:
                 raise EOFError("Wrong content length")
-            while pos[0] + clen < len(buf):
+            while pos[0] + clen > len(buf):
                 readmore()
             return status, headers, buf[pos[0]:pos[0]+clen]
 
 | 
	zorro/http.py | 
	ReplaceText(target='>' @(124,32)->(124,33)) | 
	class RequestChannel(channel.PipelinedReqChannel):
             clen = int(headers.get('Content-Length', '0'))
             if clen < 0:
                 raise EOFError("Wrong content length")
             while pos[0] + clen < len(buf):
                 readmore()
             return status, headers, buf[pos[0]:pos[0]+clen]
  | 
	class RequestChannel(channel.PipelinedReqChannel):
             clen = int(headers.get('Content-Length', '0'))
             if clen < 0:
                 raise EOFError("Wrong content length")
             while pos[0] + clen > len(buf):
                 readmore()
             return status, headers, buf[pos[0]:pos[0]+clen]
  | 
| 12 | 
	https://:@gitlab.com/eavise/brambox.git | 
	f1faeed0b52d6f1c9c9ba6da818c1656f841622c | 
	@@ -73,7 +73,7 @@ def test_multiclass(parser, df_anno_simple):
     parser = parser()
 
     with pytest.raises(ValueError) as errinfo:
-        bb.io.save(parser, df_anno_simple, 'path.txt')
+        bb.io.save(df_anno_simple, parser, 'path.txt')
     assert 'single-class problems' in str(errinfo.value)
 
 
 | 
	test/io/parser/test_anno_cvc.py | 
	ArgSwap(idxs=0<->1 @(76,8)->(76,18)) | 
	def test_multiclass(parser, df_anno_simple):
     parser = parser()
 
     with pytest.raises(ValueError) as errinfo:
         bb.io.save(parser, df_anno_simple, 'path.txt')
     assert 'single-class problems' in str(errinfo.value)
 
  | 
	def test_multiclass(parser, df_anno_simple):
     parser = parser()
 
     with pytest.raises(ValueError) as errinfo:
         bb.io.save(df_anno_simple, parser, 'path.txt')
     assert 'single-class problems' in str(errinfo.value)
 
  | 
| 13 | 
	https://:@github.com/uber/h3-py.git | 
	359924df907144c85ec323ae2804e2c0d173dfc5 | 
	@@ -631,7 +631,7 @@ def hex_ranges(h3_address_list, ring_size):
                     (1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1
             # hexRanges doesn't return distance array
             hex_range_list[ring_index].add(
-                h3_to_string(krings[i * num_hexagons + j]))
+                h3_to_string(krings[i * array_len + j]))
     return out
 
 
 | 
	h3/h3.py | 
	ReplaceText(target='array_len' @(634,40)->(634,52)) | 
	def hex_ranges(h3_address_list, ring_size):
                     (1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1
             # hexRanges doesn't return distance array
             hex_range_list[ring_index].add(
                 h3_to_string(krings[i * num_hexagons + j]))
     return out
 
  | 
	def hex_ranges(h3_address_list, ring_size):
                     (1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1
             # hexRanges doesn't return distance array
             hex_range_list[ring_index].add(
                 h3_to_string(krings[i * array_len + j]))
     return out
 
  | 
| 14 | 
	https://:@github.com/polysquare/polysquare-generic-file-linter.git | 
	e9dbb28ea30955ab59d1339c04f0710b24ba53aa | 
	@@ -388,7 +388,7 @@ def _maybe_log_technical_terms(global_options, tool_options):
             terms = set(terms_file.read().splitlines())  # suppress(PYC70)
             terms_file.seek(0)  # suppress(PYC70)
             terms_file.truncate(0)  # suppress(PYC70)
-            tech_terms = freduce(lambda x, y: x + y,
+            tech_terms = freduce(lambda x, y: x | y,
                                  _drain(log_technical_terms_to_queue))
             terms_file.write("\n".join(list(terms |  # suppress(PYC70)
                                             set(tech_terms))))
 | 
	polysquarelinter/linter.py | 
	ReplaceText(target='|' @(391,48)->(391,49)) | 
	def _maybe_log_technical_terms(global_options, tool_options):
             terms = set(terms_file.read().splitlines())  # suppress(PYC70)
             terms_file.seek(0)  # suppress(PYC70)
             terms_file.truncate(0)  # suppress(PYC70)
             tech_terms = freduce(lambda x, y: x + y,
                                  _drain(log_technical_terms_to_queue))
             terms_file.write("\n".join(list(terms |  # suppress(PYC70)
                                             set(tech_terms)))) | 
	def _maybe_log_technical_terms(global_options, tool_options):
             terms = set(terms_file.read().splitlines())  # suppress(PYC70)
             terms_file.seek(0)  # suppress(PYC70)
             terms_file.truncate(0)  # suppress(PYC70)
             tech_terms = freduce(lambda x, y: x | y,
                                  _drain(log_technical_terms_to_queue))
             terms_file.write("\n".join(list(terms |  # suppress(PYC70)
                                             set(tech_terms)))) | 
| 15 | 
	https://:@github.com/johntruckenbrodt/spatialist.git | 
	c9d552e64cd47b30156b288e035d17debea48b45 | 
	@@ -300,7 +300,7 @@ def centerdist(obj1, obj2):
 
 
 def intersect(obj1, obj2):
-    if not (isinstance(obj1, Vector) or isinstance(obj2, Vector)):
+    if not (isinstance(obj1, Vector) and isinstance(obj2, Vector)):
         raise IOError('object must be of type Vector')
     obj1.reproject(obj2.srs)
 
 | 
	pyroSAR/spatial/vector.py | 
	ReplaceText(target='and' @(303,37)->(303,39)) | 
	def centerdist(obj1, obj2):
 
 
 def intersect(obj1, obj2):
     if not (isinstance(obj1, Vector) or isinstance(obj2, Vector)):
         raise IOError('object must be of type Vector')
     obj1.reproject(obj2.srs)
  | 
	def centerdist(obj1, obj2):
 
 
 def intersect(obj1, obj2):
     if not (isinstance(obj1, Vector) and isinstance(obj2, Vector)):
         raise IOError('object must be of type Vector')
     obj1.reproject(obj2.srs)
  | 
| 16 | 
	https://:@github.com/Toblerity/Shapely.git | 
	9f1b78e6fd5f4286f210b54827bdd26661f0ee7a | 
	@@ -40,7 +40,7 @@ if __name__ == '__main__':
         ]
         
     if pattern:
-        tests = [f for f in docfiles if f.find(pattern) >= 0]
+        tests = [f for f in docfiles if f.find(pattern) == 0]
     else:
         tests = docfiles
         
 | 
	tests/runalldoctests.py | 
	ReplaceText(target='==' @(43,56)->(43,58)) | 
	if __name__ == '__main__':
         ]
         
     if pattern:
         tests = [f for f in docfiles if f.find(pattern) >= 0]
     else:
         tests = docfiles
          | 
	if __name__ == '__main__':
         ]
         
     if pattern:
         tests = [f for f in docfiles if f.find(pattern) == 0]
     else:
         tests = docfiles
          | 
| 17 | 
	https://:@github.com/Toblerity/Shapely.git | 
	d6fc8cc0e0d50b23ba0d7ca6195bc530b2f8d1b9 | 
	@@ -11,7 +11,7 @@ def halton(base):
         i = index
         while i > 0:
             result += f * (i % base)
-            i = i/base
+            i = i//base
             f = f/base
         return result
     i = 1
 | 
	shapely/tests/test_unary_union.py | 
	ReplaceText(target='//' @(14,17)->(14,18)) | 
	def halton(base):
         i = index
         while i > 0:
             result += f * (i % base)
             i = i/base
             f = f/base
         return result
     i = 1 | 
	def halton(base):
         i = index
         while i > 0:
             result += f * (i % base)
             i = i//base
             f = f/base
         return result
     i = 1 | 
| 18 | 
	https://:@github.com/Toblerity/Shapely.git | 
	5f0db7fdc052beeeef36aa1251f19175d0abeedb | 
	@@ -36,7 +36,7 @@ if version is None:
 
 # Handle UTF-8 encoding of certain text files.
 open_kwds = {}
-if sys.version_info > (3,):
+if sys.version_info >= (3,):
     open_kwds['encoding'] = 'utf-8'
 
 with open('VERSION.txt', 'w', **open_kwds) as fp:
 | 
	setup.py | 
	ReplaceText(target='>=' @(39,20)->(39,21)) | 
	if version is None:
 
 # Handle UTF-8 encoding of certain text files.
 open_kwds = {}
 if sys.version_info > (3,):
     open_kwds['encoding'] = 'utf-8'
 
 with open('VERSION.txt', 'w', **open_kwds) as fp: | 
	if version is None:
 
 # Handle UTF-8 encoding of certain text files.
 open_kwds = {}
 if sys.version_info >= (3,):
     open_kwds['encoding'] = 'utf-8'
 
 with open('VERSION.txt', 'w', **open_kwds) as fp: | 
| 19 | 
	https://:@github.com/svetlyak40wt/django-tagging-ng.git | 
	0293b78ee0274d123eb70c1f8c5c01a5b36e2b40 | 
	@@ -163,7 +163,7 @@ class TaggedItemManager(models.Manager):
         associated with a given Tag or list of Tags.
         """
         tags = get_tag_list(tags)
-        if len(tags) == 0:
+        if len(tags) == 1:
             tag = tags[0] # Optimisation for single tag
         else:
             return self.get_intersection_by_model(Model, tags)
 | 
	models.py | 
	ReplaceText(target='1' @(166,24)->(166,25)) | 
	class TaggedItemManager(models.Manager):
         associated with a given Tag or list of Tags.
         """
         tags = get_tag_list(tags)
         if len(tags) == 0:
             tag = tags[0] # Optimisation for single tag
         else:
             return self.get_intersection_by_model(Model, tags) | 
	class TaggedItemManager(models.Manager):
         associated with a given Tag or list of Tags.
         """
         tags = get_tag_list(tags)
         if len(tags) == 1:
             tag = tags[0] # Optimisation for single tag
         else:
             return self.get_intersection_by_model(Model, tags) | 
| 20 | 
	https://:@github.com/svetlyak40wt/django-tagging-ng.git | 
	3285d40e4c1de628886a7fa45a6d4cf6ed4cd7e7 | 
	@@ -163,7 +163,7 @@ class TaggedItemManager(models.Manager):
         associated with a given Tag or list of Tags.
         """
         tags = get_tag_list(tags)
-        if len(tags) == 0:
+        if len(tags) == 1:
             tag = tags[0] # Optimisation for single tag
         else:
             return self.get_intersection_by_model(Model, tags)
 | 
	models.py | 
	ReplaceText(target='1' @(166,24)->(166,25)) | 
	class TaggedItemManager(models.Manager):
         associated with a given Tag or list of Tags.
         """
         tags = get_tag_list(tags)
         if len(tags) == 0:
             tag = tags[0] # Optimisation for single tag
         else:
             return self.get_intersection_by_model(Model, tags) | 
	class TaggedItemManager(models.Manager):
         associated with a given Tag or list of Tags.
         """
         tags = get_tag_list(tags)
         if len(tags) == 1:
             tag = tags[0] # Optimisation for single tag
         else:
             return self.get_intersection_by_model(Model, tags) | 
| 21 | 
	https://:@github.com/zzzsochi/yadm.git | 
	03efd06fe95c7d84264455c4fac5c8cbb17eb4dd | 
	@@ -316,7 +316,7 @@ class QuerySet(BaseQuerySet):
 
         if data is None:
             if exc is not None:
-                raise exc(criteria)
+                raise exc(qs)
             else:
                 return None
 
 | 
	yadm/queryset.py | 
	ReplaceText(target='qs' @(319,26)->(319,34)) | 
	class QuerySet(BaseQuerySet):
 
         if data is None:
             if exc is not None:
                 raise exc(criteria)
             else:
                 return None
  | 
	class QuerySet(BaseQuerySet):
 
         if data is None:
             if exc is not None:
                 raise exc(qs)
             else:
                 return None
  | 
| 22 | 
	https://:@github.com/instacart/lore.git | 
	1c1e0efdac6b27dc111eaa93bb99317c59aaffaf | 
	@@ -196,7 +196,7 @@ class Base(object):
     def upload(self):
         self.fitting = 0
         self.save()
-        lore.io.upload(self.remote_model_path(), self.model_path())
+        lore.io.upload(self.model_path(), self.remote_model_path())
      
     @classmethod
     def download(cls, fitting=0):
 | 
	lore/models/base.py | 
	ArgSwap(idxs=0<->1 @(199,8)->(199,22)) | 
	class Base(object):
     def upload(self):
         self.fitting = 0
         self.save()
         lore.io.upload(self.remote_model_path(), self.model_path())
      
     @classmethod
     def download(cls, fitting=0): | 
	class Base(object):
     def upload(self):
         self.fitting = 0
         self.save()
         lore.io.upload(self.model_path(), self.remote_model_path())
      
     @classmethod
     def download(cls, fitting=0): | 
| 23 | 
	https://:@github.com/instacart/lore.git | 
	1c1e0efdac6b27dc111eaa93bb99317c59aaffaf | 
	@@ -78,7 +78,7 @@ class Base(lore.models.base.Base):
     
     def upload(self):
         super(Base, self).upload()
-        lore.io.upload(self.remote_weights_path(), self.weights_path())
+        lore.io.upload(self.weights_path(), self.remote_weights_path())
     
     @classmethod
     def download(cls, fitting=0):
 | 
	lore/models/keras.py | 
	ArgSwap(idxs=0<->1 @(81,8)->(81,22)) | 
	class Base(lore.models.base.Base):
     
     def upload(self):
         super(Base, self).upload()
         lore.io.upload(self.remote_weights_path(), self.weights_path())
     
     @classmethod
     def download(cls, fitting=0): | 
	class Base(lore.models.base.Base):
     
     def upload(self):
         super(Base, self).upload()
         lore.io.upload(self.weights_path(), self.remote_weights_path())
     
     @classmethod
     def download(cls, fitting=0): | 
| 24 | 
	https://:@github.com/instacart/lore.git | 
	f4ded2b3199d1c33ba6c9c79cd66b25d43c83c81 | 
	@@ -464,7 +464,7 @@ class Base(BaseEstimator):
             result = self.keras.predict(dataframe, batch_size=self.batch_size)
 
         if self.towers > 1:
-            result = numpy.mean(result, axis=0).squeeze(axis=0)
+            result = numpy.mean(result, axis=0).squeeze(axis=1)
 
         return result
 
 | 
	lore/estimators/keras.py | 
	ReplaceText(target='1' @(467,61)->(467,62)) | 
	class Base(BaseEstimator):
             result = self.keras.predict(dataframe, batch_size=self.batch_size)
 
         if self.towers > 1:
             result = numpy.mean(result, axis=0).squeeze(axis=0)
 
         return result
  | 
	class Base(BaseEstimator):
             result = self.keras.predict(dataframe, batch_size=self.batch_size)
 
         if self.towers > 1:
             result = numpy.mean(result, axis=0).squeeze(axis=1)
 
         return result
  | 
| 25 | 
	https://:@github.com/baliga-lab/cmonkey2.git | 
	3201a0e97688724450196da8fef96d283c855b3f | 
	@@ -273,7 +273,7 @@ class ClusterMembership:
             #logging.warn("cluster %s already associated with %s",
             #             str(cluster), str(column))
             pass
-        if columns not in columns:
+        if column not in columns:
             columns.append(column)
 
     def remove_cluster_from_column(self, column, cluster):
 | 
	cmonkey/membership.py | 
	ReplaceText(target='column' @(276,11)->(276,18)) | 
	class ClusterMembership:
             #logging.warn("cluster %s already associated with %s",
             #             str(cluster), str(column))
             pass
         if columns not in columns:
             columns.append(column)
 
     def remove_cluster_from_column(self, column, cluster): | 
	class ClusterMembership:
             #logging.warn("cluster %s already associated with %s",
             #             str(cluster), str(column))
             pass
         if column not in columns:
             columns.append(column)
 
     def remove_cluster_from_column(self, column, cluster): | 
| 26 | 
	https://:@github.com/baliga-lab/cmonkey2.git | 
	48d14ac785b1013354a55a37239c66433fbf19eb | 
	@@ -460,7 +460,7 @@ class ClusterMembership:
             max_score = sys.float_info.min
             for row in range(sm.num_rows()):
                 if sm_values[row][0] > max_score:
-                    max_score = sm[row][0]
+                    max_score = sm_values[row][0]
                     max_row = row
             return sm.row_names[max_row]
 
 | 
	cmonkey/membership.py | 
	ReplaceText(target='sm_values' @(463,32)->(463,34)) | 
	class ClusterMembership:
             max_score = sys.float_info.min
             for row in range(sm.num_rows()):
                 if sm_values[row][0] > max_score:
                     max_score = sm[row][0]
                     max_row = row
             return sm.row_names[max_row]
  | 
	class ClusterMembership:
             max_score = sys.float_info.min
             for row in range(sm.num_rows()):
                 if sm_values[row][0] > max_score:
                     max_score = sm_values[row][0]
                     max_row = row
             return sm.row_names[max_row]
  | 
| 27 | 
	https://:@github.com/baliga-lab/cmonkey2.git | 
	311548905a9def1cbdf63d2e0fd8a17346564742 | 
	@@ -17,7 +17,7 @@ class MicrobesOnlineTest(unittest.TestCase):  # pylint: disable-msg=R0904
         """test fixture"""
         if not os.path.exists('testcache'):
             os.mkdir('testcache')
-        self.service = mo.MicrobesOnline(mo.MICROBES_ONLINE_BASE_URL, 'testcache')
+        self.service = mo.MicrobesOnline('testcache', mo.MICROBES_ONLINE_BASE_URL)
 
     def tearDown(self):  # pylint: disable-msg=C0103
         """test cleanup"""
 | 
	test/microbes_online_test.py | 
	ArgSwap(idxs=0<->1 @(20,23)->(20,40)) | 
	class MicrobesOnlineTest(unittest.TestCase):  # pylint: disable-msg=R0904
         """test fixture"""
         if not os.path.exists('testcache'):
             os.mkdir('testcache')
         self.service = mo.MicrobesOnline(mo.MICROBES_ONLINE_BASE_URL, 'testcache')
 
     def tearDown(self):  # pylint: disable-msg=C0103
         """test cleanup""" | 
	class MicrobesOnlineTest(unittest.TestCase):  # pylint: disable-msg=R0904
         """test fixture"""
         if not os.path.exists('testcache'):
             os.mkdir('testcache')
         self.service = mo.MicrobesOnline('testcache', mo.MICROBES_ONLINE_BASE_URL)
 
     def tearDown(self):  # pylint: disable-msg=C0103
         """test cleanup""" | 
| 28 | 
	https://:@github.com/muammar/ml4chem.git | 
	dbb7de0379cb8881538d211899e4bec8794f16e3 | 
	@@ -344,7 +344,7 @@ def train(inputs, targets, model=None, data=None, optimizer=None, lr=None,
     logger.info('Training finished in {} hours {} minutes {:.2f} seconds.'
                 .format(h, m, s))
     logger.info('outputs')
-    logger.info(outputs)
+    logger.info(outputs_)
     logger.info('targets')
     logger.info(targets)
 
 | 
	mlchem/models/neuralnetwork.py | 
	ReplaceText(target='outputs_' @(347,16)->(347,23)) | 
	def train(inputs, targets, model=None, data=None, optimizer=None, lr=None,
     logger.info('Training finished in {} hours {} minutes {:.2f} seconds.'
                 .format(h, m, s))
     logger.info('outputs')
     logger.info(outputs)
     logger.info('targets')
     logger.info(targets)
  | 
	def train(inputs, targets, model=None, data=None, optimizer=None, lr=None,
     logger.info('Training finished in {} hours {} minutes {:.2f} seconds.'
                 .format(h, m, s))
     logger.info('outputs')
     logger.info(outputs_)
     logger.info('targets')
     logger.info(targets)
  | 
| 29 | 
	https://:@github.com/chris7/pyquant.git | 
	3730fbdb9789a59a65d38f5a2ae21c645086096f | 
	@@ -624,7 +624,7 @@ def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si
     best_fit = np.array(best_fit)
     peak_func = bigauss_ndim if bigauss_fit else gauss_ndim
     # Get rid of peaks with low r^2
-    if micro and r2_cutoff is not None:
+    if not micro and r2_cutoff is not None:
         final_fit = np.array([])
         for peak_index in xrange(0, len(best_fit), step_size):
 
 | 
	pyquant/peaks.py | 
	ReplaceText(target='not ' @(627,7)->(627,7)) | 
	def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si
     best_fit = np.array(best_fit)
     peak_func = bigauss_ndim if bigauss_fit else gauss_ndim
     # Get rid of peaks with low r^2
     if micro and r2_cutoff is not None:
         final_fit = np.array([])
         for peak_index in xrange(0, len(best_fit), step_size):
  | 
	def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si
     best_fit = np.array(best_fit)
     peak_func = bigauss_ndim if bigauss_fit else gauss_ndim
     # Get rid of peaks with low r^2
     if not micro and r2_cutoff is not None:
         final_fit = np.array([])
         for peak_index in xrange(0, len(best_fit), step_size):
  | 
| 30 | 
	https://:@github.com/chris7/pyquant.git | 
	4a0755563e0a36fecf1f4393554cfaf4c1615c2c | 
	@@ -615,7 +615,7 @@ def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
     # By default, cross points returns the left side
     for i in xrange(len(cross_points)):
         index = cross_points[i]
-        if index < len(cross_points):
+        if index < len(ydata):
             if ydata[index] < ydata[index+1]:
                 cross_points[i] = index+1
 
 | 
	pyquant/utils.py | 
	ReplaceText(target='ydata' @(618,23)->(618,35)) | 
	def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
     # By default, cross points returns the left side
     for i in xrange(len(cross_points)):
         index = cross_points[i]
         if index < len(cross_points):
             if ydata[index] < ydata[index+1]:
                 cross_points[i] = index+1
  | 
	def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
     # By default, cross points returns the left side
     for i in xrange(len(cross_points)):
         index = cross_points[i]
         if index < len(ydata):
             if ydata[index] < ydata[index+1]:
                 cross_points[i] = index+1
  | 
| 31 | 
	https://:@github.com/chris7/pyquant.git | 
	cd61286935d8ca64eb539851e39a98a0655ff400 | 
	@@ -611,7 +611,7 @@ def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
     ydata = np.abs(ydata_peaks)
 
     if min_peak_width is None:
-        max_peak_width = int(len(ydata) / 2)
+        min_peak_width = int(len(ydata) / 2)
         if min_peak_width > 5:
             min_peak_width = 5
 
 | 
	pyquant/utils.py | 
	ReplaceText(target='min_peak_width' @(614,8)->(614,22)) | 
	def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
     ydata = np.abs(ydata_peaks)
 
     if min_peak_width is None:
         max_peak_width = int(len(ydata) / 2)
         if min_peak_width > 5:
             min_peak_width = 5
  | 
	def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
     ydata = np.abs(ydata_peaks)
 
     if min_peak_width is None:
         min_peak_width = int(len(ydata) / 2)
         if min_peak_width > 5:
             min_peak_width = 5
  | 
| 32 | 
	https://:@github.com/ssec/sift.git | 
	24ce052cd497c42c917de06f0c89a7c5be13ab50 | 
	@@ -599,7 +599,7 @@ class ProbeGraphDisplay (object) :
                 x_point = self.workspace.get_content_point(x_uuid, point_xy)
                 format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point)
                 y_point = self.workspace.get_content_point(y_uuid, point_xy)
-                format_str, unit_str, y_point = self.document.convert_units(x_uuid, y_point)
+                format_str, unit_str, y_point = self.document.convert_units(y_uuid, y_point)
             else:
                 x_point = None
                 y_point = None
 | 
	py/cspov/view/ProbeGraphs.py | 
	ReplaceText(target='y_uuid' @(602,76)->(602,82)) | 
	class ProbeGraphDisplay (object) :
                 x_point = self.workspace.get_content_point(x_uuid, point_xy)
                 format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point)
                 y_point = self.workspace.get_content_point(y_uuid, point_xy)
                 format_str, unit_str, y_point = self.document.convert_units(x_uuid, y_point)
             else:
                 x_point = None
                 y_point = None | 
	class ProbeGraphDisplay (object) :
                 x_point = self.workspace.get_content_point(x_uuid, point_xy)
                 format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point)
                 y_point = self.workspace.get_content_point(y_uuid, point_xy)
                 format_str, unit_str, y_point = self.document.convert_units(y_uuid, y_point)
             else:
                 x_point = None
                 y_point = None | 
| 33 | 
	https://:@github.com/threatwatch/twigs.git | 
	7ee5d95178a459a8c2e8ff7855e3156e620c395c | 
	@@ -85,7 +85,7 @@ def parse_inventory(email,data,params):
             asset_map = {}
             asset_map['owner'] = email
             asset_map['host'] = host
-            asset_map['id'] = host
+            asset_map['id'] = vmuuid
             asset_map['name'] = host
             asset_map['tags'] = [ ]
             asset_map['patch_tracker'] = { } # To help remove duplicate patches
 | 
	twigs/azure.py | 
	ReplaceText(target='vmuuid' @(88,30)->(88,34)) | 
	def parse_inventory(email,data,params):
             asset_map = {}
             asset_map['owner'] = email
             asset_map['host'] = host
             asset_map['id'] = host
             asset_map['name'] = host
             asset_map['tags'] = [ ]
             asset_map['patch_tracker'] = { } # To help remove duplicate patches | 
	def parse_inventory(email,data,params):
             asset_map = {}
             asset_map['owner'] = email
             asset_map['host'] = host
             asset_map['id'] = vmuuid
             asset_map['name'] = host
             asset_map['tags'] = [ ]
             asset_map['patch_tracker'] = { } # To help remove duplicate patches | 
| 34 | 
	https://:@github.com/Keeper-Security/Commander.git | 
	9bd55c8dd48ab62759bbbb8dfcd38ab364cec2dc | 
	@@ -1550,7 +1550,7 @@ def prepare_record(params, record):
     else:
         if params.debug: print('Generated record key')
         unencrypted_key = os.urandom(32)
-        record_object['record_key'] = encrypt_aes(params.data_key, unencrypted_key)
+        record_object['record_key'] = encrypt_aes(unencrypted_key, params.data_key)
         record_object['revision'] = 0
 
     data['title'] = record.title
 | 
	keepercommander/api.py | 
	ArgSwap(idxs=0<->1 @(1553,38)->(1553,49)) | 
	def prepare_record(params, record):
     else:
         if params.debug: print('Generated record key')
         unencrypted_key = os.urandom(32)
         record_object['record_key'] = encrypt_aes(params.data_key, unencrypted_key)
         record_object['revision'] = 0
 
     data['title'] = record.title | 
	def prepare_record(params, record):
     else:
         if params.debug: print('Generated record key')
         unencrypted_key = os.urandom(32)
         record_object['record_key'] = encrypt_aes(unencrypted_key, params.data_key)
         record_object['revision'] = 0
 
     data['title'] = record.title | 
| 35 | 
	https://:@github.com/hpapaxen/rope.git | 
	27d5085b30e89095e88339c96d9940e338482106 | 
	@@ -69,7 +69,7 @@ class JobSet(object):
 
     def get_percent_done(self):
         if self.count is not None and self.count > 0:
-            percent = self.done * 100 / self.count
+            percent = self.done * 100 // self.count
             return min(percent, 100)
 
     def get_name(self):
 | 
	rope/base/taskhandle.py | 
	ReplaceText(target='//' @(72,38)->(72,39)) | 
	class JobSet(object):
 
     def get_percent_done(self):
         if self.count is not None and self.count > 0:
             percent = self.done * 100 / self.count
             return min(percent, 100)
 
     def get_name(self): | 
	class JobSet(object):
 
     def get_percent_done(self):
         if self.count is not None and self.count > 0:
             percent = self.done * 100 // self.count
             return min(percent, 100)
 
     def get_name(self): | 
| 36 | 
	https://:@github.com/hpapaxen/rope.git | 
	27d5085b30e89095e88339c96d9940e338482106 | 
	@@ -524,7 +524,7 @@ class ProgressBar(object):
         self.text['text'] = text
 
     def _draw_shape(self):
-        width = int(self.canvas['width']) * self.percent / 100
+        width = int(self.canvas['width']) * self.percent // 100
         self.canvas.create_rectangle(0, 0, width, self.canvas['height'],
                                      fill=self.color)
         total_width = self.canvas['width']
 | 
	rope/ui/uihelpers.py | 
	ReplaceText(target='//' @(527,57)->(527,58)) | 
	class ProgressBar(object):
         self.text['text'] = text
 
     def _draw_shape(self):
         width = int(self.canvas['width']) * self.percent / 100
         self.canvas.create_rectangle(0, 0, width, self.canvas['height'],
                                      fill=self.color)
         total_width = self.canvas['width'] | 
	class ProgressBar(object):
         self.text['text'] = text
 
     def _draw_shape(self):
         width = int(self.canvas['width']) * self.percent // 100
         self.canvas.create_rectangle(0, 0, width, self.canvas['height'],
                                      fill=self.color)
         total_width = self.canvas['width'] | 
| 37 | 
	https://:@github.com/hpapaxen/rope.git | 
	2720419618aceab7fba51aaa4d66f7eae005b22d | 
	@@ -129,7 +129,7 @@ class SimilarFinderTest(unittest.TestCase):
         source = 'x.a = 1\n'
         finder = similarfinder.SimilarFinder(source)
         result = list(finder.get_matches('${a} = 1'))
-        self.assertEquals(1, len(result))
+        self.assertEquals(0, len(result))
 
     def test_functions_not_matching_when_only_first_parameters(self):
         source = 'f(1, 2)\n'
 | 
	ropetest/refactor/similarfindertest.py | 
	ReplaceText(target='0' @(132,26)->(132,27)) | 
	class SimilarFinderTest(unittest.TestCase):
         source = 'x.a = 1\n'
         finder = similarfinder.SimilarFinder(source)
         result = list(finder.get_matches('${a} = 1'))
         self.assertEquals(1, len(result))
 
     def test_functions_not_matching_when_only_first_parameters(self):
         source = 'f(1, 2)\n' | 
	class SimilarFinderTest(unittest.TestCase):
         source = 'x.a = 1\n'
         finder = similarfinder.SimilarFinder(source)
         result = list(finder.get_matches('${a} = 1'))
         self.assertEquals(0, len(result))
 
     def test_functions_not_matching_when_only_first_parameters(self):
         source = 'f(1, 2)\n' | 
| 38 | 
	https://:@github.com/hpapaxen/rope.git | 
	0eb3cb58493cdaea83a4e24d47b5bd4dbd19f963 | 
	@@ -32,7 +32,7 @@ class BuiltinModule(pyobjects.AbstractModule):
         result.update(self.initial)
         for modname in self.submodules:
             name = modname.split('.')[-1]
-            result[name] = BuiltinModule(name, self.submodules)
+            result[name] = BuiltinModule(modname, self.submodules)
         return result
 
     @property
 | 
	rope/base/builtins.py | 
	ReplaceText(target='modname' @(35,41)->(35,45)) | 
	class BuiltinModule(pyobjects.AbstractModule):
         result.update(self.initial)
         for modname in self.submodules:
             name = modname.split('.')[-1]
             result[name] = BuiltinModule(name, self.submodules)
         return result
 
     @property | 
	class BuiltinModule(pyobjects.AbstractModule):
         result.update(self.initial)
         for modname in self.submodules:
             name = modname.split('.')[-1]
             result[name] = BuiltinModule(modname, self.submodules)
         return result
 
     @property | 
| 39 | 
	https://:@github.com/hpapaxen/rope.git | 
	528744bb4bc1b8076680f7c2c1bfac508ddca4f9 | 
	@@ -37,7 +37,7 @@ def relative(root, path):
         if os.path.samefile(root, path):
             return '/'.join(reversed(rel))
         parent = os.path.dirname(path)
-        if not path or parent == path:
+        if not parent or parent == path:
             break
         rel.append(os.path.basename(path))
         path = parent
 | 
	rope/base/libutils.py | 
	ReplaceText(target='parent' @(40,15)->(40,19)) | 
	def relative(root, path):
         if os.path.samefile(root, path):
             return '/'.join(reversed(rel))
         parent = os.path.dirname(path)
         if not path or parent == path:
             break
         rel.append(os.path.basename(path))
         path = parent | 
	def relative(root, path):
         if os.path.samefile(root, path):
             return '/'.join(reversed(rel))
         parent = os.path.dirname(path)
         if not parent or parent == path:
             break
         rel.append(os.path.basename(path))
         path = parent | 
| 40 | 
	https://:@github.com/benjamincrom/baseball.git | 
	6ef29729ad07458aebe709b4e42f56ecd3761ec4 | 
	@@ -111,7 +111,7 @@ def write_game_svg_and_html(game_id, game, output_path):
     html_filename = game_id + '.html'
 
     svg_text = game.get_svg_str()
-    html_text = HTML_WRAPPER.format(title=game_id, filename=html_filename)
+    html_text = HTML_WRAPPER.format(title=game_id, filename=svg_filename)
 
     output_svg_path = join(output_path, svg_filename)
     output_html_path = join(output_path, html_filename)
 | 
	fetch_game.py | 
	ReplaceText(target='svg_filename' @(114,60)->(114,73)) | 
	def write_game_svg_and_html(game_id, game, output_path):
     html_filename = game_id + '.html'
 
     svg_text = game.get_svg_str()
     html_text = HTML_WRAPPER.format(title=game_id, filename=html_filename)
 
     output_svg_path = join(output_path, svg_filename)
     output_html_path = join(output_path, html_filename) | 
	def write_game_svg_and_html(game_id, game, output_path):
     html_filename = game_id + '.html'
 
     svg_text = game.get_svg_str()
     html_text = HTML_WRAPPER.format(title=game_id, filename=svg_filename)
 
     output_svg_path = join(output_path, svg_filename)
     output_html_path = join(output_path, html_filename) | 
| 41 | 
	https://:@github.com/sporestack/bitcash.git | 
	dbc65e1b47426e0e4d286db5b27216ec36cb32cf | 
	@@ -16,7 +16,7 @@ def test_set_fee_cache_time():
 
 
 def test_get_fee():
-    assert get_fee(fast=True) != get_fee(fast=False)
+    assert get_fee(fast=True) >= get_fee(fast=False)
 
 
 class TestFeeCache:
 | 
	tests/network/test_fees.py | 
	ReplaceText(target='>=' @(19,30)->(19,32)) | 
	def test_set_fee_cache_time():
 
 
 def test_get_fee():
     assert get_fee(fast=True) != get_fee(fast=False)
 
 
 class TestFeeCache: | 
	def test_set_fee_cache_time():
 
 
 def test_get_fee():
     assert get_fee(fast=True) >= get_fee(fast=False)
 
 
 class TestFeeCache: | 
| 42 | 
	https://:@github.com/galaxy-genome-annotation/python-apollo.git | 
	53e514b619844fa1f87179d738b9d29830027300 | 
	@@ -29,7 +29,7 @@ class ApolloTestCase(unittest.TestCase):
         """
 
         org_info = wa.organisms.show_organism(org_id)
-        if 'directory' in org_info:
+        if 'directory' not in org_info:
             time.sleep(1)
             org_info = wa.organisms.show_organism(org_id)
 
 | 
	test/__init__.py | 
	ReplaceText(target=' not in ' @(32,22)->(32,26)) | 
	class ApolloTestCase(unittest.TestCase):
         """
 
         org_info = wa.organisms.show_organism(org_id)
         if 'directory' in org_info:
             time.sleep(1)
             org_info = wa.organisms.show_organism(org_id)
  | 
	class ApolloTestCase(unittest.TestCase):
         """
 
         org_info = wa.organisms.show_organism(org_id)
         if 'directory' not in org_info:
             time.sleep(1)
             org_info = wa.organisms.show_organism(org_id)
  | 
| 43 | 
	https://:@github.com/jakubplichta/grafana-dashboard-builder.git | 
	3228e6950d65b9bd347cacb56a9e85ec410b14ce | 
	@@ -35,7 +35,7 @@ class Context(object):
             formatter = string.Formatter()
             (result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand)
             while result != to_expand:
-                (result, to_expand) = (formatter.vformat(to_expand, (), self._context), result)
+                (result, to_expand) = (formatter.vformat(result, (), self._context), result)
             return result
         elif isinstance(to_expand, list):
             return [self.expand_placeholders(value) for value in to_expand]
 | 
	grafana_dashboards/context.py | 
	ReplaceText(target='result' @(38,57)->(38,66)) | 
	class Context(object):
             formatter = string.Formatter()
             (result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand)
             while result != to_expand:
                 (result, to_expand) = (formatter.vformat(to_expand, (), self._context), result)
             return result
         elif isinstance(to_expand, list):
             return [self.expand_placeholders(value) for value in to_expand] | 
	class Context(object):
             formatter = string.Formatter()
             (result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand)
             while result != to_expand:
                 (result, to_expand) = (formatter.vformat(result, (), self._context), result)
             return result
         elif isinstance(to_expand, list):
             return [self.expand_placeholders(value) for value in to_expand] | 
| 44 | 
	https://:@github.com/Phylliade/ikpy.git | 
	815dbff3a521532a7b792c309902ffea82abac85 | 
	@@ -13,7 +13,7 @@ class TestFK(unittest.TestCase):
         one_move[5] = np.pi / 4
         one_move[6] = -np.pi / 2
         one_move[4] = -np.pi / 2
-        self.test_pos = one_move
+        self.test_pos = all_zeros
 
     def test_fk_creature(self):
 
 | 
	tests/test_fk.py | 
	ReplaceText(target='all_zeros' @(16,24)->(16,32)) | 
	class TestFK(unittest.TestCase):
         one_move[5] = np.pi / 4
         one_move[6] = -np.pi / 2
         one_move[4] = -np.pi / 2
         self.test_pos = one_move
 
     def test_fk_creature(self):
  | 
	class TestFK(unittest.TestCase):
         one_move[5] = np.pi / 4
         one_move[6] = -np.pi / 2
         one_move[4] = -np.pi / 2
         self.test_pos = all_zeros
 
     def test_fk_creature(self):
  | 
| 45 | 
	https://:@github.com/tingbot/tingbot-python.git | 
	5374186675f6809faf9ce953fc35c81217348753 | 
	@@ -72,7 +72,7 @@ class RunLoop(object):
         while self.running:
             if len(self.timers) > 0:
                 try:
-                    self._wait(self.timers[0].next_fire_time)
+                    self._wait(self.timers[-1].next_fire_time)
                 except Exception as e:
                     self._error(e)
                     continue
 | 
	tingbot/run_loop.py | 
	ReplaceText(target='-1' @(75,43)->(75,44)) | 
	class RunLoop(object):
         while self.running:
             if len(self.timers) > 0:
                 try:
                     self._wait(self.timers[0].next_fire_time)
                 except Exception as e:
                     self._error(e)
                     continue | 
	class RunLoop(object):
         while self.running:
             if len(self.timers) > 0:
                 try:
                     self._wait(self.timers[-1].next_fire_time)
                 except Exception as e:
                     self._error(e)
                     continue | 
| 46 | 
	https://:@github.com/nyoka-pmml/nyoka.git | 
	8d5c0d31d0bf1e251abe686f06b614a16e5ffcfb | 
	@@ -74,7 +74,7 @@ class TestMethods(unittest.TestCase):
         self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance")
         
         ##3
-        self.assertEqual(pmml_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors)
+        self.assertEqual(pipeline_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors)
 
     
     def test_sklearn_03(self):
 | 
	nyoka/tests/skl_to_pmml_UnitTest.py | 
	ReplaceText(target='pipeline_obj' @(77,25)->(77,33)) | 
	class TestMethods(unittest.TestCase):
         self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance")
         
         ##3
         self.assertEqual(pmml_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors)
 
     
     def test_sklearn_03(self): | 
	class TestMethods(unittest.TestCase):
         self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance")
         
         ##3
         self.assertEqual(pipeline_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors)
 
     
     def test_sklearn_03(self): | 
| 47 | 
	https://:@github.com/iris-edu/pyweed.git | 
	77d919acd8d54d4879d1a34598e9e04f16fdf708 | 
	@@ -97,7 +97,7 @@ class WaveformEntry(AttribDict):
 
         self.error = None
 
-        self.start_time = self.distances.arrival + self.config.offsets[0]
+        self.start_time = self.distances.arrival - self.config.offsets[0]
         self.end_time = self.distances.arrival + self.config.offsets[1]
 
         self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_')
 | 
	pyweed/waveforms_handler.py | 
	ReplaceText(target='-' @(100,49)->(100,50)) | 
	class WaveformEntry(AttribDict):
 
         self.error = None
 
         self.start_time = self.distances.arrival + self.config.offsets[0]
         self.end_time = self.distances.arrival + self.config.offsets[1]
 
         self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_') | 
	class WaveformEntry(AttribDict):
 
         self.error = None
 
         self.start_time = self.distances.arrival - self.config.offsets[0]
         self.end_time = self.distances.arrival + self.config.offsets[1]
 
         self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_') | 
| 48 | 
	https://:@github.com/anaxilaus/coindata.git | 
	2e5067311c4eed50eed41c45f43ad63e8973e579 | 
	@@ -52,6 +52,6 @@ def dump_json(data, filepath):
 
     try:
         with open(filepath, 'w') as file:
-            json.dump(data, filepath)
+            json.dump(data, file)
     except TypeError as e:
         print("Data isn't JSON compatible.\n", e)
 | 
	coindata/utils.py | 
	ReplaceText(target='file' @(55,28)->(55,36)) | 
	def dump_json(data, filepath):
 
     try:
         with open(filepath, 'w') as file:
             json.dump(data, filepath)
     except TypeError as e:
         print("Data isn't JSON compatible.\n", e) | 
	def dump_json(data, filepath):
 
     try:
         with open(filepath, 'w') as file:
             json.dump(data, file)
     except TypeError as e:
         print("Data isn't JSON compatible.\n", e) | 
| 49 | 
	https://:@github.com/Pixelapse/pyglass.git | 
	a31e95cbc259ce61f5851d6f0d769792aaa182fe | 
	@@ -20,7 +20,7 @@ def preview(src_path):
     preview_path = thumbnail_preview(src_path)
 
   if preview_path:
-    mimetype = magic.from_file(src_path, mime=True).lower()
+    mimetype = magic.from_file(preview_path, mime=True).lower()
     if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]:
       return preview_path
 
 | 
	pyglass/quicklook/api.py | 
	ReplaceText(target='preview_path' @(23,31)->(23,39)) | 
	def preview(src_path):
     preview_path = thumbnail_preview(src_path)
 
   if preview_path:
     mimetype = magic.from_file(src_path, mime=True).lower()
     if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]:
       return preview_path
  | 
	def preview(src_path):
     preview_path = thumbnail_preview(src_path)
 
   if preview_path:
     mimetype = magic.from_file(preview_path, mime=True).lower()
     if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]:
       return preview_path
  | 
| 50 | 
	https://:@github.com/erezsh/plyplus.git | 
	8cc69bebfcb2cb0480ac66d07fe1f4b8637bba11 | 
	@@ -784,7 +784,7 @@ class _Grammar(object):
                         subtree.extend(child.tail)
                     else:
                         subtree.append(child)
-                p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) > 1 else subtree[0]
+                p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) != 1 else subtree[0]
         else:
             def p_rule(self, p):
                 p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True)
 | 
	plyplus/plyplus.py | 
	ReplaceText(target='!=' @(787,98)->(787,99)) | 
	class _Grammar(object):
                         subtree.extend(child.tail)
                     else:
                         subtree.append(child)
                 p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) > 1 else subtree[0]
         else:
             def p_rule(self, p):
                 p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True) | 
	class _Grammar(object):
                         subtree.extend(child.tail)
                     else:
                         subtree.append(child)
                 p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) != 1 else subtree[0]
         else:
             def p_rule(self, p):
                 p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True) | 
| 51 | 
	https://:@github.com/olls/graphics.git | 
	877ef2670d4ad34c4fcf951dab0922419f081531 | 
	@@ -22,7 +22,7 @@ def colorStr(text, color=WHITE):
 		return seq
 		sys.stdout.write(seq + '\n')
 	else:
-		return seq
+		return text
 		sys.stdout.write(text + '\n')
 
 if __name__ == '__main__':
 | 
	colors.py | 
	ReplaceText(target='text' @(25,9)->(25,12)) | 
	def colorStr(text, color=WHITE):
 		return seq
 		sys.stdout.write(seq + '\n')
 	else:
 		return seq
 		sys.stdout.write(text + '\n')
 
 if __name__ == '__main__': | 
	def colorStr(text, color=WHITE):
 		return seq
 		sys.stdout.write(seq + '\n')
 	else:
 		return text
 		sys.stdout.write(text + '\n')
 
 if __name__ == '__main__': | 
| 52 | 
	https://:@github.com/hkwi/twink.git | 
	5ef359deb609fc55659596a8cb327abb9c7e4653 | 
	@@ -319,7 +319,7 @@ def ofp_action_set_field(message, offset):
 	cursor = _cursor(offset)
 	offset = cursor.offset
 	
-	(type,len) = ofp_action_header(message, offset)
+	(type,len) = ofp_action_header(message, cursor)
 	field = message[cursor.offset:offset+len]
 	cursor.offset = offset+len
 	return namedtuple("ofp_action_set_field",
 | 
	twink/ofp4/parse.py | 
	ReplaceText(target='cursor' @(322,41)->(322,47)) | 
	def ofp_action_set_field(message, offset):
 	cursor = _cursor(offset)
 	offset = cursor.offset
 	
 	(type,len) = ofp_action_header(message, offset)
 	field = message[cursor.offset:offset+len]
 	cursor.offset = offset+len
 	return namedtuple("ofp_action_set_field", | 
	def ofp_action_set_field(message, offset):
 	cursor = _cursor(offset)
 	offset = cursor.offset
 	
 	(type,len) = ofp_action_header(message, cursor)
 	field = message[cursor.offset:offset+len]
 	cursor.offset = offset+len
 	return namedtuple("ofp_action_set_field", | 
| 53 | 
	https://:@github.com/hkwi/twink.git | 
	fc6f5ad63cb12f9caf5455e3179ecfd9cd9de060 | 
	@@ -27,7 +27,7 @@ def _unpack(fmt, message, offset):
 	return struct.unpack_from(fmt, message, offset)
 
 def _align(length):
-	return (length+7)/8*8
+	return (length+7)//8*8
 
 # 7.1
 def ofp_header(version, type, length, xid):
 | 
	twink/ofp4/build.py | 
	ReplaceText(target='//' @(30,18)->(30,19)) | 
	def _unpack(fmt, message, offset):
 	return struct.unpack_from(fmt, message, offset)
 
 def _align(length):
 	return (length+7)/8*8
 
 # 7.1
 def ofp_header(version, type, length, xid): | 
	def _unpack(fmt, message, offset):
 	return struct.unpack_from(fmt, message, offset)
 
 def _align(length):
 	return (length+7)//8*8
 
 # 7.1
 def ofp_header(version, type, length, xid): | 
| 54 | 
	https://:@github.com/biolab/orange3-datafusion.git | 
	54f941a66a9b369a73190dfe2007e5b6dae1803a | 
	@@ -33,7 +33,7 @@ def _find_completion(fuser, relation):
     for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type,
                                                            relation.col_type):
         if fuser_relation._id == relation._id:
-            return fuser.complete(fuser_relation)
+            return fuser.complete(relation)
     return None
 
 
 | 
	orangecontrib/datafusion/widgets/owcompletionscoring.py | 
	ReplaceText(target='relation' @(36,34)->(36,48)) | 
	def _find_completion(fuser, relation):
     for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type,
                                                            relation.col_type):
         if fuser_relation._id == relation._id:
             return fuser.complete(fuser_relation)
     return None
 
  | 
	def _find_completion(fuser, relation):
     for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type,
                                                            relation.col_type):
         if fuser_relation._id == relation._id:
             return fuser.complete(relation)
     return None
 
  | 
| 55 | 
	https://:@github.com/aiqm/torchani.git | 
	5bb6691515e5e56fbe4994b140dd40b73043a33f | 
	@@ -151,7 +151,7 @@ class PrepareInput(torch.nn.Module):
         new_tensors = []
         for t in tensors:
             new_tensors.append(t.index_select(1, reverse))
-        return (species, *tensors)
+        return (species, *new_tensors)
 
     def forward(self, species_coordinates):
         species, coordinates = species_coordinates
 | 
	torchani/aev.py | 
	ReplaceText(target='new_tensors' @(154,26)->(154,33)) | 
	class PrepareInput(torch.nn.Module):
         new_tensors = []
         for t in tensors:
             new_tensors.append(t.index_select(1, reverse))
         return (species, *tensors)
 
     def forward(self, species_coordinates):
         species, coordinates = species_coordinates | 
	class PrepareInput(torch.nn.Module):
         new_tensors = []
         for t in tensors:
             new_tensors.append(t.index_select(1, reverse))
         return (species, *new_tensors)
 
     def forward(self, species_coordinates):
         species, coordinates = species_coordinates | 
| 56 | 
	https://:@github.com/aiqm/torchani.git | 
	abc8f7f842ae4b273c6e867b392413dcadd9c921 | 
	@@ -593,7 +593,7 @@ def collate_fn(data, chunk_threshold, properties_info):
         if properties_info['padding_values'][i] is None:
             prop = torch.stack(prop)
         else:
-            prop = torch.nn.utils.rnn.pad_sequence(batch_species,
+            prop = torch.nn.utils.rnn.pad_sequence(prop,
                                                    batch_first=True,
                                                    padding_value=properties_info['padding_values'][i])
         # sort with number of atoms
 | 
	torchani/data/new.py | 
	ReplaceText(target='prop' @(596,51)->(596,64)) | 
	def collate_fn(data, chunk_threshold, properties_info):
         if properties_info['padding_values'][i] is None:
             prop = torch.stack(prop)
         else:
             prop = torch.nn.utils.rnn.pad_sequence(batch_species,
                                                    batch_first=True,
                                                    padding_value=properties_info['padding_values'][i])
         # sort with number of atoms | 
	def collate_fn(data, chunk_threshold, properties_info):
         if properties_info['padding_values'][i] is None:
             prop = torch.stack(prop)
         else:
             prop = torch.nn.utils.rnn.pad_sequence(prop,
                                                    batch_first=True,
                                                    padding_value=properties_info['padding_values'][i])
         # sort with number of atoms | 
| 57 | 
	https://:@github.com/aiqm/torchani.git | 
	c18f4a5ea1f9732cc07c8816caa401981e43dc48 | 
	@@ -274,7 +274,7 @@ def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor,
     num_atoms = species.shape[1]
     num_species_pairs = angular_length // angular_sublength
     # PBC calculation is bypassed if there are no shifts
-    if shifts.numel() == 1:
+    if shifts.numel() == 0:
         atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr)
     else:
         atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr)
 | 
	torchani/aev.py | 
	ReplaceText(target='0' @(277,25)->(277,26)) | 
	def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor,
     num_atoms = species.shape[1]
     num_species_pairs = angular_length // angular_sublength
     # PBC calculation is bypassed if there are no shifts
     if shifts.numel() == 1:
         atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr)
     else:
         atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr) | 
	def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor,
     num_atoms = species.shape[1]
     num_species_pairs = angular_length // angular_sublength
     # PBC calculation is bypassed if there are no shifts
     if shifts.numel() == 0:
         atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr)
     else:
         atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr) | 
| 58 | 
	https://:@github.com/cs207group4/cs207-FinalProject.git | 
	e7f1cc613ace275a8d259de7455ee39ca063e029 | 
	@@ -120,7 +120,7 @@ class ChemSolver:
         r.set_initial_value(y0, 0)
         self._t = [0]
         self._y = [y0]
-        while r.successful() and r.t <= t1:
+        while r.successful() and r.t < t1:
             self._t.append(r.t + dt)
             self._y.append(r.integrate(r.t + dt))
         self._t = np.array(self._t)
 | 
	pychemkin/ChemSolver.py | 
	ReplaceText(target='<' @(123,37)->(123,39)) | 
	class ChemSolver:
         r.set_initial_value(y0, 0)
         self._t = [0]
         self._y = [y0]
         while r.successful() and r.t <= t1:
             self._t.append(r.t + dt)
             self._y.append(r.integrate(r.t + dt))
         self._t = np.array(self._t) | 
	class ChemSolver:
         r.set_initial_value(y0, 0)
         self._t = [0]
         self._y = [y0]
         while r.successful() and r.t < t1:
             self._t.append(r.t + dt)
             self._y.append(r.integrate(r.t + dt))
         self._t = np.array(self._t) | 
| 59 | 
	https://:@github.com/MrLeeh/pyads.git | 
	d14fd2a7bb2d4b784a4f6a47b6981ba2a86b699c | 
	@@ -97,7 +97,7 @@ def set_local_address(ams_netid):
     else:
         ams_netid_st = ams_netid
 
-    assert isinstance(ams_netid, SAmsNetId)
+    assert isinstance(ams_netid_st, SAmsNetId)
 
     if linux:
         return adsSetLocalAddress(ams_netid_st)
 | 
	pyads/ads.py | 
	ReplaceText(target='ams_netid_st' @(100,22)->(100,31)) | 
	def set_local_address(ams_netid):
     else:
         ams_netid_st = ams_netid
 
     assert isinstance(ams_netid, SAmsNetId)
 
     if linux:
         return adsSetLocalAddress(ams_netid_st) | 
	def set_local_address(ams_netid):
     else:
         ams_netid_st = ams_netid
 
     assert isinstance(ams_netid_st, SAmsNetId)
 
     if linux:
         return adsSetLocalAddress(ams_netid_st) | 
| 60 | 
	https://:@github.com/pytorch/fairseq.git | 
	0a836276129ef71fa6c44975dd02ab70bccc496d | 
	@@ -58,7 +58,7 @@ class FConvEncoder(FairseqEncoder):
         self.projections = nn.ModuleList()
         self.convolutions = nn.ModuleList()
         for (out_channels, kernel_size) in convolutions:
-            pad = (kernel_size - 1) // 2
+            pad = (kernel_size - 1) / 2
             self.projections.append(Linear(in_channels, out_channels)
                                     if in_channels != out_channels else None)
             self.convolutions.append(
 | 
	fairseq/models/fconv.py | 
	ReplaceText(target='/' @(61,36)->(61,38)) | 
	class FConvEncoder(FairseqEncoder):
         self.projections = nn.ModuleList()
         self.convolutions = nn.ModuleList()
         for (out_channels, kernel_size) in convolutions:
             pad = (kernel_size - 1) // 2
             self.projections.append(Linear(in_channels, out_channels)
                                     if in_channels != out_channels else None)
             self.convolutions.append( | 
	class FConvEncoder(FairseqEncoder):
         self.projections = nn.ModuleList()
         self.convolutions = nn.ModuleList()
         for (out_channels, kernel_size) in convolutions:
             pad = (kernel_size - 1) / 2
             self.projections.append(Linear(in_channels, out_channels)
                                     if in_channels != out_channels else None)
             self.convolutions.append( | 
| 61 | 
	https://:@github.com/pytorch/fairseq.git | 
	f68a44359b6596997b931d2e662a899ffba9d407 | 
	@@ -62,7 +62,7 @@ class SinusoidalPositionalEmbedding(nn.Module):
         # recompute/expand embeddings if needed
         bsz, seq_len = input.size()
         max_pos = self.padding_idx + 1 + seq_len
-        if seq_len > self.weights.size(0):
+        if max_pos > self.weights.size(0):
             self.weights = SinusoidalPositionalEmbedding.get_embedding(
                 max_pos,
                 self.embedding_dim,
 | 
	fairseq/modules/sinusoidal_positional_embedding.py | 
	ReplaceText(target='max_pos' @(65,11)->(65,18)) | 
	class SinusoidalPositionalEmbedding(nn.Module):
         # recompute/expand embeddings if needed
         bsz, seq_len = input.size()
         max_pos = self.padding_idx + 1 + seq_len
         if seq_len > self.weights.size(0):
             self.weights = SinusoidalPositionalEmbedding.get_embedding(
                 max_pos,
                 self.embedding_dim, | 
	class SinusoidalPositionalEmbedding(nn.Module):
         # recompute/expand embeddings if needed
         bsz, seq_len = input.size()
         max_pos = self.padding_idx + 1 + seq_len
         if max_pos > self.weights.size(0):
             self.weights = SinusoidalPositionalEmbedding.get_embedding(
                 max_pos,
                 self.embedding_dim, | 
| 62 | 
	https://:@github.com/pytorch/fairseq.git | 
	762956a559e65e1e48df8f8b4df515d23b66fddb | 
	@@ -82,7 +82,7 @@ def main(args):
     train_meter.start()
     valid_losses = [None]
     valid_subsets = args.valid_subset.split(',')
-    while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update:
+    while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update:
         # train for one epoch
         train(args, trainer, task, epoch_itr)
 
 | 
	train.py | 
	ReplaceText(target='<' @(85,47)->(85,49)) | 
	def main(args):
     train_meter.start()
     valid_losses = [None]
     valid_subsets = args.valid_subset.split(',')
     while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update:
         # train for one epoch
         train(args, trainer, task, epoch_itr)
  | 
	def main(args):
     train_meter.start()
     valid_losses = [None]
     valid_subsets = args.valid_subset.split(',')
     while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update:
         # train for one epoch
         train(args, trainer, task, epoch_itr)
  | 
| 63 | 
	https://:@github.com/pytorch/fairseq.git | 
	e9967cd334783f5da50deadc17cf8a4fc3380171 | 
	@@ -82,7 +82,7 @@ def main(args):
     train_meter.start()
     valid_losses = [None]
     valid_subsets = args.valid_subset.split(',')
-    while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update:
+    while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update:
         # train for one epoch
         train(args, trainer, task, epoch_itr)
 
 | 
	train.py | 
	ReplaceText(target='<' @(85,47)->(85,49)) | 
	def main(args):
     train_meter.start()
     valid_losses = [None]
     valid_subsets = args.valid_subset.split(',')
     while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update:
         # train for one epoch
         train(args, trainer, task, epoch_itr)
  | 
	def main(args):
     train_meter.start()
     valid_losses = [None]
     valid_subsets = args.valid_subset.split(',')
     while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update:
         # train for one epoch
         train(args, trainer, task, epoch_itr)
  | 
| 64 | 
	https://:@github.com/pytorch/fairseq.git | 
	7bcb487aad8504043d13c9b869d555aa565a46c7 | 
	@@ -49,7 +49,7 @@ class LabelSmoothedCrossEntropyCriterion(FairseqCriterion):
         sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens']
         logging_output = {
             'loss': utils.item(loss.data) if reduce else loss.data,
-            'nll_loss': utils.item(nll_loss.data) if reduce else loss.data,
+            'nll_loss': utils.item(nll_loss.data) if reduce else nll_loss.data,
             'ntokens': sample['ntokens'],
             'sample_size': sample_size,
         }
 | 
	fairseq/criterions/label_smoothed_cross_entropy.py | 
	ReplaceText(target='nll_loss' @(52,65)->(52,69)) | 
	class LabelSmoothedCrossEntropyCriterion(FairseqCriterion):
         sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens']
         logging_output = {
             'loss': utils.item(loss.data) if reduce else loss.data,
             'nll_loss': utils.item(nll_loss.data) if reduce else loss.data,
             'ntokens': sample['ntokens'],
             'sample_size': sample_size,
         } | 
	class LabelSmoothedCrossEntropyCriterion(FairseqCriterion):
         sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens']
         logging_output = {
             'loss': utils.item(loss.data) if reduce else loss.data,
             'nll_loss': utils.item(nll_loss.data) if reduce else nll_loss.data,
             'ntokens': sample['ntokens'],
             'sample_size': sample_size,
         } | 
| 65 | 
	https://:@github.com/pytorch/fairseq.git | 
	74efc21403477d103bd426ae64c37b7a30d8f4bf | 
	@@ -154,7 +154,7 @@ class TestIncrementalDecoder(FairseqIncrementalDecoder):
                     probs[:, i, self.dictionary.eos()] = 1.0
 
         # random attention
-        attn = torch.rand(bbsz, src_len, tgt_len)
+        attn = torch.rand(bbsz, tgt_len, src_len)
 
         return Variable(probs), Variable(attn)
 
 | 
	tests/utils.py | 
	ArgSwap(idxs=1<->2 @(157,15)->(157,25)) | 
	class TestIncrementalDecoder(FairseqIncrementalDecoder):
                     probs[:, i, self.dictionary.eos()] = 1.0
 
         # random attention
         attn = torch.rand(bbsz, src_len, tgt_len)
 
         return Variable(probs), Variable(attn)
  | 
	class TestIncrementalDecoder(FairseqIncrementalDecoder):
                     probs[:, i, self.dictionary.eos()] = 1.0
 
         # random attention
         attn = torch.rand(bbsz, tgt_len, src_len)
 
         return Variable(probs), Variable(attn)
  | 
| 66 | 
	https://:@github.com/pytorch/fairseq.git | 
	dfd77717b91a6e233829735795ab49d6fd85c0b3 | 
	@@ -93,7 +93,7 @@ class CosineSchedule(FairseqLRScheduler):
             else:
                 i = math.floor(curr_updates / self.period)
                 t_i = self.period
-                t_curr = num_updates - (self.period * i)
+                t_curr = curr_updates - (self.period * i)
 
             lr_shrink = self.lr_shrink ** i
             min_lr = self.min_lr  * lr_shrink
 | 
	fairseq/optim/lr_scheduler/cosine_lr_scheduler.py | 
	ReplaceText(target='curr_updates' @(96,25)->(96,36)) | 
	class CosineSchedule(FairseqLRScheduler):
             else:
                 i = math.floor(curr_updates / self.period)
                 t_i = self.period
                 t_curr = num_updates - (self.period * i)
 
             lr_shrink = self.lr_shrink ** i
             min_lr = self.min_lr  * lr_shrink | 
	class CosineSchedule(FairseqLRScheduler):
             else:
                 i = math.floor(curr_updates / self.period)
                 t_i = self.period
                 t_curr = curr_updates - (self.period * i)
 
             lr_shrink = self.lr_shrink ** i
             min_lr = self.min_lr  * lr_shrink | 
| 67 | 
	https://:@github.com/pytorch/fairseq.git | 
	0eea6923b9d7f408e667714709b070171ac7fe05 | 
	@@ -312,7 +312,7 @@ def make_positions(tensor, padding_idx, left_pad, onnx_trace=False):
         positions = range_buf.expand_as(tensor)
         if left_pad:
             positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1)
-        return positions * mask.long() + positions * (1 - mask.long())
+        return positions * mask.long() + padding_idx * (1 - mask.long())
 
     max_pos = padding_idx + 1 + tensor.size(1)
     if not hasattr(make_positions, 'range_buf'):
 | 
	fairseq/utils.py | 
	ReplaceText(target='padding_idx' @(315,41)->(315,50)) | 
	def make_positions(tensor, padding_idx, left_pad, onnx_trace=False):
         positions = range_buf.expand_as(tensor)
         if left_pad:
             positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1)
         return positions * mask.long() + positions * (1 - mask.long())
 
     max_pos = padding_idx + 1 + tensor.size(1)
     if not hasattr(make_positions, 'range_buf'): | 
	def make_positions(tensor, padding_idx, left_pad, onnx_trace=False):
         positions = range_buf.expand_as(tensor)
         if left_pad:
             positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1)
         return positions * mask.long() + padding_idx * (1 - mask.long())
 
     max_pos = padding_idx + 1 + tensor.size(1)
     if not hasattr(make_positions, 'range_buf'): | 
| 68 | 
	https://:@github.com/pytorch/fairseq.git | 
	4d3401b09f155995cd81fd394dfa50bf65ee8e5f | 
	@@ -183,7 +183,7 @@ class Sampling(Search):
             lprobs = lprobs[:, ::beam_size, :].contiguous()
 
         # we exclude the first two vocab items, one of which is pad
-        assert self.pad == 1, 'sampling assumes the first two symbols can be ignored'
+        assert self.pad <= 1, 'sampling assumes the first two symbols can be ignored'
         lprobs_nopad = lprobs[:, :, 2:]
 
         # only sample from top-k candidates
 | 
	fairseq/search.py | 
	ReplaceText(target='<=' @(186,24)->(186,26)) | 
	class Sampling(Search):
             lprobs = lprobs[:, ::beam_size, :].contiguous()
 
         # we exclude the first two vocab items, one of which is pad
         assert self.pad == 1, 'sampling assumes the first two symbols can be ignored'
         lprobs_nopad = lprobs[:, :, 2:]
 
         # only sample from top-k candidates | 
	class Sampling(Search):
             lprobs = lprobs[:, ::beam_size, :].contiguous()
 
         # we exclude the first two vocab items, one of which is pad
         assert self.pad <= 1, 'sampling assumes the first two symbols can be ignored'
         lprobs_nopad = lprobs[:, :, 2:]
 
         # only sample from top-k candidates | 
| 69 | 
	https://:@github.com/pytorch/fairseq.git | 
	2340832fdd7acaaaf07626daa6a0cef6fda06cd1 | 
	@@ -160,7 +160,7 @@ def main(args):
                     ))
 
         # update running id counter
-        start_id += len(results)
+        start_id += len(inputs)
 
 
 def cli_main():
 | 
	interactive.py | 
	ReplaceText(target='inputs' @(163,24)->(163,31)) | 
	def main(args):
                     ))
 
         # update running id counter
         start_id += len(results)
 
 
 def cli_main(): | 
	def main(args):
                     ))
 
         # update running id counter
         start_id += len(inputs)
 
 
 def cli_main(): | 
| 70 | 
	https://:@github.com/pytorch/fairseq.git | 
	39a60b844aad67aa59267d873edeb4948f6f0af9 | 
	@@ -351,7 +351,7 @@ class LSTMDecoder(FairseqIncrementalDecoder):
             self.additional_fc = Linear(hidden_size, out_embed_dim)
         if adaptive_softmax_cutoff is not None:
             # setting adaptive_softmax dropout to dropout_out for now but can be redefined
-            self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, embed_dim, adaptive_softmax_cutoff,
+            self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, hidden_size, adaptive_softmax_cutoff,
                                                     dropout=dropout_out)
         elif not self.share_input_output_embed:
             self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out)
 | 
	fairseq/models/lstm.py | 
	ReplaceText(target='hidden_size' @(354,68)->(354,77)) | 
	class LSTMDecoder(FairseqIncrementalDecoder):
             self.additional_fc = Linear(hidden_size, out_embed_dim)
         if adaptive_softmax_cutoff is not None:
             # setting adaptive_softmax dropout to dropout_out for now but can be redefined
             self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, embed_dim, adaptive_softmax_cutoff,
                                                     dropout=dropout_out)
         elif not self.share_input_output_embed:
             self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out) | 
	class LSTMDecoder(FairseqIncrementalDecoder):
             self.additional_fc = Linear(hidden_size, out_embed_dim)
         if adaptive_softmax_cutoff is not None:
             # setting adaptive_softmax dropout to dropout_out for now but can be redefined
             self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, hidden_size, adaptive_softmax_cutoff,
                                                     dropout=dropout_out)
         elif not self.share_input_output_embed:
             self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out) | 
| 71 | 
	https://:@github.com/pytorch/fairseq.git | 
	49177c99c45f7d6e99a8f1500d16396e2d7b4519 | 
	@@ -498,7 +498,7 @@ class TransformerDecoder(FairseqIncrementalDecoder):
                         del state_dict[k]
 
         version_key = '{}.version'.format(name)
-        if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) < 2:
+        if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) <= 2:
             # earlier checkpoints did not normalize after the stack of layers
             self.layer_norm = None
             self.normalize = False
 | 
	fairseq/models/transformer.py | 
	ReplaceText(target='<=' @(501,73)->(501,74)) | 
	class TransformerDecoder(FairseqIncrementalDecoder):
                         del state_dict[k]
 
         version_key = '{}.version'.format(name)
         if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) < 2:
             # earlier checkpoints did not normalize after the stack of layers
             self.layer_norm = None
             self.normalize = False | 
	class TransformerDecoder(FairseqIncrementalDecoder):
                         del state_dict[k]
 
         version_key = '{}.version'.format(name)
         if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) <= 2:
             # earlier checkpoints did not normalize after the stack of layers
             self.layer_norm = None
             self.normalize = False | 
| 72 | 
	https://:@github.com/pytorch/fairseq.git | 
	5d7a81099462e9f19715ce5fa37c03816a750e12 | 
	@@ -412,7 +412,7 @@ class LevenshteinTransformerModel(FairseqNATModel):
             max_lens = torch.zeros_like(output_tokens).fill_(255)
         else:
             if encoder_out.encoder_padding_mask is None:
-                max_src_len = encoder_out.encoder_out.size(1)
+                max_src_len = encoder_out.encoder_out.size(0)
                 src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len)
             else:
                 src_lens = (~encoder_out.encoder_padding_mask).sum(1)
 | 
	fairseq/models/nat/levenshtein_transformer.py | 
	ReplaceText(target='0' @(415,59)->(415,60)) | 
	class LevenshteinTransformerModel(FairseqNATModel):
             max_lens = torch.zeros_like(output_tokens).fill_(255)
         else:
             if encoder_out.encoder_padding_mask is None:
                 max_src_len = encoder_out.encoder_out.size(1)
                 src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len)
             else:
                 src_lens = (~encoder_out.encoder_padding_mask).sum(1) | 
	class LevenshteinTransformerModel(FairseqNATModel):
             max_lens = torch.zeros_like(output_tokens).fill_(255)
         else:
             if encoder_out.encoder_padding_mask is None:
                 max_src_len = encoder_out.encoder_out.size(0)
                 src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len)
             else:
                 src_lens = (~encoder_out.encoder_padding_mask).sum(1) | 
| 73 | 
	https://:@github.com/pytorch/fairseq.git | 
	431d604f696a15c06fceab56b4ace271bb85e74b | 
	@@ -331,7 +331,7 @@ class SequenceGenerator(object):
                 avg_attn_scores = avg_attn_scores[0]
             if avg_attn_scores is not None:
                 if attn is None:
-                    attn = scores.new(bsz * beam_size, src_tokens.size(1), max_len + 2)
+                    attn = scores.new(bsz * beam_size, avg_attn_scores.size(1), max_len + 2)
                     attn_buf = attn.clone()
                 attn[:, :, step + 1].copy_(avg_attn_scores)
 
 | 
	fairseq/sequence_generator.py | 
	ReplaceText(target='avg_attn_scores' @(334,55)->(334,65)) | 
	class SequenceGenerator(object):
                 avg_attn_scores = avg_attn_scores[0]
             if avg_attn_scores is not None:
                 if attn is None:
                     attn = scores.new(bsz * beam_size, src_tokens.size(1), max_len + 2)
                     attn_buf = attn.clone()
                 attn[:, :, step + 1].copy_(avg_attn_scores)
  | 
	class SequenceGenerator(object):
                 avg_attn_scores = avg_attn_scores[0]
             if avg_attn_scores is not None:
                 if attn is None:
                     attn = scores.new(bsz * beam_size, avg_attn_scores.size(1), max_len + 2)
                     attn_buf = attn.clone()
                 attn[:, :, step + 1].copy_(avg_attn_scores)
  | 
| 74 | 
	https://:@github.com/pytorch/fairseq.git | 
	4f8b0643c80d6a41039ae29e94fca6b44de8791a | 
	@@ -138,7 +138,7 @@ def should_stop_early(args, valid_loss):
         return False
     else:
         should_stop_early.num_runs += 1
-        return should_stop_early.num_runs > args.patience
+        return should_stop_early.num_runs >= args.patience
 
 
 @metrics.aggregate('train')
 | 
	fairseq_cli/train.py | 
	ReplaceText(target='>=' @(141,42)->(141,43)) | 
	def should_stop_early(args, valid_loss):
         return False
     else:
         should_stop_early.num_runs += 1
         return should_stop_early.num_runs > args.patience
 
 
 @metrics.aggregate('train') | 
	def should_stop_early(args, valid_loss):
         return False
     else:
         should_stop_early.num_runs += 1
         return should_stop_early.num_runs >= args.patience
 
 
 @metrics.aggregate('train') | 
| 75 | 
	https://:@github.com/pytorch/fairseq.git | 
	9a718e29855713a51877237b2dcc25e39c234c82 | 
	@@ -110,5 +110,5 @@ class TranslationFromPretrainedBARTTask(TranslationTask):
         for s_t in src_tokens:
             s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)])
             source_tokens.append(s_t)
-        dataset = LanguagePairDataset(src_tokens, src_lengths, self.source_dictionary)
+        dataset = LanguagePairDataset(source_tokens, src_lengths, self.source_dictionary)
         return dataset
 | 
	fairseq/tasks/translation_from_pretrained_bart.py | 
	ReplaceText(target='source_tokens' @(113,38)->(113,48)) | 
	class TranslationFromPretrainedBARTTask(TranslationTask):
         for s_t in src_tokens:
             s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)])
             source_tokens.append(s_t)
         dataset = LanguagePairDataset(src_tokens, src_lengths, self.source_dictionary)
         return dataset | 
	class TranslationFromPretrainedBARTTask(TranslationTask):
         for s_t in src_tokens:
             s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)])
             source_tokens.append(s_t)
         dataset = LanguagePairDataset(source_tokens, src_lengths, self.source_dictionary)
         return dataset | 
| 76 | 
	https://:@github.com/pytorch/fairseq.git | 
	b689b6ff3ab7b806217b8aa41821bb8fc85f7cd8 | 
	@@ -264,7 +264,7 @@ class LanguagePairDataset(FairseqDataset):
                 tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]])
 
             bos = self.src_dict.bos()
-            if self.src[index][-1] != bos:
+            if self.src[index][0] != bos:
                 src_item = torch.cat([torch.LongTensor([bos]), self.src[index]])
 
         if self.remove_eos_from_source:
 | 
	fairseq/data/language_pair_dataset.py | 
	ReplaceText(target='0' @(267,31)->(267,33)) | 
	class LanguagePairDataset(FairseqDataset):
                 tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]])
 
             bos = self.src_dict.bos()
             if self.src[index][-1] != bos:
                 src_item = torch.cat([torch.LongTensor([bos]), self.src[index]])
 
         if self.remove_eos_from_source: | 
	class LanguagePairDataset(FairseqDataset):
                 tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]])
 
             bos = self.src_dict.bos()
             if self.src[index][0] != bos:
                 src_item = torch.cat([torch.LongTensor([bos]), self.src[index]])
 
         if self.remove_eos_from_source: | 
| 77 | 
	https://:@github.com/prprprus/PyMySQLPool.git | 
	66b07cdf844554245cf209a72de89bd17133269c | 
	@@ -169,7 +169,7 @@ class Pool(object):
         if self.ping_check:
             now = int(time())
             timeout = now
-            if isinstance(int, self.ping_check):
+            if isinstance(self.ping_check, int):
                 timeout = timeout - self.ping_check
             if not hasattr(c, '__ping_check_timestamp'):
                 c.__ping_check_timestamp = now
 | 
	pymysqlpool/pool.py | 
	ArgSwap(idxs=0<->1 @(172,15)->(172,25)) | 
	class Pool(object):
         if self.ping_check:
             now = int(time())
             timeout = now
             if isinstance(int, self.ping_check):
                 timeout = timeout - self.ping_check
             if not hasattr(c, '__ping_check_timestamp'):
                 c.__ping_check_timestamp = now | 
	class Pool(object):
         if self.ping_check:
             now = int(time())
             timeout = now
             if isinstance(self.ping_check, int):
                 timeout = timeout - self.ping_check
             if not hasattr(c, '__ping_check_timestamp'):
                 c.__ping_check_timestamp = now | 
| 78 | 
	https://:@github.com/dailymuse/oz.git | 
	a15adf73c721d07b9dac886fcc27145e2449563c | 
	@@ -173,7 +173,7 @@ class S3File(CDNFile):
 
     def copy(self, new_path, replace=False):
         """Uses boto to copy the file to the new path instead of uploading another file to the new key"""
-        if replace or get_file(new_path).exists():
+        if replace or not get_file(new_path).exists():
             self.key.copy(self.key.bucket, new_path)
             return True
         return False
 | 
	oz/aws_cdn/__init__.py | 
	ReplaceText(target='not ' @(176,22)->(176,22)) | 
	class S3File(CDNFile):
 
     def copy(self, new_path, replace=False):
         """Uses boto to copy the file to the new path instead of uploading another file to the new key"""
         if replace or get_file(new_path).exists():
             self.key.copy(self.key.bucket, new_path)
             return True
         return False | 
	class S3File(CDNFile):
 
     def copy(self, new_path, replace=False):
         """Uses boto to copy the file to the new path instead of uploading another file to the new key"""
         if replace or not get_file(new_path).exists():
             self.key.copy(self.key.bucket, new_path)
             return True
         return False | 
| 79 | 
	https://:@github.com/juju/amulet.git | 
	016bfab60aca89cbcb58e80f4103e371a77b06ba | 
	@@ -77,7 +77,7 @@ class Deployment(object):
             pass  # Copy the current parent directory to temp and deploy that
         elif self.charm_name:
             if charm_name == self.charm_name:
-                charm = os.getcwd()
+                charm_branch = os.getcwd()
 
         self.services[service] = {'branch': charm_branch}
         if units > 1:
 | 
	amulet/deployer.py | 
	ReplaceText(target='charm_branch' @(80,16)->(80,21)) | 
	class Deployment(object):
             pass  # Copy the current parent directory to temp and deploy that
         elif self.charm_name:
             if charm_name == self.charm_name:
                 charm = os.getcwd()
 
         self.services[service] = {'branch': charm_branch}
         if units > 1: | 
	class Deployment(object):
             pass  # Copy the current parent directory to temp and deploy that
         elif self.charm_name:
             if charm_name == self.charm_name:
                 charm_branch = os.getcwd()
 
         self.services[service] = {'branch': charm_branch}
         if units > 1: | 
| 80 | 
	https://:@github.com/gitpython-developers/gitdb.git | 
	ca829e0b341dd5c3ae1408b24702f2c75db6ec73 | 
	@@ -445,7 +445,7 @@ class DeltaApplyReader(LazyMixin):
 		
 	
 	#{ Configuration
-	if not has_perf_mod:
+	if has_perf_mod:
 		_set_cache_ = _set_cache_brute_
 	else:
 		_set_cache_ = _set_cache_too_slow_without_c
 | 
	stream.py | 
	ReplaceText(target='' @(448,4)->(448,8)) | 
	class DeltaApplyReader(LazyMixin):
 		
 	
 	#{ Configuration
 	if not has_perf_mod:
 		_set_cache_ = _set_cache_brute_
 	else:
 		_set_cache_ = _set_cache_too_slow_without_c | 
	class DeltaApplyReader(LazyMixin):
 		
 	
 	#{ Configuration
 	if has_perf_mod:
 		_set_cache_ = _set_cache_brute_
 	else:
 		_set_cache_ = _set_cache_too_slow_without_c | 
| 81 | 
	https://:@github.com/longld/peda.git | 
	82fcb5a12c92c27fc5722772a84df47b996d3d03 | 
	@@ -4746,7 +4746,7 @@ class PEDACmd(object):
 
         step = peda.intsize()
         if not peda.is_address(address): # cannot determine address
-            msg("Invalid $SP address: 0x%x" % sp, "red")
+            msg("Invalid $SP address: 0x%x" % address, "red")
             return
             for i in range(count):
                 if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)):
 | 
	peda.py | 
	ReplaceText(target='address' @(4749,46)->(4749,48)) | 
	class PEDACmd(object):
 
         step = peda.intsize()
         if not peda.is_address(address): # cannot determine address
             msg("Invalid $SP address: 0x%x" % sp, "red")
             return
             for i in range(count):
                 if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)): | 
	class PEDACmd(object):
 
         step = peda.intsize()
         if not peda.is_address(address): # cannot determine address
             msg("Invalid $SP address: 0x%x" % address, "red")
             return
             for i in range(count):
                 if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)): | 
| 82 | 
	https://:@github.com/TheGhouls/oct.git | 
	1f9ea29181962353fe0ea275cb4ba4ec9ae93142 | 
	@@ -18,7 +18,7 @@ class Report(object):
         self.set_statics()
 
     def set_statics(self):
-        if os.path.exists(self.results_dir):
+        if not os.path.exists(self.results_dir):
             return
         try:
             shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css'))
 | 
	oct/results/reportwriter.py | 
	ReplaceText(target='not ' @(21,11)->(21,11)) | 
	class Report(object):
         self.set_statics()
 
     def set_statics(self):
         if os.path.exists(self.results_dir):
             return
         try:
             shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css')) | 
	class Report(object):
         self.set_statics()
 
     def set_statics(self):
         if not os.path.exists(self.results_dir):
             return
         try:
             shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css')) | 
| 83 | 
	https://:@github.com/ASPP/pelita.git | 
	bee6872dbd95a1e526305ef39f42ac537fd2f708 | 
	@@ -105,7 +105,7 @@ def create_maze(layout_mesh):
         Mesh of lists of MazeComponents
 
     """
-    maze_mesh = Mesh(layout_mesh.height, layout_mesh.width,
+    maze_mesh = Mesh(layout_mesh.width, layout_mesh.height,
             data=[[] for i in range(len(layout_mesh))])
     for index in maze_mesh.iterkeys():
         if layout_mesh[index] == CTFUniverse.wall:
 | 
	pelita/universe.py | 
	ArgSwap(idxs=0<->1 @(108,16)->(108,20)) | 
	def create_maze(layout_mesh):
         Mesh of lists of MazeComponents
 
     """
     maze_mesh = Mesh(layout_mesh.height, layout_mesh.width,
             data=[[] for i in range(len(layout_mesh))])
     for index in maze_mesh.iterkeys():
         if layout_mesh[index] == CTFUniverse.wall: | 
	def create_maze(layout_mesh):
         Mesh of lists of MazeComponents
 
     """
     maze_mesh = Mesh(layout_mesh.width, layout_mesh.height,
             data=[[] for i in range(len(layout_mesh))])
     for index in maze_mesh.iterkeys():
         if layout_mesh[index] == CTFUniverse.wall: | 
| 84 | 
	https://:@github.com/ASPP/pelita.git | 
	508cd180dce7b72ab248211c977c8525a9c023de | 
	@@ -31,7 +31,7 @@ def __init__(self, index, initial_pos, team, homezone,
 
     @property
     def in_own_zone(self):
-        return self.homezone[0] <= self.current_pos[1] <= self.homezone[1]
+        return self.homezone[0] <= self.current_pos[0] <= self.homezone[1]
 
     def move(self, new_pos):
         self.current_pos = new_pos
 | 
	pelita/universe.py | 
	ReplaceText(target='0' @(34,52)->(34,53)) | 
	def __init__(self, index, initial_pos, team, homezone,
 
     @property
     def in_own_zone(self):
         return self.homezone[0] <= self.current_pos[1] <= self.homezone[1]
 
     def move(self, new_pos):
         self.current_pos = new_pos | 
	def __init__(self, index, initial_pos, team, homezone,
 
     @property
     def in_own_zone(self):
         return self.homezone[0] <= self.current_pos[0] <= self.homezone[1]
 
     def move(self, new_pos):
         self.current_pos = new_pos | 
| 85 | 
	https://:@github.com/ASPP/pelita.git | 
	6b76e416da2dc0d18224e47d7b176dad967e15b2 | 
	@@ -182,7 +182,7 @@ def a_star(self, initial, target):
             else:
                 seen.append(current)
                 for pos in self.adjacency[current]:
-                    heapq.heappush(to_visit, (datamodel.manhattan_dist(current, pos), (pos)))
+                    heapq.heappush(to_visit, (datamodel.manhattan_dist(target, pos), (pos)))
 
         # Now back-track using seen to determine how we got here.
         # Initialise the path with current node, i.e. position of food.
 | 
	pelita/game_master.py | 
	ReplaceText(target='target' @(185,71)->(185,78)) | 
	def a_star(self, initial, target):
             else:
                 seen.append(current)
                 for pos in self.adjacency[current]:
                     heapq.heappush(to_visit, (datamodel.manhattan_dist(current, pos), (pos)))
 
         # Now back-track using seen to determine how we got here.
         # Initialise the path with current node, i.e. position of food. | 
	def a_star(self, initial, target):
             else:
                 seen.append(current)
                 for pos in self.adjacency[current]:
                     heapq.heappush(to_visit, (datamodel.manhattan_dist(target, pos), (pos)))
 
         # Now back-track using seen to determine how we got here.
         # Initialise the path with current node, i.e. position of food. | 
| 86 | 
	https://:@github.com/ASPP/pelita.git | 
	fa2505d44ae3d3724f7fa979c0167f03bf7424f7 | 
	@@ -136,7 +136,7 @@ def play(self):
         if self.universe.teams[0].score < self.universe.teams[1].score:
             events.append(datamodel.TeamWins(1))
         elif self.universe.teams[0].score > self.universe.teams[1].score:
-            events.append(datamodel.TeamWins(1))
+            events.append(datamodel.TeamWins(0))
         else:
             events.append(datamodel.GameDraw())
         self.send_to_viewers(round_index, None, events)
 | 
	pelita/game_master.py | 
	ReplaceText(target='0' @(139,45)->(139,46)) | 
	def play(self):
         if self.universe.teams[0].score < self.universe.teams[1].score:
             events.append(datamodel.TeamWins(1))
         elif self.universe.teams[0].score > self.universe.teams[1].score:
             events.append(datamodel.TeamWins(1))
         else:
             events.append(datamodel.GameDraw())
         self.send_to_viewers(round_index, None, events) | 
	def play(self):
         if self.universe.teams[0].score < self.universe.teams[1].score:
             events.append(datamodel.TeamWins(1))
         elif self.universe.teams[0].score > self.universe.teams[1].score:
             events.append(datamodel.TeamWins(0))
         else:
             events.append(datamodel.GameDraw())
         self.send_to_viewers(round_index, None, events) | 
| 87 | 
	https://:@github.com/ASPP/pelita.git | 
	4044f845e54c2077d6896010c25fcc123fc10203 | 
	@@ -78,7 +78,7 @@ def test_equal_positions(self):
         layout = create_layout(layout_str)
         assert layout.bots == [(1, 1), (1, 1)]
         assert layout.enemy ==  [(1, 1), (1, 1)]
-        setup_test_game(layout=layout)
+        setup_test_game(layout=layout_str)
 
     def test_define_after(self):
         layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None)
 | 
	test/test_team.py | 
	ReplaceText(target='layout_str' @(81,31)->(81,37)) | 
	def test_equal_positions(self):
         layout = create_layout(layout_str)
         assert layout.bots == [(1, 1), (1, 1)]
         assert layout.enemy ==  [(1, 1), (1, 1)]
         setup_test_game(layout=layout)
 
     def test_define_after(self):
         layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None) | 
	def test_equal_positions(self):
         layout = create_layout(layout_str)
         assert layout.bots == [(1, 1), (1, 1)]
         assert layout.enemy ==  [(1, 1), (1, 1)]
         setup_test_game(layout=layout_str)
 
     def test_define_after(self):
         layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None) | 
| 88 | 
	https://:@github.com/ASPP/pelita.git | 
	6fd0a9d2af44c491c1cc6774c3a169e97e2040be | 
	@@ -398,7 +398,7 @@ def _team(self):
     @property
     def turn(self):
         """ The turn of our bot. """
-        return self.bot_index // 2
+        return self.bot_index % 2
 
     @property
     def other(self):
 | 
	pelita/player/team.py | 
	ReplaceText(target='%' @(401,30)->(401,32)) | 
	def _team(self):
     @property
     def turn(self):
         """ The turn of our bot. """
         return self.bot_index // 2
 
     @property
     def other(self): | 
	def _team(self):
     @property
     def turn(self):
         """ The turn of our bot. """
         return self.bot_index % 2
 
     @property
     def other(self): | 
| 89 | 
	https://:@github.com/iotaledger/ccurl.interface.py.git | 
	eb7f9190d24995d3f8d03a8350382ab6045a6e67 | 
	@@ -44,7 +44,7 @@ gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your
 
 mwm = 14 # target is mainnet
 
-bundle = entangled_interface.local_attach_to_tangle(pb, gta['trunkTransaction'], gta['branchTransaction'], mwm)
+bundle = entangled_interface.local_attach_to_tangle(pb, gta['branchTransaction'],gta['trunkTransaction'],  mwm)
 
 bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
 
 | 
	examples/with_entangled.py | 
	ArgSwap(idxs=1<->2 @(47,9)->(47,51)) | 
	gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your
 
 mwm = 14 # target is mainnet
 
 bundle = entangled_interface.local_attach_to_tangle(pb, gta['trunkTransaction'], gta['branchTransaction'], mwm)
 
 bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
  | 
	gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your
 
 mwm = 14 # target is mainnet
 
 bundle = entangled_interface.local_attach_to_tangle(pb, gta['branchTransaction'],gta['trunkTransaction'],  mwm)
 
 bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
  | 
| 90 | 
	https://:@github.com/softlayer/softlayer-python.git | 
	53731de7e51d31475cc224aceb0f3ff7217cdafd | 
	@@ -153,7 +153,7 @@ class NetworkManager(object):
                            ('privateResidenceFlag', private_residence),
                            ('state', state),
                            ('postalCode', postal_code)]:
-            if key is not None:
+            if value is not None:
                 update[key] = value
 
         # If there's anything to update, update it
 | 
	SoftLayer/managers/network.py | 
	ReplaceText(target='value' @(156,15)->(156,18)) | 
	class NetworkManager(object):
                            ('privateResidenceFlag', private_residence),
                            ('state', state),
                            ('postalCode', postal_code)]:
             if key is not None:
                 update[key] = value
 
         # If there's anything to update, update it | 
	class NetworkManager(object):
                            ('privateResidenceFlag', private_residence),
                            ('state', state),
                            ('postalCode', postal_code)]:
             if value is not None:
                 update[key] = value
 
         # If there's anything to update, update it | 
| 91 | 
	https://:@github.com/softlayer/softlayer-python.git | 
	dcf66e15711e47c594f20ffac7605bfc6d1a8746 | 
	@@ -15,7 +15,7 @@ import click
               type=click.Choice(['vs', 'vlan', 'server']),
               help='Firewall type',
               required=True)
[email protected]('--high-availability', '--ha',
[email protected]('--ha', '--high-availability',
               is_flag=True,
               help='High available firewall option')
 @environment.pass_env
 | 
	SoftLayer/CLI/firewall/add.py | 
	ArgSwap(idxs=0<->1 @(18,1)->(18,13)) | 
	import click
               type=click.Choice(['vs', 'vlan', 'server']),
               help='Firewall type',
               required=True)
 @click.option('--high-availability', '--ha',
               is_flag=True,
               help='High available firewall option')
 @environment.pass_env | 
	import click
               type=click.Choice(['vs', 'vlan', 'server']),
               help='Firewall type',
               required=True)
 @click.option('--ha', '--high-availability',
               is_flag=True,
               help='High available firewall option')
 @environment.pass_env | 
| 92 | 
	https://:@github.com/softlayer/softlayer-python.git | 
	f0840e302d486d6002a14419bbde85c1deedaf6a | 
	@@ -271,7 +271,7 @@ class BlockStorageManager(utils.IdentifierMixin, object):
                     package,
                     'performance_storage_iscsi'
                     ),
-                storage_utils.find_performance_space_price(package, iops),
+                storage_utils.find_performance_space_price(package, size),
                 storage_utils.find_performance_iops_price(package, size, iops),
             ]
         elif storage_type == 'storage_service_enterprise':
 | 
	SoftLayer/managers/block.py | 
	ReplaceText(target='size' @(274,68)->(274,72)) | 
	class BlockStorageManager(utils.IdentifierMixin, object):
                     package,
                     'performance_storage_iscsi'
                     ),
                 storage_utils.find_performance_space_price(package, iops),
                 storage_utils.find_performance_iops_price(package, size, iops),
             ]
         elif storage_type == 'storage_service_enterprise': | 
	class BlockStorageManager(utils.IdentifierMixin, object):
                     package,
                     'performance_storage_iscsi'
                     ),
                 storage_utils.find_performance_space_price(package, size),
                 storage_utils.find_performance_iops_price(package, size, iops),
             ]
         elif storage_type == 'storage_service_enterprise': | 
| 93 | 
	https://:@github.com/softlayer/softlayer-python.git | 
	4418057fc0e3632aba2d89b6e42494c79cadd16a | 
	@@ -367,7 +367,7 @@ class VSManager(utils.IdentifierMixin, object):
         if datacenter:
             data["datacenter"] = {"name": datacenter}
 
-        if private_vlan and public_vlan:
+        if private_vlan or public_vlan:
             network_components = self._create_network_components(public_vlan, private_vlan,
                                                                  private_subnet, public_subnet)
             data.update(network_components)
 | 
	SoftLayer/managers/vs.py | 
	ReplaceText(target='or' @(370,24)->(370,27)) | 
	class VSManager(utils.IdentifierMixin, object):
         if datacenter:
             data["datacenter"] = {"name": datacenter}
 
         if private_vlan and public_vlan:
             network_components = self._create_network_components(public_vlan, private_vlan,
                                                                  private_subnet, public_subnet)
             data.update(network_components) | 
	class VSManager(utils.IdentifierMixin, object):
         if datacenter:
             data["datacenter"] = {"name": datacenter}
 
         if private_vlan or public_vlan:
             network_components = self._create_network_components(public_vlan, private_vlan,
                                                                  private_subnet, public_subnet)
             data.update(network_components) | 
| 94 | 
	https://:@github.com/softlayer/softlayer-python.git | 
	58b27c6bf5400a717acd00b7866964ef11f36e59 | 
	@@ -87,6 +87,6 @@ def cli(env, identifier):
     for guest in guests:
         real_guest = guest.get('virtualGuest')
         member_table.add_row([
-            guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate'))
+            real_guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate'))
         ])
     env.fout(member_table)
 | 
	SoftLayer/CLI/autoscale/detail.py | 
	ReplaceText(target='real_guest' @(90,12)->(90,17)) | 
	def cli(env, identifier):
     for guest in guests:
         real_guest = guest.get('virtualGuest')
         member_table.add_row([
             guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate'))
         ])
     env.fout(member_table) | 
	def cli(env, identifier):
     for guest in guests:
         real_guest = guest.get('virtualGuest')
         member_table.add_row([
             real_guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate'))
         ])
     env.fout(member_table) | 
| 95 | 
	https://:@github.com/data-8/datascience.git | 
	fd9aceb598290fb89a0f3131c3fb39dde18ef543 | 
	@@ -446,7 +446,7 @@ class Table(collections.abc.Mapping):
         count | points
         9     | 10
         """
-        percentiles = [percentile(self[column_name], p) for column_name in self]
+        percentiles = [percentile(p, self[column_name]) for column_name in self]
         return Table(percentiles, self.column_labels)
     ##################
     # Export/Display #
 | 
	datascience/tables.py | 
	ArgSwap(idxs=0<->1 @(449,23)->(449,33)) | 
	class Table(collections.abc.Mapping):
         count | points
         9     | 10
         """
         percentiles = [percentile(self[column_name], p) for column_name in self]
         return Table(percentiles, self.column_labels)
     ##################
     # Export/Display # | 
	class Table(collections.abc.Mapping):
         count | points
         9     | 10
         """
         percentiles = [percentile(p, self[column_name]) for column_name in self]
         return Table(percentiles, self.column_labels)
     ##################
     # Export/Display # | 
| 96 | 
	https://:@github.com/data-8/datascience.git | 
	084450f127ecc490b887cad82fa43cda5f9b32fe | 
	@@ -2255,7 +2255,7 @@ class Table(collections.abc.MutableMapping):
                     space_count[labels[i]] += 1
                 return updated_labels
             return labels
-        yticks = make_unique_labels(labels)
+        yticks = make_unique_labels(yticks)
 
         print("yticks: " + str(yticks))
         print("ylabel: " + str(ylabel))
 | 
	datascience/tables.py | 
	ReplaceText(target='yticks' @(2258,36)->(2258,42)) | 
	class Table(collections.abc.MutableMapping):
                     space_count[labels[i]] += 1
                 return updated_labels
             return labels
         yticks = make_unique_labels(labels)
 
         print("yticks: " + str(yticks))
         print("ylabel: " + str(ylabel)) | 
	class Table(collections.abc.MutableMapping):
                     space_count[labels[i]] += 1
                 return updated_labels
             return labels
         yticks = make_unique_labels(yticks)
 
         print("yticks: " + str(yticks))
         print("ylabel: " + str(ylabel)) | 
| 97 | 
	https://:@github.com/dnaeon/py-vpoller.git | 
	81769f6f8d9cb0dfc8cbc39a44027afa7d459636 | 
	@@ -51,6 +51,6 @@ def task(name, required=None):
                 result = {'success': 1, 'msg': e.message}
             finally:
                 return result
-        registry.register(name=name, fn=fn, required=required)
+        registry.register(name=name, fn=wrapper, required=required)
         return wrapper
     return decorator
 | 
	src/vpoller/decorators.py | 
	ReplaceText(target='wrapper' @(54,40)->(54,42)) | 
	def task(name, required=None):
                 result = {'success': 1, 'msg': e.message}
             finally:
                 return result
         registry.register(name=name, fn=fn, required=required)
         return wrapper
     return decorator | 
	def task(name, required=None):
                 result = {'success': 1, 'msg': e.message}
             finally:
                 return result
         registry.register(name=name, fn=wrapper, required=required)
         return wrapper
     return decorator | 
| 98 | 
	https://:@github.com/enthought/qt_binder.git | 
	68381b406035f2ce9666cb8ef1ab2e8e57cf8bf8 | 
	@@ -58,4 +58,4 @@ else:
         loader = RecordingUiLoader()
         ui = loader.load(path)
 
-        return ui, ui.to_be_bound()
+        return ui, loader.to_be_bound()
 | 
	qt_binder/qt/ui_loader.py | 
	ReplaceText(target='loader' @(61,19)->(61,21)) | 
	else:
         loader = RecordingUiLoader()
         ui = loader.load(path)
 
         return ui, ui.to_be_bound() | 
	else:
         loader = RecordingUiLoader()
         ui = loader.load(path)
 
         return ui, loader.to_be_bound() | 
| 99 | 
	https://:@github.com/ggozad/behaving.git | 
	1bc546aa03f9d42ff78a0a79e0894e488edc9add | 
	@@ -40,7 +40,7 @@ def should_receive_email(context, address):
 def click_link_in_email(context, address):
     mails = context.mail.user_messages(address)
     assert mails, u'message not found'
-    mail = email.message_from_string(mails[-1])
+    mail = email.message_from_string(mails[0])
     links = URL_RE.findall(str(mail).replace('=\n', ''))
     assert links, u'link not found'
     url = links[0]
 | 
	src/behaving/mail/steps.py | 
	ReplaceText(target='0' @(43,43)->(43,45)) | 
	def should_receive_email(context, address):
 def click_link_in_email(context, address):
     mails = context.mail.user_messages(address)
     assert mails, u'message not found'
     mail = email.message_from_string(mails[-1])
     links = URL_RE.findall(str(mail).replace('=\n', ''))
     assert links, u'link not found'
     url = links[0] | 
	def should_receive_email(context, address):
 def click_link_in_email(context, address):
     mails = context.mail.user_messages(address)
     assert mails, u'message not found'
     mail = email.message_from_string(mails[0])
     links = URL_RE.findall(str(mail).replace('=\n', ''))
     assert links, u'link not found'
     url = links[0] | 
End of preview. Expand
						in Data Studio
					
See Allamanis et al., 2021 (NeurIPS 2021) for more information.
- Downloads last month
- 17
