Dataset Viewer
Unnamed: 0
int64 0
2.44k
| repo
stringlengths 32
81
| hash
stringlengths 40
40
| diff
stringlengths 113
1.17k
| old_path
stringlengths 5
84
| rewrite
stringlengths 34
79
| initial_state
stringlengths 75
980
| final_state
stringlengths 76
980
|
---|---|---|---|---|---|---|---|
0 |
https://:@github.com/emedvedev/attention-ocr.git
|
291042e7cb623c8a908e9badd132c1fa2360288c
|
@@ -465,7 +465,7 @@ class Model(object):
mh = 32
mw = math.floor(1. * w / h * mh)
img = img.resize(
- (mw, h),
+ (mw, mh),
Image.ANTIALIAS)
img_data = np.asarray(img, dtype=np.uint8)
for idx in xrange(len(output)):
|
aocr/model/model.py
|
ReplaceText(target='mh' @(468,25)->(468,26))
|
class Model(object):
mh = 32
mw = math.floor(1. * w / h * mh)
img = img.resize(
(mw, h),
Image.ANTIALIAS)
img_data = np.asarray(img, dtype=np.uint8)
for idx in xrange(len(output)):
|
class Model(object):
mh = 32
mw = math.floor(1. * w / h * mh)
img = img.resize(
(mw, mh),
Image.ANTIALIAS)
img_data = np.asarray(img, dtype=np.uint8)
for idx in xrange(len(output)):
|
1 |
https://:@github.com/emedvedev/attention-ocr.git
|
6e6593c27fe0e63118adadf11562b1c4699b14e3
|
@@ -477,7 +477,7 @@ class Model(object):
attention_orig[i] = attention[int(i/4)-1]
attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same')
attention_orig = np.maximum(attention_orig, 0.3)
- attention_out = np.zeros((h, mw))
+ attention_out = np.zeros((mh, mw))
for i in xrange(mw):
attention_out[:, i] = attention_orig[i]
if len(img_data.shape) == 3:
|
aocr/model/model.py
|
ReplaceText(target='mh' @(480,42)->(480,43))
|
class Model(object):
attention_orig[i] = attention[int(i/4)-1]
attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same')
attention_orig = np.maximum(attention_orig, 0.3)
attention_out = np.zeros((h, mw))
for i in xrange(mw):
attention_out[:, i] = attention_orig[i]
if len(img_data.shape) == 3:
|
class Model(object):
attention_orig[i] = attention[int(i/4)-1]
attention_orig = np.convolve(attention_orig, [0.199547, 0.200226, 0.200454, 0.200226, 0.199547], mode='same')
attention_orig = np.maximum(attention_orig, 0.3)
attention_out = np.zeros((mh, mw))
for i in xrange(mw):
attention_out[:, i] = attention_orig[i]
if len(img_data.shape) == 3:
|
2 |
https://:@github.com/emedvedev/attention-ocr.git
|
e741baf7170e72a974754908d323cacc0bd55247
|
@@ -133,7 +133,7 @@ class Model(object):
self.target_weights = []
for i in xrange(self.decoder_size + 1):
self.decoder_inputs.append(
- tf.tile([0], [num_images])
+ tf.tile([1], [num_images])
)
if i < self.decoder_size:
self.target_weights.append(tf.tile([1.], [num_images]))
|
aocr/model/model.py
|
ReplaceText(target='1' @(136,29)->(136,30))
|
class Model(object):
self.target_weights = []
for i in xrange(self.decoder_size + 1):
self.decoder_inputs.append(
tf.tile([0], [num_images])
)
if i < self.decoder_size:
self.target_weights.append(tf.tile([1.], [num_images]))
|
class Model(object):
self.target_weights = []
for i in xrange(self.decoder_size + 1):
self.decoder_inputs.append(
tf.tile([1], [num_images])
)
if i < self.decoder_size:
self.target_weights.append(tf.tile([1.], [num_images]))
|
3 |
https://:@github.com/matthewdowney/TogglPy.git
|
d5b630aec58d29b85ccffa527d24766eef6f61f9
|
@@ -197,7 +197,7 @@ class Toggl():
day = datetime.now().day if not day else day
hour = datetime.now().hour if not hour else hour
- timestruct = datetime(year, month, day, hour - hourdiff).isoformat() + '.000Z'
+ timestruct = datetime(year, month, day, hour + hourdiff).isoformat() + '.000Z'
data['time_entry']['start'] = timestruct
data['time_entry']['duration'] = hourduration * 3600
data['time_entry']['pid'] = projectid
|
toggl/TogglPy.py
|
ReplaceText(target='+' @(200,53)->(200,54))
|
class Toggl():
day = datetime.now().day if not day else day
hour = datetime.now().hour if not hour else hour
timestruct = datetime(year, month, day, hour - hourdiff).isoformat() + '.000Z'
data['time_entry']['start'] = timestruct
data['time_entry']['duration'] = hourduration * 3600
data['time_entry']['pid'] = projectid
|
class Toggl():
day = datetime.now().day if not day else day
hour = datetime.now().hour if not hour else hour
timestruct = datetime(year, month, day, hour + hourdiff).isoformat() + '.000Z'
data['time_entry']['start'] = timestruct
data['time_entry']['duration'] = hourduration * 3600
data['time_entry']['pid'] = projectid
|
4 |
https://:@github.com/eEcoLiDAR/eEcoLiDAR.git
|
f0b2a0b7a5fdd41887ba40b7687c8161c0faba1e
|
@@ -28,6 +28,6 @@ class Test3FeatureExtractor(AbstractFeatureExtractor):
return ['test3_a']
def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume):
- t2a, t2c = utils.get_features(targetpc, targetindex, self.requires())
+ t2a, t2c = utils.get_features(targetpc, self.requires(), targetindex)
x, y, z = utils.get_point(targetpc, targetindex)
return t2c - t2a - z # z
|
laserchicken/test_feature_extractor/feature_test23.py
|
ArgSwap(idxs=1<->2 @(31,19)->(31,37))
|
class Test3FeatureExtractor(AbstractFeatureExtractor):
return ['test3_a']
def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume):
t2a, t2c = utils.get_features(targetpc, targetindex, self.requires())
x, y, z = utils.get_point(targetpc, targetindex)
return t2c - t2a - z # z
|
class Test3FeatureExtractor(AbstractFeatureExtractor):
return ['test3_a']
def extract(self, sourcepc, neighborhood, targetpc, targetindex, volume):
t2a, t2c = utils.get_features(targetpc, self.requires(), targetindex)
x, y, z = utils.get_point(targetpc, targetindex)
return t2c - t2a - z # z
|
5 |
https://:@github.com/eEcoLiDAR/eEcoLiDAR.git
|
f0b2a0b7a5fdd41887ba40b7687c8161c0faba1e
|
@@ -31,7 +31,7 @@ class TestUtils(unittest.TestCase):
pc[keys.point]["color"] = {"type": "double", "data": cols}
pc[keys.point]["flavor"] = {"type": "double", "data": flavs}
x, y, z = utils.get_point(pc, 2)
- c, f = utils.get_features(pc, 2, ("color", "flavor"))
+ c, f = utils.get_features(pc, ("color", "flavor"), 2)
self.assertEqual(c, 0.5 * (x + y))
self.assertEqual(f, 0.5 * (x - y))
|
laserchicken/test_utils.py
|
ArgSwap(idxs=1<->2 @(34,15)->(34,33))
|
class TestUtils(unittest.TestCase):
pc[keys.point]["color"] = {"type": "double", "data": cols}
pc[keys.point]["flavor"] = {"type": "double", "data": flavs}
x, y, z = utils.get_point(pc, 2)
c, f = utils.get_features(pc, 2, ("color", "flavor"))
self.assertEqual(c, 0.5 * (x + y))
self.assertEqual(f, 0.5 * (x - y))
|
class TestUtils(unittest.TestCase):
pc[keys.point]["color"] = {"type": "double", "data": cols}
pc[keys.point]["flavor"] = {"type": "double", "data": flavs}
x, y, z = utils.get_point(pc, 2)
c, f = utils.get_features(pc, ("color", "flavor"), 2)
self.assertEqual(c, 0.5 * (x + y))
self.assertEqual(f, 0.5 * (x - y))
|
6 |
https://:@github.com/eEcoLiDAR/eEcoLiDAR.git
|
502a365efda1393b130281803702d01f7e2d1dcd
|
@@ -29,7 +29,7 @@ class TestExtractEigenValues(unittest.TestCase):
["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
- target_point_cloud[keys.provenance][0]["module"])
+ target_point_cloud[keys.provenance][1]["module"])
@staticmethod
def test_eigenvalues_of_too_few_points_results_in_0():
|
laserchicken/feature_extractor/test_eigenvals_feature_extractor.py
|
ReplaceText(target='1' @(32,61)->(32,62))
|
class TestExtractEigenValues(unittest.TestCase):
["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
target_point_cloud[keys.provenance][0]["module"])
@staticmethod
def test_eigenvalues_of_too_few_points_results_in_0():
|
class TestExtractEigenValues(unittest.TestCase):
["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
target_point_cloud[keys.provenance][1]["module"])
@staticmethod
def test_eigenvalues_of_too_few_points_results_in_0():
|
7 |
https://:@github.com/eEcoLiDAR/eEcoLiDAR.git
|
eb7b021147a60b57e5dec536bd6f118c213f0952
|
@@ -29,7 +29,7 @@ class TestExtractEigenValues(unittest.TestCase):
["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
- target_point_cloud[keys.provenance][1]["module"])
+ target_point_cloud[keys.provenance][-1]["module"])
@staticmethod
def test_eigenvalues_of_too_few_points_results_in_0():
|
laserchicken/feature_extractor/test_eigenvals_feature_extractor.py
|
ReplaceText(target='-1' @(32,61)->(32,62))
|
class TestExtractEigenValues(unittest.TestCase):
["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
target_point_cloud[keys.provenance][1]["module"])
@staticmethod
def test_eigenvalues_of_too_few_points_results_in_0():
|
class TestExtractEigenValues(unittest.TestCase):
["eigenv_1", "eigenv_2", "eigenv_3"], InfiniteCylinder(5))
self.assertEqual("laserchicken.feature_extractor.eigenvals_feature_extractor",
target_point_cloud[keys.provenance][-1]["module"])
@staticmethod
def test_eigenvalues_of_too_few_points_results_in_0():
|
8 |
https://:@github.com/VinF/deer.git
|
66ea41db02c3361f18874dea7fd97720b1b06590
|
@@ -441,7 +441,7 @@ class CircularBuffer(object):
if end == sys.maxsize:
return self._data[self._lb+start:self._ub]
- elif self._lb + end >= self._ub:
+ elif self._lb + end > self._ub:
raise IndexError()
else:
return self._data[self._lb+start:self._lb+end]
|
General_Deep_Q_RL/agent.py
|
ReplaceText(target='>' @(444,28)->(444,30))
|
class CircularBuffer(object):
if end == sys.maxsize:
return self._data[self._lb+start:self._ub]
elif self._lb + end >= self._ub:
raise IndexError()
else:
return self._data[self._lb+start:self._lb+end]
|
class CircularBuffer(object):
if end == sys.maxsize:
return self._data[self._lb+start:self._ub]
elif self._lb + end > self._ub:
raise IndexError()
else:
return self._data[self._lb+start:self._lb+end]
|
9 |
https://:@github.com/VinF/deer.git
|
fd939e272d5441d48fd7d30bf24312c0f6bc8aaa
|
@@ -176,7 +176,7 @@ class MyEnv(Environment):
# Lack of energy
if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery):
# If enough energy in the battery, use it
- self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size*self.battery_eta
+ self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size/self.battery_eta
else:
# Otherwise: use what is left and then penalty
reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh
|
General_Deep_Q_RL/environments/MG_two_storages_env.py
|
ReplaceText(target='/' @(179,126)->(179,127))
|
class MyEnv(Environment):
# Lack of energy
if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery):
# If enough energy in the battery, use it
self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size*self.battery_eta
else:
# Otherwise: use what is left and then penalty
reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh
|
class MyEnv(Environment):
# Lack of energy
if (self._lastPonctualObservation[0]*self.battery_size>Energy_needed_from_battery):
# If enough energy in the battery, use it
self._lastPonctualObservation[0]=self._lastPonctualObservation[0]-Energy_needed_from_battery/self.battery_size/self.battery_eta
else:
# Otherwise: use what is left and then penalty
reward-=(Energy_needed_from_battery-self._lastPonctualObservation[0]*self.battery_size)*2 #2euro/kWh
|
10 |
https://:@github.com/piccolbo/altair_recipes.git
|
4992dd864a317eaad641d0408f003c429ed24af6
|
@@ -6,7 +6,7 @@ from vega_datasets import data
@viz_reg_test
def test_boxplot_melted():
- return ar.boxplot(data.iris(), "species", "petalLength")
+ return ar.boxplot(data.iris(), "petalLength", "species")
@viz_reg_test
|
tests/test_boxplot.py
|
ArgSwap(idxs=1<->2 @(9,11)->(9,21))
|
from vega_datasets import data
@viz_reg_test
def test_boxplot_melted():
return ar.boxplot(data.iris(), "species", "petalLength")
@viz_reg_test
|
from vega_datasets import data
@viz_reg_test
def test_boxplot_melted():
return ar.boxplot(data.iris(), "petalLength", "species")
@viz_reg_test
|
11 |
https://:@github.com/tailhook/zorro.git
|
dcbc37d47fe2a8de029f5a2f3ae13adf52e7aace
|
@@ -121,7 +121,7 @@ class RequestChannel(channel.PipelinedReqChannel):
clen = int(headers.get('Content-Length', '0'))
if clen < 0:
raise EOFError("Wrong content length")
- while pos[0] + clen < len(buf):
+ while pos[0] + clen > len(buf):
readmore()
return status, headers, buf[pos[0]:pos[0]+clen]
|
zorro/http.py
|
ReplaceText(target='>' @(124,32)->(124,33))
|
class RequestChannel(channel.PipelinedReqChannel):
clen = int(headers.get('Content-Length', '0'))
if clen < 0:
raise EOFError("Wrong content length")
while pos[0] + clen < len(buf):
readmore()
return status, headers, buf[pos[0]:pos[0]+clen]
|
class RequestChannel(channel.PipelinedReqChannel):
clen = int(headers.get('Content-Length', '0'))
if clen < 0:
raise EOFError("Wrong content length")
while pos[0] + clen > len(buf):
readmore()
return status, headers, buf[pos[0]:pos[0]+clen]
|
12 |
https://:@gitlab.com/eavise/brambox.git
|
f1faeed0b52d6f1c9c9ba6da818c1656f841622c
|
@@ -73,7 +73,7 @@ def test_multiclass(parser, df_anno_simple):
parser = parser()
with pytest.raises(ValueError) as errinfo:
- bb.io.save(parser, df_anno_simple, 'path.txt')
+ bb.io.save(df_anno_simple, parser, 'path.txt')
assert 'single-class problems' in str(errinfo.value)
|
test/io/parser/test_anno_cvc.py
|
ArgSwap(idxs=0<->1 @(76,8)->(76,18))
|
def test_multiclass(parser, df_anno_simple):
parser = parser()
with pytest.raises(ValueError) as errinfo:
bb.io.save(parser, df_anno_simple, 'path.txt')
assert 'single-class problems' in str(errinfo.value)
|
def test_multiclass(parser, df_anno_simple):
parser = parser()
with pytest.raises(ValueError) as errinfo:
bb.io.save(df_anno_simple, parser, 'path.txt')
assert 'single-class problems' in str(errinfo.value)
|
13 |
https://:@github.com/uber/h3-py.git
|
359924df907144c85ec323ae2804e2c0d173dfc5
|
@@ -631,7 +631,7 @@ def hex_ranges(h3_address_list, ring_size):
(1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1
# hexRanges doesn't return distance array
hex_range_list[ring_index].add(
- h3_to_string(krings[i * num_hexagons + j]))
+ h3_to_string(krings[i * array_len + j]))
return out
|
h3/h3.py
|
ReplaceText(target='array_len' @(634,40)->(634,52))
|
def hex_ranges(h3_address_list, ring_size):
(1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1
# hexRanges doesn't return distance array
hex_range_list[ring_index].add(
h3_to_string(krings[i * num_hexagons + j]))
return out
|
def hex_ranges(h3_address_list, ring_size):
(1 + math.sqrt(1 + 8 * math.ceil(j / 6.0))) / 2)) - 1
# hexRanges doesn't return distance array
hex_range_list[ring_index].add(
h3_to_string(krings[i * array_len + j]))
return out
|
14 |
https://:@github.com/polysquare/polysquare-generic-file-linter.git
|
e9dbb28ea30955ab59d1339c04f0710b24ba53aa
|
@@ -388,7 +388,7 @@ def _maybe_log_technical_terms(global_options, tool_options):
terms = set(terms_file.read().splitlines()) # suppress(PYC70)
terms_file.seek(0) # suppress(PYC70)
terms_file.truncate(0) # suppress(PYC70)
- tech_terms = freduce(lambda x, y: x + y,
+ tech_terms = freduce(lambda x, y: x | y,
_drain(log_technical_terms_to_queue))
terms_file.write("\n".join(list(terms | # suppress(PYC70)
set(tech_terms))))
|
polysquarelinter/linter.py
|
ReplaceText(target='|' @(391,48)->(391,49))
|
def _maybe_log_technical_terms(global_options, tool_options):
terms = set(terms_file.read().splitlines()) # suppress(PYC70)
terms_file.seek(0) # suppress(PYC70)
terms_file.truncate(0) # suppress(PYC70)
tech_terms = freduce(lambda x, y: x + y,
_drain(log_technical_terms_to_queue))
terms_file.write("\n".join(list(terms | # suppress(PYC70)
set(tech_terms))))
|
def _maybe_log_technical_terms(global_options, tool_options):
terms = set(terms_file.read().splitlines()) # suppress(PYC70)
terms_file.seek(0) # suppress(PYC70)
terms_file.truncate(0) # suppress(PYC70)
tech_terms = freduce(lambda x, y: x | y,
_drain(log_technical_terms_to_queue))
terms_file.write("\n".join(list(terms | # suppress(PYC70)
set(tech_terms))))
|
15 |
https://:@github.com/johntruckenbrodt/spatialist.git
|
c9d552e64cd47b30156b288e035d17debea48b45
|
@@ -300,7 +300,7 @@ def centerdist(obj1, obj2):
def intersect(obj1, obj2):
- if not (isinstance(obj1, Vector) or isinstance(obj2, Vector)):
+ if not (isinstance(obj1, Vector) and isinstance(obj2, Vector)):
raise IOError('object must be of type Vector')
obj1.reproject(obj2.srs)
|
pyroSAR/spatial/vector.py
|
ReplaceText(target='and' @(303,37)->(303,39))
|
def centerdist(obj1, obj2):
def intersect(obj1, obj2):
if not (isinstance(obj1, Vector) or isinstance(obj2, Vector)):
raise IOError('object must be of type Vector')
obj1.reproject(obj2.srs)
|
def centerdist(obj1, obj2):
def intersect(obj1, obj2):
if not (isinstance(obj1, Vector) and isinstance(obj2, Vector)):
raise IOError('object must be of type Vector')
obj1.reproject(obj2.srs)
|
16 |
https://:@github.com/Toblerity/Shapely.git
|
9f1b78e6fd5f4286f210b54827bdd26661f0ee7a
|
@@ -40,7 +40,7 @@ if __name__ == '__main__':
]
if pattern:
- tests = [f for f in docfiles if f.find(pattern) >= 0]
+ tests = [f for f in docfiles if f.find(pattern) == 0]
else:
tests = docfiles
|
tests/runalldoctests.py
|
ReplaceText(target='==' @(43,56)->(43,58))
|
if __name__ == '__main__':
]
if pattern:
tests = [f for f in docfiles if f.find(pattern) >= 0]
else:
tests = docfiles
|
if __name__ == '__main__':
]
if pattern:
tests = [f for f in docfiles if f.find(pattern) == 0]
else:
tests = docfiles
|
17 |
https://:@github.com/Toblerity/Shapely.git
|
d6fc8cc0e0d50b23ba0d7ca6195bc530b2f8d1b9
|
@@ -11,7 +11,7 @@ def halton(base):
i = index
while i > 0:
result += f * (i % base)
- i = i/base
+ i = i//base
f = f/base
return result
i = 1
|
shapely/tests/test_unary_union.py
|
ReplaceText(target='//' @(14,17)->(14,18))
|
def halton(base):
i = index
while i > 0:
result += f * (i % base)
i = i/base
f = f/base
return result
i = 1
|
def halton(base):
i = index
while i > 0:
result += f * (i % base)
i = i//base
f = f/base
return result
i = 1
|
18 |
https://:@github.com/Toblerity/Shapely.git
|
5f0db7fdc052beeeef36aa1251f19175d0abeedb
|
@@ -36,7 +36,7 @@ if version is None:
# Handle UTF-8 encoding of certain text files.
open_kwds = {}
-if sys.version_info > (3,):
+if sys.version_info >= (3,):
open_kwds['encoding'] = 'utf-8'
with open('VERSION.txt', 'w', **open_kwds) as fp:
|
setup.py
|
ReplaceText(target='>=' @(39,20)->(39,21))
|
if version is None:
# Handle UTF-8 encoding of certain text files.
open_kwds = {}
if sys.version_info > (3,):
open_kwds['encoding'] = 'utf-8'
with open('VERSION.txt', 'w', **open_kwds) as fp:
|
if version is None:
# Handle UTF-8 encoding of certain text files.
open_kwds = {}
if sys.version_info >= (3,):
open_kwds['encoding'] = 'utf-8'
with open('VERSION.txt', 'w', **open_kwds) as fp:
|
19 |
https://:@github.com/svetlyak40wt/django-tagging-ng.git
|
0293b78ee0274d123eb70c1f8c5c01a5b36e2b40
|
@@ -163,7 +163,7 @@ class TaggedItemManager(models.Manager):
associated with a given Tag or list of Tags.
"""
tags = get_tag_list(tags)
- if len(tags) == 0:
+ if len(tags) == 1:
tag = tags[0] # Optimisation for single tag
else:
return self.get_intersection_by_model(Model, tags)
|
models.py
|
ReplaceText(target='1' @(166,24)->(166,25))
|
class TaggedItemManager(models.Manager):
associated with a given Tag or list of Tags.
"""
tags = get_tag_list(tags)
if len(tags) == 0:
tag = tags[0] # Optimisation for single tag
else:
return self.get_intersection_by_model(Model, tags)
|
class TaggedItemManager(models.Manager):
associated with a given Tag or list of Tags.
"""
tags = get_tag_list(tags)
if len(tags) == 1:
tag = tags[0] # Optimisation for single tag
else:
return self.get_intersection_by_model(Model, tags)
|
20 |
https://:@github.com/svetlyak40wt/django-tagging-ng.git
|
3285d40e4c1de628886a7fa45a6d4cf6ed4cd7e7
|
@@ -163,7 +163,7 @@ class TaggedItemManager(models.Manager):
associated with a given Tag or list of Tags.
"""
tags = get_tag_list(tags)
- if len(tags) == 0:
+ if len(tags) == 1:
tag = tags[0] # Optimisation for single tag
else:
return self.get_intersection_by_model(Model, tags)
|
models.py
|
ReplaceText(target='1' @(166,24)->(166,25))
|
class TaggedItemManager(models.Manager):
associated with a given Tag or list of Tags.
"""
tags = get_tag_list(tags)
if len(tags) == 0:
tag = tags[0] # Optimisation for single tag
else:
return self.get_intersection_by_model(Model, tags)
|
class TaggedItemManager(models.Manager):
associated with a given Tag or list of Tags.
"""
tags = get_tag_list(tags)
if len(tags) == 1:
tag = tags[0] # Optimisation for single tag
else:
return self.get_intersection_by_model(Model, tags)
|
21 |
https://:@github.com/zzzsochi/yadm.git
|
03efd06fe95c7d84264455c4fac5c8cbb17eb4dd
|
@@ -316,7 +316,7 @@ class QuerySet(BaseQuerySet):
if data is None:
if exc is not None:
- raise exc(criteria)
+ raise exc(qs)
else:
return None
|
yadm/queryset.py
|
ReplaceText(target='qs' @(319,26)->(319,34))
|
class QuerySet(BaseQuerySet):
if data is None:
if exc is not None:
raise exc(criteria)
else:
return None
|
class QuerySet(BaseQuerySet):
if data is None:
if exc is not None:
raise exc(qs)
else:
return None
|
22 |
https://:@github.com/instacart/lore.git
|
1c1e0efdac6b27dc111eaa93bb99317c59aaffaf
|
@@ -196,7 +196,7 @@ class Base(object):
def upload(self):
self.fitting = 0
self.save()
- lore.io.upload(self.remote_model_path(), self.model_path())
+ lore.io.upload(self.model_path(), self.remote_model_path())
@classmethod
def download(cls, fitting=0):
|
lore/models/base.py
|
ArgSwap(idxs=0<->1 @(199,8)->(199,22))
|
class Base(object):
def upload(self):
self.fitting = 0
self.save()
lore.io.upload(self.remote_model_path(), self.model_path())
@classmethod
def download(cls, fitting=0):
|
class Base(object):
def upload(self):
self.fitting = 0
self.save()
lore.io.upload(self.model_path(), self.remote_model_path())
@classmethod
def download(cls, fitting=0):
|
23 |
https://:@github.com/instacart/lore.git
|
1c1e0efdac6b27dc111eaa93bb99317c59aaffaf
|
@@ -78,7 +78,7 @@ class Base(lore.models.base.Base):
def upload(self):
super(Base, self).upload()
- lore.io.upload(self.remote_weights_path(), self.weights_path())
+ lore.io.upload(self.weights_path(), self.remote_weights_path())
@classmethod
def download(cls, fitting=0):
|
lore/models/keras.py
|
ArgSwap(idxs=0<->1 @(81,8)->(81,22))
|
class Base(lore.models.base.Base):
def upload(self):
super(Base, self).upload()
lore.io.upload(self.remote_weights_path(), self.weights_path())
@classmethod
def download(cls, fitting=0):
|
class Base(lore.models.base.Base):
def upload(self):
super(Base, self).upload()
lore.io.upload(self.weights_path(), self.remote_weights_path())
@classmethod
def download(cls, fitting=0):
|
24 |
https://:@github.com/instacart/lore.git
|
f4ded2b3199d1c33ba6c9c79cd66b25d43c83c81
|
@@ -464,7 +464,7 @@ class Base(BaseEstimator):
result = self.keras.predict(dataframe, batch_size=self.batch_size)
if self.towers > 1:
- result = numpy.mean(result, axis=0).squeeze(axis=0)
+ result = numpy.mean(result, axis=0).squeeze(axis=1)
return result
|
lore/estimators/keras.py
|
ReplaceText(target='1' @(467,61)->(467,62))
|
class Base(BaseEstimator):
result = self.keras.predict(dataframe, batch_size=self.batch_size)
if self.towers > 1:
result = numpy.mean(result, axis=0).squeeze(axis=0)
return result
|
class Base(BaseEstimator):
result = self.keras.predict(dataframe, batch_size=self.batch_size)
if self.towers > 1:
result = numpy.mean(result, axis=0).squeeze(axis=1)
return result
|
25 |
https://:@github.com/baliga-lab/cmonkey2.git
|
3201a0e97688724450196da8fef96d283c855b3f
|
@@ -273,7 +273,7 @@ class ClusterMembership:
#logging.warn("cluster %s already associated with %s",
# str(cluster), str(column))
pass
- if columns not in columns:
+ if column not in columns:
columns.append(column)
def remove_cluster_from_column(self, column, cluster):
|
cmonkey/membership.py
|
ReplaceText(target='column' @(276,11)->(276,18))
|
class ClusterMembership:
#logging.warn("cluster %s already associated with %s",
# str(cluster), str(column))
pass
if columns not in columns:
columns.append(column)
def remove_cluster_from_column(self, column, cluster):
|
class ClusterMembership:
#logging.warn("cluster %s already associated with %s",
# str(cluster), str(column))
pass
if column not in columns:
columns.append(column)
def remove_cluster_from_column(self, column, cluster):
|
26 |
https://:@github.com/baliga-lab/cmonkey2.git
|
48d14ac785b1013354a55a37239c66433fbf19eb
|
@@ -460,7 +460,7 @@ class ClusterMembership:
max_score = sys.float_info.min
for row in range(sm.num_rows()):
if sm_values[row][0] > max_score:
- max_score = sm[row][0]
+ max_score = sm_values[row][0]
max_row = row
return sm.row_names[max_row]
|
cmonkey/membership.py
|
ReplaceText(target='sm_values' @(463,32)->(463,34))
|
class ClusterMembership:
max_score = sys.float_info.min
for row in range(sm.num_rows()):
if sm_values[row][0] > max_score:
max_score = sm[row][0]
max_row = row
return sm.row_names[max_row]
|
class ClusterMembership:
max_score = sys.float_info.min
for row in range(sm.num_rows()):
if sm_values[row][0] > max_score:
max_score = sm_values[row][0]
max_row = row
return sm.row_names[max_row]
|
27 |
https://:@github.com/baliga-lab/cmonkey2.git
|
311548905a9def1cbdf63d2e0fd8a17346564742
|
@@ -17,7 +17,7 @@ class MicrobesOnlineTest(unittest.TestCase): # pylint: disable-msg=R0904
"""test fixture"""
if not os.path.exists('testcache'):
os.mkdir('testcache')
- self.service = mo.MicrobesOnline(mo.MICROBES_ONLINE_BASE_URL, 'testcache')
+ self.service = mo.MicrobesOnline('testcache', mo.MICROBES_ONLINE_BASE_URL)
def tearDown(self): # pylint: disable-msg=C0103
"""test cleanup"""
|
test/microbes_online_test.py
|
ArgSwap(idxs=0<->1 @(20,23)->(20,40))
|
class MicrobesOnlineTest(unittest.TestCase): # pylint: disable-msg=R0904
"""test fixture"""
if not os.path.exists('testcache'):
os.mkdir('testcache')
self.service = mo.MicrobesOnline(mo.MICROBES_ONLINE_BASE_URL, 'testcache')
def tearDown(self): # pylint: disable-msg=C0103
"""test cleanup"""
|
class MicrobesOnlineTest(unittest.TestCase): # pylint: disable-msg=R0904
"""test fixture"""
if not os.path.exists('testcache'):
os.mkdir('testcache')
self.service = mo.MicrobesOnline('testcache', mo.MICROBES_ONLINE_BASE_URL)
def tearDown(self): # pylint: disable-msg=C0103
"""test cleanup"""
|
28 |
https://:@github.com/muammar/ml4chem.git
|
dbb7de0379cb8881538d211899e4bec8794f16e3
|
@@ -344,7 +344,7 @@ def train(inputs, targets, model=None, data=None, optimizer=None, lr=None,
logger.info('Training finished in {} hours {} minutes {:.2f} seconds.'
.format(h, m, s))
logger.info('outputs')
- logger.info(outputs)
+ logger.info(outputs_)
logger.info('targets')
logger.info(targets)
|
mlchem/models/neuralnetwork.py
|
ReplaceText(target='outputs_' @(347,16)->(347,23))
|
def train(inputs, targets, model=None, data=None, optimizer=None, lr=None,
logger.info('Training finished in {} hours {} minutes {:.2f} seconds.'
.format(h, m, s))
logger.info('outputs')
logger.info(outputs)
logger.info('targets')
logger.info(targets)
|
def train(inputs, targets, model=None, data=None, optimizer=None, lr=None,
logger.info('Training finished in {} hours {} minutes {:.2f} seconds.'
.format(h, m, s))
logger.info('outputs')
logger.info(outputs_)
logger.info('targets')
logger.info(targets)
|
29 |
https://:@github.com/chris7/pyquant.git
|
3730fbdb9789a59a65d38f5a2ae21c645086096f
|
@@ -624,7 +624,7 @@ def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si
best_fit = np.array(best_fit)
peak_func = bigauss_ndim if bigauss_fit else gauss_ndim
# Get rid of peaks with low r^2
- if micro and r2_cutoff is not None:
+ if not micro and r2_cutoff is not None:
final_fit = np.array([])
for peak_index in xrange(0, len(best_fit), step_size):
|
pyquant/peaks.py
|
ReplaceText(target='not ' @(627,7)->(627,7))
|
def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si
best_fit = np.array(best_fit)
peak_func = bigauss_ndim if bigauss_fit else gauss_ndim
# Get rid of peaks with low r^2
if micro and r2_cutoff is not None:
final_fit = np.array([])
for peak_index in xrange(0, len(best_fit), step_size):
|
def findAllPeaks(xdata, ydata_original, min_dist=0, method=None, local_filter_si
best_fit = np.array(best_fit)
peak_func = bigauss_ndim if bigauss_fit else gauss_ndim
# Get rid of peaks with low r^2
if not micro and r2_cutoff is not None:
final_fit = np.array([])
for peak_index in xrange(0, len(best_fit), step_size):
|
30 |
https://:@github.com/chris7/pyquant.git
|
4a0755563e0a36fecf1f4393554cfaf4c1615c2c
|
@@ -615,7 +615,7 @@ def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
# By default, cross points returns the left side
for i in xrange(len(cross_points)):
index = cross_points[i]
- if index < len(cross_points):
+ if index < len(ydata):
if ydata[index] < ydata[index+1]:
cross_points[i] = index+1
|
pyquant/utils.py
|
ReplaceText(target='ydata' @(618,23)->(618,35))
|
def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
# By default, cross points returns the left side
for i in xrange(len(cross_points)):
index = cross_points[i]
if index < len(cross_points):
if ydata[index] < ydata[index+1]:
cross_points[i] = index+1
|
def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
# By default, cross points returns the left side
for i in xrange(len(cross_points)):
index = cross_points[i]
if index < len(ydata):
if ydata[index] < ydata[index+1]:
cross_points[i] = index+1
|
31 |
https://:@github.com/chris7/pyquant.git
|
cd61286935d8ca64eb539851e39a98a0655ff400
|
@@ -611,7 +611,7 @@ def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
ydata = np.abs(ydata_peaks)
if min_peak_width is None:
- max_peak_width = int(len(ydata) / 2)
+ min_peak_width = int(len(ydata) / 2)
if min_peak_width > 5:
min_peak_width = 5
|
pyquant/utils.py
|
ReplaceText(target='min_peak_width' @(614,8)->(614,22))
|
def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
ydata = np.abs(ydata_peaks)
if min_peak_width is None:
max_peak_width = int(len(ydata) / 2)
if min_peak_width > 5:
min_peak_width = 5
|
def find_peaks_derivative(xdata, ydata, ydata_peaks=None, min_slope=None, rel_pe
ydata = np.abs(ydata_peaks)
if min_peak_width is None:
min_peak_width = int(len(ydata) / 2)
if min_peak_width > 5:
min_peak_width = 5
|
32 |
https://:@github.com/ssec/sift.git
|
24ce052cd497c42c917de06f0c89a7c5be13ab50
|
@@ -599,7 +599,7 @@ class ProbeGraphDisplay (object) :
x_point = self.workspace.get_content_point(x_uuid, point_xy)
format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point)
y_point = self.workspace.get_content_point(y_uuid, point_xy)
- format_str, unit_str, y_point = self.document.convert_units(x_uuid, y_point)
+ format_str, unit_str, y_point = self.document.convert_units(y_uuid, y_point)
else:
x_point = None
y_point = None
|
py/cspov/view/ProbeGraphs.py
|
ReplaceText(target='y_uuid' @(602,76)->(602,82))
|
class ProbeGraphDisplay (object) :
x_point = self.workspace.get_content_point(x_uuid, point_xy)
format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point)
y_point = self.workspace.get_content_point(y_uuid, point_xy)
format_str, unit_str, y_point = self.document.convert_units(x_uuid, y_point)
else:
x_point = None
y_point = None
|
class ProbeGraphDisplay (object) :
x_point = self.workspace.get_content_point(x_uuid, point_xy)
format_str, unit_str, x_point = self.document.convert_units(x_uuid, x_point)
y_point = self.workspace.get_content_point(y_uuid, point_xy)
format_str, unit_str, y_point = self.document.convert_units(y_uuid, y_point)
else:
x_point = None
y_point = None
|
33 |
https://:@github.com/threatwatch/twigs.git
|
7ee5d95178a459a8c2e8ff7855e3156e620c395c
|
@@ -85,7 +85,7 @@ def parse_inventory(email,data,params):
asset_map = {}
asset_map['owner'] = email
asset_map['host'] = host
- asset_map['id'] = host
+ asset_map['id'] = vmuuid
asset_map['name'] = host
asset_map['tags'] = [ ]
asset_map['patch_tracker'] = { } # To help remove duplicate patches
|
twigs/azure.py
|
ReplaceText(target='vmuuid' @(88,30)->(88,34))
|
def parse_inventory(email,data,params):
asset_map = {}
asset_map['owner'] = email
asset_map['host'] = host
asset_map['id'] = host
asset_map['name'] = host
asset_map['tags'] = [ ]
asset_map['patch_tracker'] = { } # To help remove duplicate patches
|
def parse_inventory(email,data,params):
asset_map = {}
asset_map['owner'] = email
asset_map['host'] = host
asset_map['id'] = vmuuid
asset_map['name'] = host
asset_map['tags'] = [ ]
asset_map['patch_tracker'] = { } # To help remove duplicate patches
|
34 |
https://:@github.com/Keeper-Security/Commander.git
|
9bd55c8dd48ab62759bbbb8dfcd38ab364cec2dc
|
@@ -1550,7 +1550,7 @@ def prepare_record(params, record):
else:
if params.debug: print('Generated record key')
unencrypted_key = os.urandom(32)
- record_object['record_key'] = encrypt_aes(params.data_key, unencrypted_key)
+ record_object['record_key'] = encrypt_aes(unencrypted_key, params.data_key)
record_object['revision'] = 0
data['title'] = record.title
|
keepercommander/api.py
|
ArgSwap(idxs=0<->1 @(1553,38)->(1553,49))
|
def prepare_record(params, record):
else:
if params.debug: print('Generated record key')
unencrypted_key = os.urandom(32)
record_object['record_key'] = encrypt_aes(params.data_key, unencrypted_key)
record_object['revision'] = 0
data['title'] = record.title
|
def prepare_record(params, record):
else:
if params.debug: print('Generated record key')
unencrypted_key = os.urandom(32)
record_object['record_key'] = encrypt_aes(unencrypted_key, params.data_key)
record_object['revision'] = 0
data['title'] = record.title
|
35 |
https://:@github.com/hpapaxen/rope.git
|
27d5085b30e89095e88339c96d9940e338482106
|
@@ -69,7 +69,7 @@ class JobSet(object):
def get_percent_done(self):
if self.count is not None and self.count > 0:
- percent = self.done * 100 / self.count
+ percent = self.done * 100 // self.count
return min(percent, 100)
def get_name(self):
|
rope/base/taskhandle.py
|
ReplaceText(target='//' @(72,38)->(72,39))
|
class JobSet(object):
def get_percent_done(self):
if self.count is not None and self.count > 0:
percent = self.done * 100 / self.count
return min(percent, 100)
def get_name(self):
|
class JobSet(object):
def get_percent_done(self):
if self.count is not None and self.count > 0:
percent = self.done * 100 // self.count
return min(percent, 100)
def get_name(self):
|
36 |
https://:@github.com/hpapaxen/rope.git
|
27d5085b30e89095e88339c96d9940e338482106
|
@@ -524,7 +524,7 @@ class ProgressBar(object):
self.text['text'] = text
def _draw_shape(self):
- width = int(self.canvas['width']) * self.percent / 100
+ width = int(self.canvas['width']) * self.percent // 100
self.canvas.create_rectangle(0, 0, width, self.canvas['height'],
fill=self.color)
total_width = self.canvas['width']
|
rope/ui/uihelpers.py
|
ReplaceText(target='//' @(527,57)->(527,58))
|
class ProgressBar(object):
self.text['text'] = text
def _draw_shape(self):
width = int(self.canvas['width']) * self.percent / 100
self.canvas.create_rectangle(0, 0, width, self.canvas['height'],
fill=self.color)
total_width = self.canvas['width']
|
class ProgressBar(object):
self.text['text'] = text
def _draw_shape(self):
width = int(self.canvas['width']) * self.percent // 100
self.canvas.create_rectangle(0, 0, width, self.canvas['height'],
fill=self.color)
total_width = self.canvas['width']
|
37 |
https://:@github.com/hpapaxen/rope.git
|
2720419618aceab7fba51aaa4d66f7eae005b22d
|
@@ -129,7 +129,7 @@ class SimilarFinderTest(unittest.TestCase):
source = 'x.a = 1\n'
finder = similarfinder.SimilarFinder(source)
result = list(finder.get_matches('${a} = 1'))
- self.assertEquals(1, len(result))
+ self.assertEquals(0, len(result))
def test_functions_not_matching_when_only_first_parameters(self):
source = 'f(1, 2)\n'
|
ropetest/refactor/similarfindertest.py
|
ReplaceText(target='0' @(132,26)->(132,27))
|
class SimilarFinderTest(unittest.TestCase):
source = 'x.a = 1\n'
finder = similarfinder.SimilarFinder(source)
result = list(finder.get_matches('${a} = 1'))
self.assertEquals(1, len(result))
def test_functions_not_matching_when_only_first_parameters(self):
source = 'f(1, 2)\n'
|
class SimilarFinderTest(unittest.TestCase):
source = 'x.a = 1\n'
finder = similarfinder.SimilarFinder(source)
result = list(finder.get_matches('${a} = 1'))
self.assertEquals(0, len(result))
def test_functions_not_matching_when_only_first_parameters(self):
source = 'f(1, 2)\n'
|
38 |
https://:@github.com/hpapaxen/rope.git
|
0eb3cb58493cdaea83a4e24d47b5bd4dbd19f963
|
@@ -32,7 +32,7 @@ class BuiltinModule(pyobjects.AbstractModule):
result.update(self.initial)
for modname in self.submodules:
name = modname.split('.')[-1]
- result[name] = BuiltinModule(name, self.submodules)
+ result[name] = BuiltinModule(modname, self.submodules)
return result
@property
|
rope/base/builtins.py
|
ReplaceText(target='modname' @(35,41)->(35,45))
|
class BuiltinModule(pyobjects.AbstractModule):
result.update(self.initial)
for modname in self.submodules:
name = modname.split('.')[-1]
result[name] = BuiltinModule(name, self.submodules)
return result
@property
|
class BuiltinModule(pyobjects.AbstractModule):
result.update(self.initial)
for modname in self.submodules:
name = modname.split('.')[-1]
result[name] = BuiltinModule(modname, self.submodules)
return result
@property
|
39 |
https://:@github.com/hpapaxen/rope.git
|
528744bb4bc1b8076680f7c2c1bfac508ddca4f9
|
@@ -37,7 +37,7 @@ def relative(root, path):
if os.path.samefile(root, path):
return '/'.join(reversed(rel))
parent = os.path.dirname(path)
- if not path or parent == path:
+ if not parent or parent == path:
break
rel.append(os.path.basename(path))
path = parent
|
rope/base/libutils.py
|
ReplaceText(target='parent' @(40,15)->(40,19))
|
def relative(root, path):
if os.path.samefile(root, path):
return '/'.join(reversed(rel))
parent = os.path.dirname(path)
if not path or parent == path:
break
rel.append(os.path.basename(path))
path = parent
|
def relative(root, path):
if os.path.samefile(root, path):
return '/'.join(reversed(rel))
parent = os.path.dirname(path)
if not parent or parent == path:
break
rel.append(os.path.basename(path))
path = parent
|
40 |
https://:@github.com/benjamincrom/baseball.git
|
6ef29729ad07458aebe709b4e42f56ecd3761ec4
|
@@ -111,7 +111,7 @@ def write_game_svg_and_html(game_id, game, output_path):
html_filename = game_id + '.html'
svg_text = game.get_svg_str()
- html_text = HTML_WRAPPER.format(title=game_id, filename=html_filename)
+ html_text = HTML_WRAPPER.format(title=game_id, filename=svg_filename)
output_svg_path = join(output_path, svg_filename)
output_html_path = join(output_path, html_filename)
|
fetch_game.py
|
ReplaceText(target='svg_filename' @(114,60)->(114,73))
|
def write_game_svg_and_html(game_id, game, output_path):
html_filename = game_id + '.html'
svg_text = game.get_svg_str()
html_text = HTML_WRAPPER.format(title=game_id, filename=html_filename)
output_svg_path = join(output_path, svg_filename)
output_html_path = join(output_path, html_filename)
|
def write_game_svg_and_html(game_id, game, output_path):
html_filename = game_id + '.html'
svg_text = game.get_svg_str()
html_text = HTML_WRAPPER.format(title=game_id, filename=svg_filename)
output_svg_path = join(output_path, svg_filename)
output_html_path = join(output_path, html_filename)
|
41 |
https://:@github.com/sporestack/bitcash.git
|
dbc65e1b47426e0e4d286db5b27216ec36cb32cf
|
@@ -16,7 +16,7 @@ def test_set_fee_cache_time():
def test_get_fee():
- assert get_fee(fast=True) != get_fee(fast=False)
+ assert get_fee(fast=True) >= get_fee(fast=False)
class TestFeeCache:
|
tests/network/test_fees.py
|
ReplaceText(target='>=' @(19,30)->(19,32))
|
def test_set_fee_cache_time():
def test_get_fee():
assert get_fee(fast=True) != get_fee(fast=False)
class TestFeeCache:
|
def test_set_fee_cache_time():
def test_get_fee():
assert get_fee(fast=True) >= get_fee(fast=False)
class TestFeeCache:
|
42 |
https://:@github.com/galaxy-genome-annotation/python-apollo.git
|
53e514b619844fa1f87179d738b9d29830027300
|
@@ -29,7 +29,7 @@ class ApolloTestCase(unittest.TestCase):
"""
org_info = wa.organisms.show_organism(org_id)
- if 'directory' in org_info:
+ if 'directory' not in org_info:
time.sleep(1)
org_info = wa.organisms.show_organism(org_id)
|
test/__init__.py
|
ReplaceText(target=' not in ' @(32,22)->(32,26))
|
class ApolloTestCase(unittest.TestCase):
"""
org_info = wa.organisms.show_organism(org_id)
if 'directory' in org_info:
time.sleep(1)
org_info = wa.organisms.show_organism(org_id)
|
class ApolloTestCase(unittest.TestCase):
"""
org_info = wa.organisms.show_organism(org_id)
if 'directory' not in org_info:
time.sleep(1)
org_info = wa.organisms.show_organism(org_id)
|
43 |
https://:@github.com/jakubplichta/grafana-dashboard-builder.git
|
3228e6950d65b9bd347cacb56a9e85ec410b14ce
|
@@ -35,7 +35,7 @@ class Context(object):
formatter = string.Formatter()
(result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand)
while result != to_expand:
- (result, to_expand) = (formatter.vformat(to_expand, (), self._context), result)
+ (result, to_expand) = (formatter.vformat(result, (), self._context), result)
return result
elif isinstance(to_expand, list):
return [self.expand_placeholders(value) for value in to_expand]
|
grafana_dashboards/context.py
|
ReplaceText(target='result' @(38,57)->(38,66))
|
class Context(object):
formatter = string.Formatter()
(result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand)
while result != to_expand:
(result, to_expand) = (formatter.vformat(to_expand, (), self._context), result)
return result
elif isinstance(to_expand, list):
return [self.expand_placeholders(value) for value in to_expand]
|
class Context(object):
formatter = string.Formatter()
(result, to_expand) = (formatter.vformat(to_expand, (), self._context), to_expand)
while result != to_expand:
(result, to_expand) = (formatter.vformat(result, (), self._context), result)
return result
elif isinstance(to_expand, list):
return [self.expand_placeholders(value) for value in to_expand]
|
44 |
https://:@github.com/Phylliade/ikpy.git
|
815dbff3a521532a7b792c309902ffea82abac85
|
@@ -13,7 +13,7 @@ class TestFK(unittest.TestCase):
one_move[5] = np.pi / 4
one_move[6] = -np.pi / 2
one_move[4] = -np.pi / 2
- self.test_pos = one_move
+ self.test_pos = all_zeros
def test_fk_creature(self):
|
tests/test_fk.py
|
ReplaceText(target='all_zeros' @(16,24)->(16,32))
|
class TestFK(unittest.TestCase):
one_move[5] = np.pi / 4
one_move[6] = -np.pi / 2
one_move[4] = -np.pi / 2
self.test_pos = one_move
def test_fk_creature(self):
|
class TestFK(unittest.TestCase):
one_move[5] = np.pi / 4
one_move[6] = -np.pi / 2
one_move[4] = -np.pi / 2
self.test_pos = all_zeros
def test_fk_creature(self):
|
45 |
https://:@github.com/tingbot/tingbot-python.git
|
5374186675f6809faf9ce953fc35c81217348753
|
@@ -72,7 +72,7 @@ class RunLoop(object):
while self.running:
if len(self.timers) > 0:
try:
- self._wait(self.timers[0].next_fire_time)
+ self._wait(self.timers[-1].next_fire_time)
except Exception as e:
self._error(e)
continue
|
tingbot/run_loop.py
|
ReplaceText(target='-1' @(75,43)->(75,44))
|
class RunLoop(object):
while self.running:
if len(self.timers) > 0:
try:
self._wait(self.timers[0].next_fire_time)
except Exception as e:
self._error(e)
continue
|
class RunLoop(object):
while self.running:
if len(self.timers) > 0:
try:
self._wait(self.timers[-1].next_fire_time)
except Exception as e:
self._error(e)
continue
|
46 |
https://:@github.com/nyoka-pmml/nyoka.git
|
8d5c0d31d0bf1e251abe686f06b614a16e5ffcfb
|
@@ -74,7 +74,7 @@ class TestMethods(unittest.TestCase):
self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance")
##3
- self.assertEqual(pmml_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors)
+ self.assertEqual(pipeline_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors)
def test_sklearn_03(self):
|
nyoka/tests/skl_to_pmml_UnitTest.py
|
ReplaceText(target='pipeline_obj' @(77,25)->(77,33))
|
class TestMethods(unittest.TestCase):
self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance")
##3
self.assertEqual(pmml_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors)
def test_sklearn_03(self):
|
class TestMethods(unittest.TestCase):
self.assertEqual(pmml_obj.NearestNeighborModel[0].ComparisonMeasure.kind, "distance")
##3
self.assertEqual(pipeline_obj.steps[-1][-1].n_neighbors, pmml_obj.NearestNeighborModel[0].numberOfNeighbors)
def test_sklearn_03(self):
|
47 |
https://:@github.com/iris-edu/pyweed.git
|
77d919acd8d54d4879d1a34598e9e04f16fdf708
|
@@ -97,7 +97,7 @@ class WaveformEntry(AttribDict):
self.error = None
- self.start_time = self.distances.arrival + self.config.offsets[0]
+ self.start_time = self.distances.arrival - self.config.offsets[0]
self.end_time = self.distances.arrival + self.config.offsets[1]
self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_')
|
pyweed/waveforms_handler.py
|
ReplaceText(target='-' @(100,49)->(100,50))
|
class WaveformEntry(AttribDict):
self.error = None
self.start_time = self.distances.arrival + self.config.offsets[0]
self.end_time = self.distances.arrival + self.config.offsets[1]
self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_')
|
class WaveformEntry(AttribDict):
self.error = None
self.start_time = self.distances.arrival - self.config.offsets[0]
self.end_time = self.distances.arrival + self.config.offsets[1]
self.start_string = UTCDateTime(self.start_time).format_iris_web_service().replace(':', '_')
|
48 |
https://:@github.com/anaxilaus/coindata.git
|
2e5067311c4eed50eed41c45f43ad63e8973e579
|
@@ -52,6 +52,6 @@ def dump_json(data, filepath):
try:
with open(filepath, 'w') as file:
- json.dump(data, filepath)
+ json.dump(data, file)
except TypeError as e:
print("Data isn't JSON compatible.\n", e)
|
coindata/utils.py
|
ReplaceText(target='file' @(55,28)->(55,36))
|
def dump_json(data, filepath):
try:
with open(filepath, 'w') as file:
json.dump(data, filepath)
except TypeError as e:
print("Data isn't JSON compatible.\n", e)
|
def dump_json(data, filepath):
try:
with open(filepath, 'w') as file:
json.dump(data, file)
except TypeError as e:
print("Data isn't JSON compatible.\n", e)
|
49 |
https://:@github.com/Pixelapse/pyglass.git
|
a31e95cbc259ce61f5851d6f0d769792aaa182fe
|
@@ -20,7 +20,7 @@ def preview(src_path):
preview_path = thumbnail_preview(src_path)
if preview_path:
- mimetype = magic.from_file(src_path, mime=True).lower()
+ mimetype = magic.from_file(preview_path, mime=True).lower()
if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]:
return preview_path
|
pyglass/quicklook/api.py
|
ReplaceText(target='preview_path' @(23,31)->(23,39))
|
def preview(src_path):
preview_path = thumbnail_preview(src_path)
if preview_path:
mimetype = magic.from_file(src_path, mime=True).lower()
if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]:
return preview_path
|
def preview(src_path):
preview_path = thumbnail_preview(src_path)
if preview_path:
mimetype = magic.from_file(preview_path, mime=True).lower()
if mimetype in [ExportMimeType.PNG, ExportMimeType.PDF]:
return preview_path
|
50 |
https://:@github.com/erezsh/plyplus.git
|
8cc69bebfcb2cb0480ac66d07fe1f4b8637bba11
|
@@ -784,7 +784,7 @@ class _Grammar(object):
subtree.extend(child.tail)
else:
subtree.append(child)
- p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) > 1 else subtree[0]
+ p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) != 1 else subtree[0]
else:
def p_rule(self, p):
p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True)
|
plyplus/plyplus.py
|
ReplaceText(target='!=' @(787,98)->(787,99))
|
class _Grammar(object):
subtree.extend(child.tail)
else:
subtree.append(child)
p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) > 1 else subtree[0]
else:
def p_rule(self, p):
p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True)
|
class _Grammar(object):
subtree.extend(child.tail)
else:
subtree.append(child)
p[0] = self.tree_class(rule_name, subtree, skip_adjustments=True) if len(subtree) != 1 else subtree[0]
else:
def p_rule(self, p):
p[0] = self.tree_class(rule_name, p[1:], skip_adjustments=True)
|
51 |
https://:@github.com/olls/graphics.git
|
877ef2670d4ad34c4fcf951dab0922419f081531
|
@@ -22,7 +22,7 @@ def colorStr(text, color=WHITE):
return seq
sys.stdout.write(seq + '\n')
else:
- return seq
+ return text
sys.stdout.write(text + '\n')
if __name__ == '__main__':
|
colors.py
|
ReplaceText(target='text' @(25,9)->(25,12))
|
def colorStr(text, color=WHITE):
return seq
sys.stdout.write(seq + '\n')
else:
return seq
sys.stdout.write(text + '\n')
if __name__ == '__main__':
|
def colorStr(text, color=WHITE):
return seq
sys.stdout.write(seq + '\n')
else:
return text
sys.stdout.write(text + '\n')
if __name__ == '__main__':
|
52 |
https://:@github.com/hkwi/twink.git
|
5ef359deb609fc55659596a8cb327abb9c7e4653
|
@@ -319,7 +319,7 @@ def ofp_action_set_field(message, offset):
cursor = _cursor(offset)
offset = cursor.offset
- (type,len) = ofp_action_header(message, offset)
+ (type,len) = ofp_action_header(message, cursor)
field = message[cursor.offset:offset+len]
cursor.offset = offset+len
return namedtuple("ofp_action_set_field",
|
twink/ofp4/parse.py
|
ReplaceText(target='cursor' @(322,41)->(322,47))
|
def ofp_action_set_field(message, offset):
cursor = _cursor(offset)
offset = cursor.offset
(type,len) = ofp_action_header(message, offset)
field = message[cursor.offset:offset+len]
cursor.offset = offset+len
return namedtuple("ofp_action_set_field",
|
def ofp_action_set_field(message, offset):
cursor = _cursor(offset)
offset = cursor.offset
(type,len) = ofp_action_header(message, cursor)
field = message[cursor.offset:offset+len]
cursor.offset = offset+len
return namedtuple("ofp_action_set_field",
|
53 |
https://:@github.com/hkwi/twink.git
|
fc6f5ad63cb12f9caf5455e3179ecfd9cd9de060
|
@@ -27,7 +27,7 @@ def _unpack(fmt, message, offset):
return struct.unpack_from(fmt, message, offset)
def _align(length):
- return (length+7)/8*8
+ return (length+7)//8*8
# 7.1
def ofp_header(version, type, length, xid):
|
twink/ofp4/build.py
|
ReplaceText(target='//' @(30,18)->(30,19))
|
def _unpack(fmt, message, offset):
return struct.unpack_from(fmt, message, offset)
def _align(length):
return (length+7)/8*8
# 7.1
def ofp_header(version, type, length, xid):
|
def _unpack(fmt, message, offset):
return struct.unpack_from(fmt, message, offset)
def _align(length):
return (length+7)//8*8
# 7.1
def ofp_header(version, type, length, xid):
|
54 |
https://:@github.com/biolab/orange3-datafusion.git
|
54f941a66a9b369a73190dfe2007e5b6dae1803a
|
@@ -33,7 +33,7 @@ def _find_completion(fuser, relation):
for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type,
relation.col_type):
if fuser_relation._id == relation._id:
- return fuser.complete(fuser_relation)
+ return fuser.complete(relation)
return None
|
orangecontrib/datafusion/widgets/owcompletionscoring.py
|
ReplaceText(target='relation' @(36,34)->(36,48))
|
def _find_completion(fuser, relation):
for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type,
relation.col_type):
if fuser_relation._id == relation._id:
return fuser.complete(fuser_relation)
return None
|
def _find_completion(fuser, relation):
for fuser_relation in fuser.fusion_graph.get_relations(relation.row_type,
relation.col_type):
if fuser_relation._id == relation._id:
return fuser.complete(relation)
return None
|
55 |
https://:@github.com/aiqm/torchani.git
|
5bb6691515e5e56fbe4994b140dd40b73043a33f
|
@@ -151,7 +151,7 @@ class PrepareInput(torch.nn.Module):
new_tensors = []
for t in tensors:
new_tensors.append(t.index_select(1, reverse))
- return (species, *tensors)
+ return (species, *new_tensors)
def forward(self, species_coordinates):
species, coordinates = species_coordinates
|
torchani/aev.py
|
ReplaceText(target='new_tensors' @(154,26)->(154,33))
|
class PrepareInput(torch.nn.Module):
new_tensors = []
for t in tensors:
new_tensors.append(t.index_select(1, reverse))
return (species, *tensors)
def forward(self, species_coordinates):
species, coordinates = species_coordinates
|
class PrepareInput(torch.nn.Module):
new_tensors = []
for t in tensors:
new_tensors.append(t.index_select(1, reverse))
return (species, *new_tensors)
def forward(self, species_coordinates):
species, coordinates = species_coordinates
|
56 |
https://:@github.com/aiqm/torchani.git
|
abc8f7f842ae4b273c6e867b392413dcadd9c921
|
@@ -593,7 +593,7 @@ def collate_fn(data, chunk_threshold, properties_info):
if properties_info['padding_values'][i] is None:
prop = torch.stack(prop)
else:
- prop = torch.nn.utils.rnn.pad_sequence(batch_species,
+ prop = torch.nn.utils.rnn.pad_sequence(prop,
batch_first=True,
padding_value=properties_info['padding_values'][i])
# sort with number of atoms
|
torchani/data/new.py
|
ReplaceText(target='prop' @(596,51)->(596,64))
|
def collate_fn(data, chunk_threshold, properties_info):
if properties_info['padding_values'][i] is None:
prop = torch.stack(prop)
else:
prop = torch.nn.utils.rnn.pad_sequence(batch_species,
batch_first=True,
padding_value=properties_info['padding_values'][i])
# sort with number of atoms
|
def collate_fn(data, chunk_threshold, properties_info):
if properties_info['padding_values'][i] is None:
prop = torch.stack(prop)
else:
prop = torch.nn.utils.rnn.pad_sequence(prop,
batch_first=True,
padding_value=properties_info['padding_values'][i])
# sort with number of atoms
|
57 |
https://:@github.com/aiqm/torchani.git
|
c18f4a5ea1f9732cc07c8816caa401981e43dc48
|
@@ -274,7 +274,7 @@ def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor,
num_atoms = species.shape[1]
num_species_pairs = angular_length // angular_sublength
# PBC calculation is bypassed if there are no shifts
- if shifts.numel() == 1:
+ if shifts.numel() == 0:
atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr)
else:
atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr)
|
torchani/aev.py
|
ReplaceText(target='0' @(277,25)->(277,26))
|
def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor,
num_atoms = species.shape[1]
num_species_pairs = angular_length // angular_sublength
# PBC calculation is bypassed if there are no shifts
if shifts.numel() == 1:
atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr)
else:
atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr)
|
def compute_aev(species: Tensor, coordinates: Tensor, cell: Tensor,
num_atoms = species.shape[1]
num_species_pairs = angular_length // angular_sublength
# PBC calculation is bypassed if there are no shifts
if shifts.numel() == 0:
atom_index1, atom_index2, shifts = neighbor_pairs_nopbc(species == -1, coordinates, cell, shifts, Rcr)
else:
atom_index1, atom_index2, shifts = neighbor_pairs(species == -1, coordinates, cell, shifts, Rcr)
|
58 |
https://:@github.com/cs207group4/cs207-FinalProject.git
|
e7f1cc613ace275a8d259de7455ee39ca063e029
|
@@ -120,7 +120,7 @@ class ChemSolver:
r.set_initial_value(y0, 0)
self._t = [0]
self._y = [y0]
- while r.successful() and r.t <= t1:
+ while r.successful() and r.t < t1:
self._t.append(r.t + dt)
self._y.append(r.integrate(r.t + dt))
self._t = np.array(self._t)
|
pychemkin/ChemSolver.py
|
ReplaceText(target='<' @(123,37)->(123,39))
|
class ChemSolver:
r.set_initial_value(y0, 0)
self._t = [0]
self._y = [y0]
while r.successful() and r.t <= t1:
self._t.append(r.t + dt)
self._y.append(r.integrate(r.t + dt))
self._t = np.array(self._t)
|
class ChemSolver:
r.set_initial_value(y0, 0)
self._t = [0]
self._y = [y0]
while r.successful() and r.t < t1:
self._t.append(r.t + dt)
self._y.append(r.integrate(r.t + dt))
self._t = np.array(self._t)
|
59 |
https://:@github.com/MrLeeh/pyads.git
|
d14fd2a7bb2d4b784a4f6a47b6981ba2a86b699c
|
@@ -97,7 +97,7 @@ def set_local_address(ams_netid):
else:
ams_netid_st = ams_netid
- assert isinstance(ams_netid, SAmsNetId)
+ assert isinstance(ams_netid_st, SAmsNetId)
if linux:
return adsSetLocalAddress(ams_netid_st)
|
pyads/ads.py
|
ReplaceText(target='ams_netid_st' @(100,22)->(100,31))
|
def set_local_address(ams_netid):
else:
ams_netid_st = ams_netid
assert isinstance(ams_netid, SAmsNetId)
if linux:
return adsSetLocalAddress(ams_netid_st)
|
def set_local_address(ams_netid):
else:
ams_netid_st = ams_netid
assert isinstance(ams_netid_st, SAmsNetId)
if linux:
return adsSetLocalAddress(ams_netid_st)
|
60 |
https://:@github.com/pytorch/fairseq.git
|
0a836276129ef71fa6c44975dd02ab70bccc496d
|
@@ -58,7 +58,7 @@ class FConvEncoder(FairseqEncoder):
self.projections = nn.ModuleList()
self.convolutions = nn.ModuleList()
for (out_channels, kernel_size) in convolutions:
- pad = (kernel_size - 1) // 2
+ pad = (kernel_size - 1) / 2
self.projections.append(Linear(in_channels, out_channels)
if in_channels != out_channels else None)
self.convolutions.append(
|
fairseq/models/fconv.py
|
ReplaceText(target='/' @(61,36)->(61,38))
|
class FConvEncoder(FairseqEncoder):
self.projections = nn.ModuleList()
self.convolutions = nn.ModuleList()
for (out_channels, kernel_size) in convolutions:
pad = (kernel_size - 1) // 2
self.projections.append(Linear(in_channels, out_channels)
if in_channels != out_channels else None)
self.convolutions.append(
|
class FConvEncoder(FairseqEncoder):
self.projections = nn.ModuleList()
self.convolutions = nn.ModuleList()
for (out_channels, kernel_size) in convolutions:
pad = (kernel_size - 1) / 2
self.projections.append(Linear(in_channels, out_channels)
if in_channels != out_channels else None)
self.convolutions.append(
|
61 |
https://:@github.com/pytorch/fairseq.git
|
f68a44359b6596997b931d2e662a899ffba9d407
|
@@ -62,7 +62,7 @@ class SinusoidalPositionalEmbedding(nn.Module):
# recompute/expand embeddings if needed
bsz, seq_len = input.size()
max_pos = self.padding_idx + 1 + seq_len
- if seq_len > self.weights.size(0):
+ if max_pos > self.weights.size(0):
self.weights = SinusoidalPositionalEmbedding.get_embedding(
max_pos,
self.embedding_dim,
|
fairseq/modules/sinusoidal_positional_embedding.py
|
ReplaceText(target='max_pos' @(65,11)->(65,18))
|
class SinusoidalPositionalEmbedding(nn.Module):
# recompute/expand embeddings if needed
bsz, seq_len = input.size()
max_pos = self.padding_idx + 1 + seq_len
if seq_len > self.weights.size(0):
self.weights = SinusoidalPositionalEmbedding.get_embedding(
max_pos,
self.embedding_dim,
|
class SinusoidalPositionalEmbedding(nn.Module):
# recompute/expand embeddings if needed
bsz, seq_len = input.size()
max_pos = self.padding_idx + 1 + seq_len
if max_pos > self.weights.size(0):
self.weights = SinusoidalPositionalEmbedding.get_embedding(
max_pos,
self.embedding_dim,
|
62 |
https://:@github.com/pytorch/fairseq.git
|
762956a559e65e1e48df8f8b4df515d23b66fddb
|
@@ -82,7 +82,7 @@ def main(args):
train_meter.start()
valid_losses = [None]
valid_subsets = args.valid_subset.split(',')
- while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update:
+ while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update:
# train for one epoch
train(args, trainer, task, epoch_itr)
|
train.py
|
ReplaceText(target='<' @(85,47)->(85,49))
|
def main(args):
train_meter.start()
valid_losses = [None]
valid_subsets = args.valid_subset.split(',')
while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update:
# train for one epoch
train(args, trainer, task, epoch_itr)
|
def main(args):
train_meter.start()
valid_losses = [None]
valid_subsets = args.valid_subset.split(',')
while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update:
# train for one epoch
train(args, trainer, task, epoch_itr)
|
63 |
https://:@github.com/pytorch/fairseq.git
|
e9967cd334783f5da50deadc17cf8a4fc3380171
|
@@ -82,7 +82,7 @@ def main(args):
train_meter.start()
valid_losses = [None]
valid_subsets = args.valid_subset.split(',')
- while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update:
+ while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update:
# train for one epoch
train(args, trainer, task, epoch_itr)
|
train.py
|
ReplaceText(target='<' @(85,47)->(85,49))
|
def main(args):
train_meter.start()
valid_losses = [None]
valid_subsets = args.valid_subset.split(',')
while lr > args.min_lr and epoch_itr.epoch <= max_epoch and trainer.get_num_updates() < max_update:
# train for one epoch
train(args, trainer, task, epoch_itr)
|
def main(args):
train_meter.start()
valid_losses = [None]
valid_subsets = args.valid_subset.split(',')
while lr > args.min_lr and epoch_itr.epoch < max_epoch and trainer.get_num_updates() < max_update:
# train for one epoch
train(args, trainer, task, epoch_itr)
|
64 |
https://:@github.com/pytorch/fairseq.git
|
7bcb487aad8504043d13c9b869d555aa565a46c7
|
@@ -49,7 +49,7 @@ class LabelSmoothedCrossEntropyCriterion(FairseqCriterion):
sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens']
logging_output = {
'loss': utils.item(loss.data) if reduce else loss.data,
- 'nll_loss': utils.item(nll_loss.data) if reduce else loss.data,
+ 'nll_loss': utils.item(nll_loss.data) if reduce else nll_loss.data,
'ntokens': sample['ntokens'],
'sample_size': sample_size,
}
|
fairseq/criterions/label_smoothed_cross_entropy.py
|
ReplaceText(target='nll_loss' @(52,65)->(52,69))
|
class LabelSmoothedCrossEntropyCriterion(FairseqCriterion):
sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens']
logging_output = {
'loss': utils.item(loss.data) if reduce else loss.data,
'nll_loss': utils.item(nll_loss.data) if reduce else loss.data,
'ntokens': sample['ntokens'],
'sample_size': sample_size,
}
|
class LabelSmoothedCrossEntropyCriterion(FairseqCriterion):
sample_size = sample['target'].size(0) if self.args.sentence_avg else sample['ntokens']
logging_output = {
'loss': utils.item(loss.data) if reduce else loss.data,
'nll_loss': utils.item(nll_loss.data) if reduce else nll_loss.data,
'ntokens': sample['ntokens'],
'sample_size': sample_size,
}
|
65 |
https://:@github.com/pytorch/fairseq.git
|
74efc21403477d103bd426ae64c37b7a30d8f4bf
|
@@ -154,7 +154,7 @@ class TestIncrementalDecoder(FairseqIncrementalDecoder):
probs[:, i, self.dictionary.eos()] = 1.0
# random attention
- attn = torch.rand(bbsz, src_len, tgt_len)
+ attn = torch.rand(bbsz, tgt_len, src_len)
return Variable(probs), Variable(attn)
|
tests/utils.py
|
ArgSwap(idxs=1<->2 @(157,15)->(157,25))
|
class TestIncrementalDecoder(FairseqIncrementalDecoder):
probs[:, i, self.dictionary.eos()] = 1.0
# random attention
attn = torch.rand(bbsz, src_len, tgt_len)
return Variable(probs), Variable(attn)
|
class TestIncrementalDecoder(FairseqIncrementalDecoder):
probs[:, i, self.dictionary.eos()] = 1.0
# random attention
attn = torch.rand(bbsz, tgt_len, src_len)
return Variable(probs), Variable(attn)
|
66 |
https://:@github.com/pytorch/fairseq.git
|
dfd77717b91a6e233829735795ab49d6fd85c0b3
|
@@ -93,7 +93,7 @@ class CosineSchedule(FairseqLRScheduler):
else:
i = math.floor(curr_updates / self.period)
t_i = self.period
- t_curr = num_updates - (self.period * i)
+ t_curr = curr_updates - (self.period * i)
lr_shrink = self.lr_shrink ** i
min_lr = self.min_lr * lr_shrink
|
fairseq/optim/lr_scheduler/cosine_lr_scheduler.py
|
ReplaceText(target='curr_updates' @(96,25)->(96,36))
|
class CosineSchedule(FairseqLRScheduler):
else:
i = math.floor(curr_updates / self.period)
t_i = self.period
t_curr = num_updates - (self.period * i)
lr_shrink = self.lr_shrink ** i
min_lr = self.min_lr * lr_shrink
|
class CosineSchedule(FairseqLRScheduler):
else:
i = math.floor(curr_updates / self.period)
t_i = self.period
t_curr = curr_updates - (self.period * i)
lr_shrink = self.lr_shrink ** i
min_lr = self.min_lr * lr_shrink
|
67 |
https://:@github.com/pytorch/fairseq.git
|
0eea6923b9d7f408e667714709b070171ac7fe05
|
@@ -312,7 +312,7 @@ def make_positions(tensor, padding_idx, left_pad, onnx_trace=False):
positions = range_buf.expand_as(tensor)
if left_pad:
positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1)
- return positions * mask.long() + positions * (1 - mask.long())
+ return positions * mask.long() + padding_idx * (1 - mask.long())
max_pos = padding_idx + 1 + tensor.size(1)
if not hasattr(make_positions, 'range_buf'):
|
fairseq/utils.py
|
ReplaceText(target='padding_idx' @(315,41)->(315,50))
|
def make_positions(tensor, padding_idx, left_pad, onnx_trace=False):
positions = range_buf.expand_as(tensor)
if left_pad:
positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1)
return positions * mask.long() + positions * (1 - mask.long())
max_pos = padding_idx + 1 + tensor.size(1)
if not hasattr(make_positions, 'range_buf'):
|
def make_positions(tensor, padding_idx, left_pad, onnx_trace=False):
positions = range_buf.expand_as(tensor)
if left_pad:
positions = positions - mask.size(1) + mask.long().sum(dim=1).unsqueeze(1)
return positions * mask.long() + padding_idx * (1 - mask.long())
max_pos = padding_idx + 1 + tensor.size(1)
if not hasattr(make_positions, 'range_buf'):
|
68 |
https://:@github.com/pytorch/fairseq.git
|
4d3401b09f155995cd81fd394dfa50bf65ee8e5f
|
@@ -183,7 +183,7 @@ class Sampling(Search):
lprobs = lprobs[:, ::beam_size, :].contiguous()
# we exclude the first two vocab items, one of which is pad
- assert self.pad == 1, 'sampling assumes the first two symbols can be ignored'
+ assert self.pad <= 1, 'sampling assumes the first two symbols can be ignored'
lprobs_nopad = lprobs[:, :, 2:]
# only sample from top-k candidates
|
fairseq/search.py
|
ReplaceText(target='<=' @(186,24)->(186,26))
|
class Sampling(Search):
lprobs = lprobs[:, ::beam_size, :].contiguous()
# we exclude the first two vocab items, one of which is pad
assert self.pad == 1, 'sampling assumes the first two symbols can be ignored'
lprobs_nopad = lprobs[:, :, 2:]
# only sample from top-k candidates
|
class Sampling(Search):
lprobs = lprobs[:, ::beam_size, :].contiguous()
# we exclude the first two vocab items, one of which is pad
assert self.pad <= 1, 'sampling assumes the first two symbols can be ignored'
lprobs_nopad = lprobs[:, :, 2:]
# only sample from top-k candidates
|
69 |
https://:@github.com/pytorch/fairseq.git
|
2340832fdd7acaaaf07626daa6a0cef6fda06cd1
|
@@ -160,7 +160,7 @@ def main(args):
))
# update running id counter
- start_id += len(results)
+ start_id += len(inputs)
def cli_main():
|
interactive.py
|
ReplaceText(target='inputs' @(163,24)->(163,31))
|
def main(args):
))
# update running id counter
start_id += len(results)
def cli_main():
|
def main(args):
))
# update running id counter
start_id += len(inputs)
def cli_main():
|
70 |
https://:@github.com/pytorch/fairseq.git
|
39a60b844aad67aa59267d873edeb4948f6f0af9
|
@@ -351,7 +351,7 @@ class LSTMDecoder(FairseqIncrementalDecoder):
self.additional_fc = Linear(hidden_size, out_embed_dim)
if adaptive_softmax_cutoff is not None:
# setting adaptive_softmax dropout to dropout_out for now but can be redefined
- self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, embed_dim, adaptive_softmax_cutoff,
+ self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, hidden_size, adaptive_softmax_cutoff,
dropout=dropout_out)
elif not self.share_input_output_embed:
self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out)
|
fairseq/models/lstm.py
|
ReplaceText(target='hidden_size' @(354,68)->(354,77))
|
class LSTMDecoder(FairseqIncrementalDecoder):
self.additional_fc = Linear(hidden_size, out_embed_dim)
if adaptive_softmax_cutoff is not None:
# setting adaptive_softmax dropout to dropout_out for now but can be redefined
self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, embed_dim, adaptive_softmax_cutoff,
dropout=dropout_out)
elif not self.share_input_output_embed:
self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out)
|
class LSTMDecoder(FairseqIncrementalDecoder):
self.additional_fc = Linear(hidden_size, out_embed_dim)
if adaptive_softmax_cutoff is not None:
# setting adaptive_softmax dropout to dropout_out for now but can be redefined
self.adaptive_softmax = AdaptiveSoftmax(num_embeddings, hidden_size, adaptive_softmax_cutoff,
dropout=dropout_out)
elif not self.share_input_output_embed:
self.fc_out = Linear(out_embed_dim, num_embeddings, dropout=dropout_out)
|
71 |
https://:@github.com/pytorch/fairseq.git
|
49177c99c45f7d6e99a8f1500d16396e2d7b4519
|
@@ -498,7 +498,7 @@ class TransformerDecoder(FairseqIncrementalDecoder):
del state_dict[k]
version_key = '{}.version'.format(name)
- if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) < 2:
+ if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) <= 2:
# earlier checkpoints did not normalize after the stack of layers
self.layer_norm = None
self.normalize = False
|
fairseq/models/transformer.py
|
ReplaceText(target='<=' @(501,73)->(501,74))
|
class TransformerDecoder(FairseqIncrementalDecoder):
del state_dict[k]
version_key = '{}.version'.format(name)
if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) < 2:
# earlier checkpoints did not normalize after the stack of layers
self.layer_norm = None
self.normalize = False
|
class TransformerDecoder(FairseqIncrementalDecoder):
del state_dict[k]
version_key = '{}.version'.format(name)
if utils.item(state_dict.get(version_key, torch.Tensor([1]))[0]) <= 2:
# earlier checkpoints did not normalize after the stack of layers
self.layer_norm = None
self.normalize = False
|
72 |
https://:@github.com/pytorch/fairseq.git
|
5d7a81099462e9f19715ce5fa37c03816a750e12
|
@@ -412,7 +412,7 @@ class LevenshteinTransformerModel(FairseqNATModel):
max_lens = torch.zeros_like(output_tokens).fill_(255)
else:
if encoder_out.encoder_padding_mask is None:
- max_src_len = encoder_out.encoder_out.size(1)
+ max_src_len = encoder_out.encoder_out.size(0)
src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len)
else:
src_lens = (~encoder_out.encoder_padding_mask).sum(1)
|
fairseq/models/nat/levenshtein_transformer.py
|
ReplaceText(target='0' @(415,59)->(415,60))
|
class LevenshteinTransformerModel(FairseqNATModel):
max_lens = torch.zeros_like(output_tokens).fill_(255)
else:
if encoder_out.encoder_padding_mask is None:
max_src_len = encoder_out.encoder_out.size(1)
src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len)
else:
src_lens = (~encoder_out.encoder_padding_mask).sum(1)
|
class LevenshteinTransformerModel(FairseqNATModel):
max_lens = torch.zeros_like(output_tokens).fill_(255)
else:
if encoder_out.encoder_padding_mask is None:
max_src_len = encoder_out.encoder_out.size(0)
src_lens = encoder_out.encoder_out.new(bsz).fill_(max_src_len)
else:
src_lens = (~encoder_out.encoder_padding_mask).sum(1)
|
73 |
https://:@github.com/pytorch/fairseq.git
|
431d604f696a15c06fceab56b4ace271bb85e74b
|
@@ -331,7 +331,7 @@ class SequenceGenerator(object):
avg_attn_scores = avg_attn_scores[0]
if avg_attn_scores is not None:
if attn is None:
- attn = scores.new(bsz * beam_size, src_tokens.size(1), max_len + 2)
+ attn = scores.new(bsz * beam_size, avg_attn_scores.size(1), max_len + 2)
attn_buf = attn.clone()
attn[:, :, step + 1].copy_(avg_attn_scores)
|
fairseq/sequence_generator.py
|
ReplaceText(target='avg_attn_scores' @(334,55)->(334,65))
|
class SequenceGenerator(object):
avg_attn_scores = avg_attn_scores[0]
if avg_attn_scores is not None:
if attn is None:
attn = scores.new(bsz * beam_size, src_tokens.size(1), max_len + 2)
attn_buf = attn.clone()
attn[:, :, step + 1].copy_(avg_attn_scores)
|
class SequenceGenerator(object):
avg_attn_scores = avg_attn_scores[0]
if avg_attn_scores is not None:
if attn is None:
attn = scores.new(bsz * beam_size, avg_attn_scores.size(1), max_len + 2)
attn_buf = attn.clone()
attn[:, :, step + 1].copy_(avg_attn_scores)
|
74 |
https://:@github.com/pytorch/fairseq.git
|
4f8b0643c80d6a41039ae29e94fca6b44de8791a
|
@@ -138,7 +138,7 @@ def should_stop_early(args, valid_loss):
return False
else:
should_stop_early.num_runs += 1
- return should_stop_early.num_runs > args.patience
+ return should_stop_early.num_runs >= args.patience
@metrics.aggregate('train')
|
fairseq_cli/train.py
|
ReplaceText(target='>=' @(141,42)->(141,43))
|
def should_stop_early(args, valid_loss):
return False
else:
should_stop_early.num_runs += 1
return should_stop_early.num_runs > args.patience
@metrics.aggregate('train')
|
def should_stop_early(args, valid_loss):
return False
else:
should_stop_early.num_runs += 1
return should_stop_early.num_runs >= args.patience
@metrics.aggregate('train')
|
75 |
https://:@github.com/pytorch/fairseq.git
|
9a718e29855713a51877237b2dcc25e39c234c82
|
@@ -110,5 +110,5 @@ class TranslationFromPretrainedBARTTask(TranslationTask):
for s_t in src_tokens:
s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)])
source_tokens.append(s_t)
- dataset = LanguagePairDataset(src_tokens, src_lengths, self.source_dictionary)
+ dataset = LanguagePairDataset(source_tokens, src_lengths, self.source_dictionary)
return dataset
|
fairseq/tasks/translation_from_pretrained_bart.py
|
ReplaceText(target='source_tokens' @(113,38)->(113,48))
|
class TranslationFromPretrainedBARTTask(TranslationTask):
for s_t in src_tokens:
s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)])
source_tokens.append(s_t)
dataset = LanguagePairDataset(src_tokens, src_lengths, self.source_dictionary)
return dataset
|
class TranslationFromPretrainedBARTTask(TranslationTask):
for s_t in src_tokens:
s_t = torch.cat([s_t, s_t.new(1).fill_(src_lang_id)])
source_tokens.append(s_t)
dataset = LanguagePairDataset(source_tokens, src_lengths, self.source_dictionary)
return dataset
|
76 |
https://:@github.com/pytorch/fairseq.git
|
b689b6ff3ab7b806217b8aa41821bb8fc85f7cd8
|
@@ -264,7 +264,7 @@ class LanguagePairDataset(FairseqDataset):
tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]])
bos = self.src_dict.bos()
- if self.src[index][-1] != bos:
+ if self.src[index][0] != bos:
src_item = torch.cat([torch.LongTensor([bos]), self.src[index]])
if self.remove_eos_from_source:
|
fairseq/data/language_pair_dataset.py
|
ReplaceText(target='0' @(267,31)->(267,33))
|
class LanguagePairDataset(FairseqDataset):
tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]])
bos = self.src_dict.bos()
if self.src[index][-1] != bos:
src_item = torch.cat([torch.LongTensor([bos]), self.src[index]])
if self.remove_eos_from_source:
|
class LanguagePairDataset(FairseqDataset):
tgt_item = torch.cat([torch.LongTensor([bos]), self.tgt[index]])
bos = self.src_dict.bos()
if self.src[index][0] != bos:
src_item = torch.cat([torch.LongTensor([bos]), self.src[index]])
if self.remove_eos_from_source:
|
77 |
https://:@github.com/prprprus/PyMySQLPool.git
|
66b07cdf844554245cf209a72de89bd17133269c
|
@@ -169,7 +169,7 @@ class Pool(object):
if self.ping_check:
now = int(time())
timeout = now
- if isinstance(int, self.ping_check):
+ if isinstance(self.ping_check, int):
timeout = timeout - self.ping_check
if not hasattr(c, '__ping_check_timestamp'):
c.__ping_check_timestamp = now
|
pymysqlpool/pool.py
|
ArgSwap(idxs=0<->1 @(172,15)->(172,25))
|
class Pool(object):
if self.ping_check:
now = int(time())
timeout = now
if isinstance(int, self.ping_check):
timeout = timeout - self.ping_check
if not hasattr(c, '__ping_check_timestamp'):
c.__ping_check_timestamp = now
|
class Pool(object):
if self.ping_check:
now = int(time())
timeout = now
if isinstance(self.ping_check, int):
timeout = timeout - self.ping_check
if not hasattr(c, '__ping_check_timestamp'):
c.__ping_check_timestamp = now
|
78 |
https://:@github.com/dailymuse/oz.git
|
a15adf73c721d07b9dac886fcc27145e2449563c
|
@@ -173,7 +173,7 @@ class S3File(CDNFile):
def copy(self, new_path, replace=False):
"""Uses boto to copy the file to the new path instead of uploading another file to the new key"""
- if replace or get_file(new_path).exists():
+ if replace or not get_file(new_path).exists():
self.key.copy(self.key.bucket, new_path)
return True
return False
|
oz/aws_cdn/__init__.py
|
ReplaceText(target='not ' @(176,22)->(176,22))
|
class S3File(CDNFile):
def copy(self, new_path, replace=False):
"""Uses boto to copy the file to the new path instead of uploading another file to the new key"""
if replace or get_file(new_path).exists():
self.key.copy(self.key.bucket, new_path)
return True
return False
|
class S3File(CDNFile):
def copy(self, new_path, replace=False):
"""Uses boto to copy the file to the new path instead of uploading another file to the new key"""
if replace or not get_file(new_path).exists():
self.key.copy(self.key.bucket, new_path)
return True
return False
|
79 |
https://:@github.com/juju/amulet.git
|
016bfab60aca89cbcb58e80f4103e371a77b06ba
|
@@ -77,7 +77,7 @@ class Deployment(object):
pass # Copy the current parent directory to temp and deploy that
elif self.charm_name:
if charm_name == self.charm_name:
- charm = os.getcwd()
+ charm_branch = os.getcwd()
self.services[service] = {'branch': charm_branch}
if units > 1:
|
amulet/deployer.py
|
ReplaceText(target='charm_branch' @(80,16)->(80,21))
|
class Deployment(object):
pass # Copy the current parent directory to temp and deploy that
elif self.charm_name:
if charm_name == self.charm_name:
charm = os.getcwd()
self.services[service] = {'branch': charm_branch}
if units > 1:
|
class Deployment(object):
pass # Copy the current parent directory to temp and deploy that
elif self.charm_name:
if charm_name == self.charm_name:
charm_branch = os.getcwd()
self.services[service] = {'branch': charm_branch}
if units > 1:
|
80 |
https://:@github.com/gitpython-developers/gitdb.git
|
ca829e0b341dd5c3ae1408b24702f2c75db6ec73
|
@@ -445,7 +445,7 @@ class DeltaApplyReader(LazyMixin):
#{ Configuration
- if not has_perf_mod:
+ if has_perf_mod:
_set_cache_ = _set_cache_brute_
else:
_set_cache_ = _set_cache_too_slow_without_c
|
stream.py
|
ReplaceText(target='' @(448,4)->(448,8))
|
class DeltaApplyReader(LazyMixin):
#{ Configuration
if not has_perf_mod:
_set_cache_ = _set_cache_brute_
else:
_set_cache_ = _set_cache_too_slow_without_c
|
class DeltaApplyReader(LazyMixin):
#{ Configuration
if has_perf_mod:
_set_cache_ = _set_cache_brute_
else:
_set_cache_ = _set_cache_too_slow_without_c
|
81 |
https://:@github.com/longld/peda.git
|
82fcb5a12c92c27fc5722772a84df47b996d3d03
|
@@ -4746,7 +4746,7 @@ class PEDACmd(object):
step = peda.intsize()
if not peda.is_address(address): # cannot determine address
- msg("Invalid $SP address: 0x%x" % sp, "red")
+ msg("Invalid $SP address: 0x%x" % address, "red")
return
for i in range(count):
if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)):
|
peda.py
|
ReplaceText(target='address' @(4749,46)->(4749,48))
|
class PEDACmd(object):
step = peda.intsize()
if not peda.is_address(address): # cannot determine address
msg("Invalid $SP address: 0x%x" % sp, "red")
return
for i in range(count):
if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)):
|
class PEDACmd(object):
step = peda.intsize()
if not peda.is_address(address): # cannot determine address
msg("Invalid $SP address: 0x%x" % address, "red")
return
for i in range(count):
if not peda.execute("x/%sx 0x%x" % ("g" if step == 8 else "w", address + i*step)):
|
82 |
https://:@github.com/TheGhouls/oct.git
|
1f9ea29181962353fe0ea275cb4ba4ec9ae93142
|
@@ -18,7 +18,7 @@ class Report(object):
self.set_statics()
def set_statics(self):
- if os.path.exists(self.results_dir):
+ if not os.path.exists(self.results_dir):
return
try:
shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css'))
|
oct/results/reportwriter.py
|
ReplaceText(target='not ' @(21,11)->(21,11))
|
class Report(object):
self.set_statics()
def set_statics(self):
if os.path.exists(self.results_dir):
return
try:
shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css'))
|
class Report(object):
self.set_statics()
def set_statics(self):
if not os.path.exists(self.results_dir):
return
try:
shutil.copytree(os.path.join(self.templates_dir, 'css'), os.path.join(self.results_dir, 'css'))
|
83 |
https://:@github.com/ASPP/pelita.git
|
bee6872dbd95a1e526305ef39f42ac537fd2f708
|
@@ -105,7 +105,7 @@ def create_maze(layout_mesh):
Mesh of lists of MazeComponents
"""
- maze_mesh = Mesh(layout_mesh.height, layout_mesh.width,
+ maze_mesh = Mesh(layout_mesh.width, layout_mesh.height,
data=[[] for i in range(len(layout_mesh))])
for index in maze_mesh.iterkeys():
if layout_mesh[index] == CTFUniverse.wall:
|
pelita/universe.py
|
ArgSwap(idxs=0<->1 @(108,16)->(108,20))
|
def create_maze(layout_mesh):
Mesh of lists of MazeComponents
"""
maze_mesh = Mesh(layout_mesh.height, layout_mesh.width,
data=[[] for i in range(len(layout_mesh))])
for index in maze_mesh.iterkeys():
if layout_mesh[index] == CTFUniverse.wall:
|
def create_maze(layout_mesh):
Mesh of lists of MazeComponents
"""
maze_mesh = Mesh(layout_mesh.width, layout_mesh.height,
data=[[] for i in range(len(layout_mesh))])
for index in maze_mesh.iterkeys():
if layout_mesh[index] == CTFUniverse.wall:
|
84 |
https://:@github.com/ASPP/pelita.git
|
508cd180dce7b72ab248211c977c8525a9c023de
|
@@ -31,7 +31,7 @@ def __init__(self, index, initial_pos, team, homezone,
@property
def in_own_zone(self):
- return self.homezone[0] <= self.current_pos[1] <= self.homezone[1]
+ return self.homezone[0] <= self.current_pos[0] <= self.homezone[1]
def move(self, new_pos):
self.current_pos = new_pos
|
pelita/universe.py
|
ReplaceText(target='0' @(34,52)->(34,53))
|
def __init__(self, index, initial_pos, team, homezone,
@property
def in_own_zone(self):
return self.homezone[0] <= self.current_pos[1] <= self.homezone[1]
def move(self, new_pos):
self.current_pos = new_pos
|
def __init__(self, index, initial_pos, team, homezone,
@property
def in_own_zone(self):
return self.homezone[0] <= self.current_pos[0] <= self.homezone[1]
def move(self, new_pos):
self.current_pos = new_pos
|
85 |
https://:@github.com/ASPP/pelita.git
|
6b76e416da2dc0d18224e47d7b176dad967e15b2
|
@@ -182,7 +182,7 @@ def a_star(self, initial, target):
else:
seen.append(current)
for pos in self.adjacency[current]:
- heapq.heappush(to_visit, (datamodel.manhattan_dist(current, pos), (pos)))
+ heapq.heappush(to_visit, (datamodel.manhattan_dist(target, pos), (pos)))
# Now back-track using seen to determine how we got here.
# Initialise the path with current node, i.e. position of food.
|
pelita/game_master.py
|
ReplaceText(target='target' @(185,71)->(185,78))
|
def a_star(self, initial, target):
else:
seen.append(current)
for pos in self.adjacency[current]:
heapq.heappush(to_visit, (datamodel.manhattan_dist(current, pos), (pos)))
# Now back-track using seen to determine how we got here.
# Initialise the path with current node, i.e. position of food.
|
def a_star(self, initial, target):
else:
seen.append(current)
for pos in self.adjacency[current]:
heapq.heappush(to_visit, (datamodel.manhattan_dist(target, pos), (pos)))
# Now back-track using seen to determine how we got here.
# Initialise the path with current node, i.e. position of food.
|
86 |
https://:@github.com/ASPP/pelita.git
|
fa2505d44ae3d3724f7fa979c0167f03bf7424f7
|
@@ -136,7 +136,7 @@ def play(self):
if self.universe.teams[0].score < self.universe.teams[1].score:
events.append(datamodel.TeamWins(1))
elif self.universe.teams[0].score > self.universe.teams[1].score:
- events.append(datamodel.TeamWins(1))
+ events.append(datamodel.TeamWins(0))
else:
events.append(datamodel.GameDraw())
self.send_to_viewers(round_index, None, events)
|
pelita/game_master.py
|
ReplaceText(target='0' @(139,45)->(139,46))
|
def play(self):
if self.universe.teams[0].score < self.universe.teams[1].score:
events.append(datamodel.TeamWins(1))
elif self.universe.teams[0].score > self.universe.teams[1].score:
events.append(datamodel.TeamWins(1))
else:
events.append(datamodel.GameDraw())
self.send_to_viewers(round_index, None, events)
|
def play(self):
if self.universe.teams[0].score < self.universe.teams[1].score:
events.append(datamodel.TeamWins(1))
elif self.universe.teams[0].score > self.universe.teams[1].score:
events.append(datamodel.TeamWins(0))
else:
events.append(datamodel.GameDraw())
self.send_to_viewers(round_index, None, events)
|
87 |
https://:@github.com/ASPP/pelita.git
|
4044f845e54c2077d6896010c25fcc123fc10203
|
@@ -78,7 +78,7 @@ def test_equal_positions(self):
layout = create_layout(layout_str)
assert layout.bots == [(1, 1), (1, 1)]
assert layout.enemy == [(1, 1), (1, 1)]
- setup_test_game(layout=layout)
+ setup_test_game(layout=layout_str)
def test_define_after(self):
layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None)
|
test/test_team.py
|
ReplaceText(target='layout_str' @(81,31)->(81,37))
|
def test_equal_positions(self):
layout = create_layout(layout_str)
assert layout.bots == [(1, 1), (1, 1)]
assert layout.enemy == [(1, 1), (1, 1)]
setup_test_game(layout=layout)
def test_define_after(self):
layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None)
|
def test_equal_positions(self):
layout = create_layout(layout_str)
assert layout.bots == [(1, 1), (1, 1)]
assert layout.enemy == [(1, 1), (1, 1)]
setup_test_game(layout=layout_str)
def test_define_after(self):
layout = create_layout(self.layout, food=[(1, 1)], bots=[None, None], enemy=None)
|
88 |
https://:@github.com/ASPP/pelita.git
|
6fd0a9d2af44c491c1cc6774c3a169e97e2040be
|
@@ -398,7 +398,7 @@ def _team(self):
@property
def turn(self):
""" The turn of our bot. """
- return self.bot_index // 2
+ return self.bot_index % 2
@property
def other(self):
|
pelita/player/team.py
|
ReplaceText(target='%' @(401,30)->(401,32))
|
def _team(self):
@property
def turn(self):
""" The turn of our bot. """
return self.bot_index // 2
@property
def other(self):
|
def _team(self):
@property
def turn(self):
""" The turn of our bot. """
return self.bot_index % 2
@property
def other(self):
|
89 |
https://:@github.com/iotaledger/ccurl.interface.py.git
|
eb7f9190d24995d3f8d03a8350382ab6045a6e67
|
@@ -44,7 +44,7 @@ gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your
mwm = 14 # target is mainnet
-bundle = entangled_interface.local_attach_to_tangle(pb, gta['trunkTransaction'], gta['branchTransaction'], mwm)
+bundle = entangled_interface.local_attach_to_tangle(pb, gta['branchTransaction'],gta['trunkTransaction'], mwm)
bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
|
examples/with_entangled.py
|
ArgSwap(idxs=1<->2 @(47,9)->(47,51))
|
gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your
mwm = 14 # target is mainnet
bundle = entangled_interface.local_attach_to_tangle(pb, gta['trunkTransaction'], gta['branchTransaction'], mwm)
bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
|
gta = api.get_transactions_to_approve(depth=3) # get tips to be approved by your
mwm = 14 # target is mainnet
bundle = entangled_interface.local_attach_to_tangle(pb, gta['branchTransaction'],gta['trunkTransaction'], mwm)
bundle_trytes = [ x.as_tryte_string() for x in pb._transactions ]
|
90 |
https://:@github.com/softlayer/softlayer-python.git
|
53731de7e51d31475cc224aceb0f3ff7217cdafd
|
@@ -153,7 +153,7 @@ class NetworkManager(object):
('privateResidenceFlag', private_residence),
('state', state),
('postalCode', postal_code)]:
- if key is not None:
+ if value is not None:
update[key] = value
# If there's anything to update, update it
|
SoftLayer/managers/network.py
|
ReplaceText(target='value' @(156,15)->(156,18))
|
class NetworkManager(object):
('privateResidenceFlag', private_residence),
('state', state),
('postalCode', postal_code)]:
if key is not None:
update[key] = value
# If there's anything to update, update it
|
class NetworkManager(object):
('privateResidenceFlag', private_residence),
('state', state),
('postalCode', postal_code)]:
if value is not None:
update[key] = value
# If there's anything to update, update it
|
91 |
https://:@github.com/softlayer/softlayer-python.git
|
dcf66e15711e47c594f20ffac7605bfc6d1a8746
|
@@ -15,7 +15,7 @@ import click
type=click.Choice(['vs', 'vlan', 'server']),
help='Firewall type',
required=True)
[email protected]('--high-availability', '--ha',
[email protected]('--ha', '--high-availability',
is_flag=True,
help='High available firewall option')
@environment.pass_env
|
SoftLayer/CLI/firewall/add.py
|
ArgSwap(idxs=0<->1 @(18,1)->(18,13))
|
import click
type=click.Choice(['vs', 'vlan', 'server']),
help='Firewall type',
required=True)
@click.option('--high-availability', '--ha',
is_flag=True,
help='High available firewall option')
@environment.pass_env
|
import click
type=click.Choice(['vs', 'vlan', 'server']),
help='Firewall type',
required=True)
@click.option('--ha', '--high-availability',
is_flag=True,
help='High available firewall option')
@environment.pass_env
|
92 |
https://:@github.com/softlayer/softlayer-python.git
|
f0840e302d486d6002a14419bbde85c1deedaf6a
|
@@ -271,7 +271,7 @@ class BlockStorageManager(utils.IdentifierMixin, object):
package,
'performance_storage_iscsi'
),
- storage_utils.find_performance_space_price(package, iops),
+ storage_utils.find_performance_space_price(package, size),
storage_utils.find_performance_iops_price(package, size, iops),
]
elif storage_type == 'storage_service_enterprise':
|
SoftLayer/managers/block.py
|
ReplaceText(target='size' @(274,68)->(274,72))
|
class BlockStorageManager(utils.IdentifierMixin, object):
package,
'performance_storage_iscsi'
),
storage_utils.find_performance_space_price(package, iops),
storage_utils.find_performance_iops_price(package, size, iops),
]
elif storage_type == 'storage_service_enterprise':
|
class BlockStorageManager(utils.IdentifierMixin, object):
package,
'performance_storage_iscsi'
),
storage_utils.find_performance_space_price(package, size),
storage_utils.find_performance_iops_price(package, size, iops),
]
elif storage_type == 'storage_service_enterprise':
|
93 |
https://:@github.com/softlayer/softlayer-python.git
|
4418057fc0e3632aba2d89b6e42494c79cadd16a
|
@@ -367,7 +367,7 @@ class VSManager(utils.IdentifierMixin, object):
if datacenter:
data["datacenter"] = {"name": datacenter}
- if private_vlan and public_vlan:
+ if private_vlan or public_vlan:
network_components = self._create_network_components(public_vlan, private_vlan,
private_subnet, public_subnet)
data.update(network_components)
|
SoftLayer/managers/vs.py
|
ReplaceText(target='or' @(370,24)->(370,27))
|
class VSManager(utils.IdentifierMixin, object):
if datacenter:
data["datacenter"] = {"name": datacenter}
if private_vlan and public_vlan:
network_components = self._create_network_components(public_vlan, private_vlan,
private_subnet, public_subnet)
data.update(network_components)
|
class VSManager(utils.IdentifierMixin, object):
if datacenter:
data["datacenter"] = {"name": datacenter}
if private_vlan or public_vlan:
network_components = self._create_network_components(public_vlan, private_vlan,
private_subnet, public_subnet)
data.update(network_components)
|
94 |
https://:@github.com/softlayer/softlayer-python.git
|
58b27c6bf5400a717acd00b7866964ef11f36e59
|
@@ -87,6 +87,6 @@ def cli(env, identifier):
for guest in guests:
real_guest = guest.get('virtualGuest')
member_table.add_row([
- guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate'))
+ real_guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate'))
])
env.fout(member_table)
|
SoftLayer/CLI/autoscale/detail.py
|
ReplaceText(target='real_guest' @(90,12)->(90,17))
|
def cli(env, identifier):
for guest in guests:
real_guest = guest.get('virtualGuest')
member_table.add_row([
guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate'))
])
env.fout(member_table)
|
def cli(env, identifier):
for guest in guests:
real_guest = guest.get('virtualGuest')
member_table.add_row([
real_guest.get('id'), real_guest.get('hostname'), utils.clean_time(real_guest.get('provisionDate'))
])
env.fout(member_table)
|
95 |
https://:@github.com/data-8/datascience.git
|
fd9aceb598290fb89a0f3131c3fb39dde18ef543
|
@@ -446,7 +446,7 @@ class Table(collections.abc.Mapping):
count | points
9 | 10
"""
- percentiles = [percentile(self[column_name], p) for column_name in self]
+ percentiles = [percentile(p, self[column_name]) for column_name in self]
return Table(percentiles, self.column_labels)
##################
# Export/Display #
|
datascience/tables.py
|
ArgSwap(idxs=0<->1 @(449,23)->(449,33))
|
class Table(collections.abc.Mapping):
count | points
9 | 10
"""
percentiles = [percentile(self[column_name], p) for column_name in self]
return Table(percentiles, self.column_labels)
##################
# Export/Display #
|
class Table(collections.abc.Mapping):
count | points
9 | 10
"""
percentiles = [percentile(p, self[column_name]) for column_name in self]
return Table(percentiles, self.column_labels)
##################
# Export/Display #
|
96 |
https://:@github.com/data-8/datascience.git
|
084450f127ecc490b887cad82fa43cda5f9b32fe
|
@@ -2255,7 +2255,7 @@ class Table(collections.abc.MutableMapping):
space_count[labels[i]] += 1
return updated_labels
return labels
- yticks = make_unique_labels(labels)
+ yticks = make_unique_labels(yticks)
print("yticks: " + str(yticks))
print("ylabel: " + str(ylabel))
|
datascience/tables.py
|
ReplaceText(target='yticks' @(2258,36)->(2258,42))
|
class Table(collections.abc.MutableMapping):
space_count[labels[i]] += 1
return updated_labels
return labels
yticks = make_unique_labels(labels)
print("yticks: " + str(yticks))
print("ylabel: " + str(ylabel))
|
class Table(collections.abc.MutableMapping):
space_count[labels[i]] += 1
return updated_labels
return labels
yticks = make_unique_labels(yticks)
print("yticks: " + str(yticks))
print("ylabel: " + str(ylabel))
|
97 |
https://:@github.com/dnaeon/py-vpoller.git
|
81769f6f8d9cb0dfc8cbc39a44027afa7d459636
|
@@ -51,6 +51,6 @@ def task(name, required=None):
result = {'success': 1, 'msg': e.message}
finally:
return result
- registry.register(name=name, fn=fn, required=required)
+ registry.register(name=name, fn=wrapper, required=required)
return wrapper
return decorator
|
src/vpoller/decorators.py
|
ReplaceText(target='wrapper' @(54,40)->(54,42))
|
def task(name, required=None):
result = {'success': 1, 'msg': e.message}
finally:
return result
registry.register(name=name, fn=fn, required=required)
return wrapper
return decorator
|
def task(name, required=None):
result = {'success': 1, 'msg': e.message}
finally:
return result
registry.register(name=name, fn=wrapper, required=required)
return wrapper
return decorator
|
98 |
https://:@github.com/enthought/qt_binder.git
|
68381b406035f2ce9666cb8ef1ab2e8e57cf8bf8
|
@@ -58,4 +58,4 @@ else:
loader = RecordingUiLoader()
ui = loader.load(path)
- return ui, ui.to_be_bound()
+ return ui, loader.to_be_bound()
|
qt_binder/qt/ui_loader.py
|
ReplaceText(target='loader' @(61,19)->(61,21))
|
else:
loader = RecordingUiLoader()
ui = loader.load(path)
return ui, ui.to_be_bound()
|
else:
loader = RecordingUiLoader()
ui = loader.load(path)
return ui, loader.to_be_bound()
|
99 |
https://:@github.com/ggozad/behaving.git
|
1bc546aa03f9d42ff78a0a79e0894e488edc9add
|
@@ -40,7 +40,7 @@ def should_receive_email(context, address):
def click_link_in_email(context, address):
mails = context.mail.user_messages(address)
assert mails, u'message not found'
- mail = email.message_from_string(mails[-1])
+ mail = email.message_from_string(mails[0])
links = URL_RE.findall(str(mail).replace('=\n', ''))
assert links, u'link not found'
url = links[0]
|
src/behaving/mail/steps.py
|
ReplaceText(target='0' @(43,43)->(43,45))
|
def should_receive_email(context, address):
def click_link_in_email(context, address):
mails = context.mail.user_messages(address)
assert mails, u'message not found'
mail = email.message_from_string(mails[-1])
links = URL_RE.findall(str(mail).replace('=\n', ''))
assert links, u'link not found'
url = links[0]
|
def should_receive_email(context, address):
def click_link_in_email(context, address):
mails = context.mail.user_messages(address)
assert mails, u'message not found'
mail = email.message_from_string(mails[0])
links = URL_RE.findall(str(mail).replace('=\n', ''))
assert links, u'link not found'
url = links[0]
|
End of preview. Expand
in Data Studio
See Allamanis et al., 2021 (NeurIPS 2021) for more information.
- Downloads last month
- 70