Unnamed: 0
int64 0
2.44k
| repo
stringlengths 32
81
| hash
stringlengths 40
40
| diff
stringlengths 113
1.17k
| old_path
stringlengths 5
84
| rewrite
stringlengths 34
79
| initial_state
stringlengths 75
980
| final_state
stringlengths 76
980
|
---|---|---|---|---|---|---|---|
200 |
https://:@github.com/Timidger/Wikia.git
|
ff74a5711925274af13738f4b588a46356a1fa98
|
@@ -395,7 +395,7 @@ class WikiaPage(object):
# Remove the /revision/ fluff after the image url
image = image.partition("/revision/")[0]
image_type = mimetypes.guess_type(image)[0]
- if image_type is None:
+ if image_type is not None:
image_type = "." + image_type.split("/")[-1]
else:
image_type = ".png" # in case mimetypes.guess cant find it it will return None
|
wikia/wikia.py
|
ReplaceText(target=' is not ' @(398,23)->(398,27))
|
class WikiaPage(object):
# Remove the /revision/ fluff after the image url
image = image.partition("/revision/")[0]
image_type = mimetypes.guess_type(image)[0]
if image_type is None:
image_type = "." + image_type.split("/")[-1]
else:
image_type = ".png" # in case mimetypes.guess cant find it it will return None
|
class WikiaPage(object):
# Remove the /revision/ fluff after the image url
image = image.partition("/revision/")[0]
image_type = mimetypes.guess_type(image)[0]
if image_type is not None:
image_type = "." + image_type.split("/")[-1]
else:
image_type = ".png" # in case mimetypes.guess cant find it it will return None
|
201 |
https://:@github.com/meetshah1995/pytorch-semseg.git
|
81997cd5af36759773a67b11cf148d3876b13a69
|
@@ -156,7 +156,7 @@ def train(cfg, writer, logger):
labels_val = labels_val.to(device)
outputs = model(images_val)
- val_loss = loss_fn(input=outputs, target=labels)
+ val_loss = loss_fn(input=outputs, target=labels_val)
pred = outputs.data.max(1)[1].cpu().numpy()
gt = labels_val.data.cpu().numpy()
|
train.py
|
ReplaceText(target='labels_val' @(159,65)->(159,71))
|
def train(cfg, writer, logger):
labels_val = labels_val.to(device)
outputs = model(images_val)
val_loss = loss_fn(input=outputs, target=labels)
pred = outputs.data.max(1)[1].cpu().numpy()
gt = labels_val.data.cpu().numpy()
|
def train(cfg, writer, logger):
labels_val = labels_val.to(device)
outputs = model(images_val)
val_loss = loss_fn(input=outputs, target=labels_val)
pred = outputs.data.max(1)[1].cpu().numpy()
gt = labels_val.data.cpu().numpy()
|
202 |
https://:@github.com/meetshah1995/pytorch-semseg.git
|
801fb200547caa5b0d91b8dde56b837da029f746
|
@@ -27,7 +27,7 @@ def multi_scale_cross_entropy2d(input, target, weight=None, size_average=True, s
n_inp = len(input)
scale = 0.4
scale_weight = torch.pow(scale * torch.ones(n_inp), torch.arange(n_inp).float()).to(
- input.device
+ target.device
)
loss = 0.0
|
ptsemseg/loss/loss.py
|
ReplaceText(target='target' @(30,12)->(30,17))
|
def multi_scale_cross_entropy2d(input, target, weight=None, size_average=True, s
n_inp = len(input)
scale = 0.4
scale_weight = torch.pow(scale * torch.ones(n_inp), torch.arange(n_inp).float()).to(
input.device
)
loss = 0.0
|
def multi_scale_cross_entropy2d(input, target, weight=None, size_average=True, s
n_inp = len(input)
scale = 0.4
scale_weight = torch.pow(scale * torch.ones(n_inp), torch.arange(n_inp).float()).to(
target.device
)
loss = 0.0
|
203 |
https://:@github.com/bhmm/bhmm.git
|
34f1e00dc49a99094b5c492381d9f711e96fc4d9
|
@@ -192,7 +192,7 @@ def estimate_initial_hmm(observations, nstates, reversible=True, eps_A=None, eps
# does the count matrix too few closed sets to give nstates metastable states? Then we need a prior
if len(_tmatrix_disconnected.closed_sets(C)) < nstates:
msm_prior = 0.001
- B = msm_prior * np.eye(C_full.shape[0]) # diagonal prior
+ B = msm_prior * np.eye(C.shape[0]) # diagonal prior
B += msmtools.estimation.prior_neighbor(C, alpha=msm_prior) # neighbor prior
C_post = C + B # posterior
P_for_pcca = _tmatrix_disconnected.estimate_P(C_post, reversible=True)
|
bhmm/init/discrete.py
|
ReplaceText(target='C' @(195,31)->(195,37))
|
def estimate_initial_hmm(observations, nstates, reversible=True, eps_A=None, eps
# does the count matrix too few closed sets to give nstates metastable states? Then we need a prior
if len(_tmatrix_disconnected.closed_sets(C)) < nstates:
msm_prior = 0.001
B = msm_prior * np.eye(C_full.shape[0]) # diagonal prior
B += msmtools.estimation.prior_neighbor(C, alpha=msm_prior) # neighbor prior
C_post = C + B # posterior
P_for_pcca = _tmatrix_disconnected.estimate_P(C_post, reversible=True)
|
def estimate_initial_hmm(observations, nstates, reversible=True, eps_A=None, eps
# does the count matrix too few closed sets to give nstates metastable states? Then we need a prior
if len(_tmatrix_disconnected.closed_sets(C)) < nstates:
msm_prior = 0.001
B = msm_prior * np.eye(C.shape[0]) # diagonal prior
B += msmtools.estimation.prior_neighbor(C, alpha=msm_prior) # neighbor prior
C_post = C + B # posterior
P_for_pcca = _tmatrix_disconnected.estimate_P(C_post, reversible=True)
|
204 |
https://:@github.com/rackspace/pyrax.git
|
71a7f1924e2c53a4caac54e2b8eb985c207869c7
|
@@ -220,7 +220,7 @@ class ManagerTest(unittest.TestCase):
ret = mgr.findall(some_att="ok")
self.assertTrue(o1 in ret)
self.assertFalse(o2 in ret)
- self.assertTrue(o1 in ret)
+ self.assertTrue(o3 in ret)
mgr.list = sav
def test_findall_bad_att(self):
|
tests/unit/test_manager.py
|
ReplaceText(target='o3' @(223,24)->(223,26))
|
class ManagerTest(unittest.TestCase):
ret = mgr.findall(some_att="ok")
self.assertTrue(o1 in ret)
self.assertFalse(o2 in ret)
self.assertTrue(o1 in ret)
mgr.list = sav
def test_findall_bad_att(self):
|
class ManagerTest(unittest.TestCase):
ret = mgr.findall(some_att="ok")
self.assertTrue(o1 in ret)
self.assertFalse(o2 in ret)
self.assertTrue(o3 in ret)
mgr.list = sav
def test_findall_bad_att(self):
|
205 |
https://:@github.com/deckbsd/glouton-satnogs-data-downloader.git
|
7b3982a3ac4c7ecb086cf6ce504aa766a54f1211
|
@@ -25,7 +25,7 @@ class ObservationRepo:
self.__read_page(r.json(), self.__cmd.start_date, self.__cmd.end_date)
page += 1
- params['page'] += str(page)
+ params['page'] = str(page)
print('\ndownloading started (Ctrl + F5 to stop)...\t~( ^o^)~')
self.__create_workers_and_wait()
|
repositories/observation/observationsRepo.py
|
ReplaceText(target='=' @(28,27)->(28,29))
|
class ObservationRepo:
self.__read_page(r.json(), self.__cmd.start_date, self.__cmd.end_date)
page += 1
params['page'] += str(page)
print('\ndownloading started (Ctrl + F5 to stop)...\t~( ^o^)~')
self.__create_workers_and_wait()
|
class ObservationRepo:
self.__read_page(r.json(), self.__cmd.start_date, self.__cmd.end_date)
page += 1
params['page'] = str(page)
print('\ndownloading started (Ctrl + F5 to stop)...\t~( ^o^)~')
self.__create_workers_and_wait()
|
206 |
https://:@github.com/vmware/pyvcloud.git
|
e250911c7ba7666c42c2fa197f32744424d647ff
|
@@ -121,7 +121,7 @@ class VCA(object):
#todo: check if vcloud session can be cached as well...
vCloudSession = self.create_vCloudSession(vdcReference[1])
if vCloudSession:
- vcd = VCD(vCloudSession, serviceId, serviceId)
+ vcd = VCD(vCloudSession, serviceId, vdcId)
return vcd
return None
|
pyvcloud/vcloudair.py
|
ReplaceText(target='vdcId' @(124,52)->(124,61))
|
class VCA(object):
#todo: check if vcloud session can be cached as well...
vCloudSession = self.create_vCloudSession(vdcReference[1])
if vCloudSession:
vcd = VCD(vCloudSession, serviceId, serviceId)
return vcd
return None
|
class VCA(object):
#todo: check if vcloud session can be cached as well...
vCloudSession = self.create_vCloudSession(vdcReference[1])
if vCloudSession:
vcd = VCD(vCloudSession, serviceId, vdcId)
return vcd
return None
|
207 |
https://:@github.com/vmware/pyvcloud.git
|
0b7926b141b58afbf94168ac95f1ab3f00f14e71
|
@@ -2129,7 +2129,7 @@ class VDC(object):
policy_list = []
for policy_reference in policy_references.VdcComputePolicyReference:
policy_list.append(policy_reference)
- return policy_reference
+ return policy_list
def add_compute_policy(self, href):
"""Add a VdcComputePolicy.
|
pyvcloud/vcd/vdc.py
|
ReplaceText(target='policy_list' @(2132,15)->(2132,31))
|
class VDC(object):
policy_list = []
for policy_reference in policy_references.VdcComputePolicyReference:
policy_list.append(policy_reference)
return policy_reference
def add_compute_policy(self, href):
"""Add a VdcComputePolicy.
|
class VDC(object):
policy_list = []
for policy_reference in policy_references.VdcComputePolicyReference:
policy_list.append(policy_reference)
return policy_list
def add_compute_policy(self, href):
"""Add a VdcComputePolicy.
|
208 |
https://:@github.com/springer-math/Mathematics-of-Epidemics-on-Networks.git
|
1f069443a464955c87cfa926c0ad8c8ce66bc424
|
@@ -851,7 +851,7 @@ def get_infected_nodes(G, tau, gamma, initial_infecteds=None):
elif G.has_node(initial_infecteds):
initial_infecteds=[initial_infecteds]
H = directed_percolate_network(G, tau, gamma)
- infected_nodes = _out_component_(G, initial_infecteds)
+ infected_nodes = _out_component_(H, initial_infecteds)
return infected_nodes
|
EoN/simulation.py
|
ReplaceText(target='H' @(854,37)->(854,38))
|
def get_infected_nodes(G, tau, gamma, initial_infecteds=None):
elif G.has_node(initial_infecteds):
initial_infecteds=[initial_infecteds]
H = directed_percolate_network(G, tau, gamma)
infected_nodes = _out_component_(G, initial_infecteds)
return infected_nodes
|
def get_infected_nodes(G, tau, gamma, initial_infecteds=None):
elif G.has_node(initial_infecteds):
initial_infecteds=[initial_infecteds]
H = directed_percolate_network(G, tau, gamma)
infected_nodes = _out_component_(H, initial_infecteds)
return infected_nodes
|
209 |
https://:@github.com/waipbmtd/python-client.git
|
3d95fa3fb3d85c14d1b92da348b4f2f0b007424b
|
@@ -135,7 +135,7 @@ def show(path):
def validate_params(ctx, param, value):
if any(['=' not in item for item in value]):
raise click.BadParameter('Parameters need to be in format <field name>=<value>')
- return dict([tuple(item.split('=', 1)) for item in param])
+ return dict([tuple(item.split('=', 1)) for item in value])
def validate_inplace(ctx, param, value):
|
coreapi/commandline.py
|
ReplaceText(target='value' @(138,55)->(138,60))
|
def show(path):
def validate_params(ctx, param, value):
if any(['=' not in item for item in value]):
raise click.BadParameter('Parameters need to be in format <field name>=<value>')
return dict([tuple(item.split('=', 1)) for item in param])
def validate_inplace(ctx, param, value):
|
def show(path):
def validate_params(ctx, param, value):
if any(['=' not in item for item in value]):
raise click.BadParameter('Parameters need to be in format <field name>=<value>')
return dict([tuple(item.split('=', 1)) for item in value])
def validate_inplace(ctx, param, value):
|
210 |
https://:@github.com/waipbmtd/python-client.git
|
e630971368e28989484272ca7774af385ce3ddd6
|
@@ -436,7 +436,7 @@ def credentials_show():
@click.option('--auth', metavar="AUTH_SCHEME", help='Auth scheme to apply to the credentials string. Options: "none", "basic". Default is "none".', default='none', type=click.Choice(['none', 'basic']))
def credentials_add(domain, credentials_string, auth):
if auth == 'none':
- header = auth
+ header = credentials_string
elif auth == 'basic':
header = 'Basic ' + b64encode(credentials_string)
credentials = get_credentials()
|
coreapi/commandline.py
|
ReplaceText(target='credentials_string' @(439,17)->(439,21))
|
def credentials_show():
@click.option('--auth', metavar="AUTH_SCHEME", help='Auth scheme to apply to the credentials string. Options: "none", "basic". Default is "none".', default='none', type=click.Choice(['none', 'basic']))
def credentials_add(domain, credentials_string, auth):
if auth == 'none':
header = auth
elif auth == 'basic':
header = 'Basic ' + b64encode(credentials_string)
credentials = get_credentials()
|
def credentials_show():
@click.option('--auth', metavar="AUTH_SCHEME", help='Auth scheme to apply to the credentials string. Options: "none", "basic". Default is "none".', default='none', type=click.Choice(['none', 'basic']))
def credentials_add(domain, credentials_string, auth):
if auth == 'none':
header = credentials_string
elif auth == 'basic':
header = 'Basic ' + b64encode(credentials_string)
credentials = get_credentials()
|
211 |
https://:@github.com/openfisca/openfisca-survey-manager.git
|
b1c6e08ca7ddd546171b5623ce0ae04ed9f597fd
|
@@ -24,7 +24,7 @@ def temporary_store_decorator(config_files_directory = default_config_files_dire
'tmp_directory is not set: {!r} in {}'.format(tmp_directory, read_config_file_name)
assert os.path.isabs(tmp_directory), \
'tmp_directory should be an absolut path: {!r} in {}'.format(tmp_directory, read_config_file_name)
- if os.path.isdir(tmp_directory):
+ if not os.path.isdir(tmp_directory):
'tmp_directory does not exist: {!r} in {}. Creating it.'.format(tmp_directory, read_config_file_name)
os.makedirs(tmp_directory)
|
openfisca_survey_manager/temporary.py
|
ReplaceText(target='not ' @(27,7)->(27,7))
|
def temporary_store_decorator(config_files_directory = default_config_files_dire
'tmp_directory is not set: {!r} in {}'.format(tmp_directory, read_config_file_name)
assert os.path.isabs(tmp_directory), \
'tmp_directory should be an absolut path: {!r} in {}'.format(tmp_directory, read_config_file_name)
if os.path.isdir(tmp_directory):
'tmp_directory does not exist: {!r} in {}. Creating it.'.format(tmp_directory, read_config_file_name)
os.makedirs(tmp_directory)
|
def temporary_store_decorator(config_files_directory = default_config_files_dire
'tmp_directory is not set: {!r} in {}'.format(tmp_directory, read_config_file_name)
assert os.path.isabs(tmp_directory), \
'tmp_directory should be an absolut path: {!r} in {}'.format(tmp_directory, read_config_file_name)
if not os.path.isdir(tmp_directory):
'tmp_directory does not exist: {!r} in {}. Creating it.'.format(tmp_directory, read_config_file_name)
os.makedirs(tmp_directory)
|
212 |
https://:@github.com/ARMmbed/yotta.git
|
e948620774774810165b28968ce0324b1f5bb953
|
@@ -329,7 +329,7 @@ def satisfyTarget(name, version_required, working_directory, update_installed=No
# must rm the old target before continuing
fsutils.rmRf(target_path)
- if not v and update_installed is None:
+ if not v and update_installed is not None:
v = latestSuitableVersion(name, version_required, registry='target')
if not v:
|
yotta/lib/access.py
|
ReplaceText(target=' is not ' @(332,33)->(332,37))
|
def satisfyTarget(name, version_required, working_directory, update_installed=No
# must rm the old target before continuing
fsutils.rmRf(target_path)
if not v and update_installed is None:
v = latestSuitableVersion(name, version_required, registry='target')
if not v:
|
def satisfyTarget(name, version_required, working_directory, update_installed=No
# must rm the old target before continuing
fsutils.rmRf(target_path)
if not v and update_installed is not None:
v = latestSuitableVersion(name, version_required, registry='target')
if not v:
|
213 |
https://:@github.com/ARMmbed/yotta.git
|
06a9d7e69861044818b5a5de490336f6be65d6f1
|
@@ -371,7 +371,7 @@ def generateTest(**kwargs):
test_method.__name__ = test_name
setattr(TestCLITestGenerated, test_name, test_method)
-if not util.canBuildNatively():
+if util.canBuildNatively():
forAllReporterTests(generateTest)
else:
print('WARNING: skipping test reporter tests (cannot build natively on this platform)')
|
yotta/test/cli/test_test.py
|
ReplaceText(target='' @(374,3)->(374,7))
|
def generateTest(**kwargs):
test_method.__name__ = test_name
setattr(TestCLITestGenerated, test_name, test_method)
if not util.canBuildNatively():
forAllReporterTests(generateTest)
else:
print('WARNING: skipping test reporter tests (cannot build natively on this platform)')
|
def generateTest(**kwargs):
test_method.__name__ = test_name
setattr(TestCLITestGenerated, test_name, test_method)
if util.canBuildNatively():
forAllReporterTests(generateTest)
else:
print('WARNING: skipping test reporter tests (cannot build natively on this platform)')
|
214 |
https://:@github.com/b1naryth1ef/disco.git
|
bd75deb29adcc42f8de451d51e7dbc02eb360b1e
|
@@ -112,7 +112,7 @@ class Plugin(LoggingClass, PluginDeco):
def register_command(self, func, *args, **kwargs):
wrapped = functools.partial(self._dispatch, 'command', func)
- self.commands[func.__name__] = Command(self, func, *args, **kwargs)
+ self.commands[func.__name__] = Command(self, wrapped, *args, **kwargs)
def destroy(self):
map(lambda k: k.remove(), self._events)
|
disco/bot/plugin.py
|
ReplaceText(target='wrapped' @(115,53)->(115,57))
|
class Plugin(LoggingClass, PluginDeco):
def register_command(self, func, *args, **kwargs):
wrapped = functools.partial(self._dispatch, 'command', func)
self.commands[func.__name__] = Command(self, func, *args, **kwargs)
def destroy(self):
map(lambda k: k.remove(), self._events)
|
class Plugin(LoggingClass, PluginDeco):
def register_command(self, func, *args, **kwargs):
wrapped = functools.partial(self._dispatch, 'command', func)
self.commands[func.__name__] = Command(self, wrapped, *args, **kwargs)
def destroy(self):
map(lambda k: k.remove(), self._events)
|
215 |
https://:@github.com/b1naryth1ef/disco.git
|
c5848dbe8b66295598b4e6cad1c5ec8f5cc3a5fb
|
@@ -305,7 +305,7 @@ class Channel(SlottedModel, Permissible):
return
if self.can(self.client.state.me, Permissions.MANAGE_MESSAGES) and len(messages) > 2:
- for chunk in chunks(messages, 100):
+ for chunk in chunks(message_ids, 100):
self.client.api.channels_messages_delete_bulk(self.id, chunk)
else:
for msg in messages:
|
disco/types/channel.py
|
ReplaceText(target='message_ids' @(308,32)->(308,40))
|
class Channel(SlottedModel, Permissible):
return
if self.can(self.client.state.me, Permissions.MANAGE_MESSAGES) and len(messages) > 2:
for chunk in chunks(messages, 100):
self.client.api.channels_messages_delete_bulk(self.id, chunk)
else:
for msg in messages:
|
class Channel(SlottedModel, Permissible):
return
if self.can(self.client.state.me, Permissions.MANAGE_MESSAGES) and len(messages) > 2:
for chunk in chunks(message_ids, 100):
self.client.api.channels_messages_delete_bulk(self.id, chunk)
else:
for msg in messages:
|
216 |
https://:@github.com/yuru-yuri/manga-dl.git
|
5cbcbd37c4f904720d08074241a890851b94a7dd
|
@@ -26,7 +26,7 @@ if __name__ == '__main__':
if parse_args.cli:
cli = Cli(args)
# cli
- exit(0 if cli.status else 0)
+ exit(0 if cli.status else 1)
# else run GUI
app = QApplication(argv)
|
manga.py
|
ReplaceText(target='1' @(29,34)->(29,35))
|
if __name__ == '__main__':
if parse_args.cli:
cli = Cli(args)
# cli
exit(0 if cli.status else 0)
# else run GUI
app = QApplication(argv)
|
if __name__ == '__main__':
if parse_args.cli:
cli = Cli(args)
# cli
exit(0 if cli.status else 1)
# else run GUI
app = QApplication(argv)
|
217 |
https://:@github.com/yuru-yuri/manga-dl.git
|
f10777831c4c128a6f6a9bb1d9903889a43ad1df
|
@@ -44,7 +44,7 @@ class WebDriver:
driver_path = self._driver_path()
if not is_file(driver_path):
self.download_drivder()
- self.is_win() and chmod(driver_path, 0o755)
+ self.is_win() or chmod(driver_path, 0o755)
driver = webdriver.Chrome(executable_path=driver_path)
driver.set_window_size(500, 600)
return driver
|
src/base_classes/web_driver.py
|
ReplaceText(target='or' @(47,22)->(47,25))
|
class WebDriver:
driver_path = self._driver_path()
if not is_file(driver_path):
self.download_drivder()
self.is_win() and chmod(driver_path, 0o755)
driver = webdriver.Chrome(executable_path=driver_path)
driver.set_window_size(500, 600)
return driver
|
class WebDriver:
driver_path = self._driver_path()
if not is_file(driver_path):
self.download_drivder()
self.is_win() or chmod(driver_path, 0o755)
driver = webdriver.Chrome(executable_path=driver_path)
driver.set_window_size(500, 600)
return driver
|
218 |
https://:@github.com/veg/bioext.git
|
5fbb2b3fb52d1e59b787713f9d9521124cb2f24b
|
@@ -217,7 +217,7 @@ def _translate_gapped(seq, *args, **kwds):
elif gaps:
protein += '-' * gaps
gaps = 0
- lwr = j
+ lwr = i
if gaps:
protein += '-' * gaps
else:
|
lib/BioExt/_util.py
|
ReplaceText(target='i' @(220,18)->(220,19))
|
def _translate_gapped(seq, *args, **kwds):
elif gaps:
protein += '-' * gaps
gaps = 0
lwr = j
if gaps:
protein += '-' * gaps
else:
|
def _translate_gapped(seq, *args, **kwds):
elif gaps:
protein += '-' * gaps
gaps = 0
lwr = i
if gaps:
protein += '-' * gaps
else:
|
219 |
https://:@github.com/python-trio/trio.git
|
82253caa8bd59b5b3bbb0ba61d196289c967f838
|
@@ -371,7 +371,7 @@ def test_waitid_eintr():
sync_wait_reapable(sleeper.pid)
assert sleeper.wait(timeout=1) == -9
finally:
- if sleeper.returncode is not None:
+ if sleeper.returncode is None:
sleeper.kill()
sleeper.wait()
signal.signal(signal.SIGALRM, old_sigalrm)
|
trio/tests/test_subprocess.py
|
ReplaceText(target=' is ' @(374,29)->(374,37))
|
def test_waitid_eintr():
sync_wait_reapable(sleeper.pid)
assert sleeper.wait(timeout=1) == -9
finally:
if sleeper.returncode is not None:
sleeper.kill()
sleeper.wait()
signal.signal(signal.SIGALRM, old_sigalrm)
|
def test_waitid_eintr():
sync_wait_reapable(sleeper.pid)
assert sleeper.wait(timeout=1) == -9
finally:
if sleeper.returncode is None:
sleeper.kill()
sleeper.wait()
signal.signal(signal.SIGALRM, old_sigalrm)
|
220 |
https://:@github.com/python-trio/trio.git
|
94a587f758f0597cd790505ca7bfbec17a247fb1
|
@@ -37,7 +37,7 @@ def test_warn_deprecated(recwarn_always):
assert "water instead" in got.message.args[0]
assert "/issues/1" in got.message.args[0]
assert got.filename == filename
- assert got.lineno == lineno + 1
+ assert got.lineno == lineno - 1
def test_warn_deprecated_no_instead_or_issue(recwarn_always):
|
trio/tests/test_deprecate.py
|
ReplaceText(target='-' @(40,32)->(40,33))
|
def test_warn_deprecated(recwarn_always):
assert "water instead" in got.message.args[0]
assert "/issues/1" in got.message.args[0]
assert got.filename == filename
assert got.lineno == lineno + 1
def test_warn_deprecated_no_instead_or_issue(recwarn_always):
|
def test_warn_deprecated(recwarn_always):
assert "water instead" in got.message.args[0]
assert "/issues/1" in got.message.args[0]
assert got.filename == filename
assert got.lineno == lineno - 1
def test_warn_deprecated_no_instead_or_issue(recwarn_always):
|
221 |
https://:@github.com/python-trio/trio.git
|
8d9effc1d32b8ef2f7a7a02c1b02a4de3f5f8e3d
|
@@ -90,7 +90,7 @@ def ki_protection_enabled(frame):
if frame.f_code.co_name == "__del__":
return True
frame = frame.f_back
- return False
+ return True
def currently_ki_protected():
|
trio/_core/_ki.py
|
ReplaceText(target='True' @(93,11)->(93,16))
|
def ki_protection_enabled(frame):
if frame.f_code.co_name == "__del__":
return True
frame = frame.f_back
return False
def currently_ki_protected():
|
def ki_protection_enabled(frame):
if frame.f_code.co_name == "__del__":
return True
frame = frame.f_back
return True
def currently_ki_protected():
|
222 |
https://:@github.com/tkrajina/srtm.py.git
|
b1e1c673b613780c6a7151ab3db53461e2b668c7
|
@@ -187,7 +187,7 @@ class GeoElevationData:
for row in range(height):
for column in range(width):
latitude = latitude_from + float(row) / height * (latitude_to - latitude_from)
- longitude = longitude_from + float(column) / height * (longitude_to - longitude_from)
+ longitude = longitude_from + float(column) / width * (longitude_to - longitude_from)
elevation = self.get_elevation(latitude, longitude)
array[row,column] = elevation
|
srtm/data.py
|
ReplaceText(target='width' @(190,65)->(190,71))
|
class GeoElevationData:
for row in range(height):
for column in range(width):
latitude = latitude_from + float(row) / height * (latitude_to - latitude_from)
longitude = longitude_from + float(column) / height * (longitude_to - longitude_from)
elevation = self.get_elevation(latitude, longitude)
array[row,column] = elevation
|
class GeoElevationData:
for row in range(height):
for column in range(width):
latitude = latitude_from + float(row) / height * (latitude_to - latitude_from)
longitude = longitude_from + float(column) / width * (longitude_to - longitude_from)
elevation = self.get_elevation(latitude, longitude)
array[row,column] = elevation
|
223 |
https://:@github.com/tkrajina/srtm.py.git
|
771bb0e73e7b478603f73644a8c8f441b0e02e9f
|
@@ -208,7 +208,7 @@ class GeoElevationData:
for row in range(height):
for column in range(width):
latitude = latitude_from + float(row) / height * (latitude_to - latitude_from)
- longitude = longitude_from + float(column) / height * (longitude_to - longitude_from)
+ longitude = longitude_from + float(column) / width * (longitude_to - longitude_from)
elevation = self.get_elevation(latitude, longitude)
if elevation == None:
color = unknown_color
|
srtm/data.py
|
ReplaceText(target='width' @(211,65)->(211,71))
|
class GeoElevationData:
for row in range(height):
for column in range(width):
latitude = latitude_from + float(row) / height * (latitude_to - latitude_from)
longitude = longitude_from + float(column) / height * (longitude_to - longitude_from)
elevation = self.get_elevation(latitude, longitude)
if elevation == None:
color = unknown_color
|
class GeoElevationData:
for row in range(height):
for column in range(width):
latitude = latitude_from + float(row) / height * (latitude_to - latitude_from)
longitude = longitude_from + float(column) / width * (longitude_to - longitude_from)
elevation = self.get_elevation(latitude, longitude)
if elevation == None:
color = unknown_color
|
224 |
https://:@github.com/mwouts/nbrmd.git
|
628e3f952bb786220efc1e14b976809bf8ea96be
|
@@ -218,7 +218,7 @@ def readf(nb_file):
'Expected extensions are {}'.format(nb_file,
notebook_extensions))
with io.open(nb_file, encoding='utf-8') as fp:
- return read(nb_file, as_version=4, ext=ext)
+ return read(fp, as_version=4, ext=ext)
def writef(nb, nb_file):
|
nbrmd/nbrmd.py
|
ReplaceText(target='fp' @(221,20)->(221,27))
|
def readf(nb_file):
'Expected extensions are {}'.format(nb_file,
notebook_extensions))
with io.open(nb_file, encoding='utf-8') as fp:
return read(nb_file, as_version=4, ext=ext)
def writef(nb, nb_file):
|
def readf(nb_file):
'Expected extensions are {}'.format(nb_file,
notebook_extensions))
with io.open(nb_file, encoding='utf-8') as fp:
return read(fp, as_version=4, ext=ext)
def writef(nb, nb_file):
|
225 |
https://:@github.com/mwouts/nbrmd.git
|
366614bce3dc65c6e88bcf37c4e2cd31ea36de92
|
@@ -63,7 +63,7 @@ def test_load_save_rename_nbpy(nb_file, tmpdir):
cm.save(model=dict(type='notebook', content=nb), path=tmp_ipynb)
# rename nbpy
- cm.rename(tmp_ipynb, 'new.nb.py')
+ cm.rename(tmp_nbpy, 'new.nb.py')
assert not os.path.isfile(str(tmpdir.join(tmp_ipynb)))
assert not os.path.isfile(str(tmpdir.join(tmp_nbpy)))
|
tests/test_contentsmanager.py
|
ReplaceText(target='tmp_nbpy' @(66,14)->(66,23))
|
def test_load_save_rename_nbpy(nb_file, tmpdir):
cm.save(model=dict(type='notebook', content=nb), path=tmp_ipynb)
# rename nbpy
cm.rename(tmp_ipynb, 'new.nb.py')
assert not os.path.isfile(str(tmpdir.join(tmp_ipynb)))
assert not os.path.isfile(str(tmpdir.join(tmp_nbpy)))
|
def test_load_save_rename_nbpy(nb_file, tmpdir):
cm.save(model=dict(type='notebook', content=nb), path=tmp_ipynb)
# rename nbpy
cm.rename(tmp_nbpy, 'new.nb.py')
assert not os.path.isfile(str(tmpdir.join(tmp_ipynb)))
assert not os.path.isfile(str(tmpdir.join(tmp_nbpy)))
|
226 |
https://:@github.com/mwouts/nbrmd.git
|
171256b67818ab86f43e35065fc828234e9abf98
|
@@ -163,7 +163,7 @@ class CellExporter():
return True
if all([line.startswith('#') for line in self.source]):
return True
- if CellReader(self.ext).read(source)[1] != len(source):
+ if CellReader(self.ext).read(source)[1] < len(source):
return True
return False
|
jupytext/cell_to_text.py
|
ReplaceText(target='<' @(166,48)->(166,50))
|
class CellExporter():
return True
if all([line.startswith('#') for line in self.source]):
return True
if CellReader(self.ext).read(source)[1] != len(source):
return True
return False
|
class CellExporter():
return True
if all([line.startswith('#') for line in self.source]):
return True
if CellReader(self.ext).read(source)[1] < len(source):
return True
return False
|
227 |
https://:@github.com/mwouts/nbrmd.git
|
8048c8d0a09eab875376c7a4d1efb3cd886b8d3c
|
@@ -122,7 +122,7 @@ def metadata_and_cell_to_header(notebook, text_format, ext):
if lines_to_next_cell is None and notebook.cells:
lines_to_next_cell = pep8_lines_between_cells(header, notebook.cells[0], ext)
else:
- lines_to_next_cell = 0
+ lines_to_next_cell = 1
header.extend([''] * lines_to_next_cell)
|
jupytext/header.py
|
ReplaceText(target='1' @(125,29)->(125,30))
|
def metadata_and_cell_to_header(notebook, text_format, ext):
if lines_to_next_cell is None and notebook.cells:
lines_to_next_cell = pep8_lines_between_cells(header, notebook.cells[0], ext)
else:
lines_to_next_cell = 0
header.extend([''] * lines_to_next_cell)
|
def metadata_and_cell_to_header(notebook, text_format, ext):
if lines_to_next_cell is None and notebook.cells:
lines_to_next_cell = pep8_lines_between_cells(header, notebook.cells[0], ext)
else:
lines_to_next_cell = 1
header.extend([''] * lines_to_next_cell)
|
228 |
https://:@github.com/mwouts/nbrmd.git
|
120bebc0d37792cfa1418476d5f26809b996dca8
|
@@ -615,7 +615,7 @@ def test_sync(nb_file, tmpdir):
compare_notebooks(nb, nb2)
# ipynb must be older than py file, otherwise our Contents Manager will complain
- assert os.path.getmtime(tmp_ipynb) < os.path.getmtime(tmp_py)
+ assert os.path.getmtime(tmp_ipynb) <= os.path.getmtime(tmp_py)
@pytest.mark.parametrize('nb_file,ext',
|
tests/test_cli.py
|
ReplaceText(target='<=' @(618,39)->(618,40))
|
def test_sync(nb_file, tmpdir):
compare_notebooks(nb, nb2)
# ipynb must be older than py file, otherwise our Contents Manager will complain
assert os.path.getmtime(tmp_ipynb) < os.path.getmtime(tmp_py)
@pytest.mark.parametrize('nb_file,ext',
|
def test_sync(nb_file, tmpdir):
compare_notebooks(nb, nb2)
# ipynb must be older than py file, otherwise our Contents Manager will complain
assert os.path.getmtime(tmp_ipynb) <= os.path.getmtime(tmp_py)
@pytest.mark.parametrize('nb_file,ext',
|
229 |
https://:@github.com/mwouts/nbrmd.git
|
dfa96996445cbc7514b93337dbf94d592ba06bad
|
@@ -19,4 +19,4 @@ def test_identity_source_write_read(nb_file, ext):
R = jupytext.writes(nb1, ext)
nb2 = jupytext.reads(R, ext)
- compare_notebooks(nb1, nb2)
+ compare_notebooks(nb2, nb1)
|
tests/test_ipynb_to_R.py
|
ArgSwap(idxs=0<->1 @(22,4)->(22,21))
|
def test_identity_source_write_read(nb_file, ext):
R = jupytext.writes(nb1, ext)
nb2 = jupytext.reads(R, ext)
compare_notebooks(nb1, nb2)
|
def test_identity_source_write_read(nb_file, ext):
R = jupytext.writes(nb1, ext)
nb2 = jupytext.reads(R, ext)
compare_notebooks(nb2, nb1)
|
230 |
https://:@github.com/mwouts/nbrmd.git
|
dfa96996445cbc7514b93337dbf94d592ba06bad
|
@@ -16,4 +16,4 @@ def test_identity_source_write_read(nb_file):
py = jupytext.writes(nb1, 'py')
nb2 = jupytext.reads(py, 'py')
- compare_notebooks(nb1, nb2)
+ compare_notebooks(nb2, nb1)
|
tests/test_ipynb_to_py.py
|
ArgSwap(idxs=0<->1 @(19,4)->(19,21))
|
def test_identity_source_write_read(nb_file):
py = jupytext.writes(nb1, 'py')
nb2 = jupytext.reads(py, 'py')
compare_notebooks(nb1, nb2)
|
def test_identity_source_write_read(nb_file):
py = jupytext.writes(nb1, 'py')
nb2 = jupytext.reads(py, 'py')
compare_notebooks(nb2, nb1)
|
231 |
https://:@github.com/mwouts/nbrmd.git
|
dfa96996445cbc7514b93337dbf94d592ba06bad
|
@@ -16,4 +16,4 @@ def test_identity_source_write_read(nb_file):
rmd = jupytext.writes(nb1, 'Rmd')
nb2 = jupytext.reads(rmd, 'Rmd')
- compare_notebooks(nb1, nb2, 'Rmd')
+ compare_notebooks(nb2, nb1, 'Rmd')
|
tests/test_ipynb_to_rmd.py
|
ArgSwap(idxs=0<->1 @(19,4)->(19,21))
|
def test_identity_source_write_read(nb_file):
rmd = jupytext.writes(nb1, 'Rmd')
nb2 = jupytext.reads(rmd, 'Rmd')
compare_notebooks(nb1, nb2, 'Rmd')
|
def test_identity_source_write_read(nb_file):
rmd = jupytext.writes(nb1, 'Rmd')
nb2 = jupytext.reads(rmd, 'Rmd')
compare_notebooks(nb2, nb1, 'Rmd')
|
232 |
https://:@github.com/splunk/splunk-sdk-python.git
|
68f87378d5bac7efc97cf118f6754b5a6de73fa3
|
@@ -48,7 +48,7 @@ class EventWriter(object):
else:
self._out = TextIOWrapper(output)
- if isinstance(output, TextIOBase):
+ if isinstance(error, TextIOBase):
self._err = error
else:
self._err = TextIOWrapper(error)
|
splunklib/modularinput/event_writer.py
|
ReplaceText(target='error' @(51,22)->(51,28))
|
class EventWriter(object):
else:
self._out = TextIOWrapper(output)
if isinstance(output, TextIOBase):
self._err = error
else:
self._err = TextIOWrapper(error)
|
class EventWriter(object):
else:
self._out = TextIOWrapper(output)
if isinstance(error, TextIOBase):
self._err = error
else:
self._err = TextIOWrapper(error)
|
233 |
https://:@github.com/tgbugs/pyontutils.git
|
78db71bb8b163794ef9bafb5a0ee50453d29971a
|
@@ -56,7 +56,7 @@ def sysidpath(ignore_options=False):
)
for option in options:
if (option.exists() and
- os.access(options, os.R_OK) and
+ os.access(option, os.R_OK) and
option.stat().st_size > 0):
return option
|
pyontutils/utils.py
|
ReplaceText(target='option' @(59,26)->(59,33))
|
def sysidpath(ignore_options=False):
)
for option in options:
if (option.exists() and
os.access(options, os.R_OK) and
option.stat().st_size > 0):
return option
|
def sysidpath(ignore_options=False):
)
for option in options:
if (option.exists() and
os.access(option, os.R_OK) and
option.stat().st_size > 0):
return option
|
234 |
https://:@github.com/holgern/beem.git
|
45ae6dc0380434d3544f5588fa24b379f5a62541
|
@@ -224,7 +224,7 @@ class Blockchain(object):
else:
self.steem.rpc.get_block(blocknum, add_to_queue=True)
latest_block = blocknum
- if batches > 1:
+ if batches >= 1:
latest_block += 1
if latest_block <= head_block:
if self.steem.rpc.get_use_appbase():
|
beem/blockchain.py
|
ReplaceText(target='>=' @(227,31)->(227,32))
|
class Blockchain(object):
else:
self.steem.rpc.get_block(blocknum, add_to_queue=True)
latest_block = blocknum
if batches > 1:
latest_block += 1
if latest_block <= head_block:
if self.steem.rpc.get_use_appbase():
|
class Blockchain(object):
else:
self.steem.rpc.get_block(blocknum, add_to_queue=True)
latest_block = blocknum
if batches >= 1:
latest_block += 1
if latest_block <= head_block:
if self.steem.rpc.get_use_appbase():
|
235 |
https://:@github.com/holgern/beem.git
|
9eb381d23d582979934e0d3256447aef7d67de55
|
@@ -509,7 +509,7 @@ class Testcases(unittest.TestCase):
op_num = account.estimate_virtual_op_num(block_num, stop_diff=0.1, max_count=100)
if op_num > 0:
op_num -= 1
- self.assertTrue(op_num < i)
+ self.assertTrue(op_num <= i)
i += 1
last_block = new_block
|
tests/beem/test_account.py
|
ReplaceText(target='<=' @(512,35)->(512,36))
|
class Testcases(unittest.TestCase):
op_num = account.estimate_virtual_op_num(block_num, stop_diff=0.1, max_count=100)
if op_num > 0:
op_num -= 1
self.assertTrue(op_num < i)
i += 1
last_block = new_block
|
class Testcases(unittest.TestCase):
op_num = account.estimate_virtual_op_num(block_num, stop_diff=0.1, max_count=100)
if op_num > 0:
op_num -= 1
self.assertTrue(op_num <= i)
i += 1
last_block = new_block
|
236 |
https://:@github.com/holgern/beem.git
|
5f156fdf5a75367c32d85efb02e456abfa2719f6
|
@@ -714,6 +714,6 @@ class RecentByPath(list):
comments = []
for reply in replies:
post = state["content"][reply]
- if category is None or (category is not None and post["category"] != category):
+ if category is None or (category is not None and post["category"] == category):
comments.append(Comment(post, lazy=True, steem_instance=self.steem))
super(RecentByPath, self).__init__(comments)
|
beem/comment.py
|
ReplaceText(target='==' @(717,78)->(717,80))
|
class RecentByPath(list):
comments = []
for reply in replies:
post = state["content"][reply]
if category is None or (category is not None and post["category"] != category):
comments.append(Comment(post, lazy=True, steem_instance=self.steem))
super(RecentByPath, self).__init__(comments)
|
class RecentByPath(list):
comments = []
for reply in replies:
post = state["content"][reply]
if category is None or (category is not None and post["category"] == category):
comments.append(Comment(post, lazy=True, steem_instance=self.steem))
super(RecentByPath, self).__init__(comments)
|
237 |
https://:@github.com/holgern/beem.git
|
7b9ee9c75cbf285d18c1b4913cffe3798444ac30
|
@@ -50,7 +50,7 @@ class RC(object):
resource_count["resource_state_bytes"] += state_bytes_count
resource_count["resource_new_accounts"] = new_account_op_count
if market_op_count > 0:
- resource_count["resource_market_bytes"] = market_op_count
+ resource_count["resource_market_bytes"] = tx_size
return resource_count
def comment_dict(self, comment_dict):
|
beem/rc.py
|
ReplaceText(target='tx_size' @(53,54)->(53,69))
|
class RC(object):
resource_count["resource_state_bytes"] += state_bytes_count
resource_count["resource_new_accounts"] = new_account_op_count
if market_op_count > 0:
resource_count["resource_market_bytes"] = market_op_count
return resource_count
def comment_dict(self, comment_dict):
|
class RC(object):
resource_count["resource_state_bytes"] += state_bytes_count
resource_count["resource_new_accounts"] = new_account_op_count
if market_op_count > 0:
resource_count["resource_market_bytes"] = tx_size
return resource_count
def comment_dict(self, comment_dict):
|
238 |
https://:@github.com/holgern/beem.git
|
9e96dc84f9965b25b6b2a56a1bca7b7652b93be8
|
@@ -1456,7 +1456,7 @@ class Steem(object):
'key_auths': active_key_authority,
"address_auths": [],
'weight_threshold': 1},
- 'posting': {'account_auths': active_accounts_authority,
+ 'posting': {'account_auths': posting_accounts_authority,
'key_auths': posting_key_authority,
"address_auths": [],
'weight_threshold': 1},
|
beem/steem.py
|
ReplaceText(target='posting_accounts_authority' @(1459,41)->(1459,66))
|
class Steem(object):
'key_auths': active_key_authority,
"address_auths": [],
'weight_threshold': 1},
'posting': {'account_auths': active_accounts_authority,
'key_auths': posting_key_authority,
"address_auths": [],
'weight_threshold': 1},
|
class Steem(object):
'key_auths': active_key_authority,
"address_auths": [],
'weight_threshold': 1},
'posting': {'account_auths': posting_accounts_authority,
'key_auths': posting_key_authority,
"address_auths": [],
'weight_threshold': 1},
|
239 |
https://:@github.com/holgern/beem.git
|
8e4214f27e746be7ed01c5d644911c830bfe988a
|
@@ -135,7 +135,7 @@ class Block(BlockchainObject):
if ops_ops is None:
ops = None
else:
- ops = ops["ops"]
+ ops = ops_ops["ops"]
except ApiNotSupported:
ops = self.steem.rpc.get_ops_in_block(self.identifier, self.only_virtual_ops, api="condenser")
else:
|
beem/block.py
|
ReplaceText(target='ops_ops' @(138,30)->(138,33))
|
class Block(BlockchainObject):
if ops_ops is None:
ops = None
else:
ops = ops["ops"]
except ApiNotSupported:
ops = self.steem.rpc.get_ops_in_block(self.identifier, self.only_virtual_ops, api="condenser")
else:
|
class Block(BlockchainObject):
if ops_ops is None:
ops = None
else:
ops = ops_ops["ops"]
except ApiNotSupported:
ops = self.steem.rpc.get_ops_in_block(self.identifier, self.only_virtual_ops, api="condenser")
else:
|
240 |
https://:@github.com/holgern/beem.git
|
d19894c85464f2bc5221e15f858326c4e8efdaf3
|
@@ -367,7 +367,7 @@ class ActiveVotes(VotesObject):
elif isinstance(authorperm, string_types):
[author, permlink] = resolve_authorperm(authorperm)
if self.steem.rpc.get_use_appbase():
- self.steem.rpc.set_next_node_on_empty_reply(True)
+ self.steem.rpc.set_next_node_on_empty_reply(False)
try:
votes = self.steem.rpc.get_active_votes(author, permlink, api="condenser")
except:
|
beem/vote.py
|
ReplaceText(target='False' @(370,60)->(370,64))
|
class ActiveVotes(VotesObject):
elif isinstance(authorperm, string_types):
[author, permlink] = resolve_authorperm(authorperm)
if self.steem.rpc.get_use_appbase():
self.steem.rpc.set_next_node_on_empty_reply(True)
try:
votes = self.steem.rpc.get_active_votes(author, permlink, api="condenser")
except:
|
class ActiveVotes(VotesObject):
elif isinstance(authorperm, string_types):
[author, permlink] = resolve_authorperm(authorperm)
if self.steem.rpc.get_use_appbase():
self.steem.rpc.set_next_node_on_empty_reply(False)
try:
votes = self.steem.rpc.get_active_votes(author, permlink, api="condenser")
except:
|
241 |
https://:@github.com/holgern/beem.git
|
8f173f0ab272a57a5c036589dcc19b8543026b3e
|
@@ -780,7 +780,7 @@ def keygen(import_word_list, strength, passphrase, path, network, role, account_
t.add_row(["Key role", role])
t.add_row(["path", path])
pubkey = ledgertx.ledgertx.get_pubkey(path, request_screen_approval=False)
- aprove_key = PrettyTable(["Approve %s Key" % r])
+ aprove_key = PrettyTable(["Approve %s Key" % role])
aprove_key.align = "l"
aprove_key.add_row([format(pubkey, "STM")])
print(aprove_key)
|
beem/cli.py
|
ReplaceText(target='role' @(783,57)->(783,58))
|
def keygen(import_word_list, strength, passphrase, path, network, role, account_
t.add_row(["Key role", role])
t.add_row(["path", path])
pubkey = ledgertx.ledgertx.get_pubkey(path, request_screen_approval=False)
aprove_key = PrettyTable(["Approve %s Key" % r])
aprove_key.align = "l"
aprove_key.add_row([format(pubkey, "STM")])
print(aprove_key)
|
def keygen(import_word_list, strength, passphrase, path, network, role, account_
t.add_row(["Key role", role])
t.add_row(["path", path])
pubkey = ledgertx.ledgertx.get_pubkey(path, request_screen_approval=False)
aprove_key = PrettyTable(["Approve %s Key" % role])
aprove_key.align = "l"
aprove_key.add_row([format(pubkey, "STM")])
print(aprove_key)
|
242 |
https://:@github.com/BlackLight/platypush.git
|
56b87f343693e89131b3b30930acfeeaec4915fe
|
@@ -101,7 +101,7 @@ class MqttBackend(Backend):
format(response_topic, response))
client = get_plugin('mqtt')
- client.send_message(topic=self.topic, msg=msg, host=self.host,
+ client.send_message(topic=self.topic, msg=response, host=self.host,
port=self.port, username=self.username,
password=self.password, tls_cafile=self.tls_cafile,
tls_certfile=self.tls_certfile,
|
platypush/backend/mqtt.py
|
ReplaceText(target='response' @(104,58)->(104,61))
|
class MqttBackend(Backend):
format(response_topic, response))
client = get_plugin('mqtt')
client.send_message(topic=self.topic, msg=msg, host=self.host,
port=self.port, username=self.username,
password=self.password, tls_cafile=self.tls_cafile,
tls_certfile=self.tls_certfile,
|
class MqttBackend(Backend):
format(response_topic, response))
client = get_plugin('mqtt')
client.send_message(topic=self.topic, msg=response, host=self.host,
port=self.port, username=self.username,
password=self.password, tls_cafile=self.tls_cafile,
tls_certfile=self.tls_certfile,
|
243 |
https://:@github.com/BlackLight/platypush.git
|
c9dc1aac44ec4df76071092f4a17601d97955a38
|
@@ -47,7 +47,7 @@ class SensorEnvirophatBackend(SensorBackend):
if enabled and sensor in sensors and sensors[sensor] != self._last_read.get(sensor)
}
- self._last_read = ret
+ self._last_read = sensors
return ret
|
platypush/backend/sensor/envirophat.py
|
ReplaceText(target='sensors' @(50,26)->(50,29))
|
class SensorEnvirophatBackend(SensorBackend):
if enabled and sensor in sensors and sensors[sensor] != self._last_read.get(sensor)
}
self._last_read = ret
return ret
|
class SensorEnvirophatBackend(SensorBackend):
if enabled and sensor in sensors and sensors[sensor] != self._last_read.get(sensor)
}
self._last_read = sensors
return ret
|
244 |
https://:@github.com/BlackLight/platypush.git
|
7f440a9160619c699f7e8e58b3c0fad3c80ceaf3
|
@@ -46,7 +46,7 @@ class GpioPlugin(Plugin):
import RPi.GPIO as GPIO
with self._init_lock:
- if self._initialized or GPIO.getmode():
+ if self._initialized and GPIO.getmode():
return
GPIO.setmode(self.mode)
|
platypush/plugins/gpio/__init__.py
|
ReplaceText(target='and' @(49,33)->(49,35))
|
class GpioPlugin(Plugin):
import RPi.GPIO as GPIO
with self._init_lock:
if self._initialized or GPIO.getmode():
return
GPIO.setmode(self.mode)
|
class GpioPlugin(Plugin):
import RPi.GPIO as GPIO
with self._init_lock:
if self._initialized and GPIO.getmode():
return
GPIO.setmode(self.mode)
|
245 |
https://:@github.com/BlackLight/platypush.git
|
c26d456109fea166f8fb5da25aed1d4fb7fc94ab
|
@@ -175,7 +175,7 @@ class SensorBackend(Backend):
def process_data(self, data, new_data):
if new_data:
- self.bus.post(SensorDataChangeEvent(data=data, source=self.plugin or self.__class__.__name__))
+ self.bus.post(SensorDataChangeEvent(data=new_data, source=self.plugin or self.__class__.__name__))
def run(self):
super().run()
|
platypush/backend/sensor/__init__.py
|
ReplaceText(target='new_data' @(178,53)->(178,57))
|
class SensorBackend(Backend):
def process_data(self, data, new_data):
if new_data:
self.bus.post(SensorDataChangeEvent(data=data, source=self.plugin or self.__class__.__name__))
def run(self):
super().run()
|
class SensorBackend(Backend):
def process_data(self, data, new_data):
if new_data:
self.bus.post(SensorDataChangeEvent(data=new_data, source=self.plugin or self.__class__.__name__))
def run(self):
super().run()
|
246 |
https://:@github.com/chrisjsewell/ipypublish.git
|
cfb66fb74d48cbcc99c695b64fe6336213877e04
|
@@ -52,7 +52,7 @@ class LatexDocLinks(Preprocessor):
': {}'.format(bib))
else:
external_files.append(bib)
- resources['bibliopath'] = external_files
+ resources['bibliopath'] = bib
nb.metadata.latex_doc.bibliography = os.path.join(self.filesfolder,
os.path.basename(bib))
|
ipypublish/preprocessors/latex_doc.py
|
ReplaceText(target='bib' @(55,46)->(55,60))
|
class LatexDocLinks(Preprocessor):
': {}'.format(bib))
else:
external_files.append(bib)
resources['bibliopath'] = external_files
nb.metadata.latex_doc.bibliography = os.path.join(self.filesfolder,
os.path.basename(bib))
|
class LatexDocLinks(Preprocessor):
': {}'.format(bib))
else:
external_files.append(bib)
resources['bibliopath'] = bib
nb.metadata.latex_doc.bibliography = os.path.join(self.filesfolder,
os.path.basename(bib))
|
247 |
https://:@github.com/pazz/alot.git
|
dd7b2a15495ce5fddcac0d34c14e5ef9f1032482
|
@@ -64,7 +64,7 @@ class ThreadlineWidget(urwid.AttrMap):
mailcountstring = "(%d)" % self.thread.get_total_messages()
else:
mailcountstring = "(?)"
- datestring = pad(mailcountstring)
+ mailcountstring = pad(mailcountstring)
width = len(mailcountstring)
mailcount_w = AttrFlipWidget(urwid.Text(mailcountstring),
struct['mailcount'])
|
alot/widgets/search.py
|
ReplaceText(target='mailcountstring' @(67,12)->(67,22))
|
class ThreadlineWidget(urwid.AttrMap):
mailcountstring = "(%d)" % self.thread.get_total_messages()
else:
mailcountstring = "(?)"
datestring = pad(mailcountstring)
width = len(mailcountstring)
mailcount_w = AttrFlipWidget(urwid.Text(mailcountstring),
struct['mailcount'])
|
class ThreadlineWidget(urwid.AttrMap):
mailcountstring = "(%d)" % self.thread.get_total_messages()
else:
mailcountstring = "(?)"
mailcountstring = pad(mailcountstring)
width = len(mailcountstring)
mailcount_w = AttrFlipWidget(urwid.Text(mailcountstring),
struct['mailcount'])
|
248 |
https://:@github.com/luozhouyang/python-string-similarity.git
|
5f6717fe6d7cae48a664eaa76a47e3388ffd0cf2
|
@@ -33,7 +33,7 @@ class Levenshtein(MetricStringDistance):
if len(s0) == 0:
return len(s1)
if len(s1) == 0:
- return len(s1)
+ return len(s0)
v0 = [0] * (len(s1) + 1)
v1 = [0] * (len(s1) + 1)
|
strsimpy/levenshtein.py
|
ReplaceText(target='s0' @(36,23)->(36,25))
|
class Levenshtein(MetricStringDistance):
if len(s0) == 0:
return len(s1)
if len(s1) == 0:
return len(s1)
v0 = [0] * (len(s1) + 1)
v1 = [0] * (len(s1) + 1)
|
class Levenshtein(MetricStringDistance):
if len(s0) == 0:
return len(s1)
if len(s1) == 0:
return len(s0)
v0 = [0] * (len(s1) + 1)
v1 = [0] * (len(s1) + 1)
|
249 |
https://:@github.com/ElementsProject/lightning.git
|
84b9e3e72b2bf8590603072d709a3ea294dd5483
|
@@ -943,7 +943,7 @@ def test_logging(node_factory):
def check_new_log():
log2 = open(logpath).readlines()
- return len(log2) > 1 and log2[0].endswith("Started log due to SIGHUP\n")
+ return len(log2) > 0 and log2[0].endswith("Started log due to SIGHUP\n")
wait_for(check_new_log)
|
tests/test_misc.py
|
ReplaceText(target='0' @(946,27)->(946,28))
|
def test_logging(node_factory):
def check_new_log():
log2 = open(logpath).readlines()
return len(log2) > 1 and log2[0].endswith("Started log due to SIGHUP\n")
wait_for(check_new_log)
|
def test_logging(node_factory):
def check_new_log():
log2 = open(logpath).readlines()
return len(log2) > 0 and log2[0].endswith("Started log due to SIGHUP\n")
wait_for(check_new_log)
|
250 |
https://:@github.com/ElementsProject/lightning.git
|
c8579b99d01d8b738feadd2d9f541daba5230a63
|
@@ -305,7 +305,7 @@ other types. Since 'msgtype' is almost identical, it inherits from this too.
def read(self, io_in: BufferedIOBase, otherfields: Dict[str, Any]) -> Optional[Dict[str, Any]]:
vals = {}
for field in self.fields:
- val = field.fieldtype.read(io_in, otherfields)
+ val = field.fieldtype.read(io_in, vals)
if val is None:
# If first field fails to read, we return None.
if field == self.fields[0]:
|
contrib/pyln-proto/pyln/proto/message/message.py
|
ReplaceText(target='vals' @(308,46)->(308,57))
|
other types. Since 'msgtype' is almost identical, it inherits from this too.
def read(self, io_in: BufferedIOBase, otherfields: Dict[str, Any]) -> Optional[Dict[str, Any]]:
vals = {}
for field in self.fields:
val = field.fieldtype.read(io_in, otherfields)
if val is None:
# If first field fails to read, we return None.
if field == self.fields[0]:
|
other types. Since 'msgtype' is almost identical, it inherits from this too.
def read(self, io_in: BufferedIOBase, otherfields: Dict[str, Any]) -> Optional[Dict[str, Any]]:
vals = {}
for field in self.fields:
val = field.fieldtype.read(io_in, vals)
if val is None:
# If first field fails to read, we return None.
if field == self.fields[0]:
|
251 |
https://:@github.com/funilrys/PyFunceble.git
|
c766732abdf31c0c1ce283ee9aa2ec32f0ac7829
|
@@ -301,7 +301,7 @@ class Generate(object): # pragma: no cover
regex_blogger = ["create-blog.g?", "87065", "doesn’t exist"]
if self.tested == PyFunceble.CONFIGURATION["domain"]:
- url_to_get = "http://%s" & self.tested
+ url_to_get = "http://%s" % self.tested
else:
url_to_get = self.tested
|
PyFunceble/generate.py
|
ReplaceText(target='%' @(304,37)->(304,38))
|
class Generate(object): # pragma: no cover
regex_blogger = ["create-blog.g?", "87065", "doesn’t exist"]
if self.tested == PyFunceble.CONFIGURATION["domain"]:
url_to_get = "http://%s" & self.tested
else:
url_to_get = self.tested
|
class Generate(object): # pragma: no cover
regex_blogger = ["create-blog.g?", "87065", "doesn’t exist"]
if self.tested == PyFunceble.CONFIGURATION["domain"]:
url_to_get = "http://%s" % self.tested
else:
url_to_get = self.tested
|
252 |
https://:@github.com/funilrys/PyFunceble.git
|
08400b05f5f3e20c24c9222a9cb9216d1a76aea5
|
@@ -80,7 +80,7 @@ class Load(object): # pylint: disable=too-few-public-methods
def __init__(self, path_to_config):
self.path_to_config = path_to_config
- if path_to_config.endswith(directory_separator):
+ if not path_to_config.endswith(directory_separator):
self.path_to_config += directory_separator
self.path_to_config += PyFunceble.CONFIGURATION_FILENAME
|
PyFunceble/config.py
|
ReplaceText(target='not ' @(83,11)->(83,11))
|
class Load(object): # pylint: disable=too-few-public-methods
def __init__(self, path_to_config):
self.path_to_config = path_to_config
if path_to_config.endswith(directory_separator):
self.path_to_config += directory_separator
self.path_to_config += PyFunceble.CONFIGURATION_FILENAME
|
class Load(object): # pylint: disable=too-few-public-methods
def __init__(self, path_to_config):
self.path_to_config = path_to_config
if not path_to_config.endswith(directory_separator):
self.path_to_config += directory_separator
self.path_to_config += PyFunceble.CONFIGURATION_FILENAME
|
253 |
https://:@github.com/funilrys/PyFunceble.git
|
fb9465acb6a7124aa84b8c419c9709243e001f94
|
@@ -96,7 +96,7 @@ class Clean:
if (
number_of_tested == 0
or list_to_test[number_of_tested - 1] == list_to_test[-1]
- or number_of_tested == len(list_to_test)
+ or number_of_tested >= len(list_to_test)
):
# * If the number of tested is null,
# or
|
PyFunceble/clean.py
|
ReplaceText(target='>=' @(99,40)->(99,42))
|
class Clean:
if (
number_of_tested == 0
or list_to_test[number_of_tested - 1] == list_to_test[-1]
or number_of_tested == len(list_to_test)
):
# * If the number of tested is null,
# or
|
class Clean:
if (
number_of_tested == 0
or list_to_test[number_of_tested - 1] == list_to_test[-1]
or number_of_tested >= len(list_to_test)
):
# * If the number of tested is null,
# or
|
254 |
https://:@github.com/funilrys/PyFunceble.git
|
a9e45c83e13eca92389d0be56634cc45ef601bc1
|
@@ -92,7 +92,7 @@ class DBTypeDownloader(DownloaderBase):
f"{PyFunceble.OUTPUTS.db_type.files[PyFunceble.CONFIGURATION.db_type]}"
)
- if is_cloned_version and (
+ if not is_cloned_version and (
PyFunceble.CONFIGURATION.db_type not in not_supported_db_types
):
destination_dir_instance.delete()
|
PyFunceble/downloader/db_type.py
|
ReplaceText(target='not ' @(95,11)->(95,11))
|
class DBTypeDownloader(DownloaderBase):
f"{PyFunceble.OUTPUTS.db_type.files[PyFunceble.CONFIGURATION.db_type]}"
)
if is_cloned_version and (
PyFunceble.CONFIGURATION.db_type not in not_supported_db_types
):
destination_dir_instance.delete()
|
class DBTypeDownloader(DownloaderBase):
f"{PyFunceble.OUTPUTS.db_type.files[PyFunceble.CONFIGURATION.db_type]}"
)
if not is_cloned_version and (
PyFunceble.CONFIGURATION.db_type not in not_supported_db_types
):
destination_dir_instance.delete()
|
255 |
https://:@github.com/funilrys/PyFunceble.git
|
76bc73c57f4ac2865fc89d70ad330be2aa501c08
|
@@ -154,7 +154,7 @@ class Credential:
regex = f"{name}=.*"
if not content:
- content += f"{to_write}\n"
+ content = f"{to_write}\n"
continue
if PyFunceble.helpers.Regex(f"^{regex}").get_matching_list(
|
PyFunceble/engine/database/loader/credential.py
|
ReplaceText(target='=' @(157,28)->(157,30))
|
class Credential:
regex = f"{name}=.*"
if not content:
content += f"{to_write}\n"
continue
if PyFunceble.helpers.Regex(f"^{regex}").get_matching_list(
|
class Credential:
regex = f"{name}=.*"
if not content:
content = f"{to_write}\n"
continue
if PyFunceble.helpers.Regex(f"^{regex}").get_matching_list(
|
256 |
https://:@github.com/valassis-digital-media/conda-mirror.git
|
e0b34555d779f9b440512f850da39aa5a0e29ece
|
@@ -85,5 +85,5 @@ def test_handling_bad_package(tmpdir, repodata):
with bz2.BZ2File(bad_pkg_path, 'wb') as f:
f.write("This is a fake package".encode())
assert bad_pkg_name in os.listdir(bad_pkg_root)
- conda_mirror._validate_packages(repodata, local_repo_root)
+ conda_mirror._validate_packages(repodata, bad_pkg_root)
assert bad_pkg_name not in os.listdir(bad_pkg_root)
\ No newline at end of file
|
test/test_conda_mirror.py
|
ReplaceText(target='bad_pkg_root' @(88,46)->(88,61))
|
def test_handling_bad_package(tmpdir, repodata):
with bz2.BZ2File(bad_pkg_path, 'wb') as f:
f.write("This is a fake package".encode())
assert bad_pkg_name in os.listdir(bad_pkg_root)
conda_mirror._validate_packages(repodata, local_repo_root)
assert bad_pkg_name not in os.listdir(bad_pkg_root)
\ No newline at end of file
|
def test_handling_bad_package(tmpdir, repodata):
with bz2.BZ2File(bad_pkg_path, 'wb') as f:
f.write("This is a fake package".encode())
assert bad_pkg_name in os.listdir(bad_pkg_root)
conda_mirror._validate_packages(repodata, bad_pkg_root)
assert bad_pkg_name not in os.listdir(bad_pkg_root)
\ No newline at end of file
|
257 |
https://:@github.com/fact-project/pycustos.git
|
e2432cee5bba6dc97b30d0b640c0c92f3bdd6961
|
@@ -8,7 +8,7 @@ class Notifier(metaclass=ABCMeta):
self.categories = set(categories)
def handle_message(self, msg):
- if self.categories.intersection(msg.categories) and msg.level > self.level:
+ if self.categories.intersection(msg.categories) and msg.level >= self.level:
self.notify(msg)
@abstractmethod
|
custos/notify/base.py
|
ReplaceText(target='>=' @(11,70)->(11,71))
|
class Notifier(metaclass=ABCMeta):
self.categories = set(categories)
def handle_message(self, msg):
if self.categories.intersection(msg.categories) and msg.level > self.level:
self.notify(msg)
@abstractmethod
|
class Notifier(metaclass=ABCMeta):
self.categories = set(categories)
def handle_message(self, msg):
if self.categories.intersection(msg.categories) and msg.level >= self.level:
self.notify(msg)
@abstractmethod
|
258 |
https://:@github.com/danilobellini/audiolazy.git
|
3119b12f8f0b175b3f55aecf3053ed6a0d1a477a
|
@@ -249,4 +249,4 @@ def gammatone(freq, bandwidth):
freqs = tee(freq, 4)
resons = [resonator.z_exp, resonator.poles_exp] * 2
return CascadeFilter(reson(f, bw)
- for reson, f, bw in zip(bws, freqs, resons))
+ for reson, f, bw in zip(resons, freqs, bws))
|
audiolazy/lazy_auditory.py
|
ArgSwap(idxs=0<->2 @(252,43)->(252,46))
|
def gammatone(freq, bandwidth):
freqs = tee(freq, 4)
resons = [resonator.z_exp, resonator.poles_exp] * 2
return CascadeFilter(reson(f, bw)
for reson, f, bw in zip(bws, freqs, resons))
|
def gammatone(freq, bandwidth):
freqs = tee(freq, 4)
resons = [resonator.z_exp, resonator.poles_exp] * 2
return CascadeFilter(reson(f, bw)
for reson, f, bw in zip(resons, freqs, bws))
|
259 |
https://:@github.com/ojii/django-sekizai.git
|
b0b50e0713a90e84acabe1ab39e951421fd4aa7a
|
@@ -91,6 +91,6 @@ class CSSSingleFileFilter(BaseMinifierFilter):
mtime = os.path.getmtime(master)
for f in files:
fpath = media_url_to_filepath(f)
- if os.path.getmtime(fpath) > mtime:
+ if os.path.getmtime(fpath) >= mtime:
return True
return False
\ No newline at end of file
|
sekizai/filters/css.py
|
ReplaceText(target='>=' @(94,39)->(94,40))
|
class CSSSingleFileFilter(BaseMinifierFilter):
mtime = os.path.getmtime(master)
for f in files:
fpath = media_url_to_filepath(f)
if os.path.getmtime(fpath) > mtime:
return True
return False
\ No newline at end of file
|
class CSSSingleFileFilter(BaseMinifierFilter):
mtime = os.path.getmtime(master)
for f in files:
fpath = media_url_to_filepath(f)
if os.path.getmtime(fpath) >= mtime:
return True
return False
\ No newline at end of file
|
260 |
https://:@github.com/websocket-client/websocket-client.git
|
6410340fca47f258d50a34646138ba03b2b2783b
|
@@ -183,7 +183,7 @@ def _tunnel(sock, host, port, auth):
if status != 200:
raise WebSocketProxyException(
- "failed CONNECT via proxy status: %r" + status)
+ "failed CONNECT via proxy status: %r" % status)
return sock
|
websocket/_http.py
|
ReplaceText(target='%' @(186,50)->(186,51))
|
def _tunnel(sock, host, port, auth):
if status != 200:
raise WebSocketProxyException(
"failed CONNECT via proxy status: %r" + status)
return sock
|
def _tunnel(sock, host, port, auth):
if status != 200:
raise WebSocketProxyException(
"failed CONNECT via proxy status: %r" % status)
return sock
|
261 |
https://:@github.com/getsentry/raven-python.git
|
dedca8e5c98124f6a43a18986e142e8cb7ecc3cf
|
@@ -59,7 +59,7 @@ def setup_logging(handler, exclude=['raven', 'sentry.errors']):
Returns a boolean based on if logging was configured or not.
"""
logger = logging.getLogger()
- if handler.__class__ not in map(type, logger.handlers):
+ if handler.__class__ in map(type, logger.handlers):
return False
logger.addHandler(handler)
|
raven/conf/__init__.py
|
ReplaceText(target=' in ' @(62,24)->(62,32))
|
def setup_logging(handler, exclude=['raven', 'sentry.errors']):
Returns a boolean based on if logging was configured or not.
"""
logger = logging.getLogger()
if handler.__class__ not in map(type, logger.handlers):
return False
logger.addHandler(handler)
|
def setup_logging(handler, exclude=['raven', 'sentry.errors']):
Returns a boolean based on if logging was configured or not.
"""
logger = logging.getLogger()
if handler.__class__ in map(type, logger.handlers):
return False
logger.addHandler(handler)
|
262 |
https://:@github.com/getsentry/raven-python.git
|
f0ad0ca6a9de44128982de50c30157b779b69d71
|
@@ -16,7 +16,7 @@ class TransportRegistry(object):
self.register_transport(transport)
def register_transport(self, transport):
- if not hasattr(transport, 'scheme') and not hasattr(transport.scheme, '__iter__'):
+ if not hasattr(transport, 'scheme') or not hasattr(transport.scheme, '__iter__'):
raise AttributeError('Transport %s must have a scheme list', transport.__class__.__name__)
for scheme in transport.scheme:
|
raven/transport/registry.py
|
ReplaceText(target='or' @(19,44)->(19,47))
|
class TransportRegistry(object):
self.register_transport(transport)
def register_transport(self, transport):
if not hasattr(transport, 'scheme') and not hasattr(transport.scheme, '__iter__'):
raise AttributeError('Transport %s must have a scheme list', transport.__class__.__name__)
for scheme in transport.scheme:
|
class TransportRegistry(object):
self.register_transport(transport)
def register_transport(self, transport):
if not hasattr(transport, 'scheme') or not hasattr(transport.scheme, '__iter__'):
raise AttributeError('Transport %s must have a scheme list', transport.__class__.__name__)
for scheme in transport.scheme:
|
263 |
https://:@github.com/ilius/pyglossary.git
|
8ad925bbb4ce19207322471c4c758d1a66f0db9d
|
@@ -186,7 +186,7 @@ def write(
with open(filePathBase + ".xml", "w", encoding="utf8") as toFile:
write_header(glos, toFile, frontBackMatter)
for entryI, entry in enumerate(glos):
- if glos.isData():
+ if entry.isData():
entry.save(myResDir)
continue
|
pyglossary/plugins/appledict/__init__.py
|
ReplaceText(target='entry' @(189,6)->(189,10))
|
def write(
with open(filePathBase + ".xml", "w", encoding="utf8") as toFile:
write_header(glos, toFile, frontBackMatter)
for entryI, entry in enumerate(glos):
if glos.isData():
entry.save(myResDir)
continue
|
def write(
with open(filePathBase + ".xml", "w", encoding="utf8") as toFile:
write_header(glos, toFile, frontBackMatter)
for entryI, entry in enumerate(glos):
if entry.isData():
entry.save(myResDir)
continue
|
264 |
https://:@github.com/ilius/pyglossary.git
|
194fea3e176f5c9caf090a58bee3c5b04fd483ef
|
@@ -450,7 +450,7 @@ class Glossary(GlossaryType):
wordCount = len(reader)
except Exception:
log.exception("")
- if wordCount > 0:
+ if wordCount >= 0:
progressbar = True
if progressbar:
self.progressInit("Converting")
|
pyglossary/glossary.py
|
ReplaceText(target='>=' @(453,17)->(453,18))
|
class Glossary(GlossaryType):
wordCount = len(reader)
except Exception:
log.exception("")
if wordCount > 0:
progressbar = True
if progressbar:
self.progressInit("Converting")
|
class Glossary(GlossaryType):
wordCount = len(reader)
except Exception:
log.exception("")
if wordCount >= 0:
progressbar = True
if progressbar:
self.progressInit("Converting")
|
265 |
https://:@github.com/ilius/pyglossary.git
|
a7237c6749dcd2f2cf34b5f8ec385b6ba0b5565a
|
@@ -99,7 +99,7 @@ class TextGlossaryReader(object):
try:
wordDefi = self.nextPair()
except StopIteration as e:
- if self._fileIndex < self._fileCount + 1:
+ if self._fileIndex < self._fileCount - 1:
if self.openNextFile():
return self.__next__()
self._wordCount = self._pos
|
pyglossary/text_reader.py
|
ReplaceText(target='-' @(102,40)->(102,41))
|
class TextGlossaryReader(object):
try:
wordDefi = self.nextPair()
except StopIteration as e:
if self._fileIndex < self._fileCount + 1:
if self.openNextFile():
return self.__next__()
self._wordCount = self._pos
|
class TextGlossaryReader(object):
try:
wordDefi = self.nextPair()
except StopIteration as e:
if self._fileIndex < self._fileCount - 1:
if self.openNextFile():
return self.__next__()
self._wordCount = self._pos
|
266 |
https://:@github.com/GOVCERT-LU/eml_parser.git
|
5a129fd37081de0c3483a317e4eefced52fed44d
|
@@ -636,7 +636,7 @@ def parse_email(msg, include_raw_body=False, include_attachment_data=False):
if list_observed_dom:
bodie['domain'] = list(set(list_observed_dom))
- if list_observed_dom:
+ if list_observed_ip:
bodie['ip'] = list(set(list_observed_ip))
else:
|
eml_parser/eml_parser.py
|
ReplaceText(target='list_observed_ip' @(639,15)->(639,32))
|
def parse_email(msg, include_raw_body=False, include_attachment_data=False):
if list_observed_dom:
bodie['domain'] = list(set(list_observed_dom))
if list_observed_dom:
bodie['ip'] = list(set(list_observed_ip))
else:
|
def parse_email(msg, include_raw_body=False, include_attachment_data=False):
if list_observed_dom:
bodie['domain'] = list(set(list_observed_dom))
if list_observed_ip:
bodie['ip'] = list(set(list_observed_ip))
else:
|
267 |
https://:@github.com/noripyt/wagtail-react-streamfield.git
|
fcd366318ffd0f20f93efcc58e00658f1334899c
|
@@ -75,4 +75,4 @@ class NewListBlock(ListBlock):
_('The maximum number of items is %d') % self.meta.max_num
)
- return value
+ return result
|
wagtail_react_streamfield/blocks/list_block.py
|
ReplaceText(target='result' @(78,15)->(78,20))
|
class NewListBlock(ListBlock):
_('The maximum number of items is %d') % self.meta.max_num
)
return value
|
class NewListBlock(ListBlock):
_('The maximum number of items is %d') % self.meta.max_num
)
return result
|
268 |
https://:@github.com/noripyt/wagtail-react-streamfield.git
|
4a33dab16bc9665349e6aaa011df3a47eea25277
|
@@ -43,7 +43,7 @@ class NewBlock(Block):
else errors.as_data()[0].params.get(NON_FIELD_ERRORS, ()))
else:
non_block_errors = errors
- if help_text and non_block_errors:
+ if help_text or non_block_errors:
return render_to_string(
'wagtailadmin/block_forms/blocks_container.html',
{
|
wagtail_react_streamfield/blocks/block.py
|
ReplaceText(target='or' @(46,21)->(46,24))
|
class NewBlock(Block):
else errors.as_data()[0].params.get(NON_FIELD_ERRORS, ()))
else:
non_block_errors = errors
if help_text and non_block_errors:
return render_to_string(
'wagtailadmin/block_forms/blocks_container.html',
{
|
class NewBlock(Block):
else errors.as_data()[0].params.get(NON_FIELD_ERRORS, ()))
else:
non_block_errors = errors
if help_text or non_block_errors:
return render_to_string(
'wagtailadmin/block_forms/blocks_container.html',
{
|
269 |
https://:@github.com/PMBio/MOFA.git
|
f7cd442f209ce679001b67e2ed439016bdef6991
|
@@ -82,7 +82,7 @@ class initModel(object):
elif qmean == "pca": # Latent variables are initialised from PCA in the concatenated matrix
pca = sklearn.decomposition.PCA(n_components=self.K, copy=True, whiten=True)
- pca.fit(s.concatenate(self.data,axis=0).T)
+ pca.fit(s.concatenate(self.data,axis=1).T)
qmean = pca.components_.T
elif isinstance(qmean,s.ndarray):
|
mofa/core/init_nodes.py
|
ReplaceText(target='1' @(85,57)->(85,58))
|
class initModel(object):
elif qmean == "pca": # Latent variables are initialised from PCA in the concatenated matrix
pca = sklearn.decomposition.PCA(n_components=self.K, copy=True, whiten=True)
pca.fit(s.concatenate(self.data,axis=0).T)
qmean = pca.components_.T
elif isinstance(qmean,s.ndarray):
|
class initModel(object):
elif qmean == "pca": # Latent variables are initialised from PCA in the concatenated matrix
pca = sklearn.decomposition.PCA(n_components=self.K, copy=True, whiten=True)
pca.fit(s.concatenate(self.data,axis=1).T)
qmean = pca.components_.T
elif isinstance(qmean,s.ndarray):
|
270 |
https://:@github.com/twisted/tubes.git
|
836d6021f14ad923bb595dffb0a6ca39f0b884b5
|
@@ -168,7 +168,7 @@ class _SiphonFount(_SiphonPiece):
def _actuallyPause():
fount = self._siphon._tdrain.fount
self._siphon._pending.suspend()
- if fount is None:
+ if fount is not None:
pbpc = fount.pauseFlow()
else:
pbpc = NoPause()
|
tubes/_siphon.py
|
ReplaceText(target=' is not ' @(171,20)->(171,24))
|
class _SiphonFount(_SiphonPiece):
def _actuallyPause():
fount = self._siphon._tdrain.fount
self._siphon._pending.suspend()
if fount is None:
pbpc = fount.pauseFlow()
else:
pbpc = NoPause()
|
class _SiphonFount(_SiphonPiece):
def _actuallyPause():
fount = self._siphon._tdrain.fount
self._siphon._pending.suspend()
if fount is not None:
pbpc = fount.pauseFlow()
else:
pbpc = NoPause()
|
271 |
https://:@github.com/deschler/django-modeltranslation.git
|
656dca3e4031d8f030a0aae833fc43c0c09cb51c
|
@@ -57,7 +57,7 @@ class TranslationBaseModelAdmin(BaseModelAdmin):
else:
orig_formfield = self.formfield_for_dbfield(orig_field, **kwargs)
field.widget = deepcopy(orig_formfield.widget)
- if orig_field.null and isinstance(field.widget, (forms.TextInput, forms.Textarea)):
+ if db_field.null and isinstance(field.widget, (forms.TextInput, forms.Textarea)):
field.widget = ClearableWidgetWrapper(field.widget)
css_classes = field.widget.attrs.get('class', '').split(' ')
css_classes.append('mt')
|
modeltranslation/admin.py
|
ReplaceText(target='db_field' @(60,15)->(60,25))
|
class TranslationBaseModelAdmin(BaseModelAdmin):
else:
orig_formfield = self.formfield_for_dbfield(orig_field, **kwargs)
field.widget = deepcopy(orig_formfield.widget)
if orig_field.null and isinstance(field.widget, (forms.TextInput, forms.Textarea)):
field.widget = ClearableWidgetWrapper(field.widget)
css_classes = field.widget.attrs.get('class', '').split(' ')
css_classes.append('mt')
|
class TranslationBaseModelAdmin(BaseModelAdmin):
else:
orig_formfield = self.formfield_for_dbfield(orig_field, **kwargs)
field.widget = deepcopy(orig_formfield.widget)
if db_field.null and isinstance(field.widget, (forms.TextInput, forms.Textarea)):
field.widget = ClearableWidgetWrapper(field.widget)
css_classes = field.widget.attrs.get('class', '').split(' ')
css_classes.append('mt')
|
272 |
https://:@github.com/xcgspring/AXUI.git
|
82f7d6f7f2565d28553fda3c12bb1a2aafb5bab9
|
@@ -123,7 +123,7 @@ class AppMap(object):
UI_element_group.stop_func = self.get_func_by_name(xml_element.attrib["stop_func"])
if xml_element.attrib.has_key("identifier"):
UI_element_group.identifier_string = xml_element.attrib["identifier"]
- UI_element_group.identifier = identifier_parser.parse(UI_element.identifier_string, lexer=identifier_lexer)
+ UI_element_group.identifier = identifier_parser.parse(UI_element_group.identifier_string, lexer=identifier_lexer)
return UI_element_group
|
AXUI/XML/app_map.py
|
ReplaceText(target='UI_element_group' @(126,70)->(126,80))
|
class AppMap(object):
UI_element_group.stop_func = self.get_func_by_name(xml_element.attrib["stop_func"])
if xml_element.attrib.has_key("identifier"):
UI_element_group.identifier_string = xml_element.attrib["identifier"]
UI_element_group.identifier = identifier_parser.parse(UI_element.identifier_string, lexer=identifier_lexer)
return UI_element_group
|
class AppMap(object):
UI_element_group.stop_func = self.get_func_by_name(xml_element.attrib["stop_func"])
if xml_element.attrib.has_key("identifier"):
UI_element_group.identifier_string = xml_element.attrib["identifier"]
UI_element_group.identifier = identifier_parser.parse(UI_element_group.identifier_string, lexer=identifier_lexer)
return UI_element_group
|
273 |
https://:@github.com/docker/compose.git
|
0e19c92e82c75f821c231367b5cda88eefdf1427
|
@@ -577,7 +577,7 @@ class VolumeConfigTest(unittest.TestCase):
def test_volume_path_with_non_ascii_directory(self):
volume = u'/Füü/data:/data'
- container_path = config.resolve_volume_path(volume, ".", "test")
+ container_path = config.resolve_volume_path(".", volume, "test")
self.assertEqual(container_path, volume)
|
tests/unit/config/config_test.py
|
ArgSwap(idxs=0<->1 @(580,25)->(580,51))
|
class VolumeConfigTest(unittest.TestCase):
def test_volume_path_with_non_ascii_directory(self):
volume = u'/Füü/data:/data'
container_path = config.resolve_volume_path(volume, ".", "test")
self.assertEqual(container_path, volume)
|
class VolumeConfigTest(unittest.TestCase):
def test_volume_path_with_non_ascii_directory(self):
volume = u'/Füü/data:/data'
container_path = config.resolve_volume_path(".", volume, "test")
self.assertEqual(container_path, volume)
|
274 |
https://:@github.com/docker/compose.git
|
c4f59e731d540780a767d105bcb8d7d164ba4cd5
|
@@ -577,7 +577,7 @@ class VolumeConfigTest(unittest.TestCase):
def test_volume_path_with_non_ascii_directory(self):
volume = u'/Füü/data:/data'
- container_path = config.resolve_volume_path(volume, ".", "test")
+ container_path = config.resolve_volume_path(".", volume, "test")
self.assertEqual(container_path, volume)
|
tests/unit/config/config_test.py
|
ArgSwap(idxs=0<->1 @(580,25)->(580,51))
|
class VolumeConfigTest(unittest.TestCase):
def test_volume_path_with_non_ascii_directory(self):
volume = u'/Füü/data:/data'
container_path = config.resolve_volume_path(volume, ".", "test")
self.assertEqual(container_path, volume)
|
class VolumeConfigTest(unittest.TestCase):
def test_volume_path_with_non_ascii_directory(self):
volume = u'/Füü/data:/data'
container_path = config.resolve_volume_path(".", volume, "test")
self.assertEqual(container_path, volume)
|
275 |
https://:@github.com/ranaroussi/qtpylib.git
|
e5eac01b67e908823c80b1bdeac57dcb68dbb164
|
@@ -65,7 +65,7 @@ def _gen_symbol_group(sym):
def _gen_asset_class(sym):
sym_class = str(sym).split("_")
- if len(sym_class) > 0:
+ if len(sym_class) > 1:
return sym_class[1]
return "STK"
|
qtpylib/blotter.py
|
ReplaceText(target='1' @(68,24)->(68,25))
|
def _gen_symbol_group(sym):
def _gen_asset_class(sym):
sym_class = str(sym).split("_")
if len(sym_class) > 0:
return sym_class[1]
return "STK"
|
def _gen_symbol_group(sym):
def _gen_asset_class(sym):
sym_class = str(sym).split("_")
if len(sym_class) > 1:
return sym_class[1]
return "STK"
|
276 |
https://:@github.com/lmjohns3/theanets.git
|
507eac0d7fe007cdc93d89ca6cb0db9521908eac
|
@@ -253,7 +253,7 @@ class Network(object):
h = self.hiddens[-1]
a, b = self.weights[i].get_value(borrow=True).shape
logging.info('tied weights from layer %d: %s x %s', i, b, a)
- o = theano.shared(np.zeros((b, ), FLOAT), name='b_out{}'.format(i))
+ o = theano.shared(np.zeros((a, ), FLOAT), name='b_out{}'.format(i))
self.preacts.append(TT.dot(h, self.weights[i].T) + o)
func = self._output_func if i == 0 else self._hidden_func
self.hiddens.append(func(self.preacts[-1]))
|
theanets/feedforward.py
|
ReplaceText(target='a' @(256,44)->(256,45))
|
class Network(object):
h = self.hiddens[-1]
a, b = self.weights[i].get_value(borrow=True).shape
logging.info('tied weights from layer %d: %s x %s', i, b, a)
o = theano.shared(np.zeros((b, ), FLOAT), name='b_out{}'.format(i))
self.preacts.append(TT.dot(h, self.weights[i].T) + o)
func = self._output_func if i == 0 else self._hidden_func
self.hiddens.append(func(self.preacts[-1]))
|
class Network(object):
h = self.hiddens[-1]
a, b = self.weights[i].get_value(borrow=True).shape
logging.info('tied weights from layer %d: %s x %s', i, b, a)
o = theano.shared(np.zeros((a, ), FLOAT), name='b_out{}'.format(i))
self.preacts.append(TT.dot(h, self.weights[i].T) + o)
func = self._output_func if i == 0 else self._hidden_func
self.hiddens.append(func(self.preacts[-1]))
|
277 |
https://:@github.com/lmjohns3/theanets.git
|
62f7125e1b179df846fdcd5d213c8bca18595648
|
@@ -83,7 +83,7 @@ class SequenceDataset:
slices = [slice(None), slice(None)]
self.batches = []
i = 0
- while i + size < shape[axis]:
+ while i + size <= shape[axis]:
slices[axis] = slice(i, i + size)
self.batches.append([d[tuple(slices)] for d in data])
i += size
|
theanets/dataset.py
|
ReplaceText(target='<=' @(86,27)->(86,28))
|
class SequenceDataset:
slices = [slice(None), slice(None)]
self.batches = []
i = 0
while i + size < shape[axis]:
slices[axis] = slice(i, i + size)
self.batches.append([d[tuple(slices)] for d in data])
i += size
|
class SequenceDataset:
slices = [slice(None), slice(None)]
self.batches = []
i = 0
while i + size <= shape[axis]:
slices[axis] = slice(i, i + size)
self.batches.append([d[tuple(slices)] for d in data])
i += size
|
278 |
https://:@github.com/lmjohns3/theanets.git
|
8fa4792eaae89e4da7d041178fdcb341f5313914
|
@@ -834,4 +834,4 @@ class Classifier(Network):
k : ndarray (num-examples, )
A vector of class index values, one per row of input data.
'''
- return self.predict(x).argmax(axis=1)
+ return self.predict(x).argmax(axis=-1)
|
theanets/feedforward.py
|
ReplaceText(target='-1' @(837,43)->(837,44))
|
class Classifier(Network):
k : ndarray (num-examples, )
A vector of class index values, one per row of input data.
'''
return self.predict(x).argmax(axis=1)
|
class Classifier(Network):
k : ndarray (num-examples, )
A vector of class index values, one per row of input data.
'''
return self.predict(x).argmax(axis=-1)
|
279 |
https://:@github.com/lmjohns3/theanets.git
|
114f37976b5c00330da9c5fed75ecf7b727eff58
|
@@ -338,7 +338,7 @@ class Network(object):
if i == 0:
noise = kwargs.get('input_noise', 0)
dropout = kwargs.get('input_dropouts', 0)
- elif i == len(self.layers) - 1:
+ elif i != len(self.layers) - 1:
noise = kwargs.get('hidden_noise', 0)
dropout = kwargs.get('hidden_dropouts', 0)
out, mon, upd = layer.connect(inputs, noise=noise, dropout=dropout)
|
theanets/graph.py
|
ReplaceText(target='!=' @(341,23)->(341,25))
|
class Network(object):
if i == 0:
noise = kwargs.get('input_noise', 0)
dropout = kwargs.get('input_dropouts', 0)
elif i == len(self.layers) - 1:
noise = kwargs.get('hidden_noise', 0)
dropout = kwargs.get('hidden_dropouts', 0)
out, mon, upd = layer.connect(inputs, noise=noise, dropout=dropout)
|
class Network(object):
if i == 0:
noise = kwargs.get('input_noise', 0)
dropout = kwargs.get('input_dropouts', 0)
elif i != len(self.layers) - 1:
noise = kwargs.get('hidden_noise', 0)
dropout = kwargs.get('hidden_dropouts', 0)
out, mon, upd = layer.connect(inputs, noise=noise, dropout=dropout)
|
280 |
https://:@github.com/pyiron/pyiron_base.git
|
f9ea8a31ec8fe36704c4c8f73e127ee3e077ca3a
|
@@ -98,7 +98,7 @@ class Vasprun(object):
d["cells"] = np.array(d["cells"])
d["positions"] = np.array(d["positions"])
# Check if the parsed coordinates are in absolute/relative coordinates. If absolute, convert to relative
- if len(np.argwhere(d["positions"].flatten() > 1).flatten()) / len(d["positions"].flatten()) < 0.01:
+ if len(np.argwhere(d["positions"].flatten() > 1).flatten()) / len(d["positions"].flatten()) > 0.01:
pos_new = d["positions"].copy()
for i, pos in enumerate(pos_new):
d["positions"][i] = np.dot(pos, np.linalg.inv(d["cells"][i]))
|
pyiron_vasp/vasprun.py
|
ReplaceText(target='>' @(101,100)->(101,101))
|
class Vasprun(object):
d["cells"] = np.array(d["cells"])
d["positions"] = np.array(d["positions"])
# Check if the parsed coordinates are in absolute/relative coordinates. If absolute, convert to relative
if len(np.argwhere(d["positions"].flatten() > 1).flatten()) / len(d["positions"].flatten()) < 0.01:
pos_new = d["positions"].copy()
for i, pos in enumerate(pos_new):
d["positions"][i] = np.dot(pos, np.linalg.inv(d["cells"][i]))
|
class Vasprun(object):
d["cells"] = np.array(d["cells"])
d["positions"] = np.array(d["positions"])
# Check if the parsed coordinates are in absolute/relative coordinates. If absolute, convert to relative
if len(np.argwhere(d["positions"].flatten() > 1).flatten()) / len(d["positions"].flatten()) > 0.01:
pos_new = d["positions"].copy()
for i, pos in enumerate(pos_new):
d["positions"][i] = np.dot(pos, np.linalg.inv(d["cells"][i]))
|
281 |
https://:@github.com/pyiron/pyiron_base.git
|
5d9677db9fcbca5fce348a7208830467381cd643
|
@@ -487,7 +487,7 @@ class ParallelMaster(GenericMaster):
job_lst.append(ham._process)
ham = next(self._job_generator, None)
if ham is None and self.server.run_mode.modal:
- while ham is not None:
+ while ham is None:
time.sleep(10)
ham = next(self._job_generator, None)
else:
|
pyiron_base/objects/job/parallel.py
|
ReplaceText(target=' is ' @(490,29)->(490,37))
|
class ParallelMaster(GenericMaster):
job_lst.append(ham._process)
ham = next(self._job_generator, None)
if ham is None and self.server.run_mode.modal:
while ham is not None:
time.sleep(10)
ham = next(self._job_generator, None)
else:
|
class ParallelMaster(GenericMaster):
job_lst.append(ham._process)
ham = next(self._job_generator, None)
if ham is None and self.server.run_mode.modal:
while ham is None:
time.sleep(10)
ham = next(self._job_generator, None)
else:
|
282 |
https://:@github.com/pyiron/pyiron_base.git
|
a4ed8375fa375ff740a8334f7ce5eb5bf9d7d4a2
|
@@ -477,7 +477,7 @@ class Settings(with_metaclass(Singleton)):
else:
# SQLite is raising ugly error messages when the database directory does not exist.
if config["sql_file"] is None:
- if len(config["resource_paths"]) > 1:
+ if len(config["resource_paths"]) >= 1:
config["sql_file"] = "/".join(
[config["resource_paths"][0], "pyiron.db"]
)
|
pyiron_base/settings/generic.py
|
ReplaceText(target='>=' @(480,49)->(480,50))
|
class Settings(with_metaclass(Singleton)):
else:
# SQLite is raising ugly error messages when the database directory does not exist.
if config["sql_file"] is None:
if len(config["resource_paths"]) > 1:
config["sql_file"] = "/".join(
[config["resource_paths"][0], "pyiron.db"]
)
|
class Settings(with_metaclass(Singleton)):
else:
# SQLite is raising ugly error messages when the database directory does not exist.
if config["sql_file"] is None:
if len(config["resource_paths"]) >= 1:
config["sql_file"] = "/".join(
[config["resource_paths"][0], "pyiron.db"]
)
|
283 |
https://:@github.com/ESSolutions/django-mssql-backend.git
|
06578347a5e775088cfe3a2c2e9c528c5ffb7de0
|
@@ -201,7 +201,7 @@ class DatabaseWrapper(BaseDatabaseWrapper):
# Only append DRIVER if DATABASE_ODBC_DSN hasn't been set
cstr_parts.append('DRIVER={%s}' % driver)
if ms_drivers.match(driver) or driver == 'FreeTDS' and \
- conn_params.get('host_is_server', False):
+ options.get('host_is_server', False):
if port:
host += ';PORT=%s' % port
cstr_parts.append('SERVER=%s' % host)
|
sql_server/pyodbc/base.py
|
ReplaceText(target='options' @(204,16)->(204,27))
|
class DatabaseWrapper(BaseDatabaseWrapper):
# Only append DRIVER if DATABASE_ODBC_DSN hasn't been set
cstr_parts.append('DRIVER={%s}' % driver)
if ms_drivers.match(driver) or driver == 'FreeTDS' and \
conn_params.get('host_is_server', False):
if port:
host += ';PORT=%s' % port
cstr_parts.append('SERVER=%s' % host)
|
class DatabaseWrapper(BaseDatabaseWrapper):
# Only append DRIVER if DATABASE_ODBC_DSN hasn't been set
cstr_parts.append('DRIVER={%s}' % driver)
if ms_drivers.match(driver) or driver == 'FreeTDS' and \
options.get('host_is_server', False):
if port:
host += ';PORT=%s' % port
cstr_parts.append('SERVER=%s' % host)
|
284 |
https://:@github.com/bykof/billomapy.git
|
29dc1fcd8f491c158a713b07680ea1305748c959
|
@@ -112,7 +112,7 @@ todo_include_todos = False
# a list of builtin themes.
on_rtd = os.environ.get('READTHEDOCS', None) == 'True'
-if not on_rtd: # only import and set the theme if we're building docs locally
+if on_rtd: # only import and set the theme if we're building docs locally
import sphinx_rtd_theme
html_theme = 'sphinx_rtd_theme'
html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
|
docs/source/conf.py
|
ReplaceText(target='' @(115,3)->(115,7))
|
todo_include_todos = False
# a list of builtin themes.
on_rtd = os.environ.get('READTHEDOCS', None) == 'True'
if not on_rtd: # only import and set the theme if we're building docs locally
import sphinx_rtd_theme
html_theme = 'sphinx_rtd_theme'
html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
|
todo_include_todos = False
# a list of builtin themes.
on_rtd = os.environ.get('READTHEDOCS', None) == 'True'
if on_rtd: # only import and set the theme if we're building docs locally
import sphinx_rtd_theme
html_theme = 'sphinx_rtd_theme'
html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
|
285 |
https://:@github.com/quva-lab/artemis.git
|
26ae4a7049b0f6ac2aef7daf4c2a5184748be516
|
@@ -76,7 +76,7 @@ def get_new_positions(fixed_positions, layout, n_plots):
positions.append(fixed_positions[i])
else:
while True:
- row, col = ix/n_cols, ix%n_cols
+ row, col = ix//n_cols, ix%n_cols
if (row, col) not in taken_positions:
positions.append((row, col))
taken_positions.add((row, col))
|
artemis/plotting/expanding_subplots.py
|
ReplaceText(target='//' @(79,29)->(79,30))
|
def get_new_positions(fixed_positions, layout, n_plots):
positions.append(fixed_positions[i])
else:
while True:
row, col = ix/n_cols, ix%n_cols
if (row, col) not in taken_positions:
positions.append((row, col))
taken_positions.add((row, col))
|
def get_new_positions(fixed_positions, layout, n_plots):
positions.append(fixed_positions[i])
else:
while True:
row, col = ix//n_cols, ix%n_cols
if (row, col) not in taken_positions:
positions.append((row, col))
taken_positions.add((row, col))
|
286 |
https://:@github.com/OCR-D/core.git
|
e82299a48e4cf4677535819f81bb944ad4f64f8a
|
@@ -312,7 +312,7 @@ def validate_consistency(node, page_textequiv_consistency, page_textequiv_strate
pass # already reported in recursive call above
elif not child_poly.within(node_poly.buffer(PARENT_SLACK)):
# TODO: automatic repair?
- report.add_error(CoordinateConsistencyError(tag, child.id, file_id,
+ report.add_error(CoordinateConsistencyError(child_tag, child.id, file_id,
parent_points, child_points))
log.debug("Inconsistent coords of %s %s", child_tag, child.id)
consistent = False
|
ocrd_validators/ocrd_validators/page_validator.py
|
ReplaceText(target='child_tag' @(315,64)->(315,67))
|
def validate_consistency(node, page_textequiv_consistency, page_textequiv_strate
pass # already reported in recursive call above
elif not child_poly.within(node_poly.buffer(PARENT_SLACK)):
# TODO: automatic repair?
report.add_error(CoordinateConsistencyError(tag, child.id, file_id,
parent_points, child_points))
log.debug("Inconsistent coords of %s %s", child_tag, child.id)
consistent = False
|
def validate_consistency(node, page_textequiv_consistency, page_textequiv_strate
pass # already reported in recursive call above
elif not child_poly.within(node_poly.buffer(PARENT_SLACK)):
# TODO: automatic repair?
report.add_error(CoordinateConsistencyError(child_tag, child.id, file_id,
parent_points, child_points))
log.debug("Inconsistent coords of %s %s", child_tag, child.id)
consistent = False
|
287 |
https://:@github.com/jhuapl-boss/heaviside.git
|
3adb4902a04dc3e1d9a325ab1752c4b2d856b677
|
@@ -69,7 +69,7 @@ class ASTCompAndOr(ASTNode):
super(ASTCompAndOr, self).__init__(comp.token)
self.comps = [comp]
for c in comps:
- self.comps.append(comp)
+ self.comps.append(c)
class ASTCompAnd(ASTCompAndOr):
op = 'And'
|
heaviside/ast.py
|
ReplaceText(target='c' @(72,30)->(72,34))
|
class ASTCompAndOr(ASTNode):
super(ASTCompAndOr, self).__init__(comp.token)
self.comps = [comp]
for c in comps:
self.comps.append(comp)
class ASTCompAnd(ASTCompAndOr):
op = 'And'
|
class ASTCompAndOr(ASTNode):
super(ASTCompAndOr, self).__init__(comp.token)
self.comps = [comp]
for c in comps:
self.comps.append(c)
class ASTCompAnd(ASTCompAndOr):
op = 'And'
|
288 |
https://:@github.com/google-research/text-to-text-transfer-transformer.git
|
d2c010b48b702f22fb665d2e5723c7baaf4f6a1c
|
@@ -98,7 +98,7 @@ def main(_):
" ".join(str(i) for i in v) if FLAGS.tokenize
else v.decode("utf-8"))
else:
- v[k] = ""
+ key_to_string[k] = ""
return FLAGS.format_string.format(**key_to_string)
for shard_path in files:
|
t5/scripts/dump_task.py
|
ReplaceText(target='key_to_string' @(101,8)->(101,9))
|
def main(_):
" ".join(str(i) for i in v) if FLAGS.tokenize
else v.decode("utf-8"))
else:
v[k] = ""
return FLAGS.format_string.format(**key_to_string)
for shard_path in files:
|
def main(_):
" ".join(str(i) for i in v) if FLAGS.tokenize
else v.decode("utf-8"))
else:
key_to_string[k] = ""
return FLAGS.format_string.format(**key_to_string)
for shard_path in files:
|
289 |
https://:@github.com/lvieirajr/mongorest.git
|
470f17173b63b14ab60bbc99db7ca3b3f93afdac
|
@@ -401,4 +401,4 @@ class CoercionError(SchemaValidationError):
self['collection'] = collection
self['field'] = field
- self['coercion_type'] = coercion_type
+ self['coercion_type'] = coercion_type_repr
|
mongorest/errors.py
|
ReplaceText(target='coercion_type_repr' @(404,32)->(404,45))
|
class CoercionError(SchemaValidationError):
self['collection'] = collection
self['field'] = field
self['coercion_type'] = coercion_type
|
class CoercionError(SchemaValidationError):
self['collection'] = collection
self['field'] = field
self['coercion_type'] = coercion_type_repr
|
290 |
https://:@github.com/internetarchive/fatcat.git
|
89b729feac30a272b557542c5c4149d3611f869a
|
@@ -295,7 +295,7 @@ def container_to_elasticsearch(entity):
t['in_sherpa_romeo'] = in_sherpa_romeo
t['is_oa'] = in_doaj or in_road or is_longtail_oa or is_oa
t['is_longtail_oa'] = is_longtail_oa
- t['any_kbart'] = any_ia_sim
+ t['any_kbart'] = any_kbart
t['any_jstor'] = any_jstor
t['any_ia_sim'] = bool(any_ia_sim)
return t
|
python/fatcat_tools/transforms.py
|
ReplaceText(target='any_kbart' @(298,21)->(298,31))
|
def container_to_elasticsearch(entity):
t['in_sherpa_romeo'] = in_sherpa_romeo
t['is_oa'] = in_doaj or in_road or is_longtail_oa or is_oa
t['is_longtail_oa'] = is_longtail_oa
t['any_kbart'] = any_ia_sim
t['any_jstor'] = any_jstor
t['any_ia_sim'] = bool(any_ia_sim)
return t
|
def container_to_elasticsearch(entity):
t['in_sherpa_romeo'] = in_sherpa_romeo
t['is_oa'] = in_doaj or in_road or is_longtail_oa or is_oa
t['is_longtail_oa'] = is_longtail_oa
t['any_kbart'] = any_kbart
t['any_jstor'] = any_jstor
t['any_ia_sim'] = bool(any_ia_sim)
return t
|
291 |
https://:@github.com/internetarchive/fatcat.git
|
11dfac5f8f9ced9b56cf277d0e3adeccc572b251
|
@@ -57,7 +57,7 @@ class ArabesqueMatchImporter(EntityImporter):
eg_extra = kwargs.get('editgroup_extra', dict())
eg_extra['agent'] = eg_extra.get('agent', 'fatcat_tools.ArabesqueMatchImporter')
if kwargs.get('crawl_id'):
- eg_extra['crawl_id'] = eg_extra.get('crawl_id')
+ eg_extra['crawl_id'] = kwargs.get('crawl_id')
super().__init__(api,
editgroup_description=eg_desc,
editgroup_extra=eg_extra,
|
python/fatcat_tools/importers/arabesque.py
|
ReplaceText(target='kwargs' @(60,35)->(60,43))
|
class ArabesqueMatchImporter(EntityImporter):
eg_extra = kwargs.get('editgroup_extra', dict())
eg_extra['agent'] = eg_extra.get('agent', 'fatcat_tools.ArabesqueMatchImporter')
if kwargs.get('crawl_id'):
eg_extra['crawl_id'] = eg_extra.get('crawl_id')
super().__init__(api,
editgroup_description=eg_desc,
editgroup_extra=eg_extra,
|
class ArabesqueMatchImporter(EntityImporter):
eg_extra = kwargs.get('editgroup_extra', dict())
eg_extra['agent'] = eg_extra.get('agent', 'fatcat_tools.ArabesqueMatchImporter')
if kwargs.get('crawl_id'):
eg_extra['crawl_id'] = kwargs.get('crawl_id')
super().__init__(api,
editgroup_description=eg_desc,
editgroup_extra=eg_extra,
|
292 |
https://:@github.com/internetarchive/fatcat.git
|
7104e6dfb99717353e3819853ae61ac6387a02a1
|
@@ -128,7 +128,7 @@ class EntityUpdatesWorker(FatcatWorker):
# update release when a file changes
# TODO: fetch old revision as well, and only update
# releases for which list changed
- release_ids.extend(e['release_ids'])
+ release_ids.extend(file_entity['release_ids'])
file_dict = self.api.api_client.sanitize_for_serialization(file_entity)
file_producer.produce(
message=json.dumps(file_dict).encode('utf-8'),
|
python/fatcat_tools/workers/changelog.py
|
ReplaceText(target='file_entity' @(131,35)->(131,36))
|
class EntityUpdatesWorker(FatcatWorker):
# update release when a file changes
# TODO: fetch old revision as well, and only update
# releases for which list changed
release_ids.extend(e['release_ids'])
file_dict = self.api.api_client.sanitize_for_serialization(file_entity)
file_producer.produce(
message=json.dumps(file_dict).encode('utf-8'),
|
class EntityUpdatesWorker(FatcatWorker):
# update release when a file changes
# TODO: fetch old revision as well, and only update
# releases for which list changed
release_ids.extend(file_entity['release_ids'])
file_dict = self.api.api_client.sanitize_for_serialization(file_entity)
file_producer.produce(
message=json.dumps(file_dict).encode('utf-8'),
|
293 |
https://:@github.com/internetarchive/fatcat.git
|
80b756d6feec3f66225287b9ca73c8b02d012027
|
@@ -167,7 +167,7 @@ class IngestFileResultImporter(EntityImporter):
urls=urls,
)
if request.get('edit_extra'):
- fe.edit_extra = fatcat['edit_extra']
+ fe.edit_extra = request['edit_extra']
else:
fe.edit_extra = dict()
if request.get('ingest_request_source'):
|
python/fatcat_tools/importers/ingest.py
|
ReplaceText(target='request' @(170,28)->(170,34))
|
class IngestFileResultImporter(EntityImporter):
urls=urls,
)
if request.get('edit_extra'):
fe.edit_extra = fatcat['edit_extra']
else:
fe.edit_extra = dict()
if request.get('ingest_request_source'):
|
class IngestFileResultImporter(EntityImporter):
urls=urls,
)
if request.get('edit_extra'):
fe.edit_extra = request['edit_extra']
else:
fe.edit_extra = dict()
if request.get('ingest_request_source'):
|
294 |
https://:@github.com/pallets/flask-ext-migrate.git
|
ea6a0666e08ce176031456dced635ca3c92c1386
|
@@ -43,7 +43,7 @@ def fix_from_imports(red):
modules = node.value
if (len(modules) < 2 or
- modules[0].value != 'flask' and modules[1].value != 'ext'):
+ modules[0].value != 'flask' or modules[1].value != 'ext'):
continue
if len(modules) >= 3:
|
flask_ext_migrate/__init__.py
|
ReplaceText(target='or' @(46,44)->(46,47))
|
def fix_from_imports(red):
modules = node.value
if (len(modules) < 2 or
modules[0].value != 'flask' and modules[1].value != 'ext'):
continue
if len(modules) >= 3:
|
def fix_from_imports(red):
modules = node.value
if (len(modules) < 2 or
modules[0].value != 'flask' or modules[1].value != 'ext'):
continue
if len(modules) >= 3:
|
295 |
https://:@github.com/SpockBotMC/SpockBot.git
|
6c8e63b97e3d9617ce0ca9bea9a677902b5ef2fb
|
@@ -42,7 +42,7 @@ class SelectSocket:
else:
slist = [(self.sock,), (), (self.sock,)]
timeout = self.timer.get_timeout()
- if timeout>0:
+ if timeout>=0:
slist.append(timeout)
try:
rlist, wlist, xlist = select.select(*slist)
|
spock/plugins/core/net.py
|
ReplaceText(target='>=' @(45,12)->(45,13))
|
class SelectSocket:
else:
slist = [(self.sock,), (), (self.sock,)]
timeout = self.timer.get_timeout()
if timeout>0:
slist.append(timeout)
try:
rlist, wlist, xlist = select.select(*slist)
|
class SelectSocket:
else:
slist = [(self.sock,), (), (self.sock,)]
timeout = self.timer.get_timeout()
if timeout>=0:
slist.append(timeout)
try:
rlist, wlist, xlist = select.select(*slist)
|
296 |
https://:@github.com/SpockBotMC/SpockBot.git
|
bebe1187dd6432ec3b4b6dddb8246baef83c2203
|
@@ -15,7 +15,7 @@ class PloaderFetch:
@pl_announce('PloaderFetch')
class SettingsPlugin:
def __init__(self, ploader, kwargs):
- settings = get_settings(kwargs, kwargs.get('settings', {}))
+ settings = get_settings(kwargs.get('settings', {}), kwargs)
plugin_list = settings.get('plugins', DefaultPlugins)
plugins = []
plugin_settings = {}
|
spock/plugins/core/settings.py
|
ArgSwap(idxs=0<->1 @(18,13)->(18,25))
|
class PloaderFetch:
@pl_announce('PloaderFetch')
class SettingsPlugin:
def __init__(self, ploader, kwargs):
settings = get_settings(kwargs, kwargs.get('settings', {}))
plugin_list = settings.get('plugins', DefaultPlugins)
plugins = []
plugin_settings = {}
|
class PloaderFetch:
@pl_announce('PloaderFetch')
class SettingsPlugin:
def __init__(self, ploader, kwargs):
settings = get_settings(kwargs.get('settings', {}), kwargs)
plugin_list = settings.get('plugins', DefaultPlugins)
plugins = []
plugin_settings = {}
|
297 |
https://:@github.com/SpockBotMC/SpockBot.git
|
bebe1187dd6432ec3b4b6dddb8246baef83c2203
|
@@ -19,7 +19,7 @@ default_settings = {
class StartPlugin:
def __init__(self, ploader, settings):
- self.settings = utils.get_settings(settings, default_settings)
+ self.settings = utils.get_settings(default_settings, settings)
self.event = ploader.requires('Event')
self.net = ploader.requires('Net')
self.auth = ploader.requires('Auth')
|
spock/plugins/helpers/start.py
|
ArgSwap(idxs=0<->1 @(22,18)->(22,36))
|
default_settings = {
class StartPlugin:
def __init__(self, ploader, settings):
self.settings = utils.get_settings(settings, default_settings)
self.event = ploader.requires('Event')
self.net = ploader.requires('Net')
self.auth = ploader.requires('Auth')
|
default_settings = {
class StartPlugin:
def __init__(self, ploader, settings):
self.settings = utils.get_settings(default_settings, settings)
self.event = ploader.requires('Event')
self.net = ploader.requires('Net')
self.auth = ploader.requires('Auth')
|
298 |
https://:@github.com/hungpham2511/toppra.git
|
fefd8430ad55fb09414a1d4c6d8be41e81977842
|
@@ -19,7 +19,7 @@ def _find_left_index(ss_waypoints, s):
for i in range(1, len(ss_waypoints)):
if ss_waypoints[i - 1] <= s and s < ss_waypoints[i]:
return i - 1
- return len(ss_waypoints) - 1
+ return len(ss_waypoints) - 2
class Interpolator(object):
|
toppra/interpolator.py
|
ReplaceText(target='2' @(22,31)->(22,32))
|
def _find_left_index(ss_waypoints, s):
for i in range(1, len(ss_waypoints)):
if ss_waypoints[i - 1] <= s and s < ss_waypoints[i]:
return i - 1
return len(ss_waypoints) - 1
class Interpolator(object):
|
def _find_left_index(ss_waypoints, s):
for i in range(1, len(ss_waypoints)):
if ss_waypoints[i - 1] <= s and s < ss_waypoints[i]:
return i - 1
return len(ss_waypoints) - 2
class Interpolator(object):
|
299 |
https://:@github.com/cmrivers/epipy.git
|
c541e6f7f559c92279623aa3c2cb7f644247280e
|
@@ -34,7 +34,7 @@ def generate_example_data(cluster_size, outbreak_len, clusters, gen_time, attrib
for i in range(clusters):
cluster_letter = np.random.choice([i for i in string.ascii_uppercase if i not in used])[0]
cluster_name = 'Cluster' + cluster_letter
- used.append(cluster_name)
+ used.append(cluster_letter)
ix_rng = pd.date_range('1/1/2014', periods=outbreak_len, freq='D')
ix_date = np.random.choice(ix_rng, size=1)
|
epipy/data_generator.py
|
ReplaceText(target='cluster_letter' @(37,20)->(37,32))
|
def generate_example_data(cluster_size, outbreak_len, clusters, gen_time, attrib
for i in range(clusters):
cluster_letter = np.random.choice([i for i in string.ascii_uppercase if i not in used])[0]
cluster_name = 'Cluster' + cluster_letter
used.append(cluster_name)
ix_rng = pd.date_range('1/1/2014', periods=outbreak_len, freq='D')
ix_date = np.random.choice(ix_rng, size=1)
|
def generate_example_data(cluster_size, outbreak_len, clusters, gen_time, attrib
for i in range(clusters):
cluster_letter = np.random.choice([i for i in string.ascii_uppercase if i not in used])[0]
cluster_name = 'Cluster' + cluster_letter
used.append(cluster_letter)
ix_rng = pd.date_range('1/1/2014', periods=outbreak_len, freq='D')
ix_date = np.random.choice(ix_rng, size=1)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.