Unnamed: 0
int64 0
2.44k
| repo
stringlengths 32
81
| hash
stringlengths 40
40
| diff
stringlengths 113
1.17k
| old_path
stringlengths 5
84
| rewrite
stringlengths 34
79
| initial_state
stringlengths 75
980
| final_state
stringlengths 76
980
|
---|---|---|---|---|---|---|---|
2,300 |
https://:@github.com/yunojuno/django-s3-upload.git
|
65e49f3277f987c47d189105c6d979d666bb3004
|
@@ -45,7 +45,7 @@ def create_upload_data(content_type, source_filename, upload_to):
bucket_url = "https://%s/%s" % (endpoint, bucket)
return {
- "policy": policy,
+ "policy": encoded,
"signature": signature_b64,
"key": key,
"AWSAccessKeyId": access_key,
|
s3direct/utils.py
|
ReplaceText(target='encoded' @(48,18)->(48,24))
|
def create_upload_data(content_type, source_filename, upload_to):
bucket_url = "https://%s/%s" % (endpoint, bucket)
return {
"policy": policy,
"signature": signature_b64,
"key": key,
"AWSAccessKeyId": access_key,
|
def create_upload_data(content_type, source_filename, upload_to):
bucket_url = "https://%s/%s" % (endpoint, bucket)
return {
"policy": encoded,
"signature": signature_b64,
"key": key,
"AWSAccessKeyId": access_key,
|
2,301 |
https://:@github.com/Manticore-attic/pyfft.git
|
01cdc24797472eec5114b68f02ce8c68fa348853
|
@@ -53,7 +53,7 @@ def clFFT_ExecuteInterleaved(plan, batchSize, dir, data_in, data_out):
currWrite = 1 if numKernelsOdd else 2
for kInfo in kernelInfo:
- if isInPlace and numKernelsOdd and not inPlaceDone and kernelInfo.in_place_possible:
+ if isInPlace and numKernelsOdd and not inPlaceDone and kInfo.in_place_possible:
currWrite = currRead
inPlaceDone = True
|
pycudafft/fft_execute.py
|
ReplaceText(target='kInfo' @(56,58)->(56,68))
|
def clFFT_ExecuteInterleaved(plan, batchSize, dir, data_in, data_out):
currWrite = 1 if numKernelsOdd else 2
for kInfo in kernelInfo:
if isInPlace and numKernelsOdd and not inPlaceDone and kernelInfo.in_place_possible:
currWrite = currRead
inPlaceDone = True
|
def clFFT_ExecuteInterleaved(plan, batchSize, dir, data_in, data_out):
currWrite = 1 if numKernelsOdd else 2
for kInfo in kernelInfo:
if isInPlace and numKernelsOdd and not inPlaceDone and kInfo.in_place_possible:
currWrite = currRead
inPlaceDone = True
|
2,302 |
https://:@github.com/sepandhaghighi/pyshutdown.git
|
8b1b84e10d5acc1b1ec07ae271fdd3f61a414c87
|
@@ -12,7 +12,7 @@ def get_method():
get_method=input("Please Enter Method , Shutdown[1] , Hibernate[2] , Restart[3]")
if get_method=="2":
flag="-h"
- elif flag=="3":
+ elif get_method=="3":
flag="-r"
else:
flag="-s"
|
main.py
|
ReplaceText(target='get_method' @(15,9)->(15,13))
|
def get_method():
get_method=input("Please Enter Method , Shutdown[1] , Hibernate[2] , Restart[3]")
if get_method=="2":
flag="-h"
elif flag=="3":
flag="-r"
else:
flag="-s"
|
def get_method():
get_method=input("Please Enter Method , Shutdown[1] , Hibernate[2] , Restart[3]")
if get_method=="2":
flag="-h"
elif get_method=="3":
flag="-r"
else:
flag="-s"
|
2,303 |
https://:@github.com/cfhamlet/os-m3-engine.git
|
b6a8890df08b38e6201e3b8cc18e870a7f61fda4
|
@@ -89,7 +89,7 @@ def create(frontend_cls='os_m3_engine.ootb.StdinFrontend',
default_engine_transport_config = ENGINE_TRANSPORT_BRIDGE_CONFIG \
if backend_cls is not None else ENGINE_TRANSPORT_CONFIG
- e_transport_config = engine_backend_config
+ e_transport_config = engine_transport_config
if engine_transport_config in (ENGINE_TRANSPORT_CONFIG, ENGINE_TRANSPORT_BRIDGE_CONFIG):
e_transport_config = None
|
src/os_m3_engine/launcher.py
|
ReplaceText(target='engine_transport_config' @(92,25)->(92,46))
|
def create(frontend_cls='os_m3_engine.ootb.StdinFrontend',
default_engine_transport_config = ENGINE_TRANSPORT_BRIDGE_CONFIG \
if backend_cls is not None else ENGINE_TRANSPORT_CONFIG
e_transport_config = engine_backend_config
if engine_transport_config in (ENGINE_TRANSPORT_CONFIG, ENGINE_TRANSPORT_BRIDGE_CONFIG):
e_transport_config = None
|
def create(frontend_cls='os_m3_engine.ootb.StdinFrontend',
default_engine_transport_config = ENGINE_TRANSPORT_BRIDGE_CONFIG \
if backend_cls is not None else ENGINE_TRANSPORT_CONFIG
e_transport_config = engine_transport_config
if engine_transport_config in (ENGINE_TRANSPORT_CONFIG, ENGINE_TRANSPORT_BRIDGE_CONFIG):
e_transport_config = None
|
2,304 |
https://:@github.com/ssi-dk/bifrost.git
|
d05050feff39d057a8db6b64d31645efbf741200
|
@@ -143,7 +143,7 @@ def query_ncbi_species(species_entry):
if result is not None:
return result["ncbi_species"]
elif group_result is not None:
- return result["ncbi_species"]
+ return group_result["ncbi_species"]
else:
return None
except Exception as e:
|
lib/bifrostlib/bifrostlib/mongo_interface.py
|
ReplaceText(target='group_result' @(146,23)->(146,29))
|
def query_ncbi_species(species_entry):
if result is not None:
return result["ncbi_species"]
elif group_result is not None:
return result["ncbi_species"]
else:
return None
except Exception as e:
|
def query_ncbi_species(species_entry):
if result is not None:
return result["ncbi_species"]
elif group_result is not None:
return group_result["ncbi_species"]
else:
return None
except Exception as e:
|
2,305 |
https://:@github.com/ssi-dk/bifrost.git
|
6f801b0adf5cf2209dec1052337b3eff413d65b0
|
@@ -268,7 +268,7 @@ def update_run_report(run, n_intervals):
)
def update_rerun_form(run_name):
run_name = run_name.split("/")[0]
- if run_name == "" or hasattr(keys, "rerun"):
+ if run_name == "" or not hasattr(keys, "rerun"):
return None
run_data = import_data.get_run(run_name)
|
reporter/run_checker.py
|
ReplaceText(target='not ' @(271,25)->(271,25))
|
def update_run_report(run, n_intervals):
)
def update_rerun_form(run_name):
run_name = run_name.split("/")[0]
if run_name == "" or hasattr(keys, "rerun"):
return None
run_data = import_data.get_run(run_name)
|
def update_run_report(run, n_intervals):
)
def update_rerun_form(run_name):
run_name = run_name.split("/")[0]
if run_name == "" or not hasattr(keys, "rerun"):
return None
run_data = import_data.get_run(run_name)
|
2,306 |
https://:@github.com/ayharano/pppipam.git
|
13ee95b04c1ccf5febedcd9568e15543b8c66365
|
@@ -190,7 +190,7 @@ class AddressSpace_description_TestCase(unittest.TestCase):
description_str = "address 0 for ipv4"
self.address_space.describe(
description=description_str,
- ip_parameter=network,
+ ip_parameter=zero_ipv4,
)
self.assertEqual(
self.address_space.description(
|
tests/test_description.py
|
ReplaceText(target='zero_ipv4' @(193,25)->(193,32))
|
class AddressSpace_description_TestCase(unittest.TestCase):
description_str = "address 0 for ipv4"
self.address_space.describe(
description=description_str,
ip_parameter=network,
)
self.assertEqual(
self.address_space.description(
|
class AddressSpace_description_TestCase(unittest.TestCase):
description_str = "address 0 for ipv4"
self.address_space.describe(
description=description_str,
ip_parameter=zero_ipv4,
)
self.assertEqual(
self.address_space.description(
|
2,307 |
https://:@github.com/ayharano/pppipam.git
|
6d57637be02e5d8597a63ba7fc3fd8525a5c2952
|
@@ -263,7 +263,7 @@ class AddressSpace:
self.__parent_supernet[child] = as_network
children_of_as_network.add(child)
children_of_supernet.remove(child)
- children_of_supernet.add(as_address)
+ children_of_supernet.add(as_network)
else:
raise TypeError("ip_parameter must be a valid IP parameter")
|
pppipam/pppipam.py
|
ReplaceText(target='as_network' @(266,37)->(266,47))
|
class AddressSpace:
self.__parent_supernet[child] = as_network
children_of_as_network.add(child)
children_of_supernet.remove(child)
children_of_supernet.add(as_address)
else:
raise TypeError("ip_parameter must be a valid IP parameter")
|
class AddressSpace:
self.__parent_supernet[child] = as_network
children_of_as_network.add(child)
children_of_supernet.remove(child)
children_of_supernet.add(as_network)
else:
raise TypeError("ip_parameter must be a valid IP parameter")
|
2,308 |
https://:@github.com/ayharano/pppipam.git
|
12fedde026bcd8de6562f0d3c6e3224e529c557e
|
@@ -289,7 +289,7 @@ class AddressSpace:
self.__description[as_network] = description
described = True
- self.__parent_supernet[as_address] = supernet
+ self.__parent_supernet[as_network] = supernet
children_of_as_network = (
self.__children_ip_object.setdefault(as_network, set())
)
|
pppipam/pppipam.py
|
ReplaceText(target='as_network' @(292,35)->(292,45))
|
class AddressSpace:
self.__description[as_network] = description
described = True
self.__parent_supernet[as_address] = supernet
children_of_as_network = (
self.__children_ip_object.setdefault(as_network, set())
)
|
class AddressSpace:
self.__description[as_network] = description
described = True
self.__parent_supernet[as_network] = supernet
children_of_as_network = (
self.__children_ip_object.setdefault(as_network, set())
)
|
2,309 |
https://:@github.com/alx-k/flask-jerify.git
|
cf3942ac578dbcb0deb358bf7bfb7c230a5d1f34
|
@@ -23,7 +23,7 @@ def jerror_handler(e):
"""http://jsonapi.org/format/#errors
"""
- if not hasattr('name', e):
+ if not hasattr(e, 'name'):
raise InternalServerError(e.description)
app.logger.error(e.description)
|
flask_jerify/flask_jerify.py
|
ArgSwap(idxs=0<->1 @(26,11)->(26,18))
|
def jerror_handler(e):
"""http://jsonapi.org/format/#errors
"""
if not hasattr('name', e):
raise InternalServerError(e.description)
app.logger.error(e.description)
|
def jerror_handler(e):
"""http://jsonapi.org/format/#errors
"""
if not hasattr(e, 'name'):
raise InternalServerError(e.description)
app.logger.error(e.description)
|
2,310 |
https://:@github.com/chenliangomc/RTFMaker.git
|
f82b42a9366a5ed2cbc83ebbe25dc6f0668f75c6
|
@@ -139,7 +139,7 @@ class RTable(object):
self._table_elements['body'].append(new_row)
html_foot = getattr(obj, 'tfoot')
if html_foot:
- for a_foot in html_body.find_all('td'):
+ for a_foot in html_foot.find_all('td'):
foot_cell = {
'value': a_foot.get_text(strip=True),
}
|
RTFMaker/utils.py
|
ReplaceText(target='html_foot' @(142,30)->(142,39))
|
class RTable(object):
self._table_elements['body'].append(new_row)
html_foot = getattr(obj, 'tfoot')
if html_foot:
for a_foot in html_body.find_all('td'):
foot_cell = {
'value': a_foot.get_text(strip=True),
}
|
class RTable(object):
self._table_elements['body'].append(new_row)
html_foot = getattr(obj, 'tfoot')
if html_foot:
for a_foot in html_foot.find_all('td'):
foot_cell = {
'value': a_foot.get_text(strip=True),
}
|
2,311 |
https://:@github.com/jianlins/PyFastNER.git
|
442227d122a7feee9766e2676cee1702c7ba645b
|
@@ -324,7 +324,7 @@ class FastCNER:
self.logger.debug(
'try add matched rule ({}-{})\t{}'.format(match_begin, match_end, str(self.rule_store[rule_id])))
current_span.rule_id = rule_id
- if key in matches:
+ if key in overlap_checkers:
current_spans_list = matches[key]
overlap_checker = overlap_checkers[key]
overlapped_pos = overlap_checker.search(current_span.begin, current_span.end)
|
PyFastNER/FastCNER.py
|
ReplaceText(target='overlap_checkers' @(327,22)->(327,29))
|
class FastCNER:
self.logger.debug(
'try add matched rule ({}-{})\t{}'.format(match_begin, match_end, str(self.rule_store[rule_id])))
current_span.rule_id = rule_id
if key in matches:
current_spans_list = matches[key]
overlap_checker = overlap_checkers[key]
overlapped_pos = overlap_checker.search(current_span.begin, current_span.end)
|
class FastCNER:
self.logger.debug(
'try add matched rule ({}-{})\t{}'.format(match_begin, match_end, str(self.rule_store[rule_id])))
current_span.rule_id = rule_id
if key in overlap_checkers:
current_spans_list = matches[key]
overlap_checker = overlap_checkers[key]
overlapped_pos = overlap_checker.search(current_span.begin, current_span.end)
|
2,312 |
https://:@github.com/groupe-conseil-nutshimit-nippour/django-geoprisma.git
|
a3a4dc9a0142e237dfac5961107b7338e9ef6298
|
@@ -142,7 +142,7 @@ class FeatureServerProxyFactory(object):
def isCreate(self):
data_id = self.featureServerProxy.getID()
- return data_id is not None and self.request.body != "" and self.request.method == "POST"
+ return data_id is None and self.request.body != "" and self.request.method == "POST"
def isUpdate(self):
data_id = self.featureServerProxy.getID()
|
geoprisma/core/proxies/featureserverproxy.py
|
ReplaceText(target=' is ' @(145,22)->(145,30))
|
class FeatureServerProxyFactory(object):
def isCreate(self):
data_id = self.featureServerProxy.getID()
return data_id is not None and self.request.body != "" and self.request.method == "POST"
def isUpdate(self):
data_id = self.featureServerProxy.getID()
|
class FeatureServerProxyFactory(object):
def isCreate(self):
data_id = self.featureServerProxy.getID()
return data_id is None and self.request.body != "" and self.request.method == "POST"
def isUpdate(self):
data_id = self.featureServerProxy.getID()
|
2,313 |
https://:@github.com/collective/mr.poe.git
|
dedca8e5c98124f6a43a18986e142e8cb7ecc3cf
|
@@ -59,7 +59,7 @@ def setup_logging(handler, exclude=['raven', 'sentry.errors']):
Returns a boolean based on if logging was configured or not.
"""
logger = logging.getLogger()
- if handler.__class__ not in map(type, logger.handlers):
+ if handler.__class__ in map(type, logger.handlers):
return False
logger.addHandler(handler)
|
raven/conf/__init__.py
|
ReplaceText(target=' in ' @(62,24)->(62,32))
|
def setup_logging(handler, exclude=['raven', 'sentry.errors']):
Returns a boolean based on if logging was configured or not.
"""
logger = logging.getLogger()
if handler.__class__ not in map(type, logger.handlers):
return False
logger.addHandler(handler)
|
def setup_logging(handler, exclude=['raven', 'sentry.errors']):
Returns a boolean based on if logging was configured or not.
"""
logger = logging.getLogger()
if handler.__class__ in map(type, logger.handlers):
return False
logger.addHandler(handler)
|
2,314 |
https://:@github.com/espenmn/medialog.mobilethemeTwo.git
|
21b37308d028659d0a540289ba2e1f30340a9481
|
@@ -39,7 +39,7 @@ class Scrape(BrowserView):
parts = url.split('//', 1)
this_base_url = parts[0]+'//'+parts[1].split('/', 1)[0]
- if url not in scrape_whitelist:
+ if this_base_url not in scrape_whitelist:
return "URL domain is not in whitelist"
#get html from the requested url
|
medialog/mobilethemeTwo/views.py
|
ReplaceText(target='this_base_url' @(42,11)->(42,14))
|
class Scrape(BrowserView):
parts = url.split('//', 1)
this_base_url = parts[0]+'//'+parts[1].split('/', 1)[0]
if url not in scrape_whitelist:
return "URL domain is not in whitelist"
#get html from the requested url
|
class Scrape(BrowserView):
parts = url.split('//', 1)
this_base_url = parts[0]+'//'+parts[1].split('/', 1)[0]
if this_base_url not in scrape_whitelist:
return "URL domain is not in whitelist"
#get html from the requested url
|
2,315 |
https://:@github.com/pyrated/vinyl.git
|
dec097819ed4e5635c3d8f64754fa6f3531278cc
|
@@ -18,7 +18,7 @@ import os
import sphinx_rtd_theme
# We cannot install llvmlite on READTHEDOCS
-if os.environ.get('READTHEDOCS') != 'True':
+if os.environ.get('READTHEDOCS') == 'True':
from unittest.mock import MagicMock
class MockModule(MagicMock):
@classmethod
|
docs/source/conf.py
|
ReplaceText(target='==' @(21,33)->(21,35))
|
import os
import sphinx_rtd_theme
# We cannot install llvmlite on READTHEDOCS
if os.environ.get('READTHEDOCS') != 'True':
from unittest.mock import MagicMock
class MockModule(MagicMock):
@classmethod
|
import os
import sphinx_rtd_theme
# We cannot install llvmlite on READTHEDOCS
if os.environ.get('READTHEDOCS') == 'True':
from unittest.mock import MagicMock
class MockModule(MagicMock):
@classmethod
|
2,316 |
https://:@github.com/marcofinalist/weathervane.git
|
4ddc4538762f2bb439b55019f2fea94480654988
|
@@ -169,6 +169,6 @@ class WeatherVaneInterface(object):
bits = self.spi.read_pin(self.station_bits)
result = 0
for index, value in enumerate(bits):
- result = value * 2**index
+ result += value * 2**index
return self.STATIONS[result]
\ No newline at end of file
|
weathervane/weathervaneinterface.py
|
ReplaceText(target='+=' @(172,19)->(172,20))
|
class WeatherVaneInterface(object):
bits = self.spi.read_pin(self.station_bits)
result = 0
for index, value in enumerate(bits):
result = value * 2**index
return self.STATIONS[result]
\ No newline at end of file
|
class WeatherVaneInterface(object):
bits = self.spi.read_pin(self.station_bits)
result = 0
for index, value in enumerate(bits):
result += value * 2**index
return self.STATIONS[result]
\ No newline at end of file
|
2,317 |
https://:@github.com/ofgulban/compoda.git
|
610dfad6de73410b4b99e8e9260b06dbe4ab91b7
|
@@ -49,7 +49,7 @@ def closure(data, k=1.0):
out = np.copy(data)
for i in range(data.shape[1]):
out[:, i] = np.divide(out[:, i], data_sum)
- out = data * k
+ out = out * k
return out
|
compoda/core.py
|
ReplaceText(target='out' @(52,10)->(52,14))
|
def closure(data, k=1.0):
out = np.copy(data)
for i in range(data.shape[1]):
out[:, i] = np.divide(out[:, i], data_sum)
out = data * k
return out
|
def closure(data, k=1.0):
out = np.copy(data)
for i in range(data.shape[1]):
out[:, i] = np.divide(out[:, i], data_sum)
out = out * k
return out
|
2,318 |
https://:@github.com/monashbiomedicalimaging/arcana.git
|
bf8c927f1d0f7c4af2175ddec836c6ffea5a3858
|
@@ -88,7 +88,7 @@ class XNATSource(ArchiveSource, XNATMixin):
proj_summ_sess_name) = XNATArchive.project_summary_name(
project.id)
try:
- proc_session = xnat_login.experiments[
+ proc_session = subject.experiments[
self.session_id + XNATArchive.PROCESSED_SUFFIX]
proc_datasets = dict(
(s.type, s) for s in proc_session.scans.itervalues())
|
nianalysis/archive/xnat.py
|
ReplaceText(target='subject' @(91,31)->(91,41))
|
class XNATSource(ArchiveSource, XNATMixin):
proj_summ_sess_name) = XNATArchive.project_summary_name(
project.id)
try:
proc_session = xnat_login.experiments[
self.session_id + XNATArchive.PROCESSED_SUFFIX]
proc_datasets = dict(
(s.type, s) for s in proc_session.scans.itervalues())
|
class XNATSource(ArchiveSource, XNATMixin):
proj_summ_sess_name) = XNATArchive.project_summary_name(
project.id)
try:
proc_session = subject.experiments[
self.session_id + XNATArchive.PROCESSED_SUFFIX]
proc_datasets = dict(
(s.type, s) for s in proc_session.scans.itervalues())
|
2,319 |
https://:@github.com/monashbiomedicalimaging/arcana.git
|
31946410e68317edea4ee76c6c24441065fc93ae
|
@@ -417,7 +417,7 @@ class TestProjectInfo(BaseMultiSubjectTestCase):
proj_dir, subject.id, SUMMARY_NAME,
dataset.filename)
for session in subject.sessions:
- for dataset in subject.datasets:
+ for dataset in session.datasets:
dataset.path = os.path.join(
proj_dir, session.subject_id,
session.visit_id, dataset.filename)
|
test/unittests/archive/test_local.py
|
ReplaceText(target='session' @(420,35)->(420,42))
|
class TestProjectInfo(BaseMultiSubjectTestCase):
proj_dir, subject.id, SUMMARY_NAME,
dataset.filename)
for session in subject.sessions:
for dataset in subject.datasets:
dataset.path = os.path.join(
proj_dir, session.subject_id,
session.visit_id, dataset.filename)
|
class TestProjectInfo(BaseMultiSubjectTestCase):
proj_dir, subject.id, SUMMARY_NAME,
dataset.filename)
for session in subject.sessions:
for dataset in session.datasets:
dataset.path = os.path.join(
proj_dir, session.subject_id,
session.visit_id, dataset.filename)
|
2,320 |
https://:@github.com/monashbiomedicalimaging/arcana.git
|
6bf29a9a339f7985d310d567a66cfca085cc4708
|
@@ -268,7 +268,7 @@ class BaseArchiveSink(BaseArchiveNode):
PATH_TRAIT)
# Add input fields
for field in fields:
- assert isinstance(dataset, FieldSpec)
+ assert isinstance(field, FieldSpec)
self._add_trait(self.inputs, field.name + FIELD_SUFFIX,
field.dtype)
|
nianalysis/archive/base.py
|
ReplaceText(target='field' @(271,30)->(271,37))
|
class BaseArchiveSink(BaseArchiveNode):
PATH_TRAIT)
# Add input fields
for field in fields:
assert isinstance(dataset, FieldSpec)
self._add_trait(self.inputs, field.name + FIELD_SUFFIX,
field.dtype)
|
class BaseArchiveSink(BaseArchiveNode):
PATH_TRAIT)
# Add input fields
for field in fields:
assert isinstance(field, FieldSpec)
self._add_trait(self.inputs, field.name + FIELD_SUFFIX,
field.dtype)
|
2,321 |
https://:@github.com/monashbiomedicalimaging/arcana.git
|
8bf472a77f50efc4d28d38d4aca46200c35e87e2
|
@@ -353,7 +353,7 @@ class LocalArchive(Archive):
Dataset.from_path(
os.path.join(session_path, dname),
multiplicity=multiplicity))
- if FIELDS_FNAME in dname:
+ if FIELDS_FNAME in dnames:
fields = self.fields_from_json(os.path.join(
session_path, FIELDS_FNAME),
multiplicity=multiplicity)
|
nianalysis/archive/local.py
|
ReplaceText(target='dnames' @(356,31)->(356,36))
|
class LocalArchive(Archive):
Dataset.from_path(
os.path.join(session_path, dname),
multiplicity=multiplicity))
if FIELDS_FNAME in dname:
fields = self.fields_from_json(os.path.join(
session_path, FIELDS_FNAME),
multiplicity=multiplicity)
|
class LocalArchive(Archive):
Dataset.from_path(
os.path.join(session_path, dname),
multiplicity=multiplicity))
if FIELDS_FNAME in dnames:
fields = self.fields_from_json(os.path.join(
session_path, FIELDS_FNAME),
multiplicity=multiplicity)
|
2,322 |
https://:@github.com/monashbiomedicalimaging/arcana.git
|
b3a6e4ec34cbfc92ef62dde78f995c4075feaa46
|
@@ -424,7 +424,7 @@ class Study(object):
"is not a valid option ('{}')".format(
"', '".join(unrecognised_values), name,
self._param_error_location,
- "', '".join(switch.choices)))
+ "', '".join(spec.choices)))
if self._referenced_switches is not None:
self._referenced_switches.add(name)
return switch.value in values
|
arcana/study/base.py
|
ReplaceText(target='spec' @(427,32)->(427,38))
|
class Study(object):
"is not a valid option ('{}')".format(
"', '".join(unrecognised_values), name,
self._param_error_location,
"', '".join(switch.choices)))
if self._referenced_switches is not None:
self._referenced_switches.add(name)
return switch.value in values
|
class Study(object):
"is not a valid option ('{}')".format(
"', '".join(unrecognised_values), name,
self._param_error_location,
"', '".join(spec.choices)))
if self._referenced_switches is not None:
self._referenced_switches.add(name)
return switch.value in values
|
2,323 |
https://:@github.com/monashbiomedicalimaging/arcana.git
|
be154e732593c6c8dd36391db9048315ad74fabc
|
@@ -214,7 +214,7 @@ class Study(object):
"to\n{}.".format(e, bound_inpt, spec))
raise e
else:
- if inpt.format not in spec.valid_formats:
+ if bound_inpt.format not in spec.valid_formats:
raise ArcanaUsageError(
"Cannot pass {} as an input to {} as it is"
" not in one of the valid formats ('{}')"
|
arcana/study/base.py
|
ReplaceText(target='bound_inpt' @(217,31)->(217,35))
|
class Study(object):
"to\n{}.".format(e, bound_inpt, spec))
raise e
else:
if inpt.format not in spec.valid_formats:
raise ArcanaUsageError(
"Cannot pass {} as an input to {} as it is"
" not in one of the valid formats ('{}')"
|
class Study(object):
"to\n{}.".format(e, bound_inpt, spec))
raise e
else:
if bound_inpt.format not in spec.valid_formats:
raise ArcanaUsageError(
"Cannot pass {} as an input to {} as it is"
" not in one of the valid formats ('{}')"
|
2,324 |
https://:@github.com/monashbiomedicalimaging/arcana.git
|
21199db4a62a164140eef9c7a45966056c069541
|
@@ -116,7 +116,7 @@ class ModulesEnvironment(BaseEnvironment):
.format(req.name, local_name))
avail_versions = []
for local_ver_name in version_names:
- ver_name = self.map_version(req_range, local_ver_name)
+ ver_name = self.map_version(req, local_ver_name)
try:
avail_versions.append(
req.v(ver_name, local_name=local_name,
|
arcana/environment/modules.py
|
ReplaceText(target='req' @(119,44)->(119,53))
|
class ModulesEnvironment(BaseEnvironment):
.format(req.name, local_name))
avail_versions = []
for local_ver_name in version_names:
ver_name = self.map_version(req_range, local_ver_name)
try:
avail_versions.append(
req.v(ver_name, local_name=local_name,
|
class ModulesEnvironment(BaseEnvironment):
.format(req.name, local_name))
avail_versions = []
for local_ver_name in version_names:
ver_name = self.map_version(req, local_ver_name)
try:
avail_versions.append(
req.v(ver_name, local_name=local_name,
|
2,325 |
https://:@github.com/monashbiomedicalimaging/arcana.git
|
35edaede784a97ddf4c0f961a4d6aac1cf3fb878
|
@@ -683,7 +683,7 @@ class Study(object):
in_branch = switch.value in values
if not in_branch:
try:
- in_branch = switch.fallbacks[switch.value] in values
+ in_branch = spec.fallbacks[switch.value] in values
except KeyError:
pass
return in_branch
|
arcana/study/base.py
|
ReplaceText(target='spec' @(686,32)->(686,38))
|
class Study(object):
in_branch = switch.value in values
if not in_branch:
try:
in_branch = switch.fallbacks[switch.value] in values
except KeyError:
pass
return in_branch
|
class Study(object):
in_branch = switch.value in values
if not in_branch:
try:
in_branch = spec.fallbacks[switch.value] in values
except KeyError:
pass
return in_branch
|
2,326 |
https://:@github.com/agartland/metadataVis.git
|
469bf0c8e514d09c966c73d21b9ff0c335d12255
|
@@ -39,7 +39,7 @@ def _generateWideform(longform_df, rx=None):
for entry in rowmeta_columns:
rowmeta_dict[entry] = longform_df[entry]
- if (rx is None):
+ if (rx is not None):
ptid_md = pd.DataFrame(data=rowmeta_dict,
columns=rowmeta_dict.keys())
ptid_md = ptid_md.drop_duplicates()
|
LongformReader.py
|
ReplaceText(target=' is not ' @(42,10)->(42,14))
|
def _generateWideform(longform_df, rx=None):
for entry in rowmeta_columns:
rowmeta_dict[entry] = longform_df[entry]
if (rx is None):
ptid_md = pd.DataFrame(data=rowmeta_dict,
columns=rowmeta_dict.keys())
ptid_md = ptid_md.drop_duplicates()
|
def _generateWideform(longform_df, rx=None):
for entry in rowmeta_columns:
rowmeta_dict[entry] = longform_df[entry]
if (rx is not None):
ptid_md = pd.DataFrame(data=rowmeta_dict,
columns=rowmeta_dict.keys())
ptid_md = ptid_md.drop_duplicates()
|
2,327 |
https://:@github.com/combatopera/pyven.git
|
2deb919498af4943aa18a7281037468809bda2d0
|
@@ -22,7 +22,7 @@ class BinMix(Node):
self.blockbuf.copybuf(self.tone(self.block))
if not noiseflag:
self.blockbuf.orbuf(self.noise(self.block))
- elif noiseflag:
+ elif not noiseflag:
self.blockbuf.copybuf(self.noise(self.block))
else:
self.blockbuf.fill(0)
|
pym2149/mix.py
|
ReplaceText(target='not ' @(25,9)->(25,9))
|
class BinMix(Node):
self.blockbuf.copybuf(self.tone(self.block))
if not noiseflag:
self.blockbuf.orbuf(self.noise(self.block))
elif noiseflag:
self.blockbuf.copybuf(self.noise(self.block))
else:
self.blockbuf.fill(0)
|
class BinMix(Node):
self.blockbuf.copybuf(self.tone(self.block))
if not noiseflag:
self.blockbuf.orbuf(self.noise(self.block))
elif not noiseflag:
self.blockbuf.copybuf(self.noise(self.block))
else:
self.blockbuf.fill(0)
|
2,328 |
https://:@github.com/rolurq/flask-gulp.git
|
615013477dcfcf53b64e5cba0858f3b01e66b8fb
|
@@ -61,7 +61,7 @@ def cjsx(filename, data):
command = "%s -c -s" % (executable or 'cjsx')
if bare:
- command = ' '.join((executable, '-b'))
+ command = ' '.join((command, '-b'))
return runner(command, filename, data, '.js')
|
flask_static/extensions.py
|
ReplaceText(target='command' @(64,28)->(64,38))
|
def cjsx(filename, data):
command = "%s -c -s" % (executable or 'cjsx')
if bare:
command = ' '.join((executable, '-b'))
return runner(command, filename, data, '.js')
|
def cjsx(filename, data):
command = "%s -c -s" % (executable or 'cjsx')
if bare:
command = ' '.join((command, '-b'))
return runner(command, filename, data, '.js')
|
2,329 |
https://:@github.com/jiep/unicode.git
|
fe2f6fe46f14778cfcb74852d817d1876d171352
|
@@ -92,7 +92,7 @@ def main():
print('Similar domains to {}'.format(dom))
domains.difference_update(set(dom))
for d in domains:
- print_diff(d, args.domain)
+ print_diff(args.domain, d)
if write:
f.write(d + "\n")
if (args.check):
|
v2d/main.py
|
ArgSwap(idxs=0<->1 @(95,16)->(95,26))
|
def main():
print('Similar domains to {}'.format(dom))
domains.difference_update(set(dom))
for d in domains:
print_diff(d, args.domain)
if write:
f.write(d + "\n")
if (args.check):
|
def main():
print('Similar domains to {}'.format(dom))
domains.difference_update(set(dom))
for d in domains:
print_diff(args.domain, d)
if write:
f.write(d + "\n")
if (args.check):
|
2,330 |
https://:@gitlab.com/admintotal/django-cfdi.git
|
9f131df38460abaafb00566867ce47b522fce7fd
|
@@ -882,7 +882,7 @@ def get_xml_object(xml_text):
nominas.append(nomina_object)
- if nomina_object:
+ if nominas:
xml.complemento.nominas = nominas
xml.complemento.nomina = nominas[0]
else:
|
cfdi/utils.py
|
ReplaceText(target='nominas' @(885,15)->(885,28))
|
def get_xml_object(xml_text):
nominas.append(nomina_object)
if nomina_object:
xml.complemento.nominas = nominas
xml.complemento.nomina = nominas[0]
else:
|
def get_xml_object(xml_text):
nominas.append(nomina_object)
if nominas:
xml.complemento.nominas = nominas
xml.complemento.nomina = nominas[0]
else:
|
2,331 |
https://:@github.com/ludeeus/addonupdater.git
|
fe325d28af7bcade5a806009c943fc0afbca63b1
|
@@ -378,7 +378,7 @@ class AddonUpdater():
remote_buildfile = self.get_file_obj(buildfile)
buildfile_content = self.get_file_content(remote_buildfile)
- used_file = remote_dockerfile.split('BUILD_FROM=hassioaddons/')[1]
+ used_file = dockerfile_content.split('BUILD_FROM=hassioaddons/')[1]
used_file = used_file.split('\n')[0]
base = used_file.split(':')[1]
|
addonupdater/updater.py
|
ReplaceText(target='dockerfile_content' @(381,20)->(381,37))
|
class AddonUpdater():
remote_buildfile = self.get_file_obj(buildfile)
buildfile_content = self.get_file_content(remote_buildfile)
used_file = remote_dockerfile.split('BUILD_FROM=hassioaddons/')[1]
used_file = used_file.split('\n')[0]
base = used_file.split(':')[1]
|
class AddonUpdater():
remote_buildfile = self.get_file_obj(buildfile)
buildfile_content = self.get_file_content(remote_buildfile)
used_file = dockerfile_content.split('BUILD_FROM=hassioaddons/')[1]
used_file = used_file.split('\n')[0]
base = used_file.split(':')[1]
|
2,332 |
https://:@github.com/VictorPavlushin/netbox-netdev-inventory.git
|
00111f3155731e4bf1b380744519a89576a96a49
|
@@ -161,7 +161,7 @@ class DeviceImporter(ContextDecorator):
self._search_key_case_insensitive(interfaces, lag)
)
except KeyError:
- logger.error("%s not exist in polled interfaces", ifname)
+ logger.error("%s not exist in polled interfaces", lag)
continue
interfaces[ifname]["lag"] = real_lag_name
|
netbox_netprod_importer/importer.py
|
ReplaceText(target='lag' @(164,66)->(164,72))
|
class DeviceImporter(ContextDecorator):
self._search_key_case_insensitive(interfaces, lag)
)
except KeyError:
logger.error("%s not exist in polled interfaces", ifname)
continue
interfaces[ifname]["lag"] = real_lag_name
|
class DeviceImporter(ContextDecorator):
self._search_key_case_insensitive(interfaces, lag)
)
except KeyError:
logger.error("%s not exist in polled interfaces", lag)
continue
interfaces[ifname]["lag"] = real_lag_name
|
2,333 |
https://:@github.com/tswicegood/cbv_utils.git
|
25d2e1ce328e485ec26f4debd8f7aebc1ee6a623
|
@@ -96,7 +96,7 @@ class ProcessInlineFormsetView(ProcessFormView):
obj = form.save(commit=False)
inline_formset = self.get_inline_formset()
if inline_formset.is_valid():
- form.save()
+ obj.save()
inline_formset.save()
return self.form_valid(form, inline_formset)
return self.form_invalid(form=form, inline_formset=inline_formset)
|
cbv_utils/views.py
|
ReplaceText(target='obj' @(99,16)->(99,20))
|
class ProcessInlineFormsetView(ProcessFormView):
obj = form.save(commit=False)
inline_formset = self.get_inline_formset()
if inline_formset.is_valid():
form.save()
inline_formset.save()
return self.form_valid(form, inline_formset)
return self.form_invalid(form=form, inline_formset=inline_formset)
|
class ProcessInlineFormsetView(ProcessFormView):
obj = form.save(commit=False)
inline_formset = self.get_inline_formset()
if inline_formset.is_valid():
obj.save()
inline_formset.save()
return self.form_valid(form, inline_formset)
return self.form_invalid(form=form, inline_formset=inline_formset)
|
2,334 |
https://:@github.com/romeric/florence.git
|
aa75d1bb83ec9de5e8ee60d9a87d2a86d7293aeb
|
@@ -6010,7 +6010,7 @@ class Mesh(object):
else:
quality_func = lambda mesh: mesh.Lengths()
elif quality_assessor == "aspect_ratio":
- quality_assessor = lambda mesh: mesh.AspectRatios()
+ quality_func = lambda mesh: mesh.AspectRatios()
elif quality_assessor == "angle":
quality_func = lambda mesh: mesh.Angles()
else:
|
Florence/MeshGeneration/Mesh.py
|
ReplaceText(target='quality_func' @(6013,16)->(6013,32))
|
class Mesh(object):
else:
quality_func = lambda mesh: mesh.Lengths()
elif quality_assessor == "aspect_ratio":
quality_assessor = lambda mesh: mesh.AspectRatios()
elif quality_assessor == "angle":
quality_func = lambda mesh: mesh.Angles()
else:
|
class Mesh(object):
else:
quality_func = lambda mesh: mesh.Lengths()
elif quality_assessor == "aspect_ratio":
quality_func = lambda mesh: mesh.AspectRatios()
elif quality_assessor == "angle":
quality_func = lambda mesh: mesh.Angles()
else:
|
2,335 |
https://:@github.com/DomainGroupOSS/ml-recsys-tools.git
|
467f8e1d859af1109bb830b3c35a752baeddbb67
|
@@ -81,7 +81,7 @@ class FactorizationRecommender(BaseDFSparseRecommender):
all_metrics.plot()
self.early_stop_metrics_df = all_metrics
- self._set_epochs(epochs=epochs_max)
+ self._set_epochs(epochs=max_epoch)
if not refit_on_all:
simple_logger.info('Loading best model from checkpoint at %d epochs' % max_epoch)
self.model, self.model_checkpoint = self.model_checkpoint, None
|
ml_recsys_tools/recommenders/factorization_base.py
|
ReplaceText(target='max_epoch' @(84,32)->(84,42))
|
class FactorizationRecommender(BaseDFSparseRecommender):
all_metrics.plot()
self.early_stop_metrics_df = all_metrics
self._set_epochs(epochs=epochs_max)
if not refit_on_all:
simple_logger.info('Loading best model from checkpoint at %d epochs' % max_epoch)
self.model, self.model_checkpoint = self.model_checkpoint, None
|
class FactorizationRecommender(BaseDFSparseRecommender):
all_metrics.plot()
self.early_stop_metrics_df = all_metrics
self._set_epochs(epochs=max_epoch)
if not refit_on_all:
simple_logger.info('Loading best model from checkpoint at %d epochs' % max_epoch)
self.model, self.model_checkpoint = self.model_checkpoint, None
|
2,336 |
https://:@github.com/south-coast-science/scs_core.git
|
6ca2ac668b486924816ed461e8f40a87d82136da
|
@@ -25,7 +25,7 @@ class Filesystem(object):
if head and not os.path.exists(head):
cls.mkdir(head)
- if os.path.exists(path): # handles case of trailing /
+ if not os.path.exists(path): # handles case of trailing /
os.mkdir(path)
|
src/scs_core/sys/filesystem.py
|
ReplaceText(target='not ' @(28,11)->(28,11))
|
class Filesystem(object):
if head and not os.path.exists(head):
cls.mkdir(head)
if os.path.exists(path): # handles case of trailing /
os.mkdir(path)
|
class Filesystem(object):
if head and not os.path.exists(head):
cls.mkdir(head)
if not os.path.exists(path): # handles case of trailing /
os.mkdir(path)
|
2,337 |
https://:@github.com/south-coast-science/scs_core.git
|
653bf53c4d76b1b422d8aa11174c9a5351bbd1f6
|
@@ -81,7 +81,7 @@ class ExegeteRenderingTRhRow(JSONable):
@classmethod
def construct(cls, gas, rh, t_min, t_max, t_delta, exegete: Exegete):
- cells = [ExegeteRenderingTRhCell(t, exegete.error(gas, t, rh))
+ cells = [ExegeteRenderingTRhCell(t, exegete.error(gas, rh, t))
for t in range(t_min, t_max + 1, t_delta)]
return ExegeteRenderingTRhRow(rh, cells)
|
src/scs_core/gas/exegesis/exegete_rendering_t_rh.py
|
ArgSwap(idxs=1<->2 @(84,44)->(84,57))
|
class ExegeteRenderingTRhRow(JSONable):
@classmethod
def construct(cls, gas, rh, t_min, t_max, t_delta, exegete: Exegete):
cells = [ExegeteRenderingTRhCell(t, exegete.error(gas, t, rh))
for t in range(t_min, t_max + 1, t_delta)]
return ExegeteRenderingTRhRow(rh, cells)
|
class ExegeteRenderingTRhRow(JSONable):
@classmethod
def construct(cls, gas, rh, t_min, t_max, t_delta, exegete: Exegete):
cells = [ExegeteRenderingTRhCell(t, exegete.error(gas, rh, t))
for t in range(t_min, t_max + 1, t_delta)]
return ExegeteRenderingTRhRow(rh, cells)
|
2,338 |
https://:@github.com/south-coast-science/scs_core.git
|
653bf53c4d76b1b422d8aa11174c9a5351bbd1f6
|
@@ -43,7 +43,7 @@ print("-")
for rh in range(10, 91, 5):
for t in range(0, 46, 5):
- interpretation = exegete.interpretation('NO2', text, t, rh)
+ interpretation = exegete.interpretation('NO2', text, rh, t)
print("rh: %2d t: %2d text: %3.1f interpretation: %3.1f" % (rh, t, text, interpretation))
print("-")
|
tests/gas/exegesis/sbl1/sbl1_no2_v1_test.py
|
ArgSwap(idxs=2<->3 @(46,25)->(46,47))
|
print("-")
for rh in range(10, 91, 5):
for t in range(0, 46, 5):
interpretation = exegete.interpretation('NO2', text, t, rh)
print("rh: %2d t: %2d text: %3.1f interpretation: %3.1f" % (rh, t, text, interpretation))
print("-")
|
print("-")
for rh in range(10, 91, 5):
for t in range(0, 46, 5):
interpretation = exegete.interpretation('NO2', text, rh, t)
print("rh: %2d t: %2d text: %3.1f interpretation: %3.1f" % (rh, t, text, interpretation))
print("-")
|
2,339 |
https://:@github.com/south-coast-science/scs_core.git
|
73c3c80fcc88fce844e0e387499753667dac811e
|
@@ -68,7 +68,7 @@ class S3Manager(object):
bucket_list.append(str(inters), bucket["Name"])
inters += 1
- return bucket_list
+ return response
def retrieve_from_bucket(self, bucket_name, resource_name):
|
src/scs_core/aws/manager/s3_manager.py
|
ReplaceText(target='response' @(71,15)->(71,26))
|
class S3Manager(object):
bucket_list.append(str(inters), bucket["Name"])
inters += 1
return bucket_list
def retrieve_from_bucket(self, bucket_name, resource_name):
|
class S3Manager(object):
bucket_list.append(str(inters), bucket["Name"])
inters += 1
return response
def retrieve_from_bucket(self, bucket_name, resource_name):
|
2,340 |
https://:@github.com/south-coast-science/scs_core.git
|
ac11b0e941914f83f984d183811a83bb0a5df544
|
@@ -59,7 +59,7 @@ class AccessKey(PersistentJSONable):
@classmethod
def persistence_location(cls, host):
- return host.aws_dir(), cls.__FILENAME
+ return cls.aws_dir(), cls.__FILENAME
@classmethod
|
src/scs_core/aws/client/access_key.py
|
ReplaceText(target='cls' @(62,15)->(62,19))
|
class AccessKey(PersistentJSONable):
@classmethod
def persistence_location(cls, host):
return host.aws_dir(), cls.__FILENAME
@classmethod
|
class AccessKey(PersistentJSONable):
@classmethod
def persistence_location(cls, host):
return cls.aws_dir(), cls.__FILENAME
@classmethod
|
2,341 |
https://:@github.com/south-coast-science/scs_core.git
|
e6293b84e0b7594738b85db5a7365fdbb5770461
|
@@ -53,7 +53,7 @@ class DeviceTester(object):
delta = now - latest_pub
elapsed_minutes = delta.total_seconds() / 60
- return elapsed_minutes > self.__config.unresponsive_minutes_allowed
+ return elapsed_minutes < self.__config.unresponsive_minutes_allowed
def has_status_changed(self, s3_device_status_list):
|
src/scs_core/aws/monitor/device_tester.py
|
ReplaceText(target='<' @(56,35)->(56,36))
|
class DeviceTester(object):
delta = now - latest_pub
elapsed_minutes = delta.total_seconds() / 60
return elapsed_minutes > self.__config.unresponsive_minutes_allowed
def has_status_changed(self, s3_device_status_list):
|
class DeviceTester(object):
delta = now - latest_pub
elapsed_minutes = delta.total_seconds() / 60
return elapsed_minutes < self.__config.unresponsive_minutes_allowed
def has_status_changed(self, s3_device_status_list):
|
2,342 |
https://:@github.com/davidfstr/notifymail.git
|
585ae19eba35db57e969941ea2c340bd3be499d8
|
@@ -8,7 +8,7 @@ def is_older_than(file1, file2):
return os.path.getmtime(file1) < os.path.getmtime(file2)
# Generate README.rst if missing or out of date
-if not os.path.exists('README.rst') and is_older_than('README.rst', 'README.md'):
+if not os.path.exists('README.rst') or is_older_than('README.rst', 'README.md'):
os.system('pandoc --from=markdown --to=rst --output=README.rst README.md')
with open('README.rst') as file:
long_description = file.read()
|
setup.py
|
ReplaceText(target='or' @(11,36)->(11,39))
|
def is_older_than(file1, file2):
return os.path.getmtime(file1) < os.path.getmtime(file2)
# Generate README.rst if missing or out of date
if not os.path.exists('README.rst') and is_older_than('README.rst', 'README.md'):
os.system('pandoc --from=markdown --to=rst --output=README.rst README.md')
with open('README.rst') as file:
long_description = file.read()
|
def is_older_than(file1, file2):
return os.path.getmtime(file1) < os.path.getmtime(file2)
# Generate README.rst if missing or out of date
if not os.path.exists('README.rst') or is_older_than('README.rst', 'README.md'):
os.system('pandoc --from=markdown --to=rst --output=README.rst README.md')
with open('README.rst') as file:
long_description = file.read()
|
2,343 |
https://:@github.com/hile/oodi.git
|
169b7db094c64a5c2ca695bf93782fce0b50945f
|
@@ -26,7 +26,7 @@ class Command(ScriptCommand):
iterators = []
for path in paths:
try:
- iterators.append(IterableTrackPaths(self.script.configuration, path))
+ iterators.append(IterableTrackPaths(path, self.script.configuration))
except LibraryError as e:
self.error(e)
return iterators
|
oodi/bin/commands/base.py
|
ArgSwap(idxs=0<->1 @(29,33)->(29,51))
|
class Command(ScriptCommand):
iterators = []
for path in paths:
try:
iterators.append(IterableTrackPaths(self.script.configuration, path))
except LibraryError as e:
self.error(e)
return iterators
|
class Command(ScriptCommand):
iterators = []
for path in paths:
try:
iterators.append(IterableTrackPaths(path, self.script.configuration))
except LibraryError as e:
self.error(e)
return iterators
|
2,344 |
https://:@github.com/tpm2-software/tpm2-pytss.git
|
675d845e04810a4129e9e5b21889a0258fcad7b2
|
@@ -141,7 +141,7 @@ class BaseContextMetaClass(type):
and "uint8_t" in docstring.split()
):
return_value.append(
- to_bytearray(value.value, args[i + 1].value)
+ to_bytearray(args[i + 1].value, value.value)
)
skip = True
continue
|
tpm2_pytss/context.py
|
ArgSwap(idxs=0<->1 @(144,36)->(144,48))
|
class BaseContextMetaClass(type):
and "uint8_t" in docstring.split()
):
return_value.append(
to_bytearray(value.value, args[i + 1].value)
)
skip = True
continue
|
class BaseContextMetaClass(type):
and "uint8_t" in docstring.split()
):
return_value.append(
to_bytearray(args[i + 1].value, value.value)
)
skip = True
continue
|
2,345 |
https://:@github.com/josesho/bootstrap_contrast.git
|
1d55129dee9130374774bb1234b577e061974d93
|
@@ -1089,7 +1089,7 @@ def pairedcontrast(data, x, y, idcol, hue = None,
linestyle = 'dotted')
# Set xlimit to appropriate limits..
- newxlim = (ax_left.get_xlim()[0], xpos + 0.25)
+ newxlim = (ax_left.get_xlim()[0], xposPlusViolin + 0.25)
ax_left.set_xlim(newxlim)
# Remove left axes x-axis title.
|
bootstrapContrast/bootstrapContrast.py
|
ReplaceText(target='xposPlusViolin' @(1092,38)->(1092,42))
|
def pairedcontrast(data, x, y, idcol, hue = None,
linestyle = 'dotted')
# Set xlimit to appropriate limits..
newxlim = (ax_left.get_xlim()[0], xpos + 0.25)
ax_left.set_xlim(newxlim)
# Remove left axes x-axis title.
|
def pairedcontrast(data, x, y, idcol, hue = None,
linestyle = 'dotted')
# Set xlimit to appropriate limits..
newxlim = (ax_left.get_xlim()[0], xposPlusViolin + 0.25)
ax_left.set_xlim(newxlim)
# Remove left axes x-axis title.
|
2,346 |
https://:@github.com/josesho/bootstrap_contrast.git
|
5875648efb0994fdac3eae216fe36d51cc0f629c
|
@@ -88,7 +88,7 @@ def plotbootstrap_hubspoke(bslist, ax, violinWidth, violinOffset,
for i in range(0, len(bslist)):
bsi=bslist[i]
# array=list(bsi.items())[7][1] # Pull out the bootstrapped array.
- array=bslist['diffarray']
+ array=bsi['diffarray']
ylims.append(array)
# Then plot as violinplot.
|
bootstrapContrast/plot_bootstrap_tools.py
|
ReplaceText(target='bsi' @(91,14)->(91,20))
|
def plotbootstrap_hubspoke(bslist, ax, violinWidth, violinOffset,
for i in range(0, len(bslist)):
bsi=bslist[i]
# array=list(bsi.items())[7][1] # Pull out the bootstrapped array.
array=bslist['diffarray']
ylims.append(array)
# Then plot as violinplot.
|
def plotbootstrap_hubspoke(bslist, ax, violinWidth, violinOffset,
for i in range(0, len(bslist)):
bsi=bslist[i]
# array=list(bsi.items())[7][1] # Pull out the bootstrapped array.
array=bsi['diffarray']
ylims.append(array)
# Then plot as violinplot.
|
2,347 |
https://:@gitlab.com/harry.sky.vortex/melodiam.git
|
bbc45aeb762242382c68cfe6fa9a32e600eb3630
|
@@ -106,7 +106,7 @@ class SpotifyAPI(object):
# Update ETag if song's playback was manipulated
if song["progress_ms"] < self.current_song["progress"] or song["progress_ms"] - 10000 > self.current_song["progress"]:
self.current_song_json_updated = str(time())
- LISTEN_ALONG_API.set_current_playing_song(song_uri=song["uri"], position_ms=song["progress_ms"])
+ LISTEN_ALONG_API.set_current_playing_song(song_uri=item["uri"], position_ms=song["progress_ms"])
self.current_song["progress"] = song["progress_ms"]
self.current_song_json = json.dumps(self.current_song)
|
backend/spotify.py
|
ReplaceText(target='item' @(109,71)->(109,75))
|
class SpotifyAPI(object):
# Update ETag if song's playback was manipulated
if song["progress_ms"] < self.current_song["progress"] or song["progress_ms"] - 10000 > self.current_song["progress"]:
self.current_song_json_updated = str(time())
LISTEN_ALONG_API.set_current_playing_song(song_uri=song["uri"], position_ms=song["progress_ms"])
self.current_song["progress"] = song["progress_ms"]
self.current_song_json = json.dumps(self.current_song)
|
class SpotifyAPI(object):
# Update ETag if song's playback was manipulated
if song["progress_ms"] < self.current_song["progress"] or song["progress_ms"] - 10000 > self.current_song["progress"]:
self.current_song_json_updated = str(time())
LISTEN_ALONG_API.set_current_playing_song(song_uri=item["uri"], position_ms=song["progress_ms"])
self.current_song["progress"] = song["progress_ms"]
self.current_song_json = json.dumps(self.current_song)
|
2,348 |
https://:@gitlab.com/harry.sky.vortex/melodiam.git
|
db126e963e784dee9d11f2331622ef9eca5baf9d
|
@@ -32,7 +32,7 @@ async def get_listen_along_users_endpoint(request: StarletteRequest) -> PlainTex
for user in ListenAlong.users:
users_json += user.public_json + ","
- users_json = ']'
+ users_json += ']'
return PlainTextResponse(content=users_json, media_type="application/json")
@SERVER.route('/get_current_song', methods=['GET'])
|
backend/music/main.py
|
ReplaceText(target='+=' @(35,15)->(35,16))
|
async def get_listen_along_users_endpoint(request: StarletteRequest) -> PlainTex
for user in ListenAlong.users:
users_json += user.public_json + ","
users_json = ']'
return PlainTextResponse(content=users_json, media_type="application/json")
@SERVER.route('/get_current_song', methods=['GET'])
|
async def get_listen_along_users_endpoint(request: StarletteRequest) -> PlainTex
for user in ListenAlong.users:
users_json += user.public_json + ","
users_json += ']'
return PlainTextResponse(content=users_json, media_type="application/json")
@SERVER.route('/get_current_song', methods=['GET'])
|
2,349 |
https://:@gitlab.com/harry.sky.vortex/melodiam.git
|
9850cc42634b01312cc8754a03df8abccd2054ce
|
@@ -21,7 +21,7 @@ class ListenAlong():
@staticmethod
def _set_song(user: ListenAlongUser, song_json: str) -> None:
if user.tokens:
- status = SpotifyWebAPI.set_current_playing_song(song_json, user.tokens.access)
+ status = SpotifyWebAPI.set_current_playing_song(user.tokens.access, song_json)
if user.public.status != status:
user.public.status = status
user.public_json = json.dumps(asdict(user.public))
|
backend/music/features/listen_along.py
|
ArgSwap(idxs=0<->1 @(24,21)->(24,59))
|
class ListenAlong():
@staticmethod
def _set_song(user: ListenAlongUser, song_json: str) -> None:
if user.tokens:
status = SpotifyWebAPI.set_current_playing_song(song_json, user.tokens.access)
if user.public.status != status:
user.public.status = status
user.public_json = json.dumps(asdict(user.public))
|
class ListenAlong():
@staticmethod
def _set_song(user: ListenAlongUser, song_json: str) -> None:
if user.tokens:
status = SpotifyWebAPI.set_current_playing_song(user.tokens.access, song_json)
if user.public.status != status:
user.public.status = status
user.public_json = json.dumps(asdict(user.public))
|
2,350 |
https://:@github.com/ladybug-tools/honeybee-radiance-command.git
|
fe62562a0b228b8e46309aa88c116882247c89b9
|
@@ -18,7 +18,7 @@ def run_command(input_command, env=None, cwd=None):
if platform.system() == 'Windows':
command = input_command.replace('\'', '"')
else:
- command = command.replace('"', '\'')
+ command = input_command.replace('"', '\'')
# change cwd - Popen cwd input simply doesn't work.
cur_dir = os.getcwd()
|
honeybee_radiance_command/_command_util.py
|
ReplaceText(target='input_command' @(21,18)->(21,25))
|
def run_command(input_command, env=None, cwd=None):
if platform.system() == 'Windows':
command = input_command.replace('\'', '"')
else:
command = command.replace('"', '\'')
# change cwd - Popen cwd input simply doesn't work.
cur_dir = os.getcwd()
|
def run_command(input_command, env=None, cwd=None):
if platform.system() == 'Windows':
command = input_command.replace('\'', '"')
else:
command = input_command.replace('"', '\'')
# change cwd - Popen cwd input simply doesn't work.
cur_dir = os.getcwd()
|
2,351 |
https://:@bitbucket.org/bertrandboichon/pi.hifi.git
|
bf0316f181d6e6597722b0f3023f70a592c49d04
|
@@ -11,7 +11,7 @@ class post_install(install):
install.run(self)
print("*** Executing post install actions:")
# update mpd configuration if necessary
- if '/tmp/mpd.fifo' in open('/etc/mpd.conf').read():
+ if '/tmp/mpd.fifo' not in open('/etc/mpd.conf').read():
os.system("sudo cat /etc/fifo-mpd.conf >> /etc/mpd.conf")
os.system("sudo service mpd restart")
# update music display init script
|
setup.py
|
ReplaceText(target=' not in ' @(14,26)->(14,30))
|
class post_install(install):
install.run(self)
print("*** Executing post install actions:")
# update mpd configuration if necessary
if '/tmp/mpd.fifo' in open('/etc/mpd.conf').read():
os.system("sudo cat /etc/fifo-mpd.conf >> /etc/mpd.conf")
os.system("sudo service mpd restart")
# update music display init script
|
class post_install(install):
install.run(self)
print("*** Executing post install actions:")
# update mpd configuration if necessary
if '/tmp/mpd.fifo' not in open('/etc/mpd.conf').read():
os.system("sudo cat /etc/fifo-mpd.conf >> /etc/mpd.conf")
os.system("sudo service mpd restart")
# update music display init script
|
2,352 |
https://:@github.com/hotoffthehamster/dob.git
|
9615cad9da920842a23b9a124b1d39356ebe7d2e
|
@@ -43,7 +43,7 @@ def echo_copyright():
cur_year = str(datetime.now().year)
year_range = '2018'
if cur_year != year_range:
- year_range = '2018-{}'.format(year_range)
+ year_range = '2018-{}'.format(cur_year)
gpl3_notice_2018 = [
'{app_name} {version}'.format(
app_name=__BigName__,
|
dob/copyright.py
|
ReplaceText(target='cur_year' @(46,38)->(46,48))
|
def echo_copyright():
cur_year = str(datetime.now().year)
year_range = '2018'
if cur_year != year_range:
year_range = '2018-{}'.format(year_range)
gpl3_notice_2018 = [
'{app_name} {version}'.format(
app_name=__BigName__,
|
def echo_copyright():
cur_year = str(datetime.now().year)
year_range = '2018'
if cur_year != year_range:
year_range = '2018-{}'.format(cur_year)
gpl3_notice_2018 = [
'{app_name} {version}'.format(
app_name=__BigName__,
|
2,353 |
https://:@github.com/superadm1n/CiscoAutomationFramework.git
|
2ffcf5da63e3479f7bcae3d552f04695b36d9466
|
@@ -594,7 +594,7 @@ class IOS(TerminalCommands, CommandGetMethods):
)
#mac_table_list.append(line.split())
- if len(line.split()) >= 1:
+ if len(line.split()) > 1:
if '--' in line.split()[0]:
flag = 1
|
CiscoAutomationFramework/CiscoIOS.py
|
ReplaceText(target='>' @(597,33)->(597,35))
|
class IOS(TerminalCommands, CommandGetMethods):
)
#mac_table_list.append(line.split())
if len(line.split()) >= 1:
if '--' in line.split()[0]:
flag = 1
|
class IOS(TerminalCommands, CommandGetMethods):
)
#mac_table_list.append(line.split())
if len(line.split()) > 1:
if '--' in line.split()[0]:
flag = 1
|
2,354 |
https://:@github.com/bruth/ipipe.git
|
dc24d06d645ab03e988a066267787aff98baee32
|
@@ -20,7 +20,7 @@ class Parser(object):
output.extend(parsed)
else:
output.append(parsed)
- return parsed
+ return output
class FileParser(Parser):
|
pipes/parser.py
|
ReplaceText(target='output' @(23,15)->(23,21))
|
class Parser(object):
output.extend(parsed)
else:
output.append(parsed)
return parsed
class FileParser(Parser):
|
class Parser(object):
output.extend(parsed)
else:
output.append(parsed)
return output
class FileParser(Parser):
|
2,355 |
https://:@github.com/la-mar/permian-frac-exchange.git
|
3d0d86bfbb6871abdded1c5357dac5ca1c3ed756
|
@@ -249,7 +249,7 @@ class Parser(object):
) -> pd.Series:
try:
apply_to = apply_to or apply_on
- self.df[apply_on] = self.df[apply_on].apply(func)
+ self.df[apply_to] = self.df[apply_on].apply(func)
except KeyError as ke:
logger.debug(
MSG_PARSER_CHECK.format(op_name=self.operator.name, col_name=apply_on)
|
src/fsec/parser.py
|
ReplaceText(target='apply_to' @(252,20)->(252,28))
|
class Parser(object):
) -> pd.Series:
try:
apply_to = apply_to or apply_on
self.df[apply_on] = self.df[apply_on].apply(func)
except KeyError as ke:
logger.debug(
MSG_PARSER_CHECK.format(op_name=self.operator.name, col_name=apply_on)
|
class Parser(object):
) -> pd.Series:
try:
apply_to = apply_to or apply_on
self.df[apply_to] = self.df[apply_on].apply(func)
except KeyError as ke:
logger.debug(
MSG_PARSER_CHECK.format(op_name=self.operator.name, col_name=apply_on)
|
2,356 |
https://:@bitbucket.org/sambowers/biota.git
|
b9c2b1b28e5a5fb57b30e7c474ecf2e6f729edeb
|
@@ -304,9 +304,9 @@ def calculateTWC(tile, patch_size = 'auto', output = False, show = False):
# Extract the data
WC = woody_cover[ymin:ymax, xmin:xmax]
-
+
# If at least 50 % of data is present...
- if TWC.mask.sum() <= ((patch_size ** 2) * 0.5):
+ if WC.mask.sum() <= ((patch_size ** 2) * 0.5):
# Calculate proportion of woody cover in patch
TWC.data[n, m] = int(round((float(WC.sum()) / ((patch_size ** 2) - WC.mask.sum())) * 100))
|
biota/indices.py
|
ReplaceText(target='WC' @(309,11)->(309,14))
|
def calculateTWC(tile, patch_size = 'auto', output = False, show = False):
# Extract the data
WC = woody_cover[ymin:ymax, xmin:xmax]
# If at least 50 % of data is present...
if TWC.mask.sum() <= ((patch_size ** 2) * 0.5):
# Calculate proportion of woody cover in patch
TWC.data[n, m] = int(round((float(WC.sum()) / ((patch_size ** 2) - WC.mask.sum())) * 100))
|
def calculateTWC(tile, patch_size = 'auto', output = False, show = False):
# Extract the data
WC = woody_cover[ymin:ymax, xmin:xmax]
# If at least 50 % of data is present...
if WC.mask.sum() <= ((patch_size ** 2) * 0.5):
# Calculate proportion of woody cover in patch
TWC.data[n, m] = int(round((float(WC.sum()) / ((patch_size ** 2) - WC.mask.sum())) * 100))
|
2,357 |
https://:@github.com/yedhrab/YInstabot.git
|
c9d3f3b18656d7b19eb7b179fc7adb171bb6efe8
|
@@ -106,7 +106,7 @@ def main():
DEBUG, WAIT, NO_REFRESH, PATHS = not args.quite, args.wait, args.noRefresh, args.paths
for PATH in PATHS:
- if not os.path.isfile(PATHS):
+ if not os.path.isfile(PATH):
print(f"`{PATH}` dosyaya ait değil.")
continue
|
yinstabot/workspace.py
|
ReplaceText(target='PATH' @(109,30)->(109,35))
|
def main():
DEBUG, WAIT, NO_REFRESH, PATHS = not args.quite, args.wait, args.noRefresh, args.paths
for PATH in PATHS:
if not os.path.isfile(PATHS):
print(f"`{PATH}` dosyaya ait değil.")
continue
|
def main():
DEBUG, WAIT, NO_REFRESH, PATHS = not args.quite, args.wait, args.noRefresh, args.paths
for PATH in PATHS:
if not os.path.isfile(PATH):
print(f"`{PATH}` dosyaya ait değil.")
continue
|
2,358 |
https://:@github.com/fladi/pyrc522.git
|
83bd4bd1c169259a8e4c3e7736e2aa610f3a8691
|
@@ -48,7 +48,7 @@ class RFID(object):
self.spi.max_speed_hz = speed
GPIO.setmode(pin_mode)
- if pin_rst is None:
+ if pin_rst is not None:
GPIO.setup(pin_rst, GPIO.OUT)
GPIO.output(pin_rst, 1)
GPIO.setup(pin_irq, GPIO.IN, pull_up_down=GPIO.PUD_UP)
|
pirc522/rfid.py
|
ReplaceText(target=' is not ' @(51,18)->(51,22))
|
class RFID(object):
self.spi.max_speed_hz = speed
GPIO.setmode(pin_mode)
if pin_rst is None:
GPIO.setup(pin_rst, GPIO.OUT)
GPIO.output(pin_rst, 1)
GPIO.setup(pin_irq, GPIO.IN, pull_up_down=GPIO.PUD_UP)
|
class RFID(object):
self.spi.max_speed_hz = speed
GPIO.setmode(pin_mode)
if pin_rst is not None:
GPIO.setup(pin_rst, GPIO.OUT)
GPIO.output(pin_rst, 1)
GPIO.setup(pin_irq, GPIO.IN, pull_up_down=GPIO.PUD_UP)
|
2,359 |
https://:@github.com/ac-tuwien/pymhlib.git
|
07e69d451d8e2f665c23f31c780cfa58f583cf4f
|
@@ -63,7 +63,7 @@ def run_optimization(problem_name: str, instance_class, solution_class, default_
:param iter_cb: optional callback function that is called each iteration by some of the algorithms
:param seed: optional seed value for the random number generators; 0: random initialization
"""
- if embedded:
+ if not embedded:
add_general_arguments_and_parse_settings(default_inst_file, seed)
init_logger()
|
pymhlib/demos/common.py
|
ReplaceText(target='not ' @(66,7)->(66,7))
|
def run_optimization(problem_name: str, instance_class, solution_class, default_
:param iter_cb: optional callback function that is called each iteration by some of the algorithms
:param seed: optional seed value for the random number generators; 0: random initialization
"""
if embedded:
add_general_arguments_and_parse_settings(default_inst_file, seed)
init_logger()
|
def run_optimization(problem_name: str, instance_class, solution_class, default_
:param iter_cb: optional callback function that is called each iteration by some of the algorithms
:param seed: optional seed value for the random number generators; 0: random initialization
"""
if not embedded:
add_general_arguments_and_parse_settings(default_inst_file, seed)
init_logger()
|
2,360 |
https://:@github.com/trevorparker/vane.git
|
8b9eb3a87c10bce414017216fbd7ef333e124597
|
@@ -140,7 +140,7 @@ def _fetch_weather_json(
if (with_forecast):
forecast_url = forecast_urls[provider]
r = requests.get(
- forecast_url.format(location, units, api_key))
+ forecast_url.format(loc_parsed, units, api_key))
f = json.loads(r.text)
if (c['response']['features']['forecast'] != 1):
return {'e': 'Unable to load forecast'}
|
vane/utils.py
|
ReplaceText(target='loc_parsed' @(143,44)->(143,52))
|
def _fetch_weather_json(
if (with_forecast):
forecast_url = forecast_urls[provider]
r = requests.get(
forecast_url.format(location, units, api_key))
f = json.loads(r.text)
if (c['response']['features']['forecast'] != 1):
return {'e': 'Unable to load forecast'}
|
def _fetch_weather_json(
if (with_forecast):
forecast_url = forecast_urls[provider]
r = requests.get(
forecast_url.format(loc_parsed, units, api_key))
f = json.loads(r.text)
if (c['response']['features']['forecast'] != 1):
return {'e': 'Unable to load forecast'}
|
2,361 |
https://:@github.com/juancgvazquez/MODApy.git
|
7c64bc5452715160f8767c891f04d7a0a4848ebc
|
@@ -227,7 +227,7 @@ class Pipeline(object):
logger2.info(step.name)
args = step.args.replace(
- 'patientname', tmpdir + patientname).replace('reference', ref).replace('samplename', samplename)
+ 'patientname', tmpdir + patientname).replace('reference', ref).replace('samplename', patientname)
cmdver = step.version.replace('.', '_')
javacmds = ['GATK', 'picard', 'SnpSift', 'snpEff']
if any(javacmd in step.command for javacmd in javacmds):
|
MODApy/pipeline.py
|
ReplaceText(target='patientname' @(230,101)->(230,111))
|
class Pipeline(object):
logger2.info(step.name)
args = step.args.replace(
'patientname', tmpdir + patientname).replace('reference', ref).replace('samplename', samplename)
cmdver = step.version.replace('.', '_')
javacmds = ['GATK', 'picard', 'SnpSift', 'snpEff']
if any(javacmd in step.command for javacmd in javacmds):
|
class Pipeline(object):
logger2.info(step.name)
args = step.args.replace(
'patientname', tmpdir + patientname).replace('reference', ref).replace('samplename', patientname)
cmdver = step.version.replace('.', '_')
javacmds = ['GATK', 'picard', 'SnpSift', 'snpEff']
if any(javacmd in step.command for javacmd in javacmds):
|
2,362 |
https://:@github.com/linhd-postdata/averell.git
|
538e7f13b3b57170d94241111b416c31deb75d5c
|
@@ -99,7 +99,7 @@ def download_corpora(corpus_indices=None,
else:
url = CORPORA_SOURCES[index]["properties"]["url"]
filename = download_corpus(url, f"{folder_name}.zip")
- folder_list.append(uncompress_corpus(filename, output_folder))
+ folder_list.append(uncompress_corpus(filename, folder_path))
else:
logging.error("No corpus selected. Nothing will be downloaded")
return folder_list
|
src/averell/utils.py
|
ReplaceText(target='folder_path' @(102,63)->(102,76))
|
def download_corpora(corpus_indices=None,
else:
url = CORPORA_SOURCES[index]["properties"]["url"]
filename = download_corpus(url, f"{folder_name}.zip")
folder_list.append(uncompress_corpus(filename, output_folder))
else:
logging.error("No corpus selected. Nothing will be downloaded")
return folder_list
|
def download_corpora(corpus_indices=None,
else:
url = CORPORA_SOURCES[index]["properties"]["url"]
filename = download_corpus(url, f"{folder_name}.zip")
folder_list.append(uncompress_corpus(filename, folder_path))
else:
logging.error("No corpus selected. Nothing will be downloaded")
return folder_list
|
2,363 |
https://:@github.com/poqweur/ctec-utils.git
|
28470709205c35754325af5e817fded28921a389
|
@@ -122,7 +122,7 @@ class OraclePool(object):
result_db = cursor.execute(sql, param)
if commit:
conn.commit()
- result = result_db.rowcount
+ result = cursor.rowcount
else:
result = result_db.fetchall()
except Exception as e:
|
ctec_utils/Database.py
|
ReplaceText(target='cursor' @(125,25)->(125,34))
|
class OraclePool(object):
result_db = cursor.execute(sql, param)
if commit:
conn.commit()
result = result_db.rowcount
else:
result = result_db.fetchall()
except Exception as e:
|
class OraclePool(object):
result_db = cursor.execute(sql, param)
if commit:
conn.commit()
result = cursor.rowcount
else:
result = result_db.fetchall()
except Exception as e:
|
2,364 |
https://:@github.com/fsepy/sfeprapy.git
|
ffdf8d512d6ca26a685f58fceac853d9ac9241b6
|
@@ -137,7 +137,7 @@ def dict_flatten(dict_in: dict):
else:
dict_out[k] = dict_in[k]
- return dict_in
+ return dict_out
def main(x: dict, num_samples: int):
|
sfeprapy/func/mcs_gen.py
|
ReplaceText(target='dict_out' @(140,11)->(140,18))
|
def dict_flatten(dict_in: dict):
else:
dict_out[k] = dict_in[k]
return dict_in
def main(x: dict, num_samples: int):
|
def dict_flatten(dict_in: dict):
else:
dict_out[k] = dict_in[k]
return dict_out
def main(x: dict, num_samples: int):
|
2,365 |
https://:@github.com/galias11/nlp_model_gen.git
|
7854eb5aac80b1f6d24bbdd7d319bb9ebb4e429a
|
@@ -173,7 +173,7 @@ class ModelManagerController:
try:
Logger.log('L-0021')
custom_model = self.__initialize_custom_model()
- new_model = Model(model_id, model_name, description, author, model_name, analyzer_rule_set)
+ new_model = Model(model_id, model_name, description, author, model_id, analyzer_rule_set)
new_model.set_reference(custom_model)
Logger.log('L-0022')
self.__apply_tokenizer_exceptions(new_model, tokenizer_exceptions_path)
|
nlp_model_gen/packages/modelManager/ModelManagerController.py
|
ReplaceText(target='model_id' @(176,73)->(176,83))
|
class ModelManagerController:
try:
Logger.log('L-0021')
custom_model = self.__initialize_custom_model()
new_model = Model(model_id, model_name, description, author, model_name, analyzer_rule_set)
new_model.set_reference(custom_model)
Logger.log('L-0022')
self.__apply_tokenizer_exceptions(new_model, tokenizer_exceptions_path)
|
class ModelManagerController:
try:
Logger.log('L-0021')
custom_model = self.__initialize_custom_model()
new_model = Model(model_id, model_name, description, author, model_id, analyzer_rule_set)
new_model.set_reference(custom_model)
Logger.log('L-0022')
self.__apply_tokenizer_exceptions(new_model, tokenizer_exceptions_path)
|
2,366 |
https://:@github.com/galias11/nlp_model_gen.git
|
f03a84b2eaa77db12a4d7698bb982a2be062566b
|
@@ -104,7 +104,7 @@ class Model:
token_analyzer = Analyzer(self.__analyzer_rules_set)
for sent in doc.sents:
for token in sent:
- generated_token = Token(token.lemma_, token.is_oov, token.pos_, token.sent, token.sentiment, token.tag_, sent.text)
+ generated_token = Token(token.lemma_, token.is_oov, token.pos_, token.sent, token.sentiment, token.tag_, token.text)
token_analyzer.analyze_token(generated_token)
if not only_positives or generated_token.is_positive():
results.append(generated_token)
|
nlp_model_gen/packages/modelManager/model/Model.py
|
ReplaceText(target='token' @(107,121)->(107,125))
|
class Model:
token_analyzer = Analyzer(self.__analyzer_rules_set)
for sent in doc.sents:
for token in sent:
generated_token = Token(token.lemma_, token.is_oov, token.pos_, token.sent, token.sentiment, token.tag_, sent.text)
token_analyzer.analyze_token(generated_token)
if not only_positives or generated_token.is_positive():
results.append(generated_token)
|
class Model:
token_analyzer = Analyzer(self.__analyzer_rules_set)
for sent in doc.sents:
for token in sent:
generated_token = Token(token.lemma_, token.is_oov, token.pos_, token.sent, token.sentiment, token.tag_, token.text)
token_analyzer.analyze_token(generated_token)
if not only_positives or generated_token.is_positive():
results.append(generated_token)
|
2,367 |
https://:@github.com/axelfahy/bff.git
|
0023bab225d0c2571fb47e12a1edf19d61396a5d
|
@@ -602,7 +602,7 @@ def plot_series(df: pd.DataFrame, column: str, groupby: str = '1S',
.mean()
.resample(groupby)
.apply(sem)
- if groupby == 'S' and groupby != '1S' else
+ if groupby != 'S' and groupby != '1S' else
df[column].groupby('datetime').apply(sem))
ax.fill_between(x, df_plot - df_sem, df_plot + df_sem,
|
bff/fancy.py
|
ReplaceText(target='!=' @(605,33)->(605,35))
|
def plot_series(df: pd.DataFrame, column: str, groupby: str = '1S',
.mean()
.resample(groupby)
.apply(sem)
if groupby == 'S' and groupby != '1S' else
df[column].groupby('datetime').apply(sem))
ax.fill_between(x, df_plot - df_sem, df_plot + df_sem,
|
def plot_series(df: pd.DataFrame, column: str, groupby: str = '1S',
.mean()
.resample(groupby)
.apply(sem)
if groupby != 'S' and groupby != '1S' else
df[column].groupby('datetime').apply(sem))
ax.fill_between(x, df_plot - df_sem, df_plot + df_sem,
|
2,368 |
https://:@github.com/GearPlug/mercadolibre-python.git
|
3b0fa8eb47a81093e884a5699c2d25a1c700b1d0
|
@@ -401,7 +401,7 @@ class Client(object):
_params = {'access_token': self.access_token}
if params:
_params.update(params)
- response = requests.request(method, self.BASE_URL + endpoint, params=params, **kwargs)
+ response = requests.request(method, self.BASE_URL + endpoint, params=_params, **kwargs)
return self._parse(response)
def _parse(self, response):
|
mercadolibre/client.py
|
ReplaceText(target='_params' @(404,77)->(404,83))
|
class Client(object):
_params = {'access_token': self.access_token}
if params:
_params.update(params)
response = requests.request(method, self.BASE_URL + endpoint, params=params, **kwargs)
return self._parse(response)
def _parse(self, response):
|
class Client(object):
_params = {'access_token': self.access_token}
if params:
_params.update(params)
response = requests.request(method, self.BASE_URL + endpoint, params=_params, **kwargs)
return self._parse(response)
def _parse(self, response):
|
2,369 |
https://:@bitbucket.org/jairhul/pytransport.git
|
1b8a3ebc993ea9e74429198fa3607e1bfe537601
|
@@ -270,7 +270,7 @@ def RemoveIllegals(line):
"""
illegal = ['"', '', '(', ')']
- linelist = [element for element in line if element in illegal]
+ linelist = [element for element in line if element not in illegal]
line = _np.array(linelist)
return line
|
pytransport/_General.py
|
ReplaceText(target=' not in ' @(273,54)->(273,58))
|
def RemoveIllegals(line):
"""
illegal = ['"', '', '(', ')']
linelist = [element for element in line if element in illegal]
line = _np.array(linelist)
return line
|
def RemoveIllegals(line):
"""
illegal = ['"', '', '(', ')']
linelist = [element for element in line if element not in illegal]
line = _np.array(linelist)
return line
|
2,370 |
https://:@github.com/lsst-sqre/jupyterlabdemo.git
|
f6240ef0aeebdfe6e62c3f1bcf79a0fb085febd6
|
@@ -480,7 +480,7 @@ class LSSTSpawner(namespacedkubespawner.NamespacedKubeSpawner):
for vol in vollist:
volname = self._get_volume_name_for_mountpoint(vol["mountpoint"])
shortname = vol["mountpoint"][1:].replace("/", "-")
- if volname in already_vols:
+ if shortname in already_vols:
self.log.info(
"Volume '{}' already exists for pod.".format(volname))
continue
|
jupyterhub/sample_configs/20-spawner.py
|
ReplaceText(target='shortname' @(483,15)->(483,22))
|
class LSSTSpawner(namespacedkubespawner.NamespacedKubeSpawner):
for vol in vollist:
volname = self._get_volume_name_for_mountpoint(vol["mountpoint"])
shortname = vol["mountpoint"][1:].replace("/", "-")
if volname in already_vols:
self.log.info(
"Volume '{}' already exists for pod.".format(volname))
continue
|
class LSSTSpawner(namespacedkubespawner.NamespacedKubeSpawner):
for vol in vollist:
volname = self._get_volume_name_for_mountpoint(vol["mountpoint"])
shortname = vol["mountpoint"][1:].replace("/", "-")
if shortname in already_vols:
self.log.info(
"Volume '{}' already exists for pod.".format(volname))
continue
|
2,371 |
https://:@github.com/pyfarm/pyfarm-core.git
|
66dacd9725338a5f49d12ccee0e0ec8f9e5f8068
|
@@ -68,7 +68,7 @@ class Task(TaskModel):
def __init__(self, job, frame, parent_task=None, state=None,
priority=None, attempts=None, agent=None):
# build parent job id
- if not modelfor(job, TABLE_JOB):
+ if modelfor(job, TABLE_JOB):
jobid = job.jobid
if jobid is None:
raise ValueError("`job` with null id provided")
|
models/task.py
|
ReplaceText(target='' @(71,11)->(71,15))
|
class Task(TaskModel):
def __init__(self, job, frame, parent_task=None, state=None,
priority=None, attempts=None, agent=None):
# build parent job id
if not modelfor(job, TABLE_JOB):
jobid = job.jobid
if jobid is None:
raise ValueError("`job` with null id provided")
|
class Task(TaskModel):
def __init__(self, job, frame, parent_task=None, state=None,
priority=None, attempts=None, agent=None):
# build parent job id
if modelfor(job, TABLE_JOB):
jobid = job.jobid
if jobid is None:
raise ValueError("`job` with null id provided")
|
2,372 |
https://:@github.com/pyfarm/pyfarm-core.git
|
47a4cc9232a09974dea7f246b96d0338a4a4339b
|
@@ -116,5 +116,5 @@ class Task(TaskModel):
if priority is not None:
self.priority = priority
- if attempts is None:
+ if attempts is not None:
self.attempts = attempts
|
models/task.py
|
ReplaceText(target=' is not ' @(119,19)->(119,23))
|
class Task(TaskModel):
if priority is not None:
self.priority = priority
if attempts is None:
self.attempts = attempts
|
class Task(TaskModel):
if priority is not None:
self.priority = priority
if attempts is not None:
self.attempts = attempts
|
2,373 |
https://:@github.com/amarouane-ABDLHAK/cumulus-process-py.git
|
3955f6f5628f0b5233ad19cf54303bd164f981f1
|
@@ -85,7 +85,7 @@ class Granule(object):
m = re.match(self.inputs[f], os.path.basename(filename))
if m is not None:
# does the file exist locally
- if os.path.exists(f):
+ if os.path.exists(filename):
self.local_in[f] = filename
else:
self.remote_in[f] = filename
|
cumulus/granule.py
|
ReplaceText(target='filename' @(88,34)->(88,35))
|
class Granule(object):
m = re.match(self.inputs[f], os.path.basename(filename))
if m is not None:
# does the file exist locally
if os.path.exists(f):
self.local_in[f] = filename
else:
self.remote_in[f] = filename
|
class Granule(object):
m = re.match(self.inputs[f], os.path.basename(filename))
if m is not None:
# does the file exist locally
if os.path.exists(filename):
self.local_in[f] = filename
else:
self.remote_in[f] = filename
|
2,374 |
https://:@github.com/CodeClubLux/TopCompiler.git
|
ec1cbd020e522f8e478000d7d898003972e11490
|
@@ -14,7 +14,7 @@ class Enum(Node):
args = self.const[name]
names = [codegen.getName() for _ in args]
codegen.inFunction()
- if len(args) > 0:
+ if len(names) > 0:
codegen.append("function "+self.package+"_"+name+"(")
codegen.append(",".join(names))
codegen.append("){return ["+str(count)+","+",".join(names)+"]}")
|
AST/Enum.py
|
ReplaceText(target='names' @(17,19)->(17,23))
|
class Enum(Node):
args = self.const[name]
names = [codegen.getName() for _ in args]
codegen.inFunction()
if len(args) > 0:
codegen.append("function "+self.package+"_"+name+"(")
codegen.append(",".join(names))
codegen.append("){return ["+str(count)+","+",".join(names)+"]}")
|
class Enum(Node):
args = self.const[name]
names = [codegen.getName() for _ in args]
codegen.inFunction()
if len(names) > 0:
codegen.append("function "+self.package+"_"+name+"(")
codegen.append(",".join(names))
codegen.append("){return ["+str(count)+","+",".join(names)+"]}")
|
2,375 |
https://:@github.com/GIScience/openpoiservice.git
|
5a686db7a201b52f836e824910d9218bd2ff790b
|
@@ -171,7 +171,7 @@ class QueryBuilder(object):
if tag in filters:
- filters.append(query.c.key == tag.lower())
+ filters_list.append(query.c.key == tag.lower())
if settings['filterable'] == 'like':
filters_list.append(query.c.value.like('%' + filters[tag].lower() + '%'))
|
openpoiservice/server/api/query_builder.py
|
ReplaceText(target='filters_list' @(174,16)->(174,23))
|
class QueryBuilder(object):
if tag in filters:
filters.append(query.c.key == tag.lower())
if settings['filterable'] == 'like':
filters_list.append(query.c.value.like('%' + filters[tag].lower() + '%'))
|
class QueryBuilder(object):
if tag in filters:
filters_list.append(query.c.key == tag.lower())
if settings['filterable'] == 'like':
filters_list.append(query.c.value.like('%' + filters[tag].lower() + '%'))
|
2,376 |
https://:@github.com/theblackcat102/jieba-tw.git
|
5270ed66ff64b2001c1bf5c4ba927fec09189e33
|
@@ -366,7 +366,7 @@ class Tokenizer(object):
f = open(f, 'rb')
for lineno, ln in enumerate(f, 1):
line = ln.strip()
- if not isinstance(f, text_type):
+ if not isinstance(line, text_type):
try:
line = line.decode('utf-8').lstrip('\ufeff')
except UnicodeDecodeError:
|
jieba/__init__.py
|
ReplaceText(target='line' @(369,30)->(369,31))
|
class Tokenizer(object):
f = open(f, 'rb')
for lineno, ln in enumerate(f, 1):
line = ln.strip()
if not isinstance(f, text_type):
try:
line = line.decode('utf-8').lstrip('\ufeff')
except UnicodeDecodeError:
|
class Tokenizer(object):
f = open(f, 'rb')
for lineno, ln in enumerate(f, 1):
line = ln.strip()
if not isinstance(line, text_type):
try:
line = line.decode('utf-8').lstrip('\ufeff')
except UnicodeDecodeError:
|
2,377 |
https://:@github.com/mozilla/measure-noise.git
|
b100399b2d650a794f50c897dfb2ec3462ad814f
|
@@ -103,7 +103,7 @@ def process(
# EG https://treeherder.mozilla.org/perf.html#/graphs?highlightAlerts=1&series=mozilla-central,fee739b45f7960e4a520d8e0bd781dd9d0a3bec4,1,10&timerange=31536000
url = "https://treeherder.mozilla.org/perf.html#/graphs?" + value2url_param({
"highlightAlerts": 1,
- "series": [sig.repository, sig.id, 1, coalesce(sig.framework, sig.framework_id)],
+ "series": [sig.repository, sig.id, 1, coalesce(sig.framework_id, sig.framework)],
"timerange": 31536000,
})
|
measure_noise/analysis.py
|
ArgSwap(idxs=0<->1 @(106,46)->(106,54))
|
def process(
# EG https://treeherder.mozilla.org/perf.html#/graphs?highlightAlerts=1&series=mozilla-central,fee739b45f7960e4a520d8e0bd781dd9d0a3bec4,1,10&timerange=31536000
url = "https://treeherder.mozilla.org/perf.html#/graphs?" + value2url_param({
"highlightAlerts": 1,
"series": [sig.repository, sig.id, 1, coalesce(sig.framework, sig.framework_id)],
"timerange": 31536000,
})
|
def process(
# EG https://treeherder.mozilla.org/perf.html#/graphs?highlightAlerts=1&series=mozilla-central,fee739b45f7960e4a520d8e0bd781dd9d0a3bec4,1,10&timerange=31536000
url = "https://treeherder.mozilla.org/perf.html#/graphs?" + value2url_param({
"highlightAlerts": 1,
"series": [sig.repository, sig.id, 1, coalesce(sig.framework_id, sig.framework)],
"timerange": 31536000,
})
|
2,378 |
https://:@github.com/USGS-WiM/WIMLib.git
|
f9f74b29ed1dfc901b31e3df81f9f2459918dc4e
|
@@ -93,7 +93,7 @@ class MapLayer(object):
raise Exception(datasetPath +" doesn't exist")
#test for schema lock, before continue
trys=0
- while arcpy.TestSchemaLock(datasetPath) or trys>6:
+ while arcpy.TestSchemaLock(datasetPath) or trys<6:
time.sleep(10)
trys+=1
#next
|
WIMLib/MapLayer.py
|
ReplaceText(target='<' @(96,59)->(96,60))
|
class MapLayer(object):
raise Exception(datasetPath +" doesn't exist")
#test for schema lock, before continue
trys=0
while arcpy.TestSchemaLock(datasetPath) or trys>6:
time.sleep(10)
trys+=1
#next
|
class MapLayer(object):
raise Exception(datasetPath +" doesn't exist")
#test for schema lock, before continue
trys=0
while arcpy.TestSchemaLock(datasetPath) or trys<6:
time.sleep(10)
trys+=1
#next
|
2,379 |
https://:@github.com/xiawu/newchain-web3.py.git
|
b253f8a8d55a087800e8e5b0947e7972a1f8258d
|
@@ -31,7 +31,7 @@ def pad_right(string, chars, filler="0"):
def is_prefixed(value, prefix):
return value.startswith(
- force_bytes(prefix) if is_bytes(prefix) else force_text(prefix)
+ force_bytes(prefix) if is_bytes(value) else force_text(prefix)
)
|
web3/utils/formatting.py
|
ReplaceText(target='value' @(34,40)->(34,46))
|
def pad_right(string, chars, filler="0"):
def is_prefixed(value, prefix):
return value.startswith(
force_bytes(prefix) if is_bytes(prefix) else force_text(prefix)
)
|
def pad_right(string, chars, filler="0"):
def is_prefixed(value, prefix):
return value.startswith(
force_bytes(prefix) if is_bytes(value) else force_text(prefix)
)
|
2,380 |
https://:@github.com/xiawu/newchain-web3.py.git
|
6c2e459fbb1c3e9cf665b8138744510f2f797149
|
@@ -143,7 +143,7 @@ def outputBlockFormatter(block):
if is_array(block.get("transactions")):
for item in block["transactions"]:
- if is_string(item):
+ if not is_string(item):
item = outputTransactionFormatter(item)
return block
|
web3/formatters.py
|
ReplaceText(target='not ' @(146,15)->(146,15))
|
def outputBlockFormatter(block):
if is_array(block.get("transactions")):
for item in block["transactions"]:
if is_string(item):
item = outputTransactionFormatter(item)
return block
|
def outputBlockFormatter(block):
if is_array(block.get("transactions")):
for item in block["transactions"]:
if not is_string(item):
item = outputTransactionFormatter(item)
return block
|
2,381 |
https://:@github.com/xiawu/newchain-web3.py.git
|
9f8282b202f17f1e98a305b49657b7ff2387b85a
|
@@ -611,7 +611,7 @@ def call_contract_function(contract=None,
if transaction is None:
call_transaction = {}
else:
- call_transaction = dict(**call_transaction)
+ call_transaction = dict(**transaction)
if not arguments:
arguments = []
|
web3/contract.py
|
ReplaceText(target='transaction' @(614,34)->(614,50))
|
def call_contract_function(contract=None,
if transaction is None:
call_transaction = {}
else:
call_transaction = dict(**call_transaction)
if not arguments:
arguments = []
|
def call_contract_function(contract=None,
if transaction is None:
call_transaction = {}
else:
call_transaction = dict(**transaction)
if not arguments:
arguments = []
|
2,382 |
https://:@github.com/Elizafox/taillight.git
|
731c28874dd7aa29f59a862e6b01f4ea8010978e
|
@@ -131,7 +131,7 @@ class Signal:
with self._slots_lock:
for slot in self.slots:
if slot.function is function:
- ret.append(function)
+ ret.append(slot)
if ret:
return ret
|
taillight/signal.py
|
ReplaceText(target='slot' @(134,31)->(134,39))
|
class Signal:
with self._slots_lock:
for slot in self.slots:
if slot.function is function:
ret.append(function)
if ret:
return ret
|
class Signal:
with self._slots_lock:
for slot in self.slots:
if slot.function is function:
ret.append(slot)
if ret:
return ret
|
2,383 |
https://:@github.com/Fak3/minidjango.git
|
659ab9846e81d95bb75dbb3c00147324bf0d6541
|
@@ -22,7 +22,7 @@ def login(request):
else:
errors = {}
response = HttpResponse()
- response.session.set_test_cookie()
+ request.session.set_test_cookie()
t = template_loader.get_template('registration/login')
c = Context(request, {
'form': formfields.FormWrapper(manipulator, request.POST, errors),
|
django/views/auth/login.py
|
ReplaceText(target='request' @(25,4)->(25,12))
|
def login(request):
else:
errors = {}
response = HttpResponse()
response.session.set_test_cookie()
t = template_loader.get_template('registration/login')
c = Context(request, {
'form': formfields.FormWrapper(manipulator, request.POST, errors),
|
def login(request):
else:
errors = {}
response = HttpResponse()
request.session.set_test_cookie()
t = template_loader.get_template('registration/login')
c = Context(request, {
'form': formfields.FormWrapper(manipulator, request.POST, errors),
|
2,384 |
https://:@github.com/Fak3/minidjango.git
|
34655a3e7816d6a8e5da6b3fd613b49b454a4691
|
@@ -227,7 +227,7 @@ class DateFormat(TimeFormat):
week_number = 1
else:
j = day_of_year + (7 - weekday) + (jan1_weekday - 1)
- week_number = j / 7
+ week_number = j // 7
if jan1_weekday > 4:
week_number -= 1
return week_number
|
django/utils/dateformat.py
|
ReplaceText(target='//' @(230,32)->(230,33))
|
class DateFormat(TimeFormat):
week_number = 1
else:
j = day_of_year + (7 - weekday) + (jan1_weekday - 1)
week_number = j / 7
if jan1_weekday > 4:
week_number -= 1
return week_number
|
class DateFormat(TimeFormat):
week_number = 1
else:
j = day_of_year + (7 - weekday) + (jan1_weekday - 1)
week_number = j // 7
if jan1_weekday > 4:
week_number -= 1
return week_number
|
2,385 |
https://:@github.com/Fak3/minidjango.git
|
a97648a7e03fb95b09e888e5d59d82d57fb289b7
|
@@ -105,7 +105,7 @@ class DecoratorsTest(TestCase):
"""
def my_view(request):
return "response"
- my_view_cached = cache_page(123, my_view)
+ my_view_cached = cache_page(my_view, 123)
self.assertEqual(my_view_cached(HttpRequest()), "response")
class MethodDecoratorAdapterTests(TestCase):
|
tests/regressiontests/decorators/tests.py
|
ArgSwap(idxs=0<->1 @(108,25)->(108,35))
|
class DecoratorsTest(TestCase):
"""
def my_view(request):
return "response"
my_view_cached = cache_page(123, my_view)
self.assertEqual(my_view_cached(HttpRequest()), "response")
class MethodDecoratorAdapterTests(TestCase):
|
class DecoratorsTest(TestCase):
"""
def my_view(request):
return "response"
my_view_cached = cache_page(my_view, 123)
self.assertEqual(my_view_cached(HttpRequest()), "response")
class MethodDecoratorAdapterTests(TestCase):
|
2,386 |
https://:@github.com/Fak3/minidjango.git
|
b2050ff546da4164f90a795e55d7d8c55981783d
|
@@ -169,7 +169,7 @@ class SQLCompiler(object):
if isinstance(col, (list, tuple)):
alias, column = col
table = self.query.alias_map[alias][TABLE_NAME]
- if table in only_load and col not in only_load[table]:
+ if table in only_load and column not in only_load[table]:
continue
r = '%s.%s' % (qn(alias), qn(column))
if with_aliases:
|
django/db/models/sql/compiler.py
|
ReplaceText(target='column' @(172,46)->(172,49))
|
class SQLCompiler(object):
if isinstance(col, (list, tuple)):
alias, column = col
table = self.query.alias_map[alias][TABLE_NAME]
if table in only_load and col not in only_load[table]:
continue
r = '%s.%s' % (qn(alias), qn(column))
if with_aliases:
|
class SQLCompiler(object):
if isinstance(col, (list, tuple)):
alias, column = col
table = self.query.alias_map[alias][TABLE_NAME]
if table in only_load and column not in only_load[table]:
continue
r = '%s.%s' % (qn(alias), qn(column))
if with_aliases:
|
2,387 |
https://:@github.com/Fak3/minidjango.git
|
cfba2460370a6d1808b78e2ba0709ea5c8b7e773
|
@@ -42,7 +42,7 @@ def check_settings(base_url=None):
Checks if the staticfiles settings have sane values.
"""
- if base_url is not None:
+ if base_url is None:
base_url = settings.STATIC_URL
if not base_url:
raise ImproperlyConfigured(
|
django/contrib/staticfiles/utils.py
|
ReplaceText(target=' is ' @(45,15)->(45,23))
|
def check_settings(base_url=None):
Checks if the staticfiles settings have sane values.
"""
if base_url is not None:
base_url = settings.STATIC_URL
if not base_url:
raise ImproperlyConfigured(
|
def check_settings(base_url=None):
Checks if the staticfiles settings have sane values.
"""
if base_url is None:
base_url = settings.STATIC_URL
if not base_url:
raise ImproperlyConfigured(
|
2,388 |
https://:@github.com/Fak3/minidjango.git
|
d72d5ce8274992ce01e39f866a7a250bc459eefe
|
@@ -37,7 +37,7 @@ class GeoSQLCompiler(compiler.SQLCompiler):
if isinstance(col, (list, tuple)):
alias, column = col
table = self.query.alias_map[alias][TABLE_NAME]
- if table in only_load and col not in only_load[table]:
+ if table in only_load and column not in only_load[table]:
continue
r = self.get_field_select(field, alias, column)
if with_aliases:
|
django/contrib/gis/db/models/sql/compiler.py
|
ReplaceText(target='column' @(40,46)->(40,49))
|
class GeoSQLCompiler(compiler.SQLCompiler):
if isinstance(col, (list, tuple)):
alias, column = col
table = self.query.alias_map[alias][TABLE_NAME]
if table in only_load and col not in only_load[table]:
continue
r = self.get_field_select(field, alias, column)
if with_aliases:
|
class GeoSQLCompiler(compiler.SQLCompiler):
if isinstance(col, (list, tuple)):
alias, column = col
table = self.query.alias_map[alias][TABLE_NAME]
if table in only_load and column not in only_load[table]:
continue
r = self.get_field_select(field, alias, column)
if with_aliases:
|
2,389 |
https://:@github.com/Fak3/minidjango.git
|
6ecbac21a9017a53fe18ac81c9c1d2f28185a292
|
@@ -111,5 +111,5 @@ class OSMWidget(BaseGeometryWidget):
return 900913
def render(self, name, value, attrs=None):
- return super(self, OSMWidget).render(name, value,
+ return super(OSMWidget, self).render(name, value,
{'default_lon': self.default_lon, 'default_lat': self.default_lat})
|
django/contrib/gis/forms/widgets.py
|
ArgSwap(idxs=0<->1 @(114,15)->(114,20))
|
class OSMWidget(BaseGeometryWidget):
return 900913
def render(self, name, value, attrs=None):
return super(self, OSMWidget).render(name, value,
{'default_lon': self.default_lon, 'default_lat': self.default_lat})
|
class OSMWidget(BaseGeometryWidget):
return 900913
def render(self, name, value, attrs=None):
return super(OSMWidget, self).render(name, value,
{'default_lon': self.default_lon, 'default_lat': self.default_lat})
|
2,390 |
https://:@github.com/Fak3/minidjango.git
|
86c248aa646183ef4a1cb407bb3e4cb597272f63
|
@@ -575,7 +575,7 @@ class SQLCompiler(object):
for order, order_params in ordering_group_by:
# Even if we have seen the same SQL string, it might have
# different params, so, we add same SQL in "has params" case.
- if order not in seen or params:
+ if order not in seen or order_params:
result.append(order)
params.extend(order_params)
seen.add(order)
|
django/db/models/sql/compiler.py
|
ReplaceText(target='order_params' @(578,44)->(578,50))
|
class SQLCompiler(object):
for order, order_params in ordering_group_by:
# Even if we have seen the same SQL string, it might have
# different params, so, we add same SQL in "has params" case.
if order not in seen or params:
result.append(order)
params.extend(order_params)
seen.add(order)
|
class SQLCompiler(object):
for order, order_params in ordering_group_by:
# Even if we have seen the same SQL string, it might have
# different params, so, we add same SQL in "has params" case.
if order not in seen or order_params:
result.append(order)
params.extend(order_params)
seen.add(order)
|
2,391 |
https://:@github.com/Fak3/minidjango.git
|
fddb0131d37109c809ec391e1a134ef1d9e442a7
|
@@ -57,7 +57,7 @@ def check_password(password, encoded, setter=None, preferred='default'):
must_update = hasher.algorithm != preferred.algorithm
if not must_update:
- must_update = hasher.must_update(encoded)
+ must_update = preferred.must_update(encoded)
is_correct = hasher.verify(password, encoded)
if setter and is_correct and must_update:
setter(password)
|
django/contrib/auth/hashers.py
|
ReplaceText(target='preferred' @(60,22)->(60,28))
|
def check_password(password, encoded, setter=None, preferred='default'):
must_update = hasher.algorithm != preferred.algorithm
if not must_update:
must_update = hasher.must_update(encoded)
is_correct = hasher.verify(password, encoded)
if setter and is_correct and must_update:
setter(password)
|
def check_password(password, encoded, setter=None, preferred='default'):
must_update = hasher.algorithm != preferred.algorithm
if not must_update:
must_update = preferred.must_update(encoded)
is_correct = hasher.verify(password, encoded)
if setter and is_correct and must_update:
setter(password)
|
2,392 |
https://:@github.com/Fak3/minidjango.git
|
e8223b889aab3b5ac0c2312eb9ee2307ea635c97
|
@@ -228,7 +228,7 @@ class GenericRelationTests(TestCase):
# then wrong results are produced here as the link to b will also match
# (b and hs1 have equal pks).
self.assertEqual(qs.count(), 1)
- self.assertEqual(qs[0].links__sum, l.id)
+ self.assertEqual(qs[0].links__sum, hs1.id)
l.delete()
# Now if we don't have proper left join, we will not produce any
# results at all here.
|
tests/generic_relations_regress/tests.py
|
ReplaceText(target='hs1' @(231,43)->(231,44))
|
class GenericRelationTests(TestCase):
# then wrong results are produced here as the link to b will also match
# (b and hs1 have equal pks).
self.assertEqual(qs.count(), 1)
self.assertEqual(qs[0].links__sum, l.id)
l.delete()
# Now if we don't have proper left join, we will not produce any
# results at all here.
|
class GenericRelationTests(TestCase):
# then wrong results are produced here as the link to b will also match
# (b and hs1 have equal pks).
self.assertEqual(qs.count(), 1)
self.assertEqual(qs[0].links__sum, hs1.id)
l.delete()
# Now if we don't have proper left join, we will not produce any
# results at all here.
|
2,393 |
https://:@github.com/Fak3/minidjango.git
|
3074c5b19e2da5f7a5359c3cf3c5308eb194cdf9
|
@@ -112,7 +112,7 @@ class ClassDecoratedTestCase(ClassDecoratedTestCaseSuper):
@classmethod
def setUpClass(cls):
- super(cls, ClassDecoratedTestCase).setUpClass()
+ super(ClassDecoratedTestCase, cls).setUpClass()
cls.foo = getattr(settings, 'TEST', 'BUG')
def test_override(self):
|
tests/settings_tests/tests.py
|
ArgSwap(idxs=0<->1 @(115,8)->(115,13))
|
class ClassDecoratedTestCase(ClassDecoratedTestCaseSuper):
@classmethod
def setUpClass(cls):
super(cls, ClassDecoratedTestCase).setUpClass()
cls.foo = getattr(settings, 'TEST', 'BUG')
def test_override(self):
|
class ClassDecoratedTestCase(ClassDecoratedTestCaseSuper):
@classmethod
def setUpClass(cls):
super(ClassDecoratedTestCase, cls).setUpClass()
cls.foo = getattr(settings, 'TEST', 'BUG')
def test_override(self):
|
2,394 |
https://:@github.com/Fak3/minidjango.git
|
c2b4967e76fd671e6199e4dd54d2a2c1f096b8eb
|
@@ -23,7 +23,7 @@ def import_string(dotted_path):
return getattr(module, class_name)
except AttributeError:
msg = 'Module "%s" does not define a "%s" attribute/class' % (
- dotted_path, class_name)
+ module_path, class_name)
six.reraise(ImportError, ImportError(msg), sys.exc_info()[2])
|
django/utils/module_loading.py
|
ReplaceText(target='module_path' @(26,12)->(26,23))
|
def import_string(dotted_path):
return getattr(module, class_name)
except AttributeError:
msg = 'Module "%s" does not define a "%s" attribute/class' % (
dotted_path, class_name)
six.reraise(ImportError, ImportError(msg), sys.exc_info()[2])
|
def import_string(dotted_path):
return getattr(module, class_name)
except AttributeError:
msg = 'Module "%s" does not define a "%s" attribute/class' % (
module_path, class_name)
six.reraise(ImportError, ImportError(msg), sys.exc_info()[2])
|
2,395 |
https://:@github.com/Fak3/minidjango.git
|
abcdb237bb313d116ce2ac8e90f79f61429afc70
|
@@ -31,7 +31,7 @@ class DatabaseCreation(BaseDatabaseCreation):
try:
if verbosity >= 1:
print("Destroying old test database for alias %s..." % (
- self._get_database_display_str(target_database_name, verbosity),
+ self._get_database_display_str(verbosity, target_database_name),
))
cursor.execute("DROP DATABASE %s" % qn(target_database_name))
cursor.execute("CREATE DATABASE %s" % qn(target_database_name))
|
django/db/backends/mysql/creation.py
|
ArgSwap(idxs=0<->1 @(34,28)->(34,58))
|
class DatabaseCreation(BaseDatabaseCreation):
try:
if verbosity >= 1:
print("Destroying old test database for alias %s..." % (
self._get_database_display_str(target_database_name, verbosity),
))
cursor.execute("DROP DATABASE %s" % qn(target_database_name))
cursor.execute("CREATE DATABASE %s" % qn(target_database_name))
|
class DatabaseCreation(BaseDatabaseCreation):
try:
if verbosity >= 1:
print("Destroying old test database for alias %s..." % (
self._get_database_display_str(verbosity, target_database_name),
))
cursor.execute("DROP DATABASE %s" % qn(target_database_name))
cursor.execute("CREATE DATABASE %s" % qn(target_database_name))
|
2,396 |
https://:@github.com/Fak3/minidjango.git
|
542b7f6c50df18f2aa201cf1de81577c1bee643c
|
@@ -50,7 +50,7 @@ class SeparateDatabaseAndState(Operation):
to_state = base_state.clone()
for dbop in self.database_operations[:-(pos + 1)]:
dbop.state_forwards(app_label, to_state)
- from_state = base_state.clone()
+ from_state = to_state.clone()
database_operation.state_forwards(app_label, from_state)
database_operation.database_backwards(app_label, schema_editor, from_state, to_state)
|
django/db/migrations/operations/special.py
|
ReplaceText(target='to_state' @(53,25)->(53,35))
|
class SeparateDatabaseAndState(Operation):
to_state = base_state.clone()
for dbop in self.database_operations[:-(pos + 1)]:
dbop.state_forwards(app_label, to_state)
from_state = base_state.clone()
database_operation.state_forwards(app_label, from_state)
database_operation.database_backwards(app_label, schema_editor, from_state, to_state)
|
class SeparateDatabaseAndState(Operation):
to_state = base_state.clone()
for dbop in self.database_operations[:-(pos + 1)]:
dbop.state_forwards(app_label, to_state)
from_state = to_state.clone()
database_operation.state_forwards(app_label, from_state)
database_operation.database_backwards(app_label, schema_editor, from_state, to_state)
|
2,397 |
https://:@github.com/Fak3/minidjango.git
|
d5088f838d837fc9e3109c828f18511055f20bea
|
@@ -383,7 +383,7 @@ class CombinedExpression(Expression):
return DurationExpression(self.lhs, self.connector, self.rhs).as_sql(compiler, connection)
if (lhs_output and rhs_output and self.connector == self.SUB and
lhs_output.get_internal_type() in {'DateField', 'DateTimeField', 'TimeField'} and
- lhs_output.get_internal_type() == lhs_output.get_internal_type()):
+ lhs_output.get_internal_type() == rhs_output.get_internal_type()):
return TemporalSubtraction(self.lhs, self.rhs).as_sql(compiler, connection)
expressions = []
expression_params = []
|
django/db/models/expressions.py
|
ReplaceText(target='rhs_output' @(386,50)->(386,60))
|
class CombinedExpression(Expression):
return DurationExpression(self.lhs, self.connector, self.rhs).as_sql(compiler, connection)
if (lhs_output and rhs_output and self.connector == self.SUB and
lhs_output.get_internal_type() in {'DateField', 'DateTimeField', 'TimeField'} and
lhs_output.get_internal_type() == lhs_output.get_internal_type()):
return TemporalSubtraction(self.lhs, self.rhs).as_sql(compiler, connection)
expressions = []
expression_params = []
|
class CombinedExpression(Expression):
return DurationExpression(self.lhs, self.connector, self.rhs).as_sql(compiler, connection)
if (lhs_output and rhs_output and self.connector == self.SUB and
lhs_output.get_internal_type() in {'DateField', 'DateTimeField', 'TimeField'} and
lhs_output.get_internal_type() == rhs_output.get_internal_type()):
return TemporalSubtraction(self.lhs, self.rhs).as_sql(compiler, connection)
expressions = []
expression_params = []
|
2,398 |
https://:@github.com/Fak3/minidjango.git
|
67a6ba391bbcf1a4c6bb0c42cb17e4fc0530f6d2
|
@@ -42,7 +42,7 @@ class PasswordResetTokenGenerator:
return False
# Check the timestamp is within limit
- if (self._num_days(self._today()) - ts) >= settings.PASSWORD_RESET_TIMEOUT_DAYS:
+ if (self._num_days(self._today()) - ts) > settings.PASSWORD_RESET_TIMEOUT_DAYS:
return False
return True
|
django/contrib/auth/tokens.py
|
ReplaceText(target='>' @(45,48)->(45,50))
|
class PasswordResetTokenGenerator:
return False
# Check the timestamp is within limit
if (self._num_days(self._today()) - ts) >= settings.PASSWORD_RESET_TIMEOUT_DAYS:
return False
return True
|
class PasswordResetTokenGenerator:
return False
# Check the timestamp is within limit
if (self._num_days(self._today()) - ts) > settings.PASSWORD_RESET_TIMEOUT_DAYS:
return False
return True
|
2,399 |
https://:@github.com/Fak3/minidjango.git
|
acc8dd4142ec81def9a73507120c0262ba6b1264
|
@@ -60,7 +60,7 @@ class RWLock:
def writer_enters(self):
with self.mutex:
if self.active_writers == 0 and self.waiting_writers == 0 and self.active_readers == 0:
- self.active_writers += 1
+ self.active_writers = 1
self.can_write.release()
else:
self.waiting_writers += 1
|
django/utils/synch.py
|
ReplaceText(target='=' @(63,36)->(63,38))
|
class RWLock:
def writer_enters(self):
with self.mutex:
if self.active_writers == 0 and self.waiting_writers == 0 and self.active_readers == 0:
self.active_writers += 1
self.can_write.release()
else:
self.waiting_writers += 1
|
class RWLock:
def writer_enters(self):
with self.mutex:
if self.active_writers == 0 and self.waiting_writers == 0 and self.active_readers == 0:
self.active_writers = 1
self.can_write.release()
else:
self.waiting_writers += 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.