Unnamed: 0 int64 0 10k | function stringlengths 79 138k | label stringclasses 20
values | info stringlengths 42 261 |
|---|---|---|---|
6,300 | def setup_threading():
if sys.version_info >= (2, 7):
return
# XXX: On Python 2.5 GAE's threading.local doesn't work correctly
# with subclassing.
try:
from django.utils._threading_local import local
import threading
threading.local = local
except __HOLE__:
... | ImportError | dataset/ETHPy150Open django-nonrel/djangoappengine/djangoappengine/boot.py/setup_threading |
6,301 | def setup_project(dev_appserver_version):
from djangoappengine.utils import have_appserver, on_production_server
if have_appserver:
# This fixes a pwd import bug for os.path.expanduser().
env_ext['HOME'] = PROJECT_DIR
# The dev_appserver creates a sandbox which restricts access to
# cer... | ImportError | dataset/ETHPy150Open django-nonrel/djangoappengine/djangoappengine/boot.py/setup_project |
6,302 | def _get_templated_url(self, template, id, method=None):
try:
id_unicode = unicode(id, "UTF-8")
except __HOLE__:
id_unicode = id
id_utf8 = id_unicode.encode("UTF-8")
md5_of_url = hashlib.md5(id_utf8).hexdigest()
url = template % md5_of_url
return(u... | TypeError | dataset/ETHPy150Open Impactstory/total-impact-core/totalimpact/providers/delicious.py/Delicious._get_templated_url |
6,303 | def find_function(funcname, filename):
cre = re.compile(r'def\s+%s\s*[(]' % re.escape(funcname))
try:
fp = open(filename)
except __HOLE__:
return None
# consumer of this info expects the first line to be 1
lineno = 1
answer = None
while 1:
line = fp.readline()
... | IOError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/find_function |
6,304 | def __init__(self, completekey='tab', stdin=None, stdout=None, skip=None):
bdb.Bdb.__init__(self, skip=skip)
cmd.Cmd.__init__(self, completekey, stdin, stdout)
if stdout:
self.use_rawinput = 0
self.prompt = '(Pdb) '
self.aliases = {}
self.mainpyfile = ''
... | IOError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.__init__ |
6,305 | def handle_command_def(self,line):
"""Handles one command line during command list definition."""
cmd, arg, line = self.parseline(line)
if not cmd:
return
if cmd == 'silent':
self.commands_silent[self.commands_bnum] = True
return # continue to handle o... | AttributeError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.handle_command_def |
6,306 | def do_break(self, arg, temporary = 0):
# break [ ([filename:]lineno | function) [, "condition"] ]
if not arg:
if self.breaks: # There's at least one
print >>self.stdout, "Num Type Disp Enb Where"
for bp in bdb.Breakpoint.bpbynumber:
... | ValueError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.do_break |
6,307 | def do_enable(self, arg):
args = arg.split()
for i in args:
try:
i = int(i)
except __HOLE__:
print >>self.stdout, 'Breakpoint index %r is not a number' % i
continue
if not (0 <= i < len(bdb.Breakpoint.bpbynumber)):
... | ValueError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.do_enable |
6,308 | def do_disable(self, arg):
args = arg.split()
for i in args:
try:
i = int(i)
except __HOLE__:
print >>self.stdout, 'Breakpoint index %r is not a number' % i
continue
if not (0 <= i < len(bdb.Breakpoint.bpbynumber)):
... | ValueError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.do_disable |
6,309 | def do_condition(self, arg):
# arg is breakpoint number and condition
args = arg.split(' ', 1)
try:
bpnum = int(args[0].strip())
except ValueError:
# something went wrong
print >>self.stdout, \
'Breakpoint index %r is not a number' % ar... | IndexError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.do_condition |
6,310 | def do_ignore(self,arg):
"""arg is bp number followed by ignore count."""
args = arg.split()
try:
bpnum = int(args[0].strip())
except ValueError:
# something went wrong
print >>self.stdout, \
'Breakpoint index %r is not a number' % args... | IndexError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.do_ignore |
6,311 | def do_clear(self, arg):
"""Three possibilities, tried in this order:
clear -> clear all breaks, ask for confirmation
clear file:lineno -> clear all breaks at file:lineno
clear bpno bpno ... -> clear breakpoints by number"""
if not arg:
try:
reply = ra... | ValueError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.do_clear |
6,312 | def do_jump(self, arg):
if self.curindex + 1 != len(self.stack):
print >>self.stdout, "*** You can only jump within the bottom frame"
return
try:
arg = int(arg)
except ValueError:
print >>self.stdout, "*** The 'jump' command requires a line number.... | ValueError | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.do_jump |
6,313 | def do_list(self, arg):
self.lastcmd = 'list'
last = None
if arg:
try:
x = eval(arg, {}, {})
if type(x) == type(()):
first, last = x
first = int(first)
last = int(last)
if ... | KeyboardInterrupt | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.do_list |
6,314 | def print_stack_trace(self):
try:
for frame_lineno in self.stack:
self.print_stack_entry(frame_lineno)
except __HOLE__:
pass | KeyboardInterrupt | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/Pdb.print_stack_trace |
6,315 | def main():
if not sys.argv[1:] or sys.argv[1] in ("--help", "-h"):
print "usage: pdb.py scriptfile [arg] ..."
sys.exit(2)
mainpyfile = sys.argv[1] # Get script filename
if not os.path.exists(mainpyfile):
print 'Error:', mainpyfile, 'does not exist'
sys.exit(1)
del... | SystemExit | dataset/ETHPy150Open azoft-dev-team/imagrium/env/Lib/pdb.py/main |
6,316 | def _get_session(self, no_load=False):
"""
Lazily loads session from storage (unless "no_load" is True, when only
an empty dict is stored) and stores it in the current instance.
"""
self.accessed = True
try:
return self._session_cache
except __HOLE__:
... | AttributeError | dataset/ETHPy150Open AppScale/appscale/AppServer/lib/django-1.4/django/contrib/sessions/backends/base.py/SessionBase._get_session |
6,317 | def set_expiry(self, value):
"""
Sets a custom expiration for the session. ``value`` can be an integer,
a Python ``datetime`` or ``timedelta`` object or ``None``.
If ``value`` is an integer, the session will expire after that many
seconds of inactivity. If set to ``0`` then the ... | KeyError | dataset/ETHPy150Open AppScale/appscale/AppServer/lib/django-1.4/django/contrib/sessions/backends/base.py/SessionBase.set_expiry |
6,318 | def catcher():
for i in range(1000):
try:
someFunction()
except __HOLE__:
pass | ValueError | dataset/ETHPy150Open kayhayen/Nuitka/tests/benchmarks/micro/TryFinallyStopOver.py/catcher |
6,319 | def extract_rate_limit(self, response):
"""Extract rate limit info from response/headers.
get it just from the response, so it is relevant to the type of query we are doing"""
try:
self.rate_limit_remaining = int(response.headers['x-rate-limit-remaining'])
self.rate_limit... | KeyError | dataset/ETHPy150Open Impactstory/total-impact-webapp/totalimpactwebapp/twitter_paging.py/TwitterPager.extract_rate_limit |
6,320 | def paginated_search(self, page=1, page_handler=None,
max_pages=None, **kwargs):
"""Issue search with AppClient up to max_pages.
For kwargs requirements, see docs for birdy AppClient."""
if max_pages is None:
max_pages = self.default_max_pages
response = self.qu... | AttributeError | dataset/ETHPy150Open Impactstory/total-impact-webapp/totalimpactwebapp/twitter_paging.py/TwitterPager.paginated_search |
6,321 | def test_session_timeout(self):
requested_url = '/project/instances/'
request = self.factory.get(requested_url)
try:
timeout = settings.SESSION_TIMEOUT
except __HOLE__:
timeout = 1800
request.session['last_activity'] = int(time.time()) - (timeout + 10)
... | AttributeError | dataset/ETHPy150Open CiscoSystems/avos/horizon/test/tests/middleware.py/MiddlewareTests.test_session_timeout |
6,322 | def is_storage_local(storage):
"""
Check to see if a file storage is local.
"""
try:
storage.path('test')
except __HOLE__:
return False
return True | NotImplementedError | dataset/ETHPy150Open SmileyChris/easy-thumbnails/easy_thumbnails/utils.py/is_storage_local |
6,323 | def get_modified_time(storage, name):
"""
Get modified time from storage, ensuring the result is a timezone-aware
datetime.
"""
try:
modified_time = storage.modified_time(name)
except OSError:
return 0
except __HOLE__:
return None
if modified_time and timezone.is_... | NotImplementedError | dataset/ETHPy150Open SmileyChris/easy-thumbnails/easy_thumbnails/utils.py/get_modified_time |
6,324 | def __init__(self, folder, callback, file_pattern=["*.log"], includeSubFolder=False, excludeFileListFile=None, tail_lines=0,
sizehint=1048576):
"""Arguments:
(str) @folder:
the folder to watch
(callable) @callback:
a function which is called every... | IOError | dataset/ETHPy150Open harryliu/edwin/edwinAgent/site_packages/logwatch_glob.py/LogWatcher.__init__ |
6,325 | def __get__(self, instance, owner=None):
try:
value = FieldProperty.__get__(self, instance, owner)
except __HOLE__:
value = None
if not value:
return None
if instance is None:
return value
return self._upload_type(value) | AttributeError | dataset/ETHPy150Open amol-/depot/depot/fields/ming.py/UploadedFileProperty.__get__ |
6,326 | @staticmethod
def _close_nodes(nodepaths, get_node):
for nodepath in nodepaths:
try:
node = get_node(nodepath)
except __HOLE__:
pass
else:
if not node._v_isopen or node._v__deleting:
continue
... | KeyError | dataset/ETHPy150Open PyTables/PyTables/tables/file.py/NodeManager._close_nodes |
6,327 | def __str__(self):
"""Return a short string representation of the object tree.
Examples
--------
::
>>> f = tables.open_file('data/test.h5')
>>> print(f)
data/test.h5 (File) 'Table Benchmark'
Last modif.: 'Mon Sep 20 12:40:47 2004'
... | OSError | dataset/ETHPy150Open PyTables/PyTables/tables/file.py/File.__str__ |
6,328 | def dispatch(self, json_data, **kwargs):
'''
Verifies that the passed json encoded string
is in the correct form according to the json-rpc spec
and calls the appropriate method
Checks:
1. that the string encodes into a javascript Object (dictionary)
2.... | TypeError | dataset/ETHPy150Open fp7-ofelia/ocf/expedient/src/python/expedient/common/rpc4django/jsonrpcdispatcher.py/JSONRPCDispatcher.dispatch |
6,329 | def get_image_field_class():
try:
from sorl.thumbnail import ImageField
except __HOLE__:
from django.db.models import ImageField
return ImageField | ImportError | dataset/ETHPy150Open hovel/pybbm/pybb/compat.py/get_image_field_class |
6,330 | def get_image_field_full_name():
try:
from sorl.thumbnail import ImageField
name = 'sorl.thumbnail.fields.ImageField'
except __HOLE__:
from django.db.models import ImageField
name = 'django.db.models.fields.files.ImageField'
return name | ImportError | dataset/ETHPy150Open hovel/pybbm/pybb/compat.py/get_image_field_full_name |
6,331 | def get_atomic_func():
try:
from django.db.transaction import atomic as atomic_func
except __HOLE__:
from django.db.transaction import commit_on_success as atomic_func
return atomic_func | ImportError | dataset/ETHPy150Open hovel/pybbm/pybb/compat.py/get_atomic_func |
6,332 | def get_paginator_class():
try:
from pure_pagination import Paginator
pure_pagination = True
except __HOLE__:
# the simplest emulation of django-pure-pagination behavior
from django.core.paginator import Paginator, Page
class PageRepr(int):
def querystring(sel... | ImportError | dataset/ETHPy150Open hovel/pybbm/pybb/compat.py/get_paginator_class |
6,333 | def get_previous_transaction(self):
siblings = StockTransaction.get_ordered_transactions_for_stock(
self.case_id, self.section_id, self.product_id
).filter(report__date__lte=self.report.date).exclude(pk=self.pk)
try:
return siblings[0]
except __HOLE__:
... | IndexError | dataset/ETHPy150Open dimagi/commcare-hq/corehq/ex-submodules/casexml/apps/stock/models.py/StockTransaction.get_previous_transaction |
6,334 | @classmethod
def latest(cls, case_id, section_id, product_id):
relevant = cls.get_ordered_transactions_for_stock(case_id, section_id, product_id)
try:
return relevant.select_related()[0]
except __HOLE__:
return None | IndexError | dataset/ETHPy150Open dimagi/commcare-hq/corehq/ex-submodules/casexml/apps/stock/models.py/StockTransaction.latest |
6,335 | def update_check(settings):
"""
Check whether the dependencies are sufficient to run Eden
@ToDo: Load deployment_settings so that we can configure the update_check
- need to rework so that 000_config.py is parsed 1st
@param settings: the deployment_settings
"""
... | ImportError | dataset/ETHPy150Open sahana/eden/modules/s3_update_check.py/update_check |
6,336 | def parse_requirements(output, filepath):
"""
"""
try:
with open(filepath) as filehandle:
dependencies = filehandle.read().splitlines()
msg = ""
for dependency in dependencies:
if dependency[0] == "#":
# either a norm... | AttributeError | dataset/ETHPy150Open sahana/eden/modules/s3_update_check.py/parse_requirements |
6,337 | def s3_check_python_lib(global_mandatory, template_mandatory, template_optional, global_optional):
"""
checks for optional as well as mandatory python libraries
"""
errors = []
warnings = []
for dependency, err in global_mandatory.iteritems():
try:
if "from" i... | ImportError | dataset/ETHPy150Open sahana/eden/modules/s3_update_check.py/s3_check_python_lib |
6,338 | def get_info(self, path):
dirname, filename = os.path.split(path)
name, ext = os.path.splitext(filename)
ftype = 'file'
if os.path.isdir(path):
ftype = 'dir'
elif os.path.islink(path):
ftype = 'link'
elif ext.lower() == '.fits':
ftype =... | OSError | dataset/ETHPy150Open ejeschke/ginga/ginga/misc/plugins/FBrowser.py/FBrowser.get_info |
6,339 | def __dict__(self):
try:
return self._current_object.__dict__
except __HOLE__:
return AttributeError('__dict__') | RuntimeError | dataset/ETHPy150Open dcramer/django-indexer/indexer/utils.py/Proxy.__dict__ |
6,340 | def __repr__(self):
try:
obj = self._current_object
except __HOLE__:
return '<%s unbound>' % self.__class__.__name__
return repr(obj) | RuntimeError | dataset/ETHPy150Open dcramer/django-indexer/indexer/utils.py/Proxy.__repr__ |
6,341 | def __nonzero__(self):
try:
return bool(self._current_object)
except __HOLE__:
return False | RuntimeError | dataset/ETHPy150Open dcramer/django-indexer/indexer/utils.py/Proxy.__nonzero__ |
6,342 | def __unicode__(self):
try:
return unicode(self.__current_oject)
except __HOLE__:
return repr(self) | RuntimeError | dataset/ETHPy150Open dcramer/django-indexer/indexer/utils.py/Proxy.__unicode__ |
6,343 | def __dir__(self):
try:
return dir(self._current_object)
except __HOLE__:
return [] | RuntimeError | dataset/ETHPy150Open dcramer/django-indexer/indexer/utils.py/Proxy.__dir__ |
6,344 | def _replaceEnv(self, match):
""" Internally used method to replace found matches of _RE_ENV regular
expression with corresponding environment variable.
"""
var = match.group('var')
try:
return os.environ[var]
except __HOLE__:
raise Environmen... | KeyError | dataset/ETHPy150Open rapyuta/rce/rce-core/rce/monitor/common.py/ArgumentMixin._replaceEnv |
6,345 | def top(sort_key='cpu'):
# TODO: add sort key middleware
_key = itemgetter(sort_key)
entries = [format_dict(pd) for pd in get_process_dicts()]
# handle an apparent bug in psutil where the first call of the
# process does not return any cpu percentages. sorting by memory
# percentages instead.
... | TypeError | dataset/ETHPy150Open mahmoud/clastic/clastic/contrib/webtop/top.py/top |
6,346 | def canonize_duedate(duedate):
if duedate == None:
return '000000'
else:
try:
return duedate.strftime('%y%m%d')
except __HOLE__:
raise DueDateFormatException("Invalid type for canonize_duedate") | AttributeError | dataset/ETHPy150Open kapsiry/sikteeri/membership/reference_numbers.py/canonize_duedate |
6,347 | @register.filter
def quantity_ordered(product, order):
"""
e.g. {% if product|quantity_ordered:plata.order > 0 %} ... {% endif %}
"""
try:
return order.items.values('quantity').get(product=product)['quantity']
except __HOLE__:
return 0 | ObjectDoesNotExist | dataset/ETHPy150Open matthiask/plata/plata/shop/templatetags/plata_tags.py/quantity_ordered |
6,348 | def request(self, req):
if self.cached:
# Try once to use the cached connection; if it fails to send the
# request then discard and try again.
try:
return self.requestOnce(self.cached, req)
except http_error.RequestError, err:
err.w... | KeyboardInterrupt | dataset/ETHPy150Open sassoftware/conary/conary/lib/http/connection.py/Connection.request |
6,349 | def startSSL(self, sock):
"""If needed, start SSL on the proxy or endpoint connection."""
if not self.doSSL:
return sock
if self.caCerts:
# If cert checking is requested use m2crypto
if SSL:
return startSSLWithChecker(sock, self.caCerts, self.c... | ImportError | dataset/ETHPy150Open sassoftware/conary/conary/lib/http/connection.py/Connection.startSSL |
6,350 | def requestOnce(self, conn, req):
if self.proxy and self.proxy.userpass[0] and not self.doTunnel:
req.headers['Proxy-Authorization'] = ('Basic ' +
base64.b64encode(":".join(self.proxy.userpass)))
try:
req.sendRequest(conn, isProxied=(self.proxy is not None))
... | SystemExit | dataset/ETHPy150Open sassoftware/conary/conary/lib/http/connection.py/Connection.requestOnce |
6,351 | def get_mimetype(path, file_contents=None):
mimetypes.init([MIMEMAP])
mimetype, _ = mimetypes.guess_type(path)
if mimetype is None:
try:
import magic
if file_contents is not None:
mimetype = magic.from_buffer(file_contents, mime=True)
else:
... | ImportError | dataset/ETHPy150Open CenterForOpenScience/osf.io/website/util/mimetype.py/get_mimetype |
6,352 | def import_class(module_name):
"""
Import a module with a reasonable traceback and return that module as the
variable.
eg. dumps = import_class('simplejson.dumps')
if dumps is None:
dumps = import_class('json.dumps')
Reference: http://lucumr.pocoo.org/2011/9/21/python-import-blackbox/
... | ImportError | dataset/ETHPy150Open winhamwr/neckbeard/neckbeard/cloud_provisioners/base.py/import_class |
6,353 | def save(self):
"""Find the run with this build, or create a new one."""
try:
this_run = model.Run.objects.get(
series=self.run,
build=self.cleaned_data["build"],
)
except __HOLE__:
this_run = self.run.clone_for_series(
... | ObjectDoesNotExist | dataset/ETHPy150Open mozilla/moztrap/moztrap/view/runtests/forms.py/EnvironmentBuildSelectionForm.save |
6,354 | def readme():
try:
with open(os.path.join(os.path.dirname(__file__), 'README.rst')) as f:
readme = f.read()
except __HOLE__:
pass
pattern = re.compile(r'''
(?P<colon> : \n{2,})?
\s* \.\. [ ] code-block:: \s+ [^\n]+ \n
[ \t]* \n
(?P<block>
... | IOError | dataset/ETHPy150Open crosspop/asuka/setup.py/readme |
6,355 | def __getattr__(self, name):
try:
return self[name]
except __HOLE__:
raise AttributeError(name) | KeyError | dataset/ETHPy150Open tensorprob/tensorprob/tensorprob/optimization_result.py/OptimizationResult.__getattr__ |
6,356 | def update_dividend(self, symbol, data):
if len(data) == 0:
return
try:
self.divstore[symbol] = data
except __HOLE__:
del self.divstore[symbol]
self.divstore[symbol] = data | ValueError | dataset/ETHPy150Open yinhm/datafeed/datafeed/datastore.py/Manager.update_dividend |
6,357 | def _update(self, symbol, quotes):
"""Archive daily ohlcs, override if datasets exists.
Arguments:
symbol: Stock instrument.
quotes: numpy quotes data.
"""
i = 0
pre_ts = 0
indexes = []
for q in quotes:
# 2 hours interval should be... | TypeError | dataset/ETHPy150Open yinhm/datafeed/datafeed/datastore.py/OHLC._update |
6,358 | def _update_multi(self, symbol, quotes):
"""Archive multiday ohlcs, override if datasets exists.
Arguments:
symbol: Stock instrument.
quotes: numpy quotes data.
"""
i = 0
pre_day = None
indexes = []
indexes.append([0, len(quotes)])
for... | TypeError | dataset/ETHPy150Open yinhm/datafeed/datafeed/datastore.py/OHLC._update_multi |
6,359 | def get(self, symbol, length):
year = datetime.datetime.today().isocalendar()[0]
try:
data = self._get_year_data(symbol, year)
except KeyError:
self.handle[symbol] # test symbol existence
data = []
while True:
if len(data) >= length:
... | KeyError | dataset/ETHPy150Open yinhm/datafeed/datafeed/datastore.py/Day.get |
6,360 | def update(self, symbol, data):
"""append daily history data to daily archive.
Arguments
=========
- `symbol`: symbol.
- `npydata`: data of npy file.
"""
prev_year = None
ds = None
newdata = None
for row in data:
day = datetim... | IndexError | dataset/ETHPy150Open yinhm/datafeed/datafeed/datastore.py/Day.update |
6,361 | def _require_dataset(self, symbol):
try:
return self._dataset(symbol)
except __HOLE__:
return self.handle.create_dataset(symbol,
(self.shape_x, ),
self.DTYPE) | KeyError | dataset/ETHPy150Open yinhm/datafeed/datafeed/datastore.py/Minute._require_dataset |
6,362 | def _rewrite(self, tostore):
if self.__len__() > 0:
for key in self.keys():
try:
tostore.update(key, self.__getitem__(key))
except __HOLE__:
logging.error("Inconsistent data for %s, ignoring." % key)
self.__delit... | AssertionError | dataset/ETHPy150Open yinhm/datafeed/datafeed/datastore.py/MinuteSnapshotCache._rewrite |
6,363 | def get_pickle():
try:
import cPickle as pickle
except __HOLE__:
import pickle # noqa
return pickle | ImportError | dataset/ETHPy150Open codysoyland/django-phased/phased/utils.py/get_pickle |
6,364 | def restore_csrf_token(request, storage=None):
"""
Given the request and a the context used during the second render phase,
this wil check if there is a CSRF cookie and restores if needed, to
counteract the way the CSRF framework invalidates the CSRF token after
each request/response cycle.
"""
... | KeyError | dataset/ETHPy150Open codysoyland/django-phased/phased/utils.py/restore_csrf_token |
6,365 | def polymorphic_union(table_map, typecolname,
aliasname='p_union', cast_nulls=True):
"""Create a ``UNION`` statement used by a polymorphic mapper.
See :ref:`concrete_inheritance` for an example of how
this is used.
:param table_map: mapping of polymorphic identities to
:cla... | KeyError | dataset/ETHPy150Open goFrendiAsgard/kokoropy/kokoropy/packages/sqlalchemy/orm/util.py/polymorphic_union |
6,366 | def identity_key(*args, **kwargs):
"""Generate "identity key" tuples, as are used as keys in the
:attr:`.Session.identity_map` dictionary.
This function has several call styles:
* ``identity_key(class, ident)``
This form receives a mapped class and a primary key scalar or
tuple as an argu... | KeyError | dataset/ETHPy150Open goFrendiAsgard/kokoropy/kokoropy/packages/sqlalchemy/orm/util.py/identity_key |
6,367 | def __getattr__(self, key):
try:
_aliased_insp = self.__dict__['_aliased_insp']
except KeyError:
raise AttributeError()
else:
for base in _aliased_insp._target.__mro__:
try:
attr = object.__getattribute__(base, key)
... | AttributeError | dataset/ETHPy150Open goFrendiAsgard/kokoropy/kokoropy/packages/sqlalchemy/orm/util.py/AliasedClass.__getattr__ |
6,368 | def ReadTag(buf, pos):
"""Read a tag from the buffer, and return a (tag_bytes, new_pos) tuple."""
try:
start = pos
while ORD_MAP_AND_0X80[buf[pos]]:
pos += 1
pos += 1
return (buf[start:pos], pos)
except __HOLE__:
raise ValueError("Invalid tag")
# This function is HOT. | IndexError | dataset/ETHPy150Open google/grr/grr/lib/rdfvalues/structs.py/ReadTag |
6,369 | def Validate(self, value, **_):
"""Validates a python format representation of the value."""
# We only accept a base string, unicode object or RDFString here.
if not (value.__class__ is str or value.__class__ is unicode or
value.__class__ is rdfvalue.RDFString):
raise type_info.TypeValueEr... | AttributeError | dataset/ETHPy150Open google/grr/grr/lib/rdfvalues/structs.py/ProtoString.Validate |
6,370 | def Validate(self, value, **_):
try:
return int(value)
except __HOLE__:
raise type_info.TypeValueError("Invalid value %s for Integer" % value) | ValueError | dataset/ETHPy150Open google/grr/grr/lib/rdfvalues/structs.py/ProtoUnsignedInteger.Validate |
6,371 | def Append(self, rdf_value=utils.NotAValue, wire_format=None, **kwargs):
"""Append the value to our internal list."""
if rdf_value is utils.NotAValue:
if wire_format is None:
rdf_value = self.type_descriptor.type(**kwargs)
self.dirty = True
else:
rdf_value = None
else:
... | TypeError | dataset/ETHPy150Open google/grr/grr/lib/rdfvalues/structs.py/RepeatedFieldHelper.Append |
6,372 | def _CopyRawData(self):
new_raw_data = {}
# We need to copy all entries in _data. Those entries are tuples of
# - an object (if it has already been deserialized)
# - the serialized object (if it has been serialized)
# - the type_info.
# To copy this, it's easiest to just copy the serialized obj... | AttributeError | dataset/ETHPy150Open google/grr/grr/lib/rdfvalues/structs.py/RDFStruct._CopyRawData |
6,373 | def UnionCast(self):
union_field = getattr(self, self.union_field)
cast_field_name = str(union_field).lower()
try:
return getattr(self, cast_field_name)
except __HOLE__:
raise AttributeError("union_field not initialized.") | AttributeError | dataset/ETHPy150Open google/grr/grr/lib/rdfvalues/structs.py/RDFProtoStruct.UnionCast |
6,374 | def _guess_mimetype(self, file):
"""Guess the mimetype of an uploaded file.
Uploaded files don't necessarily have valid mimetypes provided,
so attempt to guess them when they're blank.
This only works if `file` is in the path. If it's not, or guessing
fails, we fall back to a m... | IOError | dataset/ETHPy150Open reviewboard/reviewboard/reviewboard/attachments/forms.py/UploadFileForm._guess_mimetype |
6,375 | @interactive
def _require(*names):
"""Helper for @require decorator."""
from IPython.parallel.error import UnmetDependency
user_ns = globals()
for name in names:
if name in user_ns:
continue
try:
exec('import %s'%name, user_ns)
except __HOLE__:
... | ImportError | dataset/ETHPy150Open ipython/ipython-py3k/IPython/parallel/controller/dependency.py/_require |
6,376 | def test_hyper_param_accuracy():
mp.dps = 15
As = [n+1e-10 for n in range(-5,-1)]
Bs = [n+1e-10 for n in range(-12,-5)]
assert hyper(As,Bs,10).ae(-381757055858.652671927)
assert legenp(0.5, 100, 0.25).ae(-2.4124576567211311755e+144)
assert (hyp1f1(1000,1,-100)*10**24).ae(5.2589445437370169113)
... | ValueError | dataset/ETHPy150Open fredrik-johansson/mpmath/mpmath/tests/test_functions2.py/test_hyper_param_accuracy |
6,377 | def test_issue_239():
mp.prec = 150
x = ldexp(2476979795053773,-52)
assert betainc(206, 385, 0, 0.55, 1).ae('0.99999999999999999999996570910644857895771110649954')
mp.dps = 15
try:
u = hyp2f1(-5,5,0.5,0.5)
raise AssertionError("hyp2f1(-5,5,0.5,0.5) (failed zero detection)")
excep... | ValueError | dataset/ETHPy150Open fredrik-johansson/mpmath/mpmath/tests/test_functions2.py/test_issue_239 |
6,378 | def build_deploy_docs(docs_path):
try:
from fabric.main import load_fabfile
except __HOLE__:
warn("Couldn't build fabfile.rst, fabric not installed")
return
project_template_path = path_for_import("mezzanine.project_template")
commands = load_fabfile(os.path.join(project_template... | ImportError | dataset/ETHPy150Open stephenmcd/mezzanine/mezzanine/utils/docs.py/build_deploy_docs |
6,379 | def build_changelog(docs_path, package_name="mezzanine"):
"""
Converts Mercurial commits into a changelog in RST format.
"""
project_path = os.path.join(docs_path, "..")
version_file = os.path.join(package_name, "__init__.py")
version_var = "__version__"
changelog_filename = "CHANGELOG"
... | KeyError | dataset/ETHPy150Open stephenmcd/mezzanine/mezzanine/utils/docs.py/build_changelog |
6,380 | def build_modelgraph(docs_path, package_name="mezzanine"):
"""
Creates a diagram of all the models for mezzanine and the given
package name, generates a smaller version and add it to the
docs directory for use in model-graph.rst
"""
to_path = os.path.join(docs_path, "img", "graph.png")
build... | OSError | dataset/ETHPy150Open stephenmcd/mezzanine/mezzanine/utils/docs.py/build_modelgraph |
6,381 | def _parse(self, data):
""" Parse the output from the 'mntr' 4letter word command """
h = StringIO(data)
result = {}
for line in h.readlines():
try:
key, value = self._parse_line(line)
result[key] = value
except __HOLE__:
... | ValueError | dataset/ETHPy150Open francelabs/datafari/debian7/zookeeper/src/contrib/monitoring/ganglia/zookeeper_ganglia.py/ZooKeeperServer._parse |
6,382 | def _parse_line(self, line):
try:
key, value = map(str.strip, line.split('\t'))
except ValueError:
raise ValueError('Found invalid line: %s' % line)
if not key:
raise ValueError('The key is mandatory and should not be empty')
try:
value =... | TypeError | dataset/ETHPy150Open francelabs/datafari/debian7/zookeeper/src/contrib/monitoring/ganglia/zookeeper_ganglia.py/ZooKeeperServer._parse_line |
6,383 | def do_check(nowip, condition):
result = False
import OpenSSL
import socket
import re
context = OpenSSL.SSL.Context(OpenSSL.SSL.TLSv1_METHOD)
socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
socket.settimeout(1)
connection = OpenSSL.SSL.Connection(context, socket)
if condit... | KeyboardInterrupt | dataset/ETHPy150Open futursolo/host-witness/witness.py/do_check |
6,384 | @fresh_login_required
@nocache
def reminder_history(pk_hash):
reminder = Reminder.objects.get_or_404(pk_hash=pk_hash,
owner=current_user._get_current_object())
history = ReminderHistory.objects.filter(reminder=reminder)
# Paginate
try:
history = Pagina... | ValueError | dataset/ETHPy150Open shaunduncan/breezeminder/breezeminder/views/reminder.py/reminder_history |
6,385 | def learn_sentence(self, sentence):
"""
Learn based on the input sentence.
:param sentence: Space separated sentence to apply to Markov model
:return: No relevant return data
"""
# Split the sentence into words/parts
parts = re.findall(r"\w[\w']*", sentence.lower... | KeyError | dataset/ETHPy150Open bwall/markovobfuscate/markovobfuscate/obfuscation.py/MarkovKeyState.learn_sentence |
6,386 | def load_plugins():
"""Load any plugin modules
We load plugin modules based on directories provided to us by the environment, as well as a default in our own folder.
Returns a list of module objects
"""
# This function is a little wacky, doesn't seem like we SHOULD have to do all this just to get ... | ImportError | dataset/ETHPy150Open Yelp/Testify/testify/test_program.py/load_plugins |
6,387 | def api_url_patterns():
for version, resources in API_LIST:
api = CommCareHqApi(api_name='v%d.%d' % version)
for R in resources:
api.register(R())
yield (r'^', include(api.urls))
yield url(r'^v0.1/xform_es/$', XFormES.as_domain_specific_view())
# HACK: fix circular import... | ImportError | dataset/ETHPy150Open dimagi/commcare-hq/corehq/apps/api/urls.py/api_url_patterns |
6,388 | def output_poll(self):
try:
name = self.token['name']
except __HOLE__:
return self.renderer.poll_raw(poll_txt=self.token['raw'])
else:
return self.renderer.poll(name=name) | KeyError | dataset/ETHPy150Open nitely/Spirit/spirit/core/utils/markdown/markdown.py/Markdown.output_poll |
6,389 | def create(server_):
'''
Create a single BareMetal server from a data dict.
'''
try:
# Check for required profile parameters before sending any API calls.
if server_['profile'] and config.is_profile_configured(__opts__,
__act... | AttributeError | dataset/ETHPy150Open saltstack/salt/salt/cloud/clouds/scaleway.py/create |
6,390 | def _get_node(name):
for attempt in reversed(list(range(10))):
try:
return list_nodes_full()[name]
except __HOLE__:
log.debug(
'Failed to get the data for node \'{0}\'. Remaining '
'attempts: {1}'.format(
name, attempt
... | KeyError | dataset/ETHPy150Open saltstack/salt/salt/cloud/clouds/scaleway.py/_get_node |
6,391 | def stop(self):
try:
self.logger.removeHandler(self.handlers[self.logger_name])
except __HOLE__:
pass | KeyError | dataset/ETHPy150Open nigelsmall/py2neo/py2neo/packages/neo4j/util.py/Watcher.stop |
6,392 | def _convert_ip_address(self, ip, field_name):
try:
return netaddr.IPAddress(ip)
except (netaddr.AddrFormatError, __HOLE__):
msg = (_('%(field_name)s: Invalid IP address (value=%(ip)s)')
% {'field_name': field_name, 'ip': ip})
raise forms.Validation... | ValueError | dataset/ETHPy150Open CiscoSystems/avos/openstack_dashboard/dashboards/project/networks/workflows.py/CreateSubnetDetailAction._convert_ip_address |
6,393 | def _convert_ip_network(self, network, field_name):
try:
return netaddr.IPNetwork(network)
except (netaddr.AddrFormatError, __HOLE__):
msg = (_('%(field_name)s: Invalid IP address (value=%(network)s)')
% {'field_name': field_name, 'network': network})
... | ValueError | dataset/ETHPy150Open CiscoSystems/avos/openstack_dashboard/dashboards/project/networks/workflows.py/CreateSubnetDetailAction._convert_ip_network |
6,394 | def configure(self, candidates=None):
ctx = self.ctx
if candidates is None:
if sys.platform == "win32":
candidates = ["msvc", "gxx"]
else:
candidates = ["gxx", "cxx"]
def _detect_cxx():
detected = None
sys.path.inse... | ImportError | dataset/ETHPy150Open cournape/Bento/bento/private/_yaku/yaku/tools/cxxtasks.py/CXXBuilder.configure |
6,395 | def setUp(self):
super(CollectionTestCase, self).setUp()
if self.service.splunk_version[0] >= 5 and 'modular_input_kinds' not in collections:
collections.append('modular_input_kinds') # Not supported before Splunk 5.0
else:
logging.info("Skipping modular_input_kinds; not ... | KeyError | dataset/ETHPy150Open splunk/splunk-sdk-python/tests/test_collection.py/CollectionTestCase.setUp |
6,396 | def loop(self, sleep_time=None, callback=None):
self.trigger_init()
try:
self._notifier.loop(callback)
except __HOLE__:
self._notifier.stop()
raise | KeyboardInterrupt | dataset/ETHPy150Open jeffh/sniffer/sniffer/scanner/pyinotify_scanner.py/PyINotifyScanner.loop |
6,397 | def run(self):
""" Runs the event loop
This method starts the main asyncio event loop and stays in the loop until a SIGINT or
SIGTERM is received (see `_sigint_handler`)
"""
try:
self.logger.debug("Starting event loop and calling `get_ready`")
if self.is... | KeyboardInterrupt | dataset/ETHPy150Open panoptes/POCS/panoptes/state/event.py/PanEventManager.run |
6,398 | def _test_not_equal(self, a, b):
try:
self._assert_func(a, b)
passed = True
except __HOLE__:
pass
else:
raise AssertionError("a and b are found equal but are not") | AssertionError | dataset/ETHPy150Open gkno/gkno_launcher/src/networkx/testing/tests/test_utils.py/_GenericTest._test_not_equal |
6,399 | @stable(as_of="0.4.0")
def index(self, subsequence, start=None, end=None):
"""Find position where subsequence first occurs in the sequence.
Parameters
----------
subsequence : str, Sequence, or 1D np.ndarray (np.uint8 or '\|S1')
Subsequence to search for in this sequence... | ValueError | dataset/ETHPy150Open biocore/scikit-bio/skbio/sequence/_sequence.py/Sequence.index |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.