Unnamed: 0
int64
0
10k
function
stringlengths
79
138k
label
stringclasses
20 values
info
stringlengths
42
261
4,900
def set_body_reader(self): chunked = False content_length = None for (name, value) in self.headers: if name == "CONTENT-LENGTH": content_length = value elif name == "TRANSFER-ENCODING": chunked = value.lower() == "chunked" elif name == "SEC-WEBSOCKET-KEY1": content_length = 8 if chunked: self.body = Body(ChunkedReader(self, self.unreader)) elif content_length is not None: try: content_length = int(content_length) except __HOLE__: raise InvalidHeader("CONTENT-LENGTH", req=self) if content_length < 0: raise InvalidHeader("CONTENT-LENGTH", req=self) self.body = Body(LengthReader(self.unreader, content_length)) else: self.body = Body(EOFReader(self.unreader))
ValueError
dataset/ETHPy150Open chalasr/Flask-P2P/venv/lib/python2.7/site-packages/gunicorn/http/message.py/Message.set_body_reader
4,901
def parse_proxy_protocol(self, line): bits = line.split() if len(bits) != 6: raise InvalidProxyLine(line) # Extract data proto = bits[1] s_addr = bits[2] d_addr = bits[3] # Validation if proto not in ["TCP4", "TCP6"]: raise InvalidProxyLine("protocol '%s' not supported" % proto) if proto == "TCP4": try: socket.inet_pton(socket.AF_INET, s_addr) socket.inet_pton(socket.AF_INET, d_addr) except socket.error: raise InvalidProxyLine(line) elif proto == "TCP6": try: socket.inet_pton(socket.AF_INET6, s_addr) socket.inet_pton(socket.AF_INET6, d_addr) except socket.error: raise InvalidProxyLine(line) try: s_port = int(bits[4]) d_port = int(bits[5]) except __HOLE__: raise InvalidProxyLine("invalid port %s" % line) if not ((0 <= s_port <= 65535) and (0 <= d_port <= 65535)): raise InvalidProxyLine("invalid port %s" % line) # Set data self.proxy_protocol_info = { "proxy_protocol": proto, "client_addr": s_addr, "client_port": s_port, "proxy_addr": d_addr, "proxy_port": d_port }
ValueError
dataset/ETHPy150Open chalasr/Flask-P2P/venv/lib/python2.7/site-packages/gunicorn/http/message.py/Request.parse_proxy_protocol
4,902
def form_invalid(self, form): result = super(SingleObjectCreateView, self).form_invalid(form) try: messages.error( self.request, _('Error creating new %s.') % self.extra_context['object_name'] ) except __HOLE__: messages.error(self.request, _('Error creating object.')) return result
KeyError
dataset/ETHPy150Open mayan-edms/mayan-edms/mayan/apps/common/generics.py/SingleObjectCreateView.form_invalid
4,903
def form_valid(self, form): # This overrides the original Django form_valid method self.object = form.save(commit=False) if hasattr(self, 'get_instance_extra_data'): for key, value in self.get_instance_extra_data().items(): setattr(self.object, key, value) if hasattr(self, 'get_save_extra_data'): save_extra_data = self.get_save_extra_data() else: save_extra_data = {} self.object.save(**save_extra_data) try: messages.success( self.request, _( '%s created successfully.' ) % self.extra_context['object_name'] ) except __HOLE__: messages.success( self.request, _('New object created successfully.') ) return HttpResponseRedirect(self.get_success_url())
KeyError
dataset/ETHPy150Open mayan-edms/mayan-edms/mayan/apps/common/generics.py/SingleObjectCreateView.form_valid
4,904
def delete(self, request, *args, **kwargs): try: result = super(SingleObjectDeleteView, self).delete(request, *args, **kwargs) except Exception as exception: try: messages.error( self.request, _('Error deleting %s.') % self.extra_context['object_name'] ) except KeyError: messages.error( self.request, _('Error deleting object.') ) raise exception else: try: messages.success( self.request, _( '%s deleted successfully.' ) % self.extra_context['object_name'] ) except __HOLE__: messages.success( self.request, _('Object deleted successfully.') ) return result
KeyError
dataset/ETHPy150Open mayan-edms/mayan-edms/mayan/apps/common/generics.py/SingleObjectDeleteView.delete
4,905
def form_invalid(self, form): result = super(SingleObjectEditView, self).form_invalid(form) try: messages.error( self.request, _( 'Error saving %s details.' ) % self.extra_context['object_name'] ) except __HOLE__: messages.error( self.request, _('Error saving details.') ) return result
KeyError
dataset/ETHPy150Open mayan-edms/mayan-edms/mayan/apps/common/generics.py/SingleObjectEditView.form_invalid
4,906
def form_valid(self, form): # This overrides the original Django form_valid method self.object = form.save(commit=False) if hasattr(self, 'get_instance_extra_data'): for key, value in self.get_instance_extra_data().items(): setattr(self.object, key, value) if hasattr(self, 'get_save_extra_data'): save_extra_data = self.get_save_extra_data() else: save_extra_data = {} self.object.save(**save_extra_data) try: messages.success( self.request, _( '%s details saved successfully.' ) % self.extra_context['object_name'] ) except __HOLE__: messages.success( self.request, _('Details saved successfully.') ) return HttpResponseRedirect(self.get_success_url())
KeyError
dataset/ETHPy150Open mayan-edms/mayan-edms/mayan/apps/common/generics.py/SingleObjectEditView.form_valid
4,907
def _kurtosis(a): '''wrapper for scipy.stats.kurtosis that returns nan instead of raising Error missing options ''' try: res = stats.kurtosis(a) except __HOLE__: res = np.nan return res
ValueError
dataset/ETHPy150Open statsmodels/statsmodels/statsmodels/stats/descriptivestats.py/_kurtosis
4,908
def _skew(a): '''wrapper for scipy.stats.skew that returns nan instead of raising Error missing options ''' try: res = stats.skew(a) except __HOLE__: res = np.nan return res
ValueError
dataset/ETHPy150Open statsmodels/statsmodels/statsmodels/stats/descriptivestats.py/_skew
4,909
def _is_dtype_like(self, col): """ Check whether self.dataset.[col][0] behaves like a string, numbern unknown. `numpy.lib._iotools._is_string_like` """ def string_like(): #TODO: not sure what the result is if the first item is some type of # missing value try: self.dataset[col][0] + '' except (TypeError, ValueError): return False return True def number_like(): try: self.dataset[col][0] + 1.0 except (TypeError, __HOLE__): return False return True if number_like()==True and string_like()==False: return 'number' elif number_like()==False and string_like()==True: return 'string' else: assert (number_like()==True or string_like()==True), '\ Not sure of dtype'+str(self.dataset[col][0]) #@property
ValueError
dataset/ETHPy150Open statsmodels/statsmodels/statsmodels/stats/descriptivestats.py/Describe._is_dtype_like
4,910
def tuple_values_dict(currtuple, newtuple=None): """Return dictionary of attributes with their values :param currtuple: current Tuple :param newtuple: optional Tuple with new values :return: dictionary of attributes and values """ valdict = {} for attr in currtuple.__dict__: if attr in RESERVED_ATTRIBUTE_NAMES: continue currval = getattr(currtuple, attr) if newtuple is None: valdict.update({attr: currval}) elif hasattr(newtuple, attr): newval = getattr(newtuple, attr) try: diff = currval != newval except __HOLE__: # TODO: more needed for naive vs. aware datetime/time if isinstance(newval, datetime) or isinstance(newval, time): if newval.utcoffset() is None: diff = True else: raise if diff: valdict.update({attr: newval}) if newtuple is not None: for newattr in newtuple.__dict__: if newattr in RESERVED_ATTRIBUTE_NAMES: continue if newattr not in currtuple.__dict__: valdict.update({newattr: getattr(newtuple, newattr)}) return valdict
TypeError
dataset/ETHPy150Open perseas/Pyrseas/pyrseas/relation/tuple.py/tuple_values_dict
4,911
def runtests(): if not settings.configured: settings.configure(**DEFAULT_SETTINGS) # Compatibility with Django 1.7's stricter initialization if hasattr(django, 'setup'): django.setup() print(sys.path) try: from django.test.runner import DiscoverRunner runner_class = DiscoverRunner test_args = ['djangosecure.tests'] except __HOLE__: from django.test.simple import DjangoTestSuiteRunner runner_class = DjangoTestSuiteRunner test_args = ['tests'] failures = runner_class( verbosity=1, interactive=True, failfast=False).run_tests(test_args) sys.exit(failures)
ImportError
dataset/ETHPy150Open carljm/django-secure/runtests.py/runtests
4,912
def clean(self): self.tfile.close() try: os.unlink(self.get_full_path()) except __HOLE__: pass
OSError
dataset/ETHPy150Open tooxie/shiva-server/shiva/resources/upload.py/UploadHandler.clean
4,913
def get_path(self): path = '' upload_path = app.config.get('UPLOAD_PATH') if not upload_path: raise NoUploadPathConfigError() _artist = app.config.get('UPLOAD_DEFAULT_ARTIST', '') _album = app.config.get('UPLOAD_DEFAULT_ALBUM', '') artist = slugify(self.mdma.artist) or _artist album = slugify(self.mdma.album) or _album path = os.path.join(upload_path, artist, album) try: os.makedirs(path) except __HOLE__: # File exists pass return path
OSError
dataset/ETHPy150Open tooxie/shiva-server/shiva/resources/upload.py/UploadHandler.get_path
4,914
def parseExtensionArgs(self, args): """Set the state of this request to be that expressed in these PAPE arguments @param args: The PAPE arguments without a namespace @rtype: None @raises ValueError: When the max_auth_age is not parseable as an integer """ # preferred_auth_policies is a space-separated list of policy URIs self.preferred_auth_policies = [] policies_str = args.get('preferred_auth_policies') if policies_str: for uri in policies_str.split(' '): if uri not in self.preferred_auth_policies: self.preferred_auth_policies.append(uri) # max_auth_age is base-10 integer number of seconds max_auth_age_str = args.get('max_auth_age') self.max_auth_age = None if max_auth_age_str: try: self.max_auth_age = int(max_auth_age_str) except __HOLE__: pass
ValueError
dataset/ETHPy150Open CollabQ/CollabQ/openid/extensions/pape.py/Request.parseExtensionArgs
4,915
def parseExtensionArgs(self, args, strict=False): """Parse the provider authentication policy arguments into the internal state of this object @param args: unqualified provider authentication policy arguments @param strict: Whether to raise an exception when bad data is encountered @returns: None. The data is parsed into the internal fields of this object. """ policies_str = args.get('auth_policies') if policies_str: self.auth_policies = policies_str.split(' ') nist_level_str = args.get('nist_auth_level') if nist_level_str: try: nist_level = int(nist_level_str) except ValueError: if strict: raise ValueError('nist_auth_level must be an integer between ' 'zero and four, inclusive') else: self.nist_auth_level = None else: if 0 <= nist_level < 5: self.nist_auth_level = nist_level auth_age_str = args.get('auth_age') if auth_age_str: try: auth_age = int(auth_age_str) except __HOLE__: if strict: raise else: if auth_age >= 0: self.auth_age = auth_age elif strict: raise ValueError('Auth age must be above zero')
ValueError
dataset/ETHPy150Open CollabQ/CollabQ/openid/extensions/pape.py/Response.parseExtensionArgs
4,916
def execute(self, *args, **options): if len(args) != 1: raise CommandError("You must provide a file to import!") self.site = options["site"] if not isinstance(self.site, Site): try: self.site = Site.objects.get(domain=options["site"]) except __HOLE__: raise CommandError("No site found for domain: %s" % options["site"]) filepath = args[0] if filepath == '-': rows = _get_rows(sys.stdin) _create_redirects(self.site, rows) else: try: with open(filepath) as fp: rows = _get_rows(fp) _create_redirects(self.site, rows) except IOError as e: if e.errno == errno.ENOENT: raise CommandError("The specified file does not exist!") else: raise
ObjectDoesNotExist
dataset/ETHPy150Open okfn/foundation/foundation/redirector/management/commands/import_redirect_csv.py/Command.execute
4,917
@access.public def getImageFeatures(self, params): try: import cv2 import numpy as np cv2_available = True except __HOLE__: cv2_available = False # Disabling opencv for now cv2_available = False if 'url' in params: data = requests.get(params['url'], verify=False).content else: data = str(cherrypy.request.body.read()) # Run Tika once parsed = parser.from_buffer(data) tika = {} for (k, v) in parsed["metadata"].iteritems(): k = k.lower().replace(':', '_').replace(' ', '_').replace('-', '_') tika[k] = v[0] if type(v) is list and len(v) else v tika['content'] = parsed["content"] if cv2_available: file_bytes = np.asarray(bytearray(data), dtype=np.uint8) image = cv2.imdecode(file_bytes, flags=cv2.CV_LOAD_IMAGE_UNCHANGED); if image is not None: if len(image.shape) < 3 or image.shape[2] == 1: image = cv2.cvtColor(image, cv2.COLOR_GRAY2RGB) v = cv2.calcHist([image], [0, 1, 2], None, [8, 8, 8], [0, 256, 0, 256, 0, 256]) v = v.flatten() hist = v / sum(v) tika['histogram'] = hist.tolist() return tika
ImportError
dataset/ETHPy150Open memex-explorer/image_space/imagespace/server/imagefeatures_rest.py/ImageFeatures.getImageFeatures
4,918
def get_geo_from_ip(ip_addr): try: from geoip import geolite2 return geolite2.lookup(ip_addr) except __HOLE__: return except ValueError: return
ImportError
dataset/ETHPy150Open frappe/frappe/frappe/sessions.py/get_geo_from_ip
4,919
def supported(cls, stream=sys.stdout): """ A class method that returns True if the current platform supports coloring terminal output using this method. Returns False otherwise. """ if not stream.isatty(): return False # auto color only on TTYs try: import curses except __HOLE__: return False else: try: try: return curses.tigetnum("colors") > 2 except curses.error: curses.setupterm() return curses.tigetnum("colors") > 2 except: # guess false in case of error return False
ImportError
dataset/ETHPy150Open twisted/twisted/twisted/trial/reporter.py/_AnsiColorizer.supported
4,920
def supported(cls, stream=sys.stdout): try: import win32console screenBuffer = win32console.GetStdHandle( win32console.STD_OUTPUT_HANDLE) except __HOLE__: return False import pywintypes try: screenBuffer.SetConsoleTextAttribute( win32console.FOREGROUND_RED | win32console.FOREGROUND_GREEN | win32console.FOREGROUND_BLUE) except pywintypes.error: return False else: return True
ImportError
dataset/ETHPy150Open twisted/twisted/twisted/trial/reporter.py/_Win32Colorizer.supported
4,921
@pytest.mark.hookwrapper def pytest_runtest_makereport(item, call): outcome = yield # Only run through this at the end of the test if call.when != 'call': return # Don't continue if this isn't a screenshot test if 'screenshot' not in item.fixturenames: return # Don't add screenshots if we can't create a screenshot try: screenshot = Screenshot(item=item) except __HOLE__: return report = outcome.get_result() xfail = hasattr(report, 'wasxfail') failure = (report.skipped and xfail) or (report.failed and not xfail) pytest_html = item.config.pluginmanager.getplugin('html') # Don't add screenshots if test passed if failure: diff = pytest_html.extras.image(screenshot.get_diff_as_base64(), '') base = pytest_html.extras.image(screenshot.get_base_as_base64(), '') test = pytest_html.extras.image(screenshot.get_current_as_base64(), '') # Override existing extra and logs for screenshot tests (keeps output manageable) report.extra = [test, diff, base] report.longrepr = ''
AssertionError
dataset/ETHPy150Open bokeh/bokeh/tests/integration/integration_tests_plugin.py/pytest_runtest_makereport
4,922
def class_for_kind(kind): """Return base-class responsible for implementing kind. Necessary to recover the class responsible for implementing provided kind. Args: kind: Entity kind string. Returns: Class implementation for kind. Raises: KindError when there is no implementation for kind. """ try: return _kind_map[kind] except __HOLE__: raise KindError('No implementation for kind \'%s\'' % kind)
KeyError
dataset/ETHPy150Open AppScale/appscale/AppServer/google/appengine/ext/db/__init__.py/class_for_kind
4,923
def __get__(self, model_instance, model_class): """Returns the value for this property on the given model instance. See http://docs.python.org/ref/descriptors.html for a description of the arguments to this class and what they mean.""" if model_instance is None: return self try: return getattr(model_instance, self._attr_name()) except __HOLE__: return None
AttributeError
dataset/ETHPy150Open AppScale/appscale/AppServer/google/appengine/ext/db/__init__.py/Property.__get__
4,924
def _populate_internal_entity(self, _entity_class=datastore.Entity): """Populates self._entity, saving its state to the datastore. After this method is called, calling is_saved() will return True. Returns: Populated self._entity """ self._entity = self._populate_entity(_entity_class=_entity_class) for prop in self.properties().values(): new_value = prop.get_updated_value_for_datastore(self) if new_value is not AUTO_UPDATE_UNCHANGED: self.__set_property(self._entity, prop.name, new_value) for attr in ('_key_name', '_key'): try: delattr(self, attr) except __HOLE__: pass return self._entity
AttributeError
dataset/ETHPy150Open AppScale/appscale/AppServer/google/appengine/ext/db/__init__.py/Model._populate_internal_entity
4,925
@classmethod def _load_entity_values(cls, entity): """Load dynamic properties from entity. Loads attributes which are not defined as part of the entity in to the model instance. Args: entity: Entity which contain values to search dyanmic properties for. """ entity_values = {} for prop in cls.properties().values(): if prop.name in entity: try: value = entity[prop.name] except __HOLE__: entity_values[prop.name] = [] else: if entity.is_projection(): value = prop.make_value_from_datastore_index_value(value) else: value = prop.make_value_from_datastore(value) entity_values[prop.name] = value return entity_values
KeyError
dataset/ETHPy150Open AppScale/appscale/AppServer/google/appengine/ext/db/__init__.py/Model._load_entity_values
4,926
def delete_async(models, **kwargs): """Asynchronous version of delete one or more Model instances. Identical to db.delete() except returns an asynchronous object. Call get_result() on the return value to block on the call. """ if isinstance(models, (basestring, Model, Key)): models = [models] else: try: models = iter(models) except __HOLE__: models = [models] keys = [_coerce_to_key(v) for v in models] return datastore.DeleteAsync(keys, **kwargs)
TypeError
dataset/ETHPy150Open AppScale/appscale/AppServer/google/appengine/ext/db/__init__.py/delete_async
4,927
def get(self, **kwargs): """Get first result from this. Beware: get() ignores the LIMIT clause on GQL queries. Args: kwargs: Any keyword arguments accepted by datastore_query.QueryOptions(). Returns: First result from running the query if there are any, else None. """ results = self.run(limit=1, **kwargs) try: return results.next() except __HOLE__: return None
StopIteration
dataset/ETHPy150Open AppScale/appscale/AppServer/google/appengine/ext/db/__init__.py/_BaseQuery.get
4,928
def validate(self, value): """Validate property. Returns: A valid value. Raises: BadValueError if property is not an instance of data_type. """ if value is not None and not isinstance(value, self.data_type): try: value = self.data_type(value) except __HOLE__, err: raise BadValueError('Property %s must be convertible ' 'to a %s instance (%s)' % (self.name, self.data_type.__name__, err)) value = super(UnindexedProperty, self).validate(value) if value is not None and not isinstance(value, self.data_type): raise BadValueError('Property %s must be a %s instance' % (self.name, self.data_type.__name__)) return value
TypeError
dataset/ETHPy150Open AppScale/appscale/AppServer/google/appengine/ext/db/__init__.py/UnindexedProperty.validate
4,929
def validate(self, value): """Validate ByteString property. Returns: A valid value. Raises: BadValueError if property is not instance of 'ByteString'. """ if value is not None and not isinstance(value, ByteString): try: value = ByteString(value) except __HOLE__, err: raise BadValueError('Property %s must be convertible ' 'to a ByteString instance (%s)' % (self.name, err)) value = super(ByteStringProperty, self).validate(value) if value is not None and not isinstance(value, ByteString): raise BadValueError('Property %s must be a ByteString instance' % self.name) if value is not None and len(value) > self.MAX_LENGTH: raise BadValueError( 'Property %s is %d bytes long; it must be %d or less.' % (self.name, len(value), self.MAX_LENGTH)) return value
TypeError
dataset/ETHPy150Open AppScale/appscale/AppServer/google/appengine/ext/db/__init__.py/ByteStringProperty.validate
4,930
def load_class(path): """ Load class from path. """ try: mod_name, klass_name = path.rsplit('.', 1) mod = import_module(mod_name) except AttributeError as e: raise ImproperlyConfigured('Error importing {0}: "{1}"'.format(mod_name, e)) try: klass = getattr(mod, klass_name) except __HOLE__: raise ImproperlyConfigured('Module "{0}" does not define a "{1}" class'.format(mod_name, klass_name)) return klass
AttributeError
dataset/ETHPy150Open niwinz/django-jinja/django_jinja/utils.py/load_class
4,931
def load_bytecode(self, f): """Loads bytecode from a file or file like object.""" # make sure the magic header is correct magic = f.read(len(bc_magic)) if magic != bc_magic: self.reset() return # the source code of the file changed, we need to reload checksum = pickle.load(f) if self.checksum != checksum: self.reset() return # if marshal_load fails then we need to reload try: self.code = marshal_load(f) except (EOFError, __HOLE__, TypeError): self.reset() return
ValueError
dataset/ETHPy150Open GoogleCloudPlatform/appengine-flask-skeleton/lib/jinja2/bccache.py/Bucket.load_bytecode
4,932
def _get_default_cache_dir(self): def _unsafe_dir(): raise RuntimeError('Cannot determine safe temp directory. You ' 'need to explicitly provide one.') tmpdir = tempfile.gettempdir() # On windows the temporary directory is used specific unless # explicitly forced otherwise. We can just use that. if os.name == 'nt': return tmpdir if not hasattr(os, 'getuid'): _unsafe_dir() dirname = '_jinja2-cache-%d' % os.getuid() actual_dir = os.path.join(tmpdir, dirname) try: os.mkdir(actual_dir, stat.S_IRWXU) except __HOLE__ as e: if e.errno != errno.EEXIST: raise try: os.chmod(actual_dir, stat.S_IRWXU) actual_dir_stat = os.lstat(actual_dir) if actual_dir_stat.st_uid != os.getuid() \ or not stat.S_ISDIR(actual_dir_stat.st_mode) \ or stat.S_IMODE(actual_dir_stat.st_mode) != stat.S_IRWXU: _unsafe_dir() except OSError as e: if e.errno != errno.EEXIST: raise actual_dir_stat = os.lstat(actual_dir) if actual_dir_stat.st_uid != os.getuid() \ or not stat.S_ISDIR(actual_dir_stat.st_mode) \ or stat.S_IMODE(actual_dir_stat.st_mode) != stat.S_IRWXU: _unsafe_dir() return actual_dir
OSError
dataset/ETHPy150Open GoogleCloudPlatform/appengine-flask-skeleton/lib/jinja2/bccache.py/FileSystemBytecodeCache._get_default_cache_dir
4,933
def clear(self): # imported lazily here because google app-engine doesn't support # write access on the file system and the function does not exist # normally. from os import remove files = fnmatch.filter(listdir(self.directory), self.pattern % '*') for filename in files: try: remove(path.join(self.directory, filename)) except __HOLE__: pass
OSError
dataset/ETHPy150Open GoogleCloudPlatform/appengine-flask-skeleton/lib/jinja2/bccache.py/FileSystemBytecodeCache.clear
4,934
def check_ils_booking(gadget_url_base): """ Checks the session and returns when the lab is free (in UTC). If None is returned, it means that there is no ongoing session so they can use the lab. """ ils_student_id = extract_ils_id(request.args.get('ils_student_url')) ils_teacher_id = extract_ils_id(request.args.get('ils_teacher_url')) # BOOKING_URL = url_for('.mock_golabz_booking_service', _external=True) BOOKING_URL = 'http://www.golabz.eu/rest/lab-booking/retrieve.json' try: r = requests.get(BOOKING_URL, headers = { 'Cache-Control': 'no-cache', 'Pragma': 'no-cache', }) r.raise_for_status() booking_slots = r.json() except Exception as e: traceback.print_exc() # We can not stop students using the labs if the service is temporarily down return None affected_ilss = [] oldest_affected_endtime = datetime.datetime.utcnow() for booking_slot in booking_slots: is_current_lab = False for lab_app in booking_slot.get('lab_apps', []): lab_url = lab_app.get('app_url', '') if lab_url.startswith(gadget_url_base): is_current_lab = True break if not is_current_lab: continue start_time_str = booking_slot.get('start_time', '2000-01-01T00:00:00') end_time_str = booking_slot.get('end_time', '2000-01-01T00:00:00') FORMAT = '%Y-%m-%dT%H:%M:%S' try: start_time = datetime.datetime.strptime(start_time_str, FORMAT) end_time = datetime.datetime.strptime(end_time_str, FORMAT) except __HOLE__: traceback.print_exc() continue now = datetime.datetime.utcnow() if end_time > now > start_time: ils_id = extract_ils_id(booking_slot.get('ils_url')) or 'does.not.exist' if end_time > oldest_affected_endtime: oldest_affected_endtime = end_time affected_ilss.append(ils_id) if affected_ilss: for affected_ils_id in affected_ilss: if affected_ils_id in (ils_student_id, ils_teacher_id): return None # If there are affected ILS, but we're not part of them, we're not authorized return oldest_affected_endtime else: # If there is no affected ILS, it doesn't matter if we're in the group or not return None
ValueError
dataset/ETHPy150Open gateway4labs/labmanager/labmanager/views/opensocial.py/check_ils_booking
4,935
def test_config_from_envvar(self): env = os.environ try: os.environ = {} app = flask.Flask(__name__) try: app.config.from_envvar('FOO_SETTINGS') except __HOLE__, e: self.assert_("'FOO_SETTINGS' is not set" in str(e)) else: self.assert_(0, 'expected exception') self.assert_(not app.config.from_envvar('FOO_SETTINGS', silent=True)) os.environ = {'FOO_SETTINGS': __file__.rsplit('.', 1)[0] + '.py'} self.assert_(app.config.from_envvar('FOO_SETTINGS')) self.common_object_test(app) finally: os.environ = env
RuntimeError
dataset/ETHPy150Open baseblack/ReproWeb/3rdParty/python/flask/testsuite/config.py/ConfigTestCase.test_config_from_envvar
4,936
def test_config_from_envvar_missing(self): env = os.environ try: os.environ = {'FOO_SETTINGS': 'missing.cfg'} try: app = flask.Flask(__name__) app.config.from_envvar('FOO_SETTINGS') except __HOLE__, e: msg = str(e) self.assert_(msg.startswith('[Errno 2] Unable to load configuration ' 'file (No such file or directory):')) self.assert_(msg.endswith("missing.cfg'")) else: self.fail('expected IOError') self.assertFalse(app.config.from_envvar('FOO_SETTINGS', silent=True)) finally: os.environ = env
IOError
dataset/ETHPy150Open baseblack/ReproWeb/3rdParty/python/flask/testsuite/config.py/ConfigTestCase.test_config_from_envvar_missing
4,937
def test_config_missing(self): app = flask.Flask(__name__) try: app.config.from_pyfile('missing.cfg') except __HOLE__, e: msg = str(e) self.assert_(msg.startswith('[Errno 2] Unable to load configuration ' 'file (No such file or directory):')) self.assert_(msg.endswith("missing.cfg'")) else: self.assert_(0, 'expected config') self.assert_(not app.config.from_pyfile('missing.cfg', silent=True))
IOError
dataset/ETHPy150Open baseblack/ReproWeb/3rdParty/python/flask/testsuite/config.py/ConfigTestCase.test_config_missing
4,938
def test_explicit_instance_paths(self): here = os.path.abspath(os.path.dirname(__file__)) try: flask.Flask(__name__, instance_path='instance') except __HOLE__, e: self.assert_('must be absolute' in str(e)) else: self.fail('Expected value error') app = flask.Flask(__name__, instance_path=here) self.assert_equal(app.instance_path, here)
ValueError
dataset/ETHPy150Open baseblack/ReproWeb/3rdParty/python/flask/testsuite/config.py/InstanceTestCase.test_explicit_instance_paths
4,939
def _get_impl(): """Delay import of rpc_backend until configuration is loaded.""" global _RPCIMPL if _RPCIMPL is None: try: _RPCIMPL = importutils.import_module(CONF.rpc_backend) except __HOLE__: # For backwards compatibility with older nova config. impl = CONF.rpc_backend.replace('nova.rpc', 'nova.openstack.common.rpc') _RPCIMPL = importutils.import_module(impl) return _RPCIMPL
ImportError
dataset/ETHPy150Open Havate/havate-openstack/proto-build/gui/horizon/Horizon_GUI/openstack_dashboard/openstack/common/rpc/__init__.py/_get_impl
4,940
@classmethod def __get_pp_ip(cls, addr_family, ip_string, which): try: packed = socket.inet_pton(addr_family, ip_string.decode('ascii')) return socket.inet_ntop(addr_family, packed) except (__HOLE__, socket.error): msg = 'Invalid proxy protocol {0} IP format'.format(which) raise AssertionError(msg)
UnicodeDecodeError
dataset/ETHPy150Open slimta/python-slimta/slimta/util/proxyproto.py/ProxyProtocolV1.__get_pp_ip
4,941
@classmethod def __get_pp_port(cls, port_string, which): try: port_num = int(port_string) except __HOLE__: msg = 'Invalid proxy protocol {0} port format'.format(which) raise AssertionError(msg) assert port_num >= 0 and port_num <= 65535, \ 'Proxy protocol {0} port out of range'.format(which) return port_num
ValueError
dataset/ETHPy150Open slimta/python-slimta/slimta/util/proxyproto.py/ProxyProtocolV1.__get_pp_port
4,942
def handle(self, sock, addr): """Intercepts calls to :meth:`~slimta.edge.EdgeServer.handle`, reads the proxy protocol header, and then resumes the original call. """ try: src_addr, _ = self.process_pp_v1(sock, b'') except __HOLE__ as exc: log.proxyproto_invalid(sock, exc) src_addr = invalid_pp_source_address else: log.proxyproto_success(sock, src_addr) return super(ProxyProtocolV1, self).handle(sock, src_addr)
AssertionError
dataset/ETHPy150Open slimta/python-slimta/slimta/util/proxyproto.py/ProxyProtocolV1.handle
4,943
def handle(self, sock, addr): """Intercepts calls to :meth:`~slimta.edge.EdgeServer.handle`, reads the proxy protocol header, and then resumes the original call. """ try: src_addr, _ = self.process_pp_v2(sock, b'') except LocalConnection: log.proxyproto_local(sock) return except __HOLE__ as exc: log.proxyproto_invalid(sock, exc) src_addr = invalid_pp_source_address else: log.proxyproto_success(sock, src_addr) return super(ProxyProtocolV2, self).handle(sock, src_addr)
AssertionError
dataset/ETHPy150Open slimta/python-slimta/slimta/util/proxyproto.py/ProxyProtocolV2.handle
4,944
def handle(self, sock, addr): """Intercepts calls to :meth:`~slimta.edge.EdgeServer.handle`, reads the proxy protocol header, and then resumes the original call. """ try: initial = self.__read_pp_initial(sock) if initial.startswith(b'PROXY '): src_addr, _ = ProxyProtocolV1.process_pp_v1(sock, initial) elif initial == b'\r\n\r\n\x00\r\nQ': src_addr, _ = ProxyProtocolV2.process_pp_v2(sock, initial) else: raise AssertionError('Invalid proxy protocol signature') except LocalConnection: log.proxyproto_local(sock) return except __HOLE__ as exc: log.proxyproto_invalid(sock, exc) src_addr = invalid_pp_source_address else: log.proxyproto_success(sock, src_addr) return super(ProxyProtocol, self).handle(sock, src_addr) # vim:et:fdm=marker:sts=4:sw=4:ts=4
AssertionError
dataset/ETHPy150Open slimta/python-slimta/slimta/util/proxyproto.py/ProxyProtocol.handle
4,945
def prompt_for_yn(prompt_str, default=None): if default == True: prompt = prompt_str + ' [Y/n]: ' elif default == False: prompt = prompt_str + ' [y/N]: ' else: prompt = prompt_str + ' [y/n]: ' while True: try: value = input(prompt) except __HOLE__: print('') exit(1) except EOFError: print('') exit(1) if value != '': if value.lower()[0] == 'y': return True elif value.lower()[0] == 'n': return False else: print('Error: unrecognized response') elif default is not None: return default
KeyboardInterrupt
dataset/ETHPy150Open dnanexus/dx-toolkit/src/python/dxpy/cli/__init__.py/prompt_for_yn
4,946
def access_token(self, token): """Return request for access token value""" try: return self.get_querystring(self.ACCESS_TOKEN_URL, auth=self.oauth_auth(token)) except __HOLE__ as err: # Evernote returns a 401 error when AuthCanceled if err.response.status_code == 401: raise AuthCanceled(self, response=err.response) else: raise
HTTPError
dataset/ETHPy150Open omab/python-social-auth/social/backends/evernote.py/EvernoteOAuth.access_token
4,947
def remove_file(self, filename): """Remove a file, noop if file does not exist.""" # Unlike os.remove, if the file does not exist, # then this method does nothing. try: os.remove(filename) except __HOLE__: pass
OSError
dataset/ETHPy150Open boto/boto3/boto3/s3/transfer.py/OSUtils.remove_file
4,948
def __getattr__(self, attr): try: return self.__getitem__(attr) except __HOLE__: return super(EnhancedDictionary, self).__getattr__(attr)
KeyError
dataset/ETHPy150Open kivy/kivy/kivy/input/motionevent.py/EnhancedDictionary.__getattr__
4,949
def read_parameter_file(path, name): try: raw = misc.read_file(path, name + filename_extension) except __HOLE__: raw = "{}" cooked = ast.literal_eval(raw) return cooked
IOError
dataset/ETHPy150Open laforest/Octavo/Generator/Misc/parameters_misc.py/read_parameter_file
4,950
def parse_cmdline(entries): parameters = {} if len(entries) == 0: return parameters if entries[0] == "-f" and filename_extension in entries[1]: path, name = os.path.split(entries[1]) name = name.replace(filename_extension, '') parameters = read_parameter_file(path, name) del entries[0:2] for entry in entries: parts = entry.split('=') assert len(parts) == 2, "Incorrect parameter formating for {0}. Use 'key=value'.".format(entry) key, value = tuple(parts) try: value = ast.literal_eval(value) except __HOLE__: if not (value == "ON" or value == "OFF"): raise ValueError, value parameters.update({key:value}) return parameters
ValueError
dataset/ETHPy150Open laforest/Octavo/Generator/Misc/parameters_misc.py/parse_cmdline
4,951
def test_connection(self): class DocA(Document): structure = { "doc_a":{'foo':int}, } self.connection.register([DocA]) assertion = True try: DocA.connection except AttributeError: assertion = True assert assertion try: DocA.db except __HOLE__: assertion = True assert assertion try: DocA.collection except AttributeError: assertion = True assert assertion assert self.col.DocA.connection == Connection("localhost", 27017) assert self.col.DocA.collection == Connection("localhost", 27017)['test']['mongokit'] assert self.col.DocA.db == Connection("localhost", 27017)['test']
AttributeError
dataset/ETHPy150Open namlook/mongokit/tests/test_api.py/ApiTestCase.test_connection
4,952
def test_rewind(self): class MyDoc(Document): structure = { 'foo':int, } self.connection.register([MyDoc]) for i in range(10): doc = self.col.MyDoc() doc['foo'] = i doc.save() cur = self.col.MyDoc.find() for i in cur: assert isinstance(i, MyDoc), type(MyDoc) try: cur.next() except __HOLE__: pass cur.rewind() for i in cur: assert isinstance(i, MyDoc), type(MyDoc) for i in cur.rewind(): assert isinstance(i, MyDoc), type(MyDoc)
StopIteration
dataset/ETHPy150Open namlook/mongokit/tests/test_api.py/ApiTestCase.test_rewind
4,953
def test_database_name_filled(self): failed = False @self.connection.register class MyDoc(Document): __database__ = 'mydoc' structure = { 'foo':int, } try: doc = self.connection.MyDoc() except __HOLE__, e: failed = True self.assertEqual(str(e), 'MyDoc: __collection__ attribute not ' 'found. You cannot specify the `__database__` attribute ' 'without the `__collection__` attribute') self.assertEqual(failed, True) @self.connection.register class MyDoc(Document): __database__ = 'test' __collection__ = 'mydoc' structure = { 'foo':int, } # test directly from a connection mydoc = self.connection.MyDoc() mydoc['foo'] = 3 mydoc.save() self.assertEqual(mydoc.collection.name, 'mydoc') self.assertEqual(mydoc.collection.database.name, 'test') self.assertEqual(self.col.MyDoc.find_one(), None) raw_doc = self.connection.MyDoc.find_one() self.assertEqual(raw_doc['foo'], 3) self.assertEqual(raw_doc, mydoc) assert isinstance(raw_doc, MyDoc) # test directly from a database mydoc = self.connection.othertest.MyDoc() mydoc['foo'] = 3 mydoc.save() self.assertEqual(mydoc.collection.name, 'mydoc') self.assertEqual(mydoc.collection.database.name, 'othertest') self.assertEqual(self.col.MyDoc.find_one(), None) raw_doc = self.connection.othertest.MyDoc.find_one() self.assertEqual(raw_doc['foo'], 3) self.assertEqual(raw_doc, mydoc) assert isinstance(raw_doc, MyDoc) # and still can use it via a collection mydoc = self.col.MyDoc() mydoc['foo'] = 3 mydoc.save() self.assertEqual(mydoc.collection.name, 'mongokit') self.assertEqual(mydoc.collection.database.name, 'test') raw_doc = self.col.MyDoc.find_one() self.assertEqual(raw_doc['foo'], 3) self.assertEqual(raw_doc, mydoc) assert isinstance(raw_doc, MyDoc)
AttributeError
dataset/ETHPy150Open namlook/mongokit/tests/test_api.py/ApiTestCase.test_database_name_filled
4,954
def test_unwrapped_cursor(self): self.assertEqual(self.col.count(), 0) doc_id = self.col.save({}, safe=True) self.assertEqual(self.col.count(), 1) try: self.col.find(_id=doc_id)[0] except __HOLE__: self.fail("Cursor.__getitem__ raised TypeError unexpectedly!")
TypeError
dataset/ETHPy150Open namlook/mongokit/tests/test_api.py/ApiTestCase.test_unwrapped_cursor
4,955
def force_unicode(s): try: return unicode(s) except __HOLE__: return str(s)
NameError
dataset/ETHPy150Open ffunenga/pipgh/pipgh/tools.py/force_unicode
4,956
def __init__(self, node, level=0): self.level = level def is_str(s): try: return type(s) == unicode except __HOLE__: return type(s) == str _test = lambda i: (type(i[1]) == dict or (type(i[1]) != dict and not normalize(i[1]).startswith('http'))) self.children = [n for n in node.items()] # if _test(n)] self.children.sort(key=lambda i: i[0])
NameError
dataset/ETHPy150Open ffunenga/pipgh/pipgh/tools.py/ShowNode.__init__
4,957
def __enter__(self): try: os.mkdir(self.path) except __HOLE__ as e: if e.errno != 17: # File exists raise shutil.rmtree(self.path) os.mkdir(self.path) finally: os.chdir(self.path) return self
OSError
dataset/ETHPy150Open ffunenga/pipgh/pipgh/tools.py/TempDirContext.__enter__
4,958
def __init__(self, inner_node): """ Instantiated with a inner_node as an argument This node comes from the from the do_include function """ self.inner_node = inner_node # attempting to get the tmplate name # depends whether it is an IncludeNode or ConstantIncludeNode try: self.template_name = inner_node.template_name except __HOLE__: self.template_name = inner_node.template.name
AttributeError
dataset/ETHPy150Open haystack/eyebrowse-server/common/templatetags/jstemplate.py/JsTemplateBaseNode.__init__
4,959
def _limit_inf(expr, n): try: return Limit(expr, n, S.Infinity).doit(deep=False, sequence=False) except (__HOLE__, PoleError): return None
NotImplementedError
dataset/ETHPy150Open sympy/sympy/sympy/series/limitseq.py/_limit_inf
4,960
def merge(dict_1, dict_2, func=lambda x, y: x + y): """doc""" res = dict_1.copy() # "= dict(d1)" for lists of tuples for key, val in dict_2.iteritems(): # ".. in d2" for lists of tuple try: res[key] = func(res[key], val) except __HOLE__: res[key] = val return res
KeyError
dataset/ETHPy150Open columbia/libtrack/libtrack/parser/src/telesphorus/helpers/dict_utils.py/merge
4,961
def isDBPort(host, port, db, timeout=10): if host in (RDFTURTLEFILE_HOSTNAME, RDFXMLFILE_HOSTNAME): return True # determine if postgres port t = 2 while t < timeout: try: conn = urllib.request.urlopen("http://{0}:{1}/{2}/status".format(host, port or '80', db)) return True # success but doesn't need password except __HOLE__: return False # success, this is really a postgres socket, wants user name except URLError: return False # something is there but not postgres except socket.timeout: t = t + 2 # relax - try again with longer timeout return False # deferred rdflib import
HTTPError
dataset/ETHPy150Open Arelle/Arelle/arelle/plugin/xbrlDB/XbrlSemanticRdfDB.py/isDBPort
4,962
def __init__(self, modelXbrl, user, password, host, port, database, timeout): try: initRdflibNamespaces() except __HOLE__: raise XRDBException("xrdfDB:MissingRdflib", _("Rdflib is not installed or is not available in this build")) self.modelXbrl = modelXbrl self.disclosureSystem = modelXbrl.modelManager.disclosureSystem #self.conn = RexProConnection(host, int(port or '8182'), (database or 'emptygraph'), # user=user, password=password) self.isRdfTurtleFile = host == RDFTURTLEFILE_HOSTNAME self.isRdfXmlFile = host == RDFXMLFILE_HOSTNAME if self.isRdfTurtleFile or self.isRdfXmlFile: self.turtleFile = database else: connectionUrl = "http://{0}:{1}".format(host, port or '80') self.url = connectionUrl if database: self.url += '/' + database # Create an OpenerDirector with support for Basic HTTP Authentication... auth_handler = urllib.request.HTTPBasicAuthHandler() if user: auth_handler.add_password(realm=None, uri=connectionUrl, user=user, passwd=password) self.conn = urllib.request.build_opener(auth_handler) self.timeout = timeout or 60 self.verticePropTypes = {}
ImportError
dataset/ETHPy150Open Arelle/Arelle/arelle/plugin/xbrlDB/XbrlSemanticRdfDB.py/XbrlSemanticRdfDatabaseConnection.__init__
4,963
def execute(self, activity, graph=None, query=None): if graph is not None: headers = {'User-agent': 'Arelle/1.0', 'Accept': 'application/sparql-results+json', 'Content-Type': "text/turtle; charset='UTF-8'"} data = graph.serialize(format='pretty-xml' if self.isRdfXmlFile else 'turtle', encoding='utf=8') elif query is not None: headers = {'User-agent': 'Arelle/1.0', 'Accept': 'application/sparql-results+json'} data = ("query=" + query).encode('utf-8') else: return None # turtle may be mixture of line strings and strings with \n-separated lines if TRACERDFFILE: with io.open(TRACERDFFILE, "ab") as fh: fh.write(b"\n\n>>> sent: \n") fh.write(data) if (self.isRdfTurtleFile or self.isRdfXmlFile) and data is not None: with io.open(self.turtleFile, "ab") as fh: fh.write(data) return None if graph is not None or query is not None: url = self.url + "/sparql" request = urllib.request.Request(url, data=data, headers=headers) try: with self.conn.open(request, timeout=self.timeout) as fp: results = fp.read().decode('utf-8') try: results = json.loads(results) except __HOLE__: pass # leave results as string except HTTPError as err: results = err.fp.read().decode('utf-8') if TRACERDFFILE: with io.open(TRACERDFFILE, "a", encoding='utf-8') as fh: fh.write("\n\n>>> received: \n{0}".format(str(results))) if isinstance(results, str) and query is not None: parser = etree.HTMLParser() htmlDoc = etree.parse(io.StringIO(results), parser) body = htmlDoc.find("//body") if body is not None: error = "".join(text for text in body.itertext()) else: error = results raise XRDBException("xrdfDB:DatabaseError", _("%(activity)s not successful: %(error)s"), activity=activity, error=error) return results
ValueError
dataset/ETHPy150Open Arelle/Arelle/arelle/plugin/xbrlDB/XbrlSemanticRdfDB.py/XbrlSemanticRdfDatabaseConnection.execute
4,964
def identifyPreexistingDocuments(self): self.existingDocumentUris = set() if not (self.isRdfTurtleFile or self.isRdfXmlFile): docFilters = [] for modelDocument in self.modelXbrl.urlDocs.values(): if modelDocument.type == Type.SCHEMA: docFilters.append('STR(?doc) = "{}"'.format(UrlUtil.ensureUrl(modelDocument.uri))) results = self.execute( "select", query=""" PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX DTS: <http://xbrl.org/2013/rdf/DTS/> SELECT distinct ?doc WHERE { ?doc rdf:type DTS:Document FILTER( """ + '\n|| '.join(docFilters) + ") .}") try: for result in results['results']['bindings']: doc = result['doc'] if doc.get('type') == 'uri': self.existingDocumentUris.add(doc['value']) except __HOLE__: pass # no existingDocumentUris
KeyError
dataset/ETHPy150Open Arelle/Arelle/arelle/plugin/xbrlDB/XbrlSemanticRdfDB.py/XbrlSemanticRdfDatabaseConnection.identifyPreexistingDocuments
4,965
@classmethod def _create_ip_ranges_on_notation(cls, instance): """Create IP-address ranges basing on 'notation' field of 'meta' field :param instance: NetworkGroup instance :type instance: models.NetworkGroup :return: None """ notation = instance.meta.get("notation") if notation: try: if notation == 'cidr': cls._update_range_from_cidr( instance, IPNetwork(instance.cidr).cidr, instance.meta.get('use_gateway')) elif notation == 'ip_ranges' and instance.meta.get("ip_range"): cls._set_ip_ranges(instance, [instance.meta["ip_range"]]) else: raise errors.CannotCreate() except ( errors.CannotCreate, IndexError, __HOLE__ ): raise errors.CannotCreate( "IPAddrRange object cannot be created for network '{0}' " "with notation='{1}', ip_range='{2}'".format( instance.name, instance.meta.get('notation'), instance.meta.get('ip_range')) )
TypeError
dataset/ETHPy150Open openstack/fuel-web/nailgun/nailgun/extensions/network_manager/objects/network_group.py/NetworkGroup._create_ip_ranges_on_notation
4,966
def __call__(self, text, append_newline=1): if not self.print_it: return if append_newline: text = text + '\n' try: sys.stdout.write(unicode(text)) except __HOLE__: # Stdout might be connected to a pipe that has been closed # by now. The most likely reason for the pipe being closed # is that the user has press ctrl-c. It this is the case, # then SCons is currently shutdown. We therefore ignore # IOError's here so that SCons can continue and shutdown # properly so that the .sconsign is correctly written # before SCons exits. pass
IOError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/DisplayEngine.__call__
4,967
def to_String_for_signature(obj, to_String_for_subst=to_String_for_subst, AttributeError=AttributeError): try: f = obj.for_signature except __HOLE__: return to_String_for_subst(obj) else: return f() # The SCons "semi-deep" copy. # # This makes separate copies of lists (including UserList objects) # dictionaries (including UserDict objects) and tuples, but just copies # references to anything else it finds. # # A special case is any object that has a __semi_deepcopy__() method, # which we invoke to create the copy. Currently only used by # BuilderDict to actually prevent the copy operation (as invalid on that object) # # The dispatch table approach used here is a direct rip-off from the # normal Python copy module.
AttributeError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/to_String_for_signature
4,968
def WhereIs(file, path=None, pathext=None, reject=[]): if path is None: try: path = os.environ['PATH'] except __HOLE__: return None if is_String(path): path = path.split(os.pathsep) if pathext is None: try: pathext = os.environ['PATHEXT'] except KeyError: pathext = '.COM;.EXE;.BAT;.CMD' if is_String(pathext): pathext = pathext.split(os.pathsep) for ext in pathext: if ext.lower() == file[-len(ext):].lower(): pathext = [''] break if not is_List(reject) and not is_Tuple(reject): reject = [reject] for dir in path: f = os.path.join(dir, file) for ext in pathext: fext = f + ext if os.path.isfile(fext): try: reject.index(fext) except ValueError: return os.path.normpath(fext) continue return None
KeyError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/WhereIs
4,969
def WhereIs(file, path=None, pathext=None, reject=[]): if path is None: try: path = os.environ['PATH'] except KeyError: return None if is_String(path): path = path.split(os.pathsep) if pathext is None: pathext = ['.exe', '.cmd'] for ext in pathext: if ext.lower() == file[-len(ext):].lower(): pathext = [''] break if not is_List(reject) and not is_Tuple(reject): reject = [reject] for dir in path: f = os.path.join(dir, file) for ext in pathext: fext = f + ext if os.path.isfile(fext): try: reject.index(fext) except __HOLE__: return os.path.normpath(fext) continue return None
ValueError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/WhereIs
4,970
def WhereIs(file, path=None, pathext=None, reject=[]): import stat if path is None: try: path = os.environ['PATH'] except __HOLE__: return None if is_String(path): path = path.split(os.pathsep) if not is_List(reject) and not is_Tuple(reject): reject = [reject] for d in path: f = os.path.join(d, file) if os.path.isfile(f): try: st = os.stat(f) except OSError: # os.stat() raises OSError, not IOError if the file # doesn't exist, so in this case we let IOError get # raised so as to not mask possibly serious disk or # network issues. continue if stat.S_IMODE(st[stat.ST_MODE]) & 0111: try: reject.index(f) except ValueError: return os.path.normpath(f) continue return None
KeyError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/WhereIs
4,971
def popitem(self): try: key = self._keys[-1] except __HOLE__: raise KeyError('dictionary is empty') val = self[key] del self[key] return (key, val)
IndexError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/OrderedDict.popitem
4,972
def __call__(self, env, source, ext=None): if ext is None: try: ext = source[0].suffix except __HOLE__: ext = "" try: return self[ext] except KeyError: # Try to perform Environment substitution on the keys of # the dictionary before giving up. s_dict = {} for (k,v) in self.items(): if k is not None: s_k = env.subst(k) if s_k in s_dict: # We only raise an error when variables point # to the same suffix. If one suffix is literal # and a variable suffix contains this literal, # the literal wins and we don't raise an error. raise KeyError(s_dict[s_k][0], k, s_k) s_dict[s_k] = (k,v) try: return s_dict[ext][1] except KeyError: try: return self[None] except KeyError: return None
IndexError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/Selector.__call__
4,973
def unique(s): """Return a list of the elements in s, but without duplicates. For example, unique([1,2,3,1,2,3]) is some permutation of [1,2,3], unique("abcabc") some permutation of ["a", "b", "c"], and unique(([1, 2], [2, 3], [1, 2])) some permutation of [[2, 3], [1, 2]]. For best speed, all sequence elements should be hashable. Then unique() will usually work in linear time. If not possible, the sequence elements should enjoy a total ordering, and if list(s).sort() doesn't raise TypeError it's assumed that they do enjoy a total ordering. Then unique() will usually work in O(N*log2(N)) time. If that's not possible either, the sequence elements must support equality-testing. Then unique() will usually work in quadratic time. """ n = len(s) if n == 0: return [] # Try using a dict first, as that's the fastest and will usually # work. If it doesn't work, it will usually fail quickly, so it # usually doesn't cost much to *try* it. It requires that all the # sequence elements be hashable, and support equality comparison. u = {} try: for x in s: u[x] = 1 except TypeError: pass # move on to the next method else: return list(u.keys()) del u # We can't hash all the elements. Second fastest is to sort, # which brings the equal elements together; then duplicates are # easy to weed out in a single pass. # NOTE: Python's list.sort() was designed to be efficient in the # presence of many duplicate elements. This isn't true of all # sort functions in all languages or libraries, so this approach # is more effective in Python than it may be elsewhere. try: t = sorted(s) except __HOLE__: pass # move on to the next method else: assert n > 0 last = t[0] lasti = i = 1 while i < n: if t[i] != last: t[lasti] = last = t[i] lasti = lasti + 1 i = i + 1 return t[:lasti] del t # Brute force is all that's left. u = [] for x in s: if x not in u: u.append(x) return u # From Alex Martelli, # http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/52560 # ASPN: Python Cookbook: Remove duplicates from a sequence # First comment, dated 2001/10/13. # (Also in the printed Python Cookbook.)
TypeError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/unique
4,974
def write(self, arg): try: self.file.write(arg) self.file.flush() except __HOLE__: # Stdout might be connected to a pipe that has been closed # by now. The most likely reason for the pipe being closed # is that the user has press ctrl-c. It this is the case, # then SCons is currently shutdown. We therefore ignore # IOError's here so that SCons can continue and shutdown # properly so that the .sconsign is correctly written # before SCons exits. pass
IOError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/Unbuffered.write
4,975
def silent_intern(x): """ Perform sys.intern() on the passed argument and return the result. If the input is ineligible (e.g. a unicode string) the original argument is returned and no exception is thrown. """ try: return sys.intern(x) except __HOLE__: return x # From Dinu C. Gherman, # Python Cookbook, second edition, recipe 6.17, p. 277. # Also: # http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/68205 # ASPN: Python Cookbook: Null Object Design Pattern #TODO??? class Null(object):
TypeError
dataset/ETHPy150Open kayhayen/Nuitka/nuitka/build/inline_copy/lib/scons-2.3.2/SCons/Util.py/silent_intern
4,976
def sync(self): '''Copy disk image to path, regenerate OVF descriptor''' self.copy_disk() self.ovf_descriptor = self.new_ovf_descriptor() ovf_xml = self.ovf_descriptor.generate_ovf_xml() try: os.makedirs(os.path.dirname(self.ovf_path)) except __HOLE__, e: if "File exists" not in e: raise ovf_xml.write(self.ovf_path)
OSError
dataset/ETHPy150Open redhat-imaging/imagefactory/imagefactory_plugins/ovfcommon/ovfcommon.py/OVFPackage.sync
4,977
def get_jellyroll_items(parser, token): """ Load jellyroll ``Item`` objects into the context.In the simplest mode, the most recent N items will be returned. :: {# Fetch 10 most recent items #} {% get_jellyroll_items limit 10 as items %} Newer items will be first in the list (i.e. ordered by timstamp descending) unless you ask for them to be reversed:: {# Fetch the 10 earliest items #} {% get_jellyroll_items limit 10 reversed as items %} You can also fetch items between two given dates:: {% get_jellyroll_items between "2007-01-01" and "2007-01-31" as items %} Dates can be in any format understood by ``dateutil.parser``, and can of course be variables. Items must be limited in some way; you must either pass ``limit`` or ``between``. Dates can also be the magic strings "now" or "today":: {% get_jellyroll_items between "2007-01-01" and "today" as items %} You can also limit items by type:: {# Fetch the 10 most recent videos and photos #} {% get_jellyroll_items oftype video oftype photo limit 10 as items %} ``oftype`` can be given as many times as you like; only those types will be returned. The arguments to ``oftype`` are the lowercased class names of jellyroll'd items. You can similarly exclude types using ``excludetype``:: {# Fetch the 10 most items that are not videos #} {% get_jellyroll_items excludetype video limit 10 as items %} You can give ``excludetype`` as many times as you like, but it is an error to use both ``oftype`` and ``excludetype`` in the same tag invocation. """ # Parse out the arguments bits = token.split_contents() tagname = bits[0] bits = iter(bits[1:]) args = {} for bit in bits: try: if bit == "limit": try: args["limit"] = int(bits.next()) except ValueError: raise template.TemplateSyntaxError("%r tag: 'limit' requires an integer argument" % tagname) elif bit == "between": args["start"] = bits.next() and_ = bits.next() args["end"] = bits.next() if and_ != "and": raise template.TemplateSyntaxError("%r tag: expected 'and' in 'between' clause, but got %r" % (tagname, and_)) elif bit == "oftype": args.setdefault("oftypes", []).append(bits.next()) elif bit == "excludetype": args.setdefault("excludetypes", []).append(bits.next()) elif bit == "reversed": args["reversed"] = True elif bit == "as": args["asvar"] = bits.next() else: raise template.TemplateSyntaxError("%r tag: unknown argument: %r" % (tagname, bit)) except StopIteration: raise template.TemplateSyntaxError("%r tag: an out of arguments when parsing %r clause" % (tagname, bit)) # Make sure "as" was given if "asvar" not in args: raise template.TemplateSyntaxError("%r tag: missing 'as'" % tagname) # Either "limit" or "between" has to be specified if "limit" not in args and ("start" not in args or "end" not in args): raise template.TemplateSyntaxError("%r tag: 'limit' or a full 'between' clause is required" % tagname) # It's an error to have both "oftype" and "excludetype" if "oftype" in args and "excludetype" in args: raise template.TemplateSyntaxError("%r tag: can't handle both 'oftype' and 'excludetype'" % tagname) # Each of the "oftype" and "excludetype" arguments has be a valid model for arg in ("oftypes", "excludetypes"): if arg in args: model_list = [] for name in args[arg]: try: model_list.append(Item.objects.models_by_name[name]) except __HOLE__: raise template.TemplateSyntaxError("%r tag: invalid model name: %r" % (tagname, name)) args[arg] = model_list return GetJellyrollItemsNode(**args)
KeyError
dataset/ETHPy150Open jacobian-archive/jellyroll/src/jellyroll/templatetags/jellyroll.py/get_jellyroll_items
4,978
def resolve_date(self, d, context): """Resolve start/end, handling literals""" try: d = template.resolve_variable(d, context) except template.VariableDoesNotExist: return None # Handle date objects if isinstance(d, (datetime.date, datetime.datetime)): return d # Assume literals or strings if d == "now": return datetime.datetime.now() if d == "today": return datetime.date.today() try: return dateutil.parser.parse(d) except __HOLE__: return None
ValueError
dataset/ETHPy150Open jacobian-archive/jellyroll/src/jellyroll/templatetags/jellyroll.py/GetJellyrollItemsNode.resolve_date
4,979
def delete(self): """Removes from disk atomically, can not then subsequently call read(), write() or addChildren() """ os.remove(self.getJobFileName()) #This is the atomic operation, if this file is not present the job is deleted. dirToRemove = self.jobDir while 1: head, tail = os.path.split(dirToRemove) if re.match("t[0-9]+$", tail): command = "rm -rf %s" % dirToRemove else: command = "rm -rf %s/*" % dirToRemove #We're at the root try: system(command) except __HOLE__: pass #This is not a big deal, as we expect collisions dirToRemove = head try: if len(os.listdir(dirToRemove)) != 0: break except os.error: #In case stuff went wrong, but as this is not critical we let it slide break
RuntimeError
dataset/ETHPy150Open benedictpaten/jobTree/src/job.py/Job.delete
4,980
def main(nick=None, user=None, dest=None): if nick is None: # pragma: no cover try: nick = sys.argv[1] except __HOLE__: print(__doc__) sys.exit(-1) dest = dest or os.getcwd() user = user or os.environ.get('USER', 'mynick') config = dict(nick=nick, user=user, hash='${hash}') data = get_template('config.ini', **config) with open(os.path.join(dest, 'config.ini'), 'w') as fd: fd.write(data) data = get_template('plugin.py', **config) filename = os.path.join(dest, '{nick}_plugin.py'.format(**config)) with open(filename, 'w') as fd: fd.write(data)
IndexError
dataset/ETHPy150Open gawel/irc3/irc3/template/__init__.py/main
4,981
def _setConfigParamsInPDU(self, pdu, kwargs): """Check for PDU's mandatory parameters and try to set remaining unset params (in kwargs) from the config default values """ for param in pdu.mandatoryParams: if param not in kwargs: try: pdu.params[param] = getattr(self.config, param) except __HOLE__: pdu.params[param] = None return pdu
AttributeError
dataset/ETHPy150Open jookies/jasmin/jasmin/protocols/smpp/operations.py/SMPPOperationFactory._setConfigParamsInPDU
4,982
def SubmitSM(self, short_message, data_coding=0, **kwargs): """Depending on the short_message length, this method will return a classical SubmitSM or a serie of linked SubmitSMs (parted message) """ kwargs['short_message'] = short_message kwargs['data_coding'] = data_coding # Possible data_coding values : 0,1,2,3,4,5,6,7,8,9,10,13,14 # Set the max short message length depending on the # coding (7, 8 or 16 bits) if kwargs['data_coding'] in [3, 6, 7, 10]: # 8 bit coding bits = 8 maxSmLength = 140 slicedMaxSmLength = maxSmLength - 6 elif kwargs['data_coding'] in [2, 4, 5, 8, 9, 13, 14]: # 16 bit coding bits = 16 maxSmLength = 70 slicedMaxSmLength = maxSmLength - 3 else: # 7 bit coding is the default # for data_coding in [0, 1] or any other invalid value bits = 7 maxSmLength = 160 slicedMaxSmLength = 153 longMessage = kwargs['short_message'] if bits == 16: smLength = len(kwargs['short_message']) / 2 else: smLength = len(kwargs['short_message']) # if SM is longer than maxSmLength, build multiple SubmitSMs # and link them if smLength > maxSmLength: total_segments = int(math.ceil(smLength / float(slicedMaxSmLength))) # Obey to configured longContentMaxParts if total_segments > self.long_content_max_parts: total_segments = self.long_content_max_parts msg_ref_num = self.claimLongMsgRefNum() for i in range(total_segments): segment_seqnum = i+1 # Keep in memory previous PDU in order to set nextPdu in it later try: tmpPdu previousPdu = tmpPdu except __HOLE__: previousPdu = None if bits == 16: kwargs['short_message'] = longMessage[slicedMaxSmLength*i*2:slicedMaxSmLength*(i+1)*2] else: kwargs['short_message'] = longMessage[slicedMaxSmLength*i:slicedMaxSmLength*(i+1)] tmpPdu = self._setConfigParamsInPDU(SubmitSM(**kwargs), kwargs) if self.long_content_split == 'sar': # Slice short_message and create the PDU using SAR options tmpPdu.params['sar_total_segments'] = total_segments tmpPdu.params['sar_segment_seqnum'] = segment_seqnum tmpPdu.params['sar_msg_ref_num'] = msg_ref_num elif self.long_content_split == 'udh': # Slice short_message and create the PDU using UDH options tmpPdu.params['esm_class'] = EsmClass( EsmClassMode.DEFAULT, EsmClassType.DEFAULT, [EsmClassGsmFeatures.UDHI_INDICATOR_SET]) if segment_seqnum < total_segments: tmpPdu.params['more_messages_to_send'] = MoreMessagesToSend.MORE_MESSAGES else: tmpPdu.params['more_messages_to_send'] = MoreMessagesToSend.NO_MORE_MESSAGES # UDH composition: udh = [] # Length of User Data Header udh.append(struct.pack('!B', 5)) # Information Element Identifier, equal to 00 # (Concatenated short messages, 8-bit reference number) udh.append(struct.pack('!B', 0)) # Length of the header, excluding the first two fields; equal to 03 udh.append(struct.pack('!B', 3)) udh.append(struct.pack('!B', msg_ref_num)) udh.append(struct.pack('!B', total_segments)) udh.append(struct.pack('!B', segment_seqnum)) tmpPdu.params['short_message'] = ''.join(udh) + kwargs['short_message'] # - The first PDU is the one we return back # - sar_msg_ref_num takes the seqnum of the initial submit_sm if i == 0: pdu = tmpPdu # PDU chaining if previousPdu is not None: previousPdu.nextPdu = tmpPdu else: pdu = self._setConfigParamsInPDU(SubmitSM(**kwargs), kwargs) return pdu
NameError
dataset/ETHPy150Open jookies/jasmin/jasmin/protocols/smpp/operations.py/SMPPOperationFactory.SubmitSM
4,983
def _remove_file(filepath): if not filepath: return try: os.remove(filepath) except __HOLE__: pass
OSError
dataset/ETHPy150Open openstack/cloudbase-init/cloudbaseinit/tests/plugins/common/test_execcmd.py/_remove_file
4,984
def __getitem__(self, index): while 1: try: file = self.files[self.index] self.index = self.index + 1 except __HOLE__: self.index = 0 self.directory = self.stack.pop() self.files = os.listdir(self.directory) else: fullname = os.path.join(self.directory, file) if os.path.isdir(fullname) and not os.path.islink(fullname): self.stack.append(fullname) if not (file.startswith('.') or file.startswith('#') or file.endswith('~')) \ and fnmatch.fnmatch(file, self.pattern): return fullname
IndexError
dataset/ETHPy150Open uwdata/termite-data-server/web2py/extras/build_web2py/setup_app.py/reglob.__getitem__
4,985
def send(self, data_to_send): """ Immediately sends the data passed in to :func:`service_endpoint_uri`. If the service request fails, the passed in items are pushed back to the :func:`queue`. Args: data_to_send (Array): an array of :class:`contracts.Envelope` objects to send to the service. """ request_payload = json.dumps([ a.write() for a in data_to_send ]) request = HTTPClient.Request(self._service_endpoint_uri, bytearray(request_payload, 'utf-8'), { 'Accept': 'application/json', 'Content-Type' : 'application/json; charset=utf-8' }) try: response = HTTPClient.urlopen(request) status_code = response.getcode() if 200 <= status_code < 300: return except __HOLE__ as e: if e.getcode() == 400: return except Exception as e: pass # Add our unsent data back on to the queue for data in data_to_send: self._queue.put(data)
HTTPError
dataset/ETHPy150Open Microsoft/ApplicationInsights-Python/applicationinsights/channel/SenderBase.py/SenderBase.send
4,986
@background.setter def background(self, value): self._color = value self.refresh() # propagate changes to every clone if self.editor: for clone in self.editor.clones: try: clone.modes.get(self.__class__).background = value except __HOLE__: # this should never happen since we're working with clones pass
KeyError
dataset/ETHPy150Open OpenCobolIDE/OpenCobolIDE/open_cobol_ide/extlibs/pyqode/core/modes/line_highlighter.py/LineHighlighterMode.background
4,987
def convert_datestr(datestr): """Converts dates like `12/31/2010` into datetime objects.""" try: return dateutil.parser.parse(datestr) except (__HOLE__, ValueError, AttributeError): return None
TypeError
dataset/ETHPy150Open buildingenergy/buildingenergy-platform/seed/utils/time.py/convert_datestr
4,988
def patch_translation_field(self, db_field, field, **kwargs): if db_field.name in self.trans_opts.fields: if field.required: field.required = False field.blank = True self._orig_was_required['%s.%s' % (db_field.model._meta, db_field.name)] = True # For every localized field copy the widget from the original field # and add a css class to identify a modeltranslation widget. try: orig_field = db_field.translated_field except __HOLE__: pass else: orig_formfield = self.formfield_for_dbfield(orig_field, **kwargs) field.widget = deepcopy(orig_formfield.widget) if orig_field.name in self.both_empty_values_fields: from modeltranslation.forms import NullableField, NullCharField form_class = field.__class__ if issubclass(form_class, NullCharField): # NullableField don't work with NullCharField form_class.__bases__ = tuple( b for b in form_class.__bases__ if b != NullCharField) field.__class__ = type( 'Nullable%s' % form_class.__name__, (NullableField, form_class), {}) if ( ( db_field.empty_value == 'both' or orig_field.name in self.both_empty_values_fields ) and isinstance(field.widget, (forms.TextInput, forms.Textarea)) ): field.widget = ClearableWidgetWrapper(field.widget) css_classes = field.widget.attrs.get('class', '').split(' ') css_classes.append('mt') # Add localized fieldname css class css_classes.append(build_css_class(db_field.name, 'mt-field')) # Add mt-bidi css class if language is bidirectional if(get_language_bidi(db_field.language)): css_classes.append('mt-bidi') if db_field.language == mt_settings.DEFAULT_LANGUAGE: # Add another css class to identify a default modeltranslation widget css_classes.append('mt-default') if (orig_formfield.required or self._orig_was_required.get( '%s.%s' % (orig_field.model._meta, orig_field.name))): # In case the original form field was required, make the # default translation field required instead. orig_formfield.required = False orig_formfield.blank = True field.required = True field.blank = False # Hide clearable widget for required fields if isinstance(field.widget, ClearableWidgetWrapper): field.widget = field.widget.widget field.widget.attrs['class'] = ' '.join(css_classes)
AttributeError
dataset/ETHPy150Open deschler/django-modeltranslation/modeltranslation/admin.py/TranslationBaseModelAdmin.patch_translation_field
4,989
def __init__(self, search_query): if not isinstance(search_query, str): raise ValueError("Expected str for 1st parameter (search_query).") self.avail_files = OrderedDict() self.visible_files = None self.adj_tokens = {False: None, True: None} self.vf_iter = None self.search_params = {"maxResults": 10, "order": self.preferences["order"]} # relevance by default self.yts_opts = dict() parsed = self.__searchParser(search_query) # search params self.search_params.update(parsed[0]) # YTa options _pref = deepcopy(self.preferences) # new object, just to not affect other intances. if 'metadata' in parsed[1]: try: meta_list = parsed[1]['metadata'].split(',') except __HOLE__: meta_list = [] if 'desc' in meta_list: _pref['metadata']['desc'] = True else: _pref['metadata']['desc'] = False if 'thumb' in meta_list: _pref['metadata']['thumb'] = True else: _pref['metadata']['thumb'] = False self.preferences = _pref # YTs options self.yts_opts = parsed[1] if 'audio' in parsed[1] and 'video' not in parsed[1]: self.yts_opts['video'] = False if 'video' in parsed[1] and 'audio' not in parsed[1]: self.yts_opts['audio'] = False try: self.__getChannelId() except requests.exceptions.ConnectionError: raise ConnectionError if parsed[0].get("publishedBefore"): self.search_params["publishedBefore"] += "T00:00:00Z" if parsed[0].get("publishedAfter"): self.search_params["publishedAfter"] += "T00:00:00Z"
AttributeError
dataset/ETHPy150Open rasguanabana/ytfs/ytfs/actions.py/YTActions.__init__
4,990
def __getChannelId(self): """ Obtain channel id for channel name, if present in ``self.search_params``. """ if not self.search_params.get("channelId"): return api_fixed_url = "https://www.googleapis.com/youtube/v3/channels?part=id&maxResults=1&fields=items%2Fid&" url = api_fixed_url + urlencode({"key": self.api_key, "forUsername": self.search_params["channelId"]}) get = requests.get(url).json() try: self.search_params["channelId"] = get['items'][0]['id'] return # got it except IndexError: pass # try searching now... api_fixed_url = "https://www.googleapis.com/youtube/v3/search?part=snippet&type=channel&fields=items%2Fid&" url = api_fixed_url + urlencode({"key": self.api_key, "q": self.search_params['channelId']}) get = requests.get(url).json() try: self.search_params["channelId"] = get['items'][0]['id']['channelId'] except __HOLE__: del self.search_params["channelId"] # channel not found
IndexError
dataset/ETHPy150Open rasguanabana/ytfs/ytfs/actions.py/YTActions.__getChannelId
4,991
def __searchParser(self, query): """ Parse `query` for advanced search options. Parameters ---------- query : str Search query to parse. Besides a search query, user can specify additional search parameters and YTFS specific options. Syntax: Additional search parameters: ``option:value``. if `value` contains spaces, then surround it with parentheses; available parameters: `channel`, `max`, `before`, `after`, `order`. YTFS options: specify options between ``[`` and ``]``; Available options: `a`, `v`, `f`, `P`, `s`, `m`. If an option takes a parameter, then specify it beetween parentheses. Examples: ``channel:foo search query``, ``my favourite music [a]``, ``channel:(the famous funny cats channel) [vf(240)P] funny cats max:20``. Invalid parameters/options are ignored. Returns ------- params : tuple Tuple: 0 - dictionary of url GET parameters; 1 - dictionary of YTStor options. """ ret = dict() parse_params = True buf = "" ptr = "" p_avail = ("channel", "max", "before", "after", "order") opts = dict() par_open = False translate = { 'a': 'audio', 'v': 'video', 'f': 'format', 's': 'stream', 'P': 'stream', 'm': 'metadata', 'max': 'maxResults', 'channel': 'channelId', 'before': 'publishedBefore', 'after': 'publishedAfter', 'order': 'order', '': 'q' } for i in query+' ': if parse_params: if not par_open: if i == ' ': # flush buf try: if ret.get(translate[ptr]): ret[ translate[ptr] ] += ' ' else: ret[ translate[ptr] ] = '' ret[ translate[ptr] ] += buf except KeyError: pass ptr = "" buf = "" elif i == ':' and buf in p_avail: ptr = buf buf = "" elif not buf and i == '[': # buf must be empty parse_params = False ptr = "" elif i != '(': buf += i elif not (par_open == 1 and i == ')'): buf += i if i == '(': par_open += 1 if par_open > 0 and i == ')': par_open -= 1 else: if i == ']': parse_params = True par_open = False ptr = "" buf = "" elif ptr and not par_open and i == '(': par_open = True elif par_open: if i == ')': try: opts[ translate[ptr] ] = buf except KeyError: pass par_open = False buf = "" else: buf += i elif i.isalpha(): ptr = i try: opts[ translate[ptr] ] = not i.isupper() except __HOLE__: pass return (ret, opts)
KeyError
dataset/ETHPy150Open rasguanabana/ytfs/ytfs/actions.py/YTActions.__searchParser
4,992
def updateResults(self, forward=None): """ Reload search results or load another "page". Parameters ---------- forward : bool or None, optional Whether move forwards or backwards (``True`` or ``False``). If ``None``, then first page is loaded. """ # this choses data we need. files = lambda x: { i['snippet']['title'].replace('/', '\\'): YTStor( {'yid': i['id']['videoId'], 'pub_date': i['snippet']['publishedAt']}, opts=self.yts_opts) for i in x['items'] } descs = lambda x: { (i['snippet']['title'].replace('/', '\\') + '.txt'): YTMetaStor( { 'title': i['snippet']['title'], 'yid': i['id']['videoId'], 'desc': i['snippet']['description'], 'channel': i['snippet']['channelTitle'], 'pub_date': i['snippet']['publishedAt'] }, opts=dict() ) for i in x['items'] } thumbs = lambda x: { (i['snippet']['title'].replace('/', '\\') + '.jpg'): YTMetaStor( {'url': i['snippet']['thumbnails']['high']['url'], 'pub_date': i['snippet']['publishedAt']}, opts=dict() ) for i in x['items'] } try: if self.adj_tokens[forward] is None: # in case someone would somehow cross boundary. forward = None except KeyError: pass recv = None try: try: data = self.avail_files[ self.adj_tokens[forward] ] # maybe data is already available locally. except __HOLE__: recv = self.__search( self.adj_tokens[forward] ) # nope, we have to search. except KeyError: # wrong index in adj_tokens if forward is None: recv = self.__search() else: raise ValueError("Valid values for forward are True, False or None (default).") if recv is not None: _d = files(recv) if self.preferences['metadata']['desc']: _d.update(descs(recv)) if self.preferences['metadata']['thumb']: _d.update(thumbs(recv)) data = (None, _d) # little format unification. if len(self.avail_files) > 4: pop = self.avail_files.popitem(False) # get rid of the oldest data. for s in pop[1][1].values(): s.clean() adj_t = deepcopy(self.adj_tokens) # this will we write to avail_files, now we update self.adj_tokens. if data[0] is None: # get tokens from obtained results. try: self.adj_tokens[False] = recv['prevPageToken'] except KeyError: self.adj_tokens[False] = None try: self.adj_tokens[True] = recv['nextPageToken'] except KeyError: self.adj_tokens[True] = None else: # already in avail_files. self.adj_tokens = data[0] if forward is not None: # backup last results in avail_files: self.avail_files[ self.adj_tokens[not forward] ] = (adj_t, self.visible_files) self.visible_files = data[1]
KeyError
dataset/ETHPy150Open rasguanabana/ytfs/ytfs/actions.py/YTActions.updateResults
4,993
def act(self): try: address = self.command[0] self.window.buffer.read(address) except __HOLE__: pass
IndexError
dataset/ETHPy150Open elmar-hinz/Python.Vii/vii/Interpreter.py/EditAction.act
4,994
def act(self): try: address = self.command[0] self.window.buffer.write(address) except __HOLE__: self.window.buffer.write()
IndexError
dataset/ETHPy150Open elmar-hinz/Python.Vii/vii/Interpreter.py/WriteAction.act
4,995
def get_class_from_string(path, default=None): """ Return the class specified by the string. IE: django.contrib.auth.models.User Will return the user class or cause an ImportError """ try: from importlib import import_module except __HOLE__: from django.utils.importlib import import_module i = path.rfind('.') module, attr = path[:i], path[i + 1:] mod = import_module(module) try: return getattr(mod, attr) except AttributeError: if default: return default else: raise ImportError( 'Cannot import name {} (from {})'.format(attr, mod))
ImportError
dataset/ETHPy150Open GetStream/stream-django/stream_django/utils.py/get_class_from_string
4,996
def tno_can_add(self, node, add_object): """ Returns whether a given object is droppable on the node. """ from mayavi.core.filter import Filter try: if issubclass(add_object, Filter) or \ issubclass(add_object, ModuleManager): return True except __HOLE__: if isinstance(add_object, Filter) or \ isinstance(add_object, ModuleManager): return True return False
TypeError
dataset/ETHPy150Open enthought/mayavi/mayavi/core/source.py/Source.tno_can_add
4,997
def action(self): ''' Should only run once to cleanup stale lane uxd files. ''' if not is_windows() and self.opts.value.get('sock_dir'): sockdirpath = os.path.abspath(self.opts.value['sock_dir']) console.concise("Cleaning up uxd files in {0}\n".format(sockdirpath)) protecteds = self.opts.value.get('raet_cleanup_protecteds', []) for name in os.listdir(sockdirpath): path = os.path.join(sockdirpath, name) if os.path.isdir(path): continue root, ext = os.path.splitext(name) if ext != '.uxd': continue if not all(root.partition('.')): continue if path in protecteds: continue try: os.unlink(path) console.concise("Removed {0}\n".format(path)) except __HOLE__: console.concise("Failed removing {0}\n".format(path)) raise
OSError
dataset/ETHPy150Open saltstack/salt/salt/daemons/flo/core.py/SaltRaetCleanup.action
4,998
def _process_road_rxmsg(self, msg, sender): ''' Send to the right queue msg is the message body dict sender is the unique name of the remote estate that sent the message ''' try: s_estate, s_yard, s_share = msg['route']['src'] d_estate, d_yard, d_share = msg['route']['dst'] except (__HOLE__, IndexError): log.error('Received invalid message: {0}'.format(msg)) return if s_estate is None: # drop return log.debug("**** Road Router rxMsg **** id={0} estate={1} yard={2}\n" " msg= {3}\n".format( self.opts.value['id'], self.road_stack.value.local.name, self.lane_stack.value.local.name, msg)) if d_estate is not None and d_estate != self.road_stack.value.local.name: log.error( 'Road Router Received message for wrong estate: {0}'.format(d_estate)) return if d_yard is not None: # Meant for another yard, send it off! if d_yard in self.lane_stack.value.nameRemotes: self.lane_stack.value.transmit(msg, self.lane_stack.value.nameRemotes[d_yard].uid) return if d_share is None: # No queue destination! log.error('Received message without share: {0}'.format(msg)) return elif d_share == 'event_fire': # rebroadcast events from other masters self.event.value.append(msg) #log.debug("\n**** Event Fire \n {0}\n".format(msg)) return elif d_share == 'local_cmd': # Refuse local commands over the wire log.error('Received local command remotely! Ignoring: {0}'.format(msg)) return elif d_share == 'remote_cmd': # Send it to a remote worker if 'load' in msg: role = self.road_stack.value.nameRemotes[sender].role msg['load']['id'] = role # sender # should this be role XXXX self.lane_stack.value.transmit(msg, self.lane_stack.value.fetchUidByName(next(self.workers.value)))
ValueError
dataset/ETHPy150Open saltstack/salt/salt/daemons/flo/core.py/SaltRaetRouterMaster._process_road_rxmsg
4,999
def _process_lane_rxmsg(self, msg, sender): ''' Send uxd messages tot he right queue or forward them to the correct yard etc. msg is message body dict sender is unique name of remote that sent the message ''' try: s_estate, s_yard, s_share = msg['route']['src'] d_estate, d_yard, d_share = msg['route']['dst'] except (ValueError, __HOLE__): log.error('Lane Router Received invalid message: {0}'.format(msg)) return if s_yard is None: return # drop message if s_estate is None: # substitute local estate s_estate = self.road_stack.value.local.name msg['route']['src'] = (s_estate, s_yard, s_share) log.debug("**** Lane Router rxMsg **** id={0} estate={1} yard={2}\n" " msg={3}\n".format( self.opts.value['id'], self.road_stack.value.local.name, self.lane_stack.value.local.name, msg)) if d_estate is None: pass elif d_estate != self.road_stack.value.local.name: # Forward to the correct estate if d_estate in self.road_stack.value.nameRemotes: self.road_stack.value.message(msg, self.road_stack.value.nameRemotes[d_estate].uid) return if d_share == 'pub_ret': # only publish to available minions msg['return']['ret']['minions'] = self._availablize(msg['return']['ret']['minions']) if msg.get('__worker_verify') == self.worker_verify.value: self.publish.value.append(msg) if d_yard is None: pass elif d_yard != self.lane_stack.value.local.name: # Meant for another yard, send it off! if d_yard in self.lane_stack.value.nameRemotes: self.lane_stack.value.transmit(msg, self.lane_stack.value.nameRemotes[d_yard].uid) return if d_share is None: # No queue destination! log.error('Lane Router Received message without share: {0}'.format(msg)) return elif d_share == 'local_cmd': self.lane_stack.value.transmit(msg, self.lane_stack.value.fetchUidByName(next(self.workers.value))) elif d_share == 'event_req': self.event_req.value.append(msg) #log.debug("\n**** Event Subscribe \n {0}\n".format(msg)) elif d_share == 'event_fire': self.event.value.append(msg) #log.debug("\n**** Event Fire \n {0}\n".format(msg)) elif d_share == 'presence_req': self.presence_req.value.append(msg) #log.debug("\n**** Presence Request \n {0}\n".format(msg)) elif d_share == 'stats_req': self.stats_req.value.append(msg) #log.debug("\n**** Stats Request \n {0}\n".format(msg))
IndexError
dataset/ETHPy150Open saltstack/salt/salt/daemons/flo/core.py/SaltRaetRouterMaster._process_lane_rxmsg