repo
string
path
string
func_name
string
original_string
string
language
string
code
string
code_tokens
sequence
docstring
string
docstring_tokens
sequence
sha
string
url
string
partition
string
christophertbrown/bioscripts
ctbBio/nr_fasta.py
append_index_id
def append_index_id(id, ids): """ add index to id to make it unique wrt ids """ index = 1 mod = '%s_%s' % (id, index) while mod in ids: index += 1 mod = '%s_%s' % (id, index) ids.append(mod) return mod, ids
python
def append_index_id(id, ids): """ add index to id to make it unique wrt ids """ index = 1 mod = '%s_%s' % (id, index) while mod in ids: index += 1 mod = '%s_%s' % (id, index) ids.append(mod) return mod, ids
[ "def", "append_index_id", "(", "id", ",", "ids", ")", ":", "index", "=", "1", "mod", "=", "'%s_%s'", "%", "(", "id", ",", "index", ")", "while", "mod", "in", "ids", ":", "index", "+=", "1", "mod", "=", "'%s_%s'", "%", "(", "id", ",", "index", ")", "ids", ".", "append", "(", "mod", ")", "return", "mod", ",", "ids" ]
add index to id to make it unique wrt ids
[ "add", "index", "to", "id", "to", "make", "it", "unique", "wrt", "ids" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/nr_fasta.py#L11-L21
train
christophertbrown/bioscripts
ctbBio/nr_fasta.py
de_rep
def de_rep(fastas, append_index, return_original = False): """ de-replicate fastas based on sequence names """ ids = [] for fasta in fastas: for seq in parse_fasta(fasta): header = seq[0].split('>')[1].split() id = header[0] if id not in ids: ids.append(id) if return_original is True: yield [header, seq] else: yield seq elif append_index == True: new, ids = append_index_id(id, ids) if return_original is True: yield [header, ['>%s %s' % (new, ' '.join(header[1::])), seq[1]]] else: yield ['>%s %s' % (new, ' '.join(header[1::])), seq[1]]
python
def de_rep(fastas, append_index, return_original = False): """ de-replicate fastas based on sequence names """ ids = [] for fasta in fastas: for seq in parse_fasta(fasta): header = seq[0].split('>')[1].split() id = header[0] if id not in ids: ids.append(id) if return_original is True: yield [header, seq] else: yield seq elif append_index == True: new, ids = append_index_id(id, ids) if return_original is True: yield [header, ['>%s %s' % (new, ' '.join(header[1::])), seq[1]]] else: yield ['>%s %s' % (new, ' '.join(header[1::])), seq[1]]
[ "def", "de_rep", "(", "fastas", ",", "append_index", ",", "return_original", "=", "False", ")", ":", "ids", "=", "[", "]", "for", "fasta", "in", "fastas", ":", "for", "seq", "in", "parse_fasta", "(", "fasta", ")", ":", "header", "=", "seq", "[", "0", "]", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "split", "(", ")", "id", "=", "header", "[", "0", "]", "if", "id", "not", "in", "ids", ":", "ids", ".", "append", "(", "id", ")", "if", "return_original", "is", "True", ":", "yield", "[", "header", ",", "seq", "]", "else", ":", "yield", "seq", "elif", "append_index", "==", "True", ":", "new", ",", "ids", "=", "append_index_id", "(", "id", ",", "ids", ")", "if", "return_original", "is", "True", ":", "yield", "[", "header", ",", "[", "'>%s %s'", "%", "(", "new", ",", "' '", ".", "join", "(", "header", "[", "1", ":", ":", "]", ")", ")", ",", "seq", "[", "1", "]", "]", "]", "else", ":", "yield", "[", "'>%s %s'", "%", "(", "new", ",", "' '", ".", "join", "(", "header", "[", "1", ":", ":", "]", ")", ")", ",", "seq", "[", "1", "]", "]" ]
de-replicate fastas based on sequence names
[ "de", "-", "replicate", "fastas", "based", "on", "sequence", "names" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/nr_fasta.py#L23-L43
train
e-dard/postcodes
postcodes.py
get
def get(postcode): """ Request data associated with `postcode`. :param postcode: the postcode to search for. The postcode may contain spaces (they will be removed). :returns: a dict of the nearest postcode's data or None if no postcode data is found. """ postcode = quote(postcode.replace(' ', '')) url = '%s/postcode/%s.json' % (END_POINT, postcode) return _get_json_resp(url)
python
def get(postcode): """ Request data associated with `postcode`. :param postcode: the postcode to search for. The postcode may contain spaces (they will be removed). :returns: a dict of the nearest postcode's data or None if no postcode data is found. """ postcode = quote(postcode.replace(' ', '')) url = '%s/postcode/%s.json' % (END_POINT, postcode) return _get_json_resp(url)
[ "def", "get", "(", "postcode", ")", ":", "postcode", "=", "quote", "(", "postcode", ".", "replace", "(", "' '", ",", "''", ")", ")", "url", "=", "'%s/postcode/%s.json'", "%", "(", "END_POINT", ",", "postcode", ")", "return", "_get_json_resp", "(", "url", ")" ]
Request data associated with `postcode`. :param postcode: the postcode to search for. The postcode may contain spaces (they will be removed). :returns: a dict of the nearest postcode's data or None if no postcode data is found.
[ "Request", "data", "associated", "with", "postcode", "." ]
d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005
https://github.com/e-dard/postcodes/blob/d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005/postcodes.py#L22-L34
train
e-dard/postcodes
postcodes.py
get_from_postcode
def get_from_postcode(postcode, distance): """ Request all postcode data within `distance` miles of `postcode`. :param postcode: the postcode to search for. The postcode may contain spaces (they will be removed). :param distance: distance in miles to `postcode`. :returns: a list of dicts containing postcode data within the specified distance or `None` if `postcode` is not valid. """ postcode = quote(postcode.replace(' ', '')) return _get_from(distance, 'postcode=%s' % postcode)
python
def get_from_postcode(postcode, distance): """ Request all postcode data within `distance` miles of `postcode`. :param postcode: the postcode to search for. The postcode may contain spaces (they will be removed). :param distance: distance in miles to `postcode`. :returns: a list of dicts containing postcode data within the specified distance or `None` if `postcode` is not valid. """ postcode = quote(postcode.replace(' ', '')) return _get_from(distance, 'postcode=%s' % postcode)
[ "def", "get_from_postcode", "(", "postcode", ",", "distance", ")", ":", "postcode", "=", "quote", "(", "postcode", ".", "replace", "(", "' '", ",", "''", ")", ")", "return", "_get_from", "(", "distance", ",", "'postcode=%s'", "%", "postcode", ")" ]
Request all postcode data within `distance` miles of `postcode`. :param postcode: the postcode to search for. The postcode may contain spaces (they will be removed). :param distance: distance in miles to `postcode`. :returns: a list of dicts containing postcode data within the specified distance or `None` if `postcode` is not valid.
[ "Request", "all", "postcode", "data", "within", "distance", "miles", "of", "postcode", "." ]
d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005
https://github.com/e-dard/postcodes/blob/d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005/postcodes.py#L56-L69
train
e-dard/postcodes
postcodes.py
PostCoder._check_point
def _check_point(self, lat, lng): """ Checks if latitude and longitude correct """ if abs(lat) > 90 or abs(lng) > 180: msg = "Illegal lat and/or lng, (%s, %s) provided." % (lat, lng) raise IllegalPointException(msg)
python
def _check_point(self, lat, lng): """ Checks if latitude and longitude correct """ if abs(lat) > 90 or abs(lng) > 180: msg = "Illegal lat and/or lng, (%s, %s) provided." % (lat, lng) raise IllegalPointException(msg)
[ "def", "_check_point", "(", "self", ",", "lat", ",", "lng", ")", ":", "if", "abs", "(", "lat", ")", ">", "90", "or", "abs", "(", "lng", ")", ">", "180", ":", "msg", "=", "\"Illegal lat and/or lng, (%s, %s) provided.\"", "%", "(", "lat", ",", "lng", ")", "raise", "IllegalPointException", "(", "msg", ")" ]
Checks if latitude and longitude correct
[ "Checks", "if", "latitude", "and", "longitude", "correct" ]
d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005
https://github.com/e-dard/postcodes/blob/d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005/postcodes.py#L123-L127
train
e-dard/postcodes
postcodes.py
PostCoder._lookup
def _lookup(self, skip_cache, fun, *args, **kwargs): """ Checks for cached responses, before requesting from web-service """ if args not in self.cache or skip_cache: self.cache[args] = fun(*args, **kwargs) return self.cache[args]
python
def _lookup(self, skip_cache, fun, *args, **kwargs): """ Checks for cached responses, before requesting from web-service """ if args not in self.cache or skip_cache: self.cache[args] = fun(*args, **kwargs) return self.cache[args]
[ "def", "_lookup", "(", "self", ",", "skip_cache", ",", "fun", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "if", "args", "not", "in", "self", ".", "cache", "or", "skip_cache", ":", "self", ".", "cache", "[", "args", "]", "=", "fun", "(", "*", "args", ",", "*", "*", "kwargs", ")", "return", "self", ".", "cache", "[", "args", "]" ]
Checks for cached responses, before requesting from web-service
[ "Checks", "for", "cached", "responses", "before", "requesting", "from", "web", "-", "service" ]
d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005
https://github.com/e-dard/postcodes/blob/d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005/postcodes.py#L129-L136
train
e-dard/postcodes
postcodes.py
PostCoder.get_nearest
def get_nearest(self, lat, lng, skip_cache=False): """ Calls `postcodes.get_nearest` but checks correctness of `lat` and `long`, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a dict of the nearest postcode's data. """ lat, lng = float(lat), float(lng) self._check_point(lat, lng) return self._lookup(skip_cache, get_nearest, lat, lng)
python
def get_nearest(self, lat, lng, skip_cache=False): """ Calls `postcodes.get_nearest` but checks correctness of `lat` and `long`, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a dict of the nearest postcode's data. """ lat, lng = float(lat), float(lng) self._check_point(lat, lng) return self._lookup(skip_cache, get_nearest, lat, lng)
[ "def", "get_nearest", "(", "self", ",", "lat", ",", "lng", ",", "skip_cache", "=", "False", ")", ":", "lat", ",", "lng", "=", "float", "(", "lat", ")", ",", "float", "(", "lng", ")", "self", ".", "_check_point", "(", "lat", ",", "lng", ")", "return", "self", ".", "_lookup", "(", "skip_cache", ",", "get_nearest", ",", "lat", ",", "lng", ")" ]
Calls `postcodes.get_nearest` but checks correctness of `lat` and `long`, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a dict of the nearest postcode's data.
[ "Calls", "postcodes", ".", "get_nearest", "but", "checks", "correctness", "of", "lat", "and", "long", "and", "by", "default", "utilises", "a", "local", "cache", "." ]
d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005
https://github.com/e-dard/postcodes/blob/d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005/postcodes.py#L152-L167
train
e-dard/postcodes
postcodes.py
PostCoder.get_from_postcode
def get_from_postcode(self, postcode, distance, skip_cache=False): """ Calls `postcodes.get_from_postcode` but checks correctness of `distance`, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a list of dicts containing postcode data within the specified distance. """ distance = float(distance) if distance < 0: raise IllegalDistanceException("Distance must not be negative") # remove spaces and change case here due to caching postcode = postcode.lower().replace(' ', '') return self._lookup(skip_cache, get_from_postcode, postcode, float(distance))
python
def get_from_postcode(self, postcode, distance, skip_cache=False): """ Calls `postcodes.get_from_postcode` but checks correctness of `distance`, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a list of dicts containing postcode data within the specified distance. """ distance = float(distance) if distance < 0: raise IllegalDistanceException("Distance must not be negative") # remove spaces and change case here due to caching postcode = postcode.lower().replace(' ', '') return self._lookup(skip_cache, get_from_postcode, postcode, float(distance))
[ "def", "get_from_postcode", "(", "self", ",", "postcode", ",", "distance", ",", "skip_cache", "=", "False", ")", ":", "distance", "=", "float", "(", "distance", ")", "if", "distance", "<", "0", ":", "raise", "IllegalDistanceException", "(", "\"Distance must not be negative\"", ")", "# remove spaces and change case here due to caching", "postcode", "=", "postcode", ".", "lower", "(", ")", ".", "replace", "(", "' '", ",", "''", ")", "return", "self", ".", "_lookup", "(", "skip_cache", ",", "get_from_postcode", ",", "postcode", ",", "float", "(", "distance", ")", ")" ]
Calls `postcodes.get_from_postcode` but checks correctness of `distance`, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a list of dicts containing postcode data within the specified distance.
[ "Calls", "postcodes", ".", "get_from_postcode", "but", "checks", "correctness", "of", "distance", "and", "by", "default", "utilises", "a", "local", "cache", "." ]
d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005
https://github.com/e-dard/postcodes/blob/d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005/postcodes.py#L169-L189
train
e-dard/postcodes
postcodes.py
PostCoder.get_from_geo
def get_from_geo(self, lat, lng, distance, skip_cache=False): """ Calls `postcodes.get_from_geo` but checks the correctness of all arguments, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a list of dicts containing postcode data within the specified distance. """ # remove spaces and change case here due to caching lat, lng, distance = float(lat), float(lng), float(distance) if distance < 0: raise IllegalDistanceException("Distance must not be negative") self._check_point(lat, lng) return self._lookup(skip_cache, get_from_geo, lat, lng, distance)
python
def get_from_geo(self, lat, lng, distance, skip_cache=False): """ Calls `postcodes.get_from_geo` but checks the correctness of all arguments, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a list of dicts containing postcode data within the specified distance. """ # remove spaces and change case here due to caching lat, lng, distance = float(lat), float(lng), float(distance) if distance < 0: raise IllegalDistanceException("Distance must not be negative") self._check_point(lat, lng) return self._lookup(skip_cache, get_from_geo, lat, lng, distance)
[ "def", "get_from_geo", "(", "self", ",", "lat", ",", "lng", ",", "distance", ",", "skip_cache", "=", "False", ")", ":", "# remove spaces and change case here due to caching", "lat", ",", "lng", ",", "distance", "=", "float", "(", "lat", ")", ",", "float", "(", "lng", ")", ",", "float", "(", "distance", ")", "if", "distance", "<", "0", ":", "raise", "IllegalDistanceException", "(", "\"Distance must not be negative\"", ")", "self", ".", "_check_point", "(", "lat", ",", "lng", ")", "return", "self", ".", "_lookup", "(", "skip_cache", ",", "get_from_geo", ",", "lat", ",", "lng", ",", "distance", ")" ]
Calls `postcodes.get_from_geo` but checks the correctness of all arguments, and by default utilises a local cache. :param skip_cache: optional argument specifying whether to skip the cache and make an explicit request. :raises IllegalPointException: if the latitude or longitude are out of bounds. :returns: a list of dicts containing postcode data within the specified distance.
[ "Calls", "postcodes", ".", "get_from_geo", "but", "checks", "the", "correctness", "of", "all", "arguments", "and", "by", "default", "utilises", "a", "local", "cache", "." ]
d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005
https://github.com/e-dard/postcodes/blob/d63c47b4ecd765bc2e4e6ba34bc0b8a796f44005/postcodes.py#L191-L210
train
christophertbrown/bioscripts
ctbBio/rRNA_insertions.py
insertions_from_masked
def insertions_from_masked(seq): """ get coordinates of insertions from insertion-masked sequence """ insertions = [] prev = True for i, base in enumerate(seq): if base.isupper() and prev is True: insertions.append([]) prev = False elif base.islower(): insertions[-1].append(i) prev = True return [[min(i), max(i)] for i in insertions if i != []]
python
def insertions_from_masked(seq): """ get coordinates of insertions from insertion-masked sequence """ insertions = [] prev = True for i, base in enumerate(seq): if base.isupper() and prev is True: insertions.append([]) prev = False elif base.islower(): insertions[-1].append(i) prev = True return [[min(i), max(i)] for i in insertions if i != []]
[ "def", "insertions_from_masked", "(", "seq", ")", ":", "insertions", "=", "[", "]", "prev", "=", "True", "for", "i", ",", "base", "in", "enumerate", "(", "seq", ")", ":", "if", "base", ".", "isupper", "(", ")", "and", "prev", "is", "True", ":", "insertions", ".", "append", "(", "[", "]", ")", "prev", "=", "False", "elif", "base", ".", "islower", "(", ")", ":", "insertions", "[", "-", "1", "]", ".", "append", "(", "i", ")", "prev", "=", "True", "return", "[", "[", "min", "(", "i", ")", ",", "max", "(", "i", ")", "]", "for", "i", "in", "insertions", "if", "i", "!=", "[", "]", "]" ]
get coordinates of insertions from insertion-masked sequence
[ "get", "coordinates", "of", "insertions", "from", "insertion", "-", "masked", "sequence" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rRNA_insertions.py#L15-L28
train
christophertbrown/bioscripts
ctbBio/rRNA_insertions.py
seq_info
def seq_info(names, id2names, insertions, sequences): """ get insertion information from header """ seqs = {} # seqs[id] = [gene, model, [[i-gene_pos, i-model_pos, i-length, iseq, [orfs], [introns]], ...]] for name in names: id = id2names[name] gene = name.split('fromHMM::', 1)[0].rsplit(' ', 1)[1] model = name.split('fromHMM::', 1)[1].split('=', 1)[1].split()[0] i_gene_pos = insertions[id] # coordinates of each insertion wrt gene i_model_pos = name.split('fromHMM::', 1)[1].split('model-pos(ins-len)=')[1].split()[0].split(';') # model overlap i_info = [] for i, ins in enumerate(i_gene_pos): model_pos = i_model_pos[i].split('-')[1].split('(')[0] length = i_model_pos[i].split('(')[1].split(')')[0] iheader = '>%s_%s insertion::seq=%s type=insertion strand=n/a gene-pos=%s-%s model-pos=%s'\ % (id, (i + 1), (i + 1), ins[0], ins[1], model_pos) iseq = sequences[id][1][ins[0]:(ins[1] + 1)] iseq = [iheader, iseq] info = [ins, model_pos, length, iseq, [], []] i_info.append(info) seqs[id] = [gene, model, i_info] return seqs
python
def seq_info(names, id2names, insertions, sequences): """ get insertion information from header """ seqs = {} # seqs[id] = [gene, model, [[i-gene_pos, i-model_pos, i-length, iseq, [orfs], [introns]], ...]] for name in names: id = id2names[name] gene = name.split('fromHMM::', 1)[0].rsplit(' ', 1)[1] model = name.split('fromHMM::', 1)[1].split('=', 1)[1].split()[0] i_gene_pos = insertions[id] # coordinates of each insertion wrt gene i_model_pos = name.split('fromHMM::', 1)[1].split('model-pos(ins-len)=')[1].split()[0].split(';') # model overlap i_info = [] for i, ins in enumerate(i_gene_pos): model_pos = i_model_pos[i].split('-')[1].split('(')[0] length = i_model_pos[i].split('(')[1].split(')')[0] iheader = '>%s_%s insertion::seq=%s type=insertion strand=n/a gene-pos=%s-%s model-pos=%s'\ % (id, (i + 1), (i + 1), ins[0], ins[1], model_pos) iseq = sequences[id][1][ins[0]:(ins[1] + 1)] iseq = [iheader, iseq] info = [ins, model_pos, length, iseq, [], []] i_info.append(info) seqs[id] = [gene, model, i_info] return seqs
[ "def", "seq_info", "(", "names", ",", "id2names", ",", "insertions", ",", "sequences", ")", ":", "seqs", "=", "{", "}", "# seqs[id] = [gene, model, [[i-gene_pos, i-model_pos, i-length, iseq, [orfs], [introns]], ...]]", "for", "name", "in", "names", ":", "id", "=", "id2names", "[", "name", "]", "gene", "=", "name", ".", "split", "(", "'fromHMM::'", ",", "1", ")", "[", "0", "]", ".", "rsplit", "(", "' '", ",", "1", ")", "[", "1", "]", "model", "=", "name", ".", "split", "(", "'fromHMM::'", ",", "1", ")", "[", "1", "]", ".", "split", "(", "'='", ",", "1", ")", "[", "1", "]", ".", "split", "(", ")", "[", "0", "]", "i_gene_pos", "=", "insertions", "[", "id", "]", "# coordinates of each insertion wrt gene", "i_model_pos", "=", "name", ".", "split", "(", "'fromHMM::'", ",", "1", ")", "[", "1", "]", ".", "split", "(", "'model-pos(ins-len)='", ")", "[", "1", "]", ".", "split", "(", ")", "[", "0", "]", ".", "split", "(", "';'", ")", "# model overlap", "i_info", "=", "[", "]", "for", "i", ",", "ins", "in", "enumerate", "(", "i_gene_pos", ")", ":", "model_pos", "=", "i_model_pos", "[", "i", "]", ".", "split", "(", "'-'", ")", "[", "1", "]", ".", "split", "(", "'('", ")", "[", "0", "]", "length", "=", "i_model_pos", "[", "i", "]", ".", "split", "(", "'('", ")", "[", "1", "]", ".", "split", "(", "')'", ")", "[", "0", "]", "iheader", "=", "'>%s_%s insertion::seq=%s type=insertion strand=n/a gene-pos=%s-%s model-pos=%s'", "%", "(", "id", ",", "(", "i", "+", "1", ")", ",", "(", "i", "+", "1", ")", ",", "ins", "[", "0", "]", ",", "ins", "[", "1", "]", ",", "model_pos", ")", "iseq", "=", "sequences", "[", "id", "]", "[", "1", "]", "[", "ins", "[", "0", "]", ":", "(", "ins", "[", "1", "]", "+", "1", ")", "]", "iseq", "=", "[", "iheader", ",", "iseq", "]", "info", "=", "[", "ins", ",", "model_pos", ",", "length", ",", "iseq", ",", "[", "]", ",", "[", "]", "]", "i_info", ".", "append", "(", "info", ")", "seqs", "[", "id", "]", "=", "[", "gene", ",", "model", ",", "i_info", "]", "return", "seqs" ]
get insertion information from header
[ "get", "insertion", "information", "from", "header" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rRNA_insertions.py#L64-L86
train
christophertbrown/bioscripts
ctbBio/rRNA_insertions.py
check_overlap
def check_overlap(pos, ins, thresh): """ make sure thresh % feature is contained within insertion """ ins_pos = ins[0] ins_len = ins[2] ol = overlap(ins_pos, pos) feat_len = pos[1] - pos[0] + 1 # print float(ol) / float(feat_len) if float(ol) / float(feat_len) >= thresh: return True return False
python
def check_overlap(pos, ins, thresh): """ make sure thresh % feature is contained within insertion """ ins_pos = ins[0] ins_len = ins[2] ol = overlap(ins_pos, pos) feat_len = pos[1] - pos[0] + 1 # print float(ol) / float(feat_len) if float(ol) / float(feat_len) >= thresh: return True return False
[ "def", "check_overlap", "(", "pos", ",", "ins", ",", "thresh", ")", ":", "ins_pos", "=", "ins", "[", "0", "]", "ins_len", "=", "ins", "[", "2", "]", "ol", "=", "overlap", "(", "ins_pos", ",", "pos", ")", "feat_len", "=", "pos", "[", "1", "]", "-", "pos", "[", "0", "]", "+", "1", "# print float(ol) / float(feat_len)", "if", "float", "(", "ol", ")", "/", "float", "(", "feat_len", ")", ">=", "thresh", ":", "return", "True", "return", "False" ]
make sure thresh % feature is contained within insertion
[ "make", "sure", "thresh", "%", "feature", "is", "contained", "within", "insertion" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rRNA_insertions.py#L91-L102
train
christophertbrown/bioscripts
ctbBio/rRNA_insertions.py
max_insertion
def max_insertion(seqs, gene, domain): """ length of largest insertion """ seqs = [i[2] for i in list(seqs.values()) if i[2] != [] and i[0] == gene and i[1] == domain] lengths = [] for seq in seqs: for ins in seq: lengths.append(int(ins[2])) if lengths == []: return 100 return max(lengths)
python
def max_insertion(seqs, gene, domain): """ length of largest insertion """ seqs = [i[2] for i in list(seqs.values()) if i[2] != [] and i[0] == gene and i[1] == domain] lengths = [] for seq in seqs: for ins in seq: lengths.append(int(ins[2])) if lengths == []: return 100 return max(lengths)
[ "def", "max_insertion", "(", "seqs", ",", "gene", ",", "domain", ")", ":", "seqs", "=", "[", "i", "[", "2", "]", "for", "i", "in", "list", "(", "seqs", ".", "values", "(", ")", ")", "if", "i", "[", "2", "]", "!=", "[", "]", "and", "i", "[", "0", "]", "==", "gene", "and", "i", "[", "1", "]", "==", "domain", "]", "lengths", "=", "[", "]", "for", "seq", "in", "seqs", ":", "for", "ins", "in", "seq", ":", "lengths", ".", "append", "(", "int", "(", "ins", "[", "2", "]", ")", ")", "if", "lengths", "==", "[", "]", ":", "return", "100", "return", "max", "(", "lengths", ")" ]
length of largest insertion
[ "length", "of", "largest", "insertion" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rRNA_insertions.py#L305-L316
train
christophertbrown/bioscripts
ctbBio/rRNA_insertions.py
model_length
def model_length(gene, domain): """ get length of model """ if gene == '16S': domain2max = {'E_coli_K12': int(1538), 'bacteria': int(1689), 'archaea': int(1563), 'eukarya': int(2652)} return domain2max[domain] elif gene == '23S': domain2max = {'E_coli_K12': int(2903), 'bacteria': int(3146), 'archaea': int(3774), 'eukarya': int(9079)} return domain2max[domain] else: print(sys.stderr, '# length unknown for gene: %s, domain: %s' % (gene, domain)) exit()
python
def model_length(gene, domain): """ get length of model """ if gene == '16S': domain2max = {'E_coli_K12': int(1538), 'bacteria': int(1689), 'archaea': int(1563), 'eukarya': int(2652)} return domain2max[domain] elif gene == '23S': domain2max = {'E_coli_K12': int(2903), 'bacteria': int(3146), 'archaea': int(3774), 'eukarya': int(9079)} return domain2max[domain] else: print(sys.stderr, '# length unknown for gene: %s, domain: %s' % (gene, domain)) exit()
[ "def", "model_length", "(", "gene", ",", "domain", ")", ":", "if", "gene", "==", "'16S'", ":", "domain2max", "=", "{", "'E_coli_K12'", ":", "int", "(", "1538", ")", ",", "'bacteria'", ":", "int", "(", "1689", ")", ",", "'archaea'", ":", "int", "(", "1563", ")", ",", "'eukarya'", ":", "int", "(", "2652", ")", "}", "return", "domain2max", "[", "domain", "]", "elif", "gene", "==", "'23S'", ":", "domain2max", "=", "{", "'E_coli_K12'", ":", "int", "(", "2903", ")", ",", "'bacteria'", ":", "int", "(", "3146", ")", ",", "'archaea'", ":", "int", "(", "3774", ")", ",", "'eukarya'", ":", "int", "(", "9079", ")", "}", "return", "domain2max", "[", "domain", "]", "else", ":", "print", "(", "sys", ".", "stderr", ",", "'# length unknown for gene: %s, domain: %s'", "%", "(", "gene", ",", "domain", ")", ")", "exit", "(", ")" ]
get length of model
[ "get", "length", "of", "model" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rRNA_insertions.py#L318-L330
train
christophertbrown/bioscripts
ctbBio/rRNA_insertions.py
setup_markers
def setup_markers(seqs): """ setup unique marker for every orf annotation - change size if necessary """ family2marker = {} # family2marker[family] = [marker, size] markers = cycle(['^', 'p', '*', '+', 'x', 'd', '|', 'v', '>', '<', '8']) size = 60 families = [] for seq in list(seqs.values()): for insertion in seq[2]: for family in list(insertion[-1].values()): if family not in families: families.append(family) for family in families: marker = next(markers) if marker == '^': size = size * 0.5 family2marker[family] = [marker, size] return family2marker
python
def setup_markers(seqs): """ setup unique marker for every orf annotation - change size if necessary """ family2marker = {} # family2marker[family] = [marker, size] markers = cycle(['^', 'p', '*', '+', 'x', 'd', '|', 'v', '>', '<', '8']) size = 60 families = [] for seq in list(seqs.values()): for insertion in seq[2]: for family in list(insertion[-1].values()): if family not in families: families.append(family) for family in families: marker = next(markers) if marker == '^': size = size * 0.5 family2marker[family] = [marker, size] return family2marker
[ "def", "setup_markers", "(", "seqs", ")", ":", "family2marker", "=", "{", "}", "# family2marker[family] = [marker, size]", "markers", "=", "cycle", "(", "[", "'^'", ",", "'p'", ",", "'*'", ",", "'+'", ",", "'x'", ",", "'d'", ",", "'|'", ",", "'v'", ",", "'>'", ",", "'<'", ",", "'8'", "]", ")", "size", "=", "60", "families", "=", "[", "]", "for", "seq", "in", "list", "(", "seqs", ".", "values", "(", ")", ")", ":", "for", "insertion", "in", "seq", "[", "2", "]", ":", "for", "family", "in", "list", "(", "insertion", "[", "-", "1", "]", ".", "values", "(", ")", ")", ":", "if", "family", "not", "in", "families", ":", "families", ".", "append", "(", "family", ")", "for", "family", "in", "families", ":", "marker", "=", "next", "(", "markers", ")", "if", "marker", "==", "'^'", ":", "size", "=", "size", "*", "0.5", "family2marker", "[", "family", "]", "=", "[", "marker", ",", "size", "]", "return", "family2marker" ]
setup unique marker for every orf annotation - change size if necessary
[ "setup", "unique", "marker", "for", "every", "orf", "annotation", "-", "change", "size", "if", "necessary" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rRNA_insertions.py#L332-L351
train
christophertbrown/bioscripts
ctbBio/rRNA_insertions.py
plot_by_gene_and_domain
def plot_by_gene_and_domain(name, seqs, tax, id2name): """ plot insertions for each gene and domain """ for gene in set([seq[0] for seq in list(seqs.values())]): for domain in set([seq[1] for seq in list(seqs.values())]): plot_insertions(name, seqs, gene, domain, tax, id2name)
python
def plot_by_gene_and_domain(name, seqs, tax, id2name): """ plot insertions for each gene and domain """ for gene in set([seq[0] for seq in list(seqs.values())]): for domain in set([seq[1] for seq in list(seqs.values())]): plot_insertions(name, seqs, gene, domain, tax, id2name)
[ "def", "plot_by_gene_and_domain", "(", "name", ",", "seqs", ",", "tax", ",", "id2name", ")", ":", "for", "gene", "in", "set", "(", "[", "seq", "[", "0", "]", "for", "seq", "in", "list", "(", "seqs", ".", "values", "(", ")", ")", "]", ")", ":", "for", "domain", "in", "set", "(", "[", "seq", "[", "1", "]", "for", "seq", "in", "list", "(", "seqs", ".", "values", "(", ")", ")", "]", ")", ":", "plot_insertions", "(", "name", ",", "seqs", ",", "gene", ",", "domain", ",", "tax", ",", "id2name", ")" ]
plot insertions for each gene and domain
[ "plot", "insertions", "for", "each", "gene", "and", "domain" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rRNA_insertions.py#L551-L557
train
christophertbrown/bioscripts
ctbBio/neto.py
get_descriptions
def get_descriptions(fastas): """ get the description for each ORF """ id2desc = {} for fasta in fastas: for seq in parse_fasta(fasta): header = seq[0].split('>')[1].split(' ') id = header[0] if len(header) > 1: desc = ' '.join(header[1:]) else: desc = 'n/a' length = float(len([i for i in seq[1].strip() if i != '*'])) id2desc[id] = [fasta, desc, length] return id2desc
python
def get_descriptions(fastas): """ get the description for each ORF """ id2desc = {} for fasta in fastas: for seq in parse_fasta(fasta): header = seq[0].split('>')[1].split(' ') id = header[0] if len(header) > 1: desc = ' '.join(header[1:]) else: desc = 'n/a' length = float(len([i for i in seq[1].strip() if i != '*'])) id2desc[id] = [fasta, desc, length] return id2desc
[ "def", "get_descriptions", "(", "fastas", ")", ":", "id2desc", "=", "{", "}", "for", "fasta", "in", "fastas", ":", "for", "seq", "in", "parse_fasta", "(", "fasta", ")", ":", "header", "=", "seq", "[", "0", "]", ".", "split", "(", "'>'", ")", "[", "1", "]", ".", "split", "(", "' '", ")", "id", "=", "header", "[", "0", "]", "if", "len", "(", "header", ")", ">", "1", ":", "desc", "=", "' '", ".", "join", "(", "header", "[", "1", ":", "]", ")", "else", ":", "desc", "=", "'n/a'", "length", "=", "float", "(", "len", "(", "[", "i", "for", "i", "in", "seq", "[", "1", "]", ".", "strip", "(", ")", "if", "i", "!=", "'*'", "]", ")", ")", "id2desc", "[", "id", "]", "=", "[", "fasta", ",", "desc", ",", "length", "]", "return", "id2desc" ]
get the description for each ORF
[ "get", "the", "description", "for", "each", "ORF" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/neto.py#L37-L52
train
christophertbrown/bioscripts
ctbBio/neto.py
print_genome_matrix
def print_genome_matrix(hits, fastas, id2desc, file_name): """ optimize later? slow ... should combine with calculate_threshold module """ out = open(file_name, 'w') fastas = sorted(fastas) print('## percent identity between genomes', file=out) print('# - \t %s' % ('\t'.join(fastas)), file=out) for fasta in fastas: line = [fasta] for other in fastas: if other == fasta: average = '-' else: average = numpy.average([hits[fasta][other][i][3] for i in hits[fasta][other]]) line.append(str(average)) print('\t'.join(line), file=out) print('', file=out) print('## percent of orfs that are orthologous between genomes', file=out) print('# - \t %s' % ('\t'.join(fastas)), file=out) for fasta in fastas: line = [fasta] for other in fastas: if other == fasta: percent = '-' else: orthologs = float(len(hits[fasta][other])) orfs = float(len([i for i in id2desc if id2desc[i][0] == fasta])) percent = float(orthologs / orfs) * 100 line.append(str(percent)) print('\t'.join(line), file=out)
python
def print_genome_matrix(hits, fastas, id2desc, file_name): """ optimize later? slow ... should combine with calculate_threshold module """ out = open(file_name, 'w') fastas = sorted(fastas) print('## percent identity between genomes', file=out) print('# - \t %s' % ('\t'.join(fastas)), file=out) for fasta in fastas: line = [fasta] for other in fastas: if other == fasta: average = '-' else: average = numpy.average([hits[fasta][other][i][3] for i in hits[fasta][other]]) line.append(str(average)) print('\t'.join(line), file=out) print('', file=out) print('## percent of orfs that are orthologous between genomes', file=out) print('# - \t %s' % ('\t'.join(fastas)), file=out) for fasta in fastas: line = [fasta] for other in fastas: if other == fasta: percent = '-' else: orthologs = float(len(hits[fasta][other])) orfs = float(len([i for i in id2desc if id2desc[i][0] == fasta])) percent = float(orthologs / orfs) * 100 line.append(str(percent)) print('\t'.join(line), file=out)
[ "def", "print_genome_matrix", "(", "hits", ",", "fastas", ",", "id2desc", ",", "file_name", ")", ":", "out", "=", "open", "(", "file_name", ",", "'w'", ")", "fastas", "=", "sorted", "(", "fastas", ")", "print", "(", "'## percent identity between genomes'", ",", "file", "=", "out", ")", "print", "(", "'# - \\t %s'", "%", "(", "'\\t'", ".", "join", "(", "fastas", ")", ")", ",", "file", "=", "out", ")", "for", "fasta", "in", "fastas", ":", "line", "=", "[", "fasta", "]", "for", "other", "in", "fastas", ":", "if", "other", "==", "fasta", ":", "average", "=", "'-'", "else", ":", "average", "=", "numpy", ".", "average", "(", "[", "hits", "[", "fasta", "]", "[", "other", "]", "[", "i", "]", "[", "3", "]", "for", "i", "in", "hits", "[", "fasta", "]", "[", "other", "]", "]", ")", "line", ".", "append", "(", "str", "(", "average", ")", ")", "print", "(", "'\\t'", ".", "join", "(", "line", ")", ",", "file", "=", "out", ")", "print", "(", "''", ",", "file", "=", "out", ")", "print", "(", "'## percent of orfs that are orthologous between genomes'", ",", "file", "=", "out", ")", "print", "(", "'# - \\t %s'", "%", "(", "'\\t'", ".", "join", "(", "fastas", ")", ")", ",", "file", "=", "out", ")", "for", "fasta", "in", "fastas", ":", "line", "=", "[", "fasta", "]", "for", "other", "in", "fastas", ":", "if", "other", "==", "fasta", ":", "percent", "=", "'-'", "else", ":", "orthologs", "=", "float", "(", "len", "(", "hits", "[", "fasta", "]", "[", "other", "]", ")", ")", "orfs", "=", "float", "(", "len", "(", "[", "i", "for", "i", "in", "id2desc", "if", "id2desc", "[", "i", "]", "[", "0", "]", "==", "fasta", "]", ")", ")", "percent", "=", "float", "(", "orthologs", "/", "orfs", ")", "*", "100", "line", ".", "append", "(", "str", "(", "percent", ")", ")", "print", "(", "'\\t'", ".", "join", "(", "line", ")", ",", "file", "=", "out", ")" ]
optimize later? slow ... should combine with calculate_threshold module
[ "optimize", "later?", "slow", "...", "should", "combine", "with", "calculate_threshold", "module" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/neto.py#L140-L171
train
christophertbrown/bioscripts
ctbBio/neto.py
self_compare
def self_compare(fastas, id2desc, algorithm): """ compare genome to self to get the best possible bit score for each ORF """ for fasta in fastas: blast = open(search(fasta, fasta, method = algorithm, alignment = 'local')) for hit in best_blast(blast, 1): id, bit = hit[0].split()[0], float(hit[-1]) id2desc[id].append(bit) return id2desc
python
def self_compare(fastas, id2desc, algorithm): """ compare genome to self to get the best possible bit score for each ORF """ for fasta in fastas: blast = open(search(fasta, fasta, method = algorithm, alignment = 'local')) for hit in best_blast(blast, 1): id, bit = hit[0].split()[0], float(hit[-1]) id2desc[id].append(bit) return id2desc
[ "def", "self_compare", "(", "fastas", ",", "id2desc", ",", "algorithm", ")", ":", "for", "fasta", "in", "fastas", ":", "blast", "=", "open", "(", "search", "(", "fasta", ",", "fasta", ",", "method", "=", "algorithm", ",", "alignment", "=", "'local'", ")", ")", "for", "hit", "in", "best_blast", "(", "blast", ",", "1", ")", ":", "id", ",", "bit", "=", "hit", "[", "0", "]", ".", "split", "(", ")", "[", "0", "]", ",", "float", "(", "hit", "[", "-", "1", "]", ")", "id2desc", "[", "id", "]", ".", "append", "(", "bit", ")", "return", "id2desc" ]
compare genome to self to get the best possible bit score for each ORF
[ "compare", "genome", "to", "self", "to", "get", "the", "best", "possible", "bit", "score", "for", "each", "ORF" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/neto.py#L240-L249
train
christophertbrown/bioscripts
ctbBio/neto.py
calc_thresholds
def calc_thresholds(rbh, file_name, thresholds = [False, False, False, False], stdevs = 2): """ if thresholds are not specififed, calculate based on the distribution of normalized bit scores """ calc_threshold = thresholds[-1] norm_threshold = {} for pair in itertools.permutations([i for i in rbh], 2): if pair[0] not in norm_threshold: norm_threshold[pair[0]] = {} norm_threshold[pair[0]][pair[1]] = {} out = open(file_name, 'w') print('#### summary of rbh comparisons\n', file=out) comparisons = [] for genome in rbh: for compare in rbh[genome]: pair = ''.join(sorted([genome, compare])) if pair in comparisons: continue comparisons.append(pair) scores = {'percent identity': [], 'e-value': [], 'bit score': [], 'normalized bit score': [], 'alignment length fraction': []} print('### blast between %s and %s\n' % (genome, compare), file=out) for id in rbh[genome][compare]: pident, length_fraction, e, bit, norm_bit = rbh[genome][compare][id][3:] scores['percent identity'].append(pident) scores['alignment length fraction'].append(length_fraction) scores['e-value'].append(e) scores['bit score'].append(bit) scores['normalized bit score'].append(norm_bit) if calc_threshold is True: norms = scores['normalized bit score'] average = numpy.average(norms) std = numpy.std(norms) normal_thresh = average - (std * stdevs) print('## average normalized bit score: %s' % average, file=out) print('## standard deviation of normalized bit scores: %s' % std, file=out) print('## normalized bit score threshold set to: %s\n' % (normal_thresh), file=out) norm_threshold[genome][compare], norm_threshold[compare][genome] = normal_thresh, normal_thresh for score in scores: print('## %s' % (score), file=out) if len(scores[score]) > 0: print('## average: %s' % numpy.average(scores[score]), file=out) # hist = histogram(scores[score], []) # for line in hist: # print >> out, line print('', file=out) out.close() if calc_threshold is True: return thresholds[0:-1] + [norm_threshold] else: return thresholds
python
def calc_thresholds(rbh, file_name, thresholds = [False, False, False, False], stdevs = 2): """ if thresholds are not specififed, calculate based on the distribution of normalized bit scores """ calc_threshold = thresholds[-1] norm_threshold = {} for pair in itertools.permutations([i for i in rbh], 2): if pair[0] not in norm_threshold: norm_threshold[pair[0]] = {} norm_threshold[pair[0]][pair[1]] = {} out = open(file_name, 'w') print('#### summary of rbh comparisons\n', file=out) comparisons = [] for genome in rbh: for compare in rbh[genome]: pair = ''.join(sorted([genome, compare])) if pair in comparisons: continue comparisons.append(pair) scores = {'percent identity': [], 'e-value': [], 'bit score': [], 'normalized bit score': [], 'alignment length fraction': []} print('### blast between %s and %s\n' % (genome, compare), file=out) for id in rbh[genome][compare]: pident, length_fraction, e, bit, norm_bit = rbh[genome][compare][id][3:] scores['percent identity'].append(pident) scores['alignment length fraction'].append(length_fraction) scores['e-value'].append(e) scores['bit score'].append(bit) scores['normalized bit score'].append(norm_bit) if calc_threshold is True: norms = scores['normalized bit score'] average = numpy.average(norms) std = numpy.std(norms) normal_thresh = average - (std * stdevs) print('## average normalized bit score: %s' % average, file=out) print('## standard deviation of normalized bit scores: %s' % std, file=out) print('## normalized bit score threshold set to: %s\n' % (normal_thresh), file=out) norm_threshold[genome][compare], norm_threshold[compare][genome] = normal_thresh, normal_thresh for score in scores: print('## %s' % (score), file=out) if len(scores[score]) > 0: print('## average: %s' % numpy.average(scores[score]), file=out) # hist = histogram(scores[score], []) # for line in hist: # print >> out, line print('', file=out) out.close() if calc_threshold is True: return thresholds[0:-1] + [norm_threshold] else: return thresholds
[ "def", "calc_thresholds", "(", "rbh", ",", "file_name", ",", "thresholds", "=", "[", "False", ",", "False", ",", "False", ",", "False", "]", ",", "stdevs", "=", "2", ")", ":", "calc_threshold", "=", "thresholds", "[", "-", "1", "]", "norm_threshold", "=", "{", "}", "for", "pair", "in", "itertools", ".", "permutations", "(", "[", "i", "for", "i", "in", "rbh", "]", ",", "2", ")", ":", "if", "pair", "[", "0", "]", "not", "in", "norm_threshold", ":", "norm_threshold", "[", "pair", "[", "0", "]", "]", "=", "{", "}", "norm_threshold", "[", "pair", "[", "0", "]", "]", "[", "pair", "[", "1", "]", "]", "=", "{", "}", "out", "=", "open", "(", "file_name", ",", "'w'", ")", "print", "(", "'#### summary of rbh comparisons\\n'", ",", "file", "=", "out", ")", "comparisons", "=", "[", "]", "for", "genome", "in", "rbh", ":", "for", "compare", "in", "rbh", "[", "genome", "]", ":", "pair", "=", "''", ".", "join", "(", "sorted", "(", "[", "genome", ",", "compare", "]", ")", ")", "if", "pair", "in", "comparisons", ":", "continue", "comparisons", ".", "append", "(", "pair", ")", "scores", "=", "{", "'percent identity'", ":", "[", "]", ",", "'e-value'", ":", "[", "]", ",", "'bit score'", ":", "[", "]", ",", "'normalized bit score'", ":", "[", "]", ",", "'alignment length fraction'", ":", "[", "]", "}", "print", "(", "'### blast between %s and %s\\n'", "%", "(", "genome", ",", "compare", ")", ",", "file", "=", "out", ")", "for", "id", "in", "rbh", "[", "genome", "]", "[", "compare", "]", ":", "pident", ",", "length_fraction", ",", "e", ",", "bit", ",", "norm_bit", "=", "rbh", "[", "genome", "]", "[", "compare", "]", "[", "id", "]", "[", "3", ":", "]", "scores", "[", "'percent identity'", "]", ".", "append", "(", "pident", ")", "scores", "[", "'alignment length fraction'", "]", ".", "append", "(", "length_fraction", ")", "scores", "[", "'e-value'", "]", ".", "append", "(", "e", ")", "scores", "[", "'bit score'", "]", ".", "append", "(", "bit", ")", "scores", "[", "'normalized bit score'", "]", ".", "append", "(", "norm_bit", ")", "if", "calc_threshold", "is", "True", ":", "norms", "=", "scores", "[", "'normalized bit score'", "]", "average", "=", "numpy", ".", "average", "(", "norms", ")", "std", "=", "numpy", ".", "std", "(", "norms", ")", "normal_thresh", "=", "average", "-", "(", "std", "*", "stdevs", ")", "print", "(", "'## average normalized bit score: %s'", "%", "average", ",", "file", "=", "out", ")", "print", "(", "'## standard deviation of normalized bit scores: %s'", "%", "std", ",", "file", "=", "out", ")", "print", "(", "'## normalized bit score threshold set to: %s\\n'", "%", "(", "normal_thresh", ")", ",", "file", "=", "out", ")", "norm_threshold", "[", "genome", "]", "[", "compare", "]", ",", "norm_threshold", "[", "compare", "]", "[", "genome", "]", "=", "normal_thresh", ",", "normal_thresh", "for", "score", "in", "scores", ":", "print", "(", "'## %s'", "%", "(", "score", ")", ",", "file", "=", "out", ")", "if", "len", "(", "scores", "[", "score", "]", ")", ">", "0", ":", "print", "(", "'## average: %s'", "%", "numpy", ".", "average", "(", "scores", "[", "score", "]", ")", ",", "file", "=", "out", ")", "# hist = histogram(scores[score], [])", "# for line in hist:", "# print >> out, line", "print", "(", "''", ",", "file", "=", "out", ")", "out", ".", "close", "(", ")", "if", "calc_threshold", "is", "True", ":", "return", "thresholds", "[", "0", ":", "-", "1", "]", "+", "[", "norm_threshold", "]", "else", ":", "return", "thresholds" ]
if thresholds are not specififed, calculate based on the distribution of normalized bit scores
[ "if", "thresholds", "are", "not", "specififed", "calculate", "based", "on", "the", "distribution", "of", "normalized", "bit", "scores" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/neto.py#L303-L352
train
christophertbrown/bioscripts
ctbBio/neto.py
neto
def neto(fastas, algorithm = 'usearch', e = 0.01, bit = 40, length = .65, norm_bit = False): """ make and split a rbh network """ thresholds = [e, bit, length, norm_bit] id2desc = get_descriptions(fastas) # get [fasta, description, length] for ORF id id2desc = self_compare(fastas, id2desc, algorithm) # get best possible bit score for each ORF # (comparing with itself) [fasta, description, length, bestbit] hits = compare_genomes(fastas, id2desc, algorithm) # pair wise genome comparisons {genome: {id: [match_type = 'rbh' or 'fbh', scores]}} calc_thresholds(hits, file_name = 'fbh.scores.summary.txt') rbh_network(id2desc, hits, file_name = 'fbh.network.edges.txt') hits, rbh = find_rbh(hits, id2desc) # remove hits that are not reciprocal best blast hits thresholds = calc_thresholds(rbh, 'rbh.scores.summary.txt', thresholds) # print rbh score summary to rbh_score_summary.txt and # calculate normalized bit score cutoff for each pair of # genomes, if desired g = rbh_network(id2desc, rbh, file_name = 'rbh.network.edges.txt') filtered_g, filtered_rbh = rbh_network(id2desc, rbh, 'rbh.filtered.network.edges.txt', thresholds) calc_thresholds(filtered_rbh, file_name = 'rbh.filtered.scores.summary.txt') print_summary(filtered_g, fastas, id2desc, file_name = 'rbh.filtered.network.nodes.txt') print_network_matrix(filtered_g, fastas, id2desc, file_name = 'rbh.filtered.network.matrix.txt') print_genome_matrix(filtered_rbh, fastas, id2desc, file_name = 'rbh.filtered.network.genome_matrix.txt') split_g = split_network(filtered_g, id2desc, file_name = 'rbh.filtered.split.network.edges.txt') print_summary(split_g, fastas, id2desc, file_name = 'rbh.filtered.split.network.nodes.txt') print_network_matrix(split_g, fastas, id2desc, file_name = 'rbh.filtered.split.network.matrix.txt') return split_g
python
def neto(fastas, algorithm = 'usearch', e = 0.01, bit = 40, length = .65, norm_bit = False): """ make and split a rbh network """ thresholds = [e, bit, length, norm_bit] id2desc = get_descriptions(fastas) # get [fasta, description, length] for ORF id id2desc = self_compare(fastas, id2desc, algorithm) # get best possible bit score for each ORF # (comparing with itself) [fasta, description, length, bestbit] hits = compare_genomes(fastas, id2desc, algorithm) # pair wise genome comparisons {genome: {id: [match_type = 'rbh' or 'fbh', scores]}} calc_thresholds(hits, file_name = 'fbh.scores.summary.txt') rbh_network(id2desc, hits, file_name = 'fbh.network.edges.txt') hits, rbh = find_rbh(hits, id2desc) # remove hits that are not reciprocal best blast hits thresholds = calc_thresholds(rbh, 'rbh.scores.summary.txt', thresholds) # print rbh score summary to rbh_score_summary.txt and # calculate normalized bit score cutoff for each pair of # genomes, if desired g = rbh_network(id2desc, rbh, file_name = 'rbh.network.edges.txt') filtered_g, filtered_rbh = rbh_network(id2desc, rbh, 'rbh.filtered.network.edges.txt', thresholds) calc_thresholds(filtered_rbh, file_name = 'rbh.filtered.scores.summary.txt') print_summary(filtered_g, fastas, id2desc, file_name = 'rbh.filtered.network.nodes.txt') print_network_matrix(filtered_g, fastas, id2desc, file_name = 'rbh.filtered.network.matrix.txt') print_genome_matrix(filtered_rbh, fastas, id2desc, file_name = 'rbh.filtered.network.genome_matrix.txt') split_g = split_network(filtered_g, id2desc, file_name = 'rbh.filtered.split.network.edges.txt') print_summary(split_g, fastas, id2desc, file_name = 'rbh.filtered.split.network.nodes.txt') print_network_matrix(split_g, fastas, id2desc, file_name = 'rbh.filtered.split.network.matrix.txt') return split_g
[ "def", "neto", "(", "fastas", ",", "algorithm", "=", "'usearch'", ",", "e", "=", "0.01", ",", "bit", "=", "40", ",", "length", "=", ".65", ",", "norm_bit", "=", "False", ")", ":", "thresholds", "=", "[", "e", ",", "bit", ",", "length", ",", "norm_bit", "]", "id2desc", "=", "get_descriptions", "(", "fastas", ")", "# get [fasta, description, length] for ORF id", "id2desc", "=", "self_compare", "(", "fastas", ",", "id2desc", ",", "algorithm", ")", "# get best possible bit score for each ORF ", "# (comparing with itself) [fasta, description, length, bestbit]", "hits", "=", "compare_genomes", "(", "fastas", ",", "id2desc", ",", "algorithm", ")", "# pair wise genome comparisons {genome: {id: [match_type = 'rbh' or 'fbh', scores]}}", "calc_thresholds", "(", "hits", ",", "file_name", "=", "'fbh.scores.summary.txt'", ")", "rbh_network", "(", "id2desc", ",", "hits", ",", "file_name", "=", "'fbh.network.edges.txt'", ")", "hits", ",", "rbh", "=", "find_rbh", "(", "hits", ",", "id2desc", ")", "# remove hits that are not reciprocal best blast hits", "thresholds", "=", "calc_thresholds", "(", "rbh", ",", "'rbh.scores.summary.txt'", ",", "thresholds", ")", "# print rbh score summary to rbh_score_summary.txt and", "# calculate normalized bit score cutoff for each pair of", "# genomes, if desired", "g", "=", "rbh_network", "(", "id2desc", ",", "rbh", ",", "file_name", "=", "'rbh.network.edges.txt'", ")", "filtered_g", ",", "filtered_rbh", "=", "rbh_network", "(", "id2desc", ",", "rbh", ",", "'rbh.filtered.network.edges.txt'", ",", "thresholds", ")", "calc_thresholds", "(", "filtered_rbh", ",", "file_name", "=", "'rbh.filtered.scores.summary.txt'", ")", "print_summary", "(", "filtered_g", ",", "fastas", ",", "id2desc", ",", "file_name", "=", "'rbh.filtered.network.nodes.txt'", ")", "print_network_matrix", "(", "filtered_g", ",", "fastas", ",", "id2desc", ",", "file_name", "=", "'rbh.filtered.network.matrix.txt'", ")", "print_genome_matrix", "(", "filtered_rbh", ",", "fastas", ",", "id2desc", ",", "file_name", "=", "'rbh.filtered.network.genome_matrix.txt'", ")", "split_g", "=", "split_network", "(", "filtered_g", ",", "id2desc", ",", "file_name", "=", "'rbh.filtered.split.network.edges.txt'", ")", "print_summary", "(", "split_g", ",", "fastas", ",", "id2desc", ",", "file_name", "=", "'rbh.filtered.split.network.nodes.txt'", ")", "print_network_matrix", "(", "split_g", ",", "fastas", ",", "id2desc", ",", "file_name", "=", "'rbh.filtered.split.network.matrix.txt'", ")", "return", "split_g" ]
make and split a rbh network
[ "make", "and", "split", "a", "rbh", "network" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/neto.py#L416-L445
train
consbio/gis-metadata-parser
gis_metadata/iso_metadata_parser.py
IsoParser._parse_raster_info
def _parse_raster_info(self, prop=RASTER_INFO): """ Collapses multiple dimensions into a single raster_info complex struct """ raster_info = {}.fromkeys(_iso_definitions[prop], u'') # Ensure conversion of lists to newlines is in place raster_info['dimensions'] = get_default_for_complex_sub( prop=prop, subprop='dimensions', value=parse_property(self._xml_tree, None, self._data_map, '_ri_num_dims'), xpath=self._data_map['_ri_num_dims'] ) xpath_root = self._get_xroot_for(prop) xpath_map = self._data_structures[prop] for dimension in parse_complex_list(self._xml_tree, xpath_root, xpath_map, RASTER_DIMS): dimension_type = dimension['type'].lower() if dimension_type == 'vertical': raster_info['vertical_count'] = dimension['size'] elif dimension_type == 'column': raster_info['column_count'] = dimension['size'] raster_info['x_resolution'] = u' '.join(dimension[k] for k in ['value', 'units']).strip() elif dimension_type == 'row': raster_info['row_count'] = dimension['size'] raster_info['y_resolution'] = u' '.join(dimension[k] for k in ['value', 'units']).strip() return raster_info if any(raster_info[k] for k in raster_info) else {}
python
def _parse_raster_info(self, prop=RASTER_INFO): """ Collapses multiple dimensions into a single raster_info complex struct """ raster_info = {}.fromkeys(_iso_definitions[prop], u'') # Ensure conversion of lists to newlines is in place raster_info['dimensions'] = get_default_for_complex_sub( prop=prop, subprop='dimensions', value=parse_property(self._xml_tree, None, self._data_map, '_ri_num_dims'), xpath=self._data_map['_ri_num_dims'] ) xpath_root = self._get_xroot_for(prop) xpath_map = self._data_structures[prop] for dimension in parse_complex_list(self._xml_tree, xpath_root, xpath_map, RASTER_DIMS): dimension_type = dimension['type'].lower() if dimension_type == 'vertical': raster_info['vertical_count'] = dimension['size'] elif dimension_type == 'column': raster_info['column_count'] = dimension['size'] raster_info['x_resolution'] = u' '.join(dimension[k] for k in ['value', 'units']).strip() elif dimension_type == 'row': raster_info['row_count'] = dimension['size'] raster_info['y_resolution'] = u' '.join(dimension[k] for k in ['value', 'units']).strip() return raster_info if any(raster_info[k] for k in raster_info) else {}
[ "def", "_parse_raster_info", "(", "self", ",", "prop", "=", "RASTER_INFO", ")", ":", "raster_info", "=", "{", "}", ".", "fromkeys", "(", "_iso_definitions", "[", "prop", "]", ",", "u''", ")", "# Ensure conversion of lists to newlines is in place", "raster_info", "[", "'dimensions'", "]", "=", "get_default_for_complex_sub", "(", "prop", "=", "prop", ",", "subprop", "=", "'dimensions'", ",", "value", "=", "parse_property", "(", "self", ".", "_xml_tree", ",", "None", ",", "self", ".", "_data_map", ",", "'_ri_num_dims'", ")", ",", "xpath", "=", "self", ".", "_data_map", "[", "'_ri_num_dims'", "]", ")", "xpath_root", "=", "self", ".", "_get_xroot_for", "(", "prop", ")", "xpath_map", "=", "self", ".", "_data_structures", "[", "prop", "]", "for", "dimension", "in", "parse_complex_list", "(", "self", ".", "_xml_tree", ",", "xpath_root", ",", "xpath_map", ",", "RASTER_DIMS", ")", ":", "dimension_type", "=", "dimension", "[", "'type'", "]", ".", "lower", "(", ")", "if", "dimension_type", "==", "'vertical'", ":", "raster_info", "[", "'vertical_count'", "]", "=", "dimension", "[", "'size'", "]", "elif", "dimension_type", "==", "'column'", ":", "raster_info", "[", "'column_count'", "]", "=", "dimension", "[", "'size'", "]", "raster_info", "[", "'x_resolution'", "]", "=", "u' '", ".", "join", "(", "dimension", "[", "k", "]", "for", "k", "in", "[", "'value'", ",", "'units'", "]", ")", ".", "strip", "(", ")", "elif", "dimension_type", "==", "'row'", ":", "raster_info", "[", "'row_count'", "]", "=", "dimension", "[", "'size'", "]", "raster_info", "[", "'y_resolution'", "]", "=", "u' '", ".", "join", "(", "dimension", "[", "k", "]", "for", "k", "in", "[", "'value'", ",", "'units'", "]", ")", ".", "strip", "(", ")", "return", "raster_info", "if", "any", "(", "raster_info", "[", "k", "]", "for", "k", "in", "raster_info", ")", "else", "{", "}" ]
Collapses multiple dimensions into a single raster_info complex struct
[ "Collapses", "multiple", "dimensions", "into", "a", "single", "raster_info", "complex", "struct" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/iso_metadata_parser.py#L472-L502
train
consbio/gis-metadata-parser
gis_metadata/iso_metadata_parser.py
IsoParser._update_raster_info
def _update_raster_info(self, **update_props): """ Derives multiple dimensions from a single raster_info complex struct """ tree_to_update = update_props['tree_to_update'] prop = update_props['prop'] values = update_props.pop('values') # Update number of dimensions at raster_info root (applies to all dimensions below) xroot, xpath = None, self._data_map['_ri_num_dims'] raster_info = [update_property(tree_to_update, xroot, xpath, prop, values.get('dimensions', u''))] # Derive vertical, longitude, and latitude dimensions from raster_info xpath_root = self._get_xroot_for(prop) xpath_map = self._data_structures[prop] v_dimension = {} if values.get('vertical_count'): v_dimension = v_dimension.fromkeys(xpath_map, u'') v_dimension['type'] = 'vertical' v_dimension['size'] = values.get('vertical_count', u'') x_dimension = {} if values.get('column_count') or values.get('x_resolution'): x_dimension = x_dimension.fromkeys(xpath_map, u'') x_dimension['type'] = 'column' x_dimension['size'] = values.get('column_count', u'') x_dimension['value'] = values.get('x_resolution', u'') y_dimension = {} if values.get('row_count') or values.get('y_resolution'): y_dimension = y_dimension.fromkeys(xpath_map, u'') y_dimension['type'] = 'row' y_dimension['size'] = values.get('row_count', u'') y_dimension['value'] = values.get('y_resolution', u'') # Update derived dimensions as complex list, and append affected elements for return update_props['prop'] = RASTER_DIMS update_props['values'] = [v_dimension, x_dimension, y_dimension] raster_info += update_complex_list(xpath_root=xpath_root, xpath_map=xpath_map, **update_props) return raster_info
python
def _update_raster_info(self, **update_props): """ Derives multiple dimensions from a single raster_info complex struct """ tree_to_update = update_props['tree_to_update'] prop = update_props['prop'] values = update_props.pop('values') # Update number of dimensions at raster_info root (applies to all dimensions below) xroot, xpath = None, self._data_map['_ri_num_dims'] raster_info = [update_property(tree_to_update, xroot, xpath, prop, values.get('dimensions', u''))] # Derive vertical, longitude, and latitude dimensions from raster_info xpath_root = self._get_xroot_for(prop) xpath_map = self._data_structures[prop] v_dimension = {} if values.get('vertical_count'): v_dimension = v_dimension.fromkeys(xpath_map, u'') v_dimension['type'] = 'vertical' v_dimension['size'] = values.get('vertical_count', u'') x_dimension = {} if values.get('column_count') or values.get('x_resolution'): x_dimension = x_dimension.fromkeys(xpath_map, u'') x_dimension['type'] = 'column' x_dimension['size'] = values.get('column_count', u'') x_dimension['value'] = values.get('x_resolution', u'') y_dimension = {} if values.get('row_count') or values.get('y_resolution'): y_dimension = y_dimension.fromkeys(xpath_map, u'') y_dimension['type'] = 'row' y_dimension['size'] = values.get('row_count', u'') y_dimension['value'] = values.get('y_resolution', u'') # Update derived dimensions as complex list, and append affected elements for return update_props['prop'] = RASTER_DIMS update_props['values'] = [v_dimension, x_dimension, y_dimension] raster_info += update_complex_list(xpath_root=xpath_root, xpath_map=xpath_map, **update_props) return raster_info
[ "def", "_update_raster_info", "(", "self", ",", "*", "*", "update_props", ")", ":", "tree_to_update", "=", "update_props", "[", "'tree_to_update'", "]", "prop", "=", "update_props", "[", "'prop'", "]", "values", "=", "update_props", ".", "pop", "(", "'values'", ")", "# Update number of dimensions at raster_info root (applies to all dimensions below)", "xroot", ",", "xpath", "=", "None", ",", "self", ".", "_data_map", "[", "'_ri_num_dims'", "]", "raster_info", "=", "[", "update_property", "(", "tree_to_update", ",", "xroot", ",", "xpath", ",", "prop", ",", "values", ".", "get", "(", "'dimensions'", ",", "u''", ")", ")", "]", "# Derive vertical, longitude, and latitude dimensions from raster_info", "xpath_root", "=", "self", ".", "_get_xroot_for", "(", "prop", ")", "xpath_map", "=", "self", ".", "_data_structures", "[", "prop", "]", "v_dimension", "=", "{", "}", "if", "values", ".", "get", "(", "'vertical_count'", ")", ":", "v_dimension", "=", "v_dimension", ".", "fromkeys", "(", "xpath_map", ",", "u''", ")", "v_dimension", "[", "'type'", "]", "=", "'vertical'", "v_dimension", "[", "'size'", "]", "=", "values", ".", "get", "(", "'vertical_count'", ",", "u''", ")", "x_dimension", "=", "{", "}", "if", "values", ".", "get", "(", "'column_count'", ")", "or", "values", ".", "get", "(", "'x_resolution'", ")", ":", "x_dimension", "=", "x_dimension", ".", "fromkeys", "(", "xpath_map", ",", "u''", ")", "x_dimension", "[", "'type'", "]", "=", "'column'", "x_dimension", "[", "'size'", "]", "=", "values", ".", "get", "(", "'column_count'", ",", "u''", ")", "x_dimension", "[", "'value'", "]", "=", "values", ".", "get", "(", "'x_resolution'", ",", "u''", ")", "y_dimension", "=", "{", "}", "if", "values", ".", "get", "(", "'row_count'", ")", "or", "values", ".", "get", "(", "'y_resolution'", ")", ":", "y_dimension", "=", "y_dimension", ".", "fromkeys", "(", "xpath_map", ",", "u''", ")", "y_dimension", "[", "'type'", "]", "=", "'row'", "y_dimension", "[", "'size'", "]", "=", "values", ".", "get", "(", "'row_count'", ",", "u''", ")", "y_dimension", "[", "'value'", "]", "=", "values", ".", "get", "(", "'y_resolution'", ",", "u''", ")", "# Update derived dimensions as complex list, and append affected elements for return", "update_props", "[", "'prop'", "]", "=", "RASTER_DIMS", "update_props", "[", "'values'", "]", "=", "[", "v_dimension", ",", "x_dimension", ",", "y_dimension", "]", "raster_info", "+=", "update_complex_list", "(", "xpath_root", "=", "xpath_root", ",", "xpath_map", "=", "xpath_map", ",", "*", "*", "update_props", ")", "return", "raster_info" ]
Derives multiple dimensions from a single raster_info complex struct
[ "Derives", "multiple", "dimensions", "from", "a", "single", "raster_info", "complex", "struct" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/iso_metadata_parser.py#L622-L666
train
consbio/gis-metadata-parser
gis_metadata/iso_metadata_parser.py
IsoParser._trim_xpath
def _trim_xpath(self, xpath, prop): """ Removes primitive type tags from an XPATH """ xroot = self._get_xroot_for(prop) if xroot is None and isinstance(xpath, string_types): xtags = xpath.split(XPATH_DELIM) if xtags[-1] in _iso_tag_primitives: xroot = XPATH_DELIM.join(xtags[:-1]) return xroot
python
def _trim_xpath(self, xpath, prop): """ Removes primitive type tags from an XPATH """ xroot = self._get_xroot_for(prop) if xroot is None and isinstance(xpath, string_types): xtags = xpath.split(XPATH_DELIM) if xtags[-1] in _iso_tag_primitives: xroot = XPATH_DELIM.join(xtags[:-1]) return xroot
[ "def", "_trim_xpath", "(", "self", ",", "xpath", ",", "prop", ")", ":", "xroot", "=", "self", ".", "_get_xroot_for", "(", "prop", ")", "if", "xroot", "is", "None", "and", "isinstance", "(", "xpath", ",", "string_types", ")", ":", "xtags", "=", "xpath", ".", "split", "(", "XPATH_DELIM", ")", "if", "xtags", "[", "-", "1", "]", "in", "_iso_tag_primitives", ":", "xroot", "=", "XPATH_DELIM", ".", "join", "(", "xtags", "[", ":", "-", "1", "]", ")", "return", "xroot" ]
Removes primitive type tags from an XPATH
[ "Removes", "primitive", "type", "tags", "from", "an", "XPATH" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/iso_metadata_parser.py#L691-L702
train
scottrice/pysteam
pysteam/shortcuts.py
shortcut_app_id
def shortcut_app_id(shortcut): """ Generates the app id for a given shortcut. Steam uses app ids as a unique identifier for games, but since shortcuts dont have a canonical serverside representation they need to be generated on the fly. The important part about this function is that it will generate the same app id as Steam does for a given shortcut """ algorithm = Crc(width = 32, poly = 0x04C11DB7, reflect_in = True, xor_in = 0xffffffff, reflect_out = True, xor_out = 0xffffffff) crc_input = ''.join([shortcut.exe,shortcut.name]) high_32 = algorithm.bit_by_bit(crc_input) | 0x80000000 full_64 = (high_32 << 32) | 0x02000000 return str(full_64)
python
def shortcut_app_id(shortcut): """ Generates the app id for a given shortcut. Steam uses app ids as a unique identifier for games, but since shortcuts dont have a canonical serverside representation they need to be generated on the fly. The important part about this function is that it will generate the same app id as Steam does for a given shortcut """ algorithm = Crc(width = 32, poly = 0x04C11DB7, reflect_in = True, xor_in = 0xffffffff, reflect_out = True, xor_out = 0xffffffff) crc_input = ''.join([shortcut.exe,shortcut.name]) high_32 = algorithm.bit_by_bit(crc_input) | 0x80000000 full_64 = (high_32 << 32) | 0x02000000 return str(full_64)
[ "def", "shortcut_app_id", "(", "shortcut", ")", ":", "algorithm", "=", "Crc", "(", "width", "=", "32", ",", "poly", "=", "0x04C11DB7", ",", "reflect_in", "=", "True", ",", "xor_in", "=", "0xffffffff", ",", "reflect_out", "=", "True", ",", "xor_out", "=", "0xffffffff", ")", "crc_input", "=", "''", ".", "join", "(", "[", "shortcut", ".", "exe", ",", "shortcut", ".", "name", "]", ")", "high_32", "=", "algorithm", ".", "bit_by_bit", "(", "crc_input", ")", "|", "0x80000000", "full_64", "=", "(", "high_32", "<<", "32", ")", "|", "0x02000000", "return", "str", "(", "full_64", ")" ]
Generates the app id for a given shortcut. Steam uses app ids as a unique identifier for games, but since shortcuts dont have a canonical serverside representation they need to be generated on the fly. The important part about this function is that it will generate the same app id as Steam does for a given shortcut
[ "Generates", "the", "app", "id", "for", "a", "given", "shortcut", ".", "Steam", "uses", "app", "ids", "as", "a", "unique", "identifier", "for", "games", "but", "since", "shortcuts", "dont", "have", "a", "canonical", "serverside", "representation", "they", "need", "to", "be", "generated", "on", "the", "fly", ".", "The", "important", "part", "about", "this", "function", "is", "that", "it", "will", "generate", "the", "same", "app", "id", "as", "Steam", "does", "for", "a", "given", "shortcut" ]
1eb2254b5235a053a953e596fa7602d0b110245d
https://github.com/scottrice/pysteam/blob/1eb2254b5235a053a953e596fa7602d0b110245d/pysteam/shortcuts.py#L9-L21
train
mkouhei/bootstrap-py
bootstrap_py/vcs.py
VCS._config
def _config(self): """Execute git config.""" cfg_wr = self.repo.config_writer() cfg_wr.add_section('user') cfg_wr.set_value('user', 'name', self.metadata.author) cfg_wr.set_value('user', 'email', self.metadata.email) cfg_wr.release()
python
def _config(self): """Execute git config.""" cfg_wr = self.repo.config_writer() cfg_wr.add_section('user') cfg_wr.set_value('user', 'name', self.metadata.author) cfg_wr.set_value('user', 'email', self.metadata.email) cfg_wr.release()
[ "def", "_config", "(", "self", ")", ":", "cfg_wr", "=", "self", ".", "repo", ".", "config_writer", "(", ")", "cfg_wr", ".", "add_section", "(", "'user'", ")", "cfg_wr", ".", "set_value", "(", "'user'", ",", "'name'", ",", "self", ".", "metadata", ".", "author", ")", "cfg_wr", ".", "set_value", "(", "'user'", ",", "'email'", ",", "self", ".", "metadata", ".", "email", ")", "cfg_wr", ".", "release", "(", ")" ]
Execute git config.
[ "Execute", "git", "config", "." ]
95d56ed98ef409fd9f019dc352fd1c3711533275
https://github.com/mkouhei/bootstrap-py/blob/95d56ed98ef409fd9f019dc352fd1c3711533275/bootstrap_py/vcs.py#L35-L41
train
mkouhei/bootstrap-py
bootstrap_py/vcs.py
VCS._remote_add
def _remote_add(self): """Execute git remote add.""" self.repo.create_remote( 'origin', 'git@github.com:{username}/{repo}.git'.format( username=self.metadata.username, repo=self.metadata.name))
python
def _remote_add(self): """Execute git remote add.""" self.repo.create_remote( 'origin', 'git@github.com:{username}/{repo}.git'.format( username=self.metadata.username, repo=self.metadata.name))
[ "def", "_remote_add", "(", "self", ")", ":", "self", ".", "repo", ".", "create_remote", "(", "'origin'", ",", "'git@github.com:{username}/{repo}.git'", ".", "format", "(", "username", "=", "self", ".", "metadata", ".", "username", ",", "repo", "=", "self", ".", "metadata", ".", "name", ")", ")" ]
Execute git remote add.
[ "Execute", "git", "remote", "add", "." ]
95d56ed98ef409fd9f019dc352fd1c3711533275
https://github.com/mkouhei/bootstrap-py/blob/95d56ed98ef409fd9f019dc352fd1c3711533275/bootstrap_py/vcs.py#L47-L53
train
deep-compute/basescript
basescript/basescript.py
BaseScript.start
def start(self): ''' Starts execution of the script ''' # invoke the appropriate sub-command as requested from command-line try: self.args.func() except SystemExit as e: if e.code != 0: raise except KeyboardInterrupt: self.log.warning("exited via keyboard interrupt") except: self.log.exception("exited start function") # set exit code so we know it did not end successfully # TODO different exit codes based on signals ? finally: self._flush_metrics_q.put(None, block=True) self._flush_metrics_q.put(None, block=True, timeout=1) self.log.debug("exited_successfully")
python
def start(self): ''' Starts execution of the script ''' # invoke the appropriate sub-command as requested from command-line try: self.args.func() except SystemExit as e: if e.code != 0: raise except KeyboardInterrupt: self.log.warning("exited via keyboard interrupt") except: self.log.exception("exited start function") # set exit code so we know it did not end successfully # TODO different exit codes based on signals ? finally: self._flush_metrics_q.put(None, block=True) self._flush_metrics_q.put(None, block=True, timeout=1) self.log.debug("exited_successfully")
[ "def", "start", "(", "self", ")", ":", "# invoke the appropriate sub-command as requested from command-line", "try", ":", "self", ".", "args", ".", "func", "(", ")", "except", "SystemExit", "as", "e", ":", "if", "e", ".", "code", "!=", "0", ":", "raise", "except", "KeyboardInterrupt", ":", "self", ".", "log", ".", "warning", "(", "\"exited via keyboard interrupt\"", ")", "except", ":", "self", ".", "log", ".", "exception", "(", "\"exited start function\"", ")", "# set exit code so we know it did not end successfully", "# TODO different exit codes based on signals ?", "finally", ":", "self", ".", "_flush_metrics_q", ".", "put", "(", "None", ",", "block", "=", "True", ")", "self", ".", "_flush_metrics_q", ".", "put", "(", "None", ",", "block", "=", "True", ",", "timeout", "=", "1", ")", "self", ".", "log", ".", "debug", "(", "\"exited_successfully\"", ")" ]
Starts execution of the script
[ "Starts", "execution", "of", "the", "script" ]
f7233963c5291530fcb2444a7f45b556e6407b90
https://github.com/deep-compute/basescript/blob/f7233963c5291530fcb2444a7f45b556e6407b90/basescript/basescript.py#L67-L87
train
deep-compute/basescript
basescript/basescript.py
BaseScript.define_baseargs
def define_baseargs(self, parser): ''' Define basic command-line arguments required by the script. @parser is a parser object created using the `argparse` module. returns: None ''' parser.add_argument('--name', default=sys.argv[0], help='Name to identify this instance') parser.add_argument('--log-level', default=None, help='Logging level as picked from the logging module') parser.add_argument('--log-format', default=None, # TODO add more formats choices=("json", "pretty",), help=("Force the format of the logs. By default, if the " "command is from a terminal, print colorful logs. " "Otherwise print json."), ) parser.add_argument('--log-file', default=None, help='Writes logs to log file if specified, default: %(default)s', ) parser.add_argument('--quiet', default=False, action="store_true", help='if true, does not print logs to stderr, default: %(default)s', ) parser.add_argument('--metric-grouping-interval', default=None, type=int, help='To group metrics based on time interval ex:10 i.e;(10 sec)', ) parser.add_argument('--debug', default=False, action="store_true", help='To run the code in debug mode', )
python
def define_baseargs(self, parser): ''' Define basic command-line arguments required by the script. @parser is a parser object created using the `argparse` module. returns: None ''' parser.add_argument('--name', default=sys.argv[0], help='Name to identify this instance') parser.add_argument('--log-level', default=None, help='Logging level as picked from the logging module') parser.add_argument('--log-format', default=None, # TODO add more formats choices=("json", "pretty",), help=("Force the format of the logs. By default, if the " "command is from a terminal, print colorful logs. " "Otherwise print json."), ) parser.add_argument('--log-file', default=None, help='Writes logs to log file if specified, default: %(default)s', ) parser.add_argument('--quiet', default=False, action="store_true", help='if true, does not print logs to stderr, default: %(default)s', ) parser.add_argument('--metric-grouping-interval', default=None, type=int, help='To group metrics based on time interval ex:10 i.e;(10 sec)', ) parser.add_argument('--debug', default=False, action="store_true", help='To run the code in debug mode', )
[ "def", "define_baseargs", "(", "self", ",", "parser", ")", ":", "parser", ".", "add_argument", "(", "'--name'", ",", "default", "=", "sys", ".", "argv", "[", "0", "]", ",", "help", "=", "'Name to identify this instance'", ")", "parser", ".", "add_argument", "(", "'--log-level'", ",", "default", "=", "None", ",", "help", "=", "'Logging level as picked from the logging module'", ")", "parser", ".", "add_argument", "(", "'--log-format'", ",", "default", "=", "None", ",", "# TODO add more formats", "choices", "=", "(", "\"json\"", ",", "\"pretty\"", ",", ")", ",", "help", "=", "(", "\"Force the format of the logs. By default, if the \"", "\"command is from a terminal, print colorful logs. \"", "\"Otherwise print json.\"", ")", ",", ")", "parser", ".", "add_argument", "(", "'--log-file'", ",", "default", "=", "None", ",", "help", "=", "'Writes logs to log file if specified, default: %(default)s'", ",", ")", "parser", ".", "add_argument", "(", "'--quiet'", ",", "default", "=", "False", ",", "action", "=", "\"store_true\"", ",", "help", "=", "'if true, does not print logs to stderr, default: %(default)s'", ",", ")", "parser", ".", "add_argument", "(", "'--metric-grouping-interval'", ",", "default", "=", "None", ",", "type", "=", "int", ",", "help", "=", "'To group metrics based on time interval ex:10 i.e;(10 sec)'", ",", ")", "parser", ".", "add_argument", "(", "'--debug'", ",", "default", "=", "False", ",", "action", "=", "\"store_true\"", ",", "help", "=", "'To run the code in debug mode'", ",", ")" ]
Define basic command-line arguments required by the script. @parser is a parser object created using the `argparse` module. returns: None
[ "Define", "basic", "command", "-", "line", "arguments", "required", "by", "the", "script", "." ]
f7233963c5291530fcb2444a7f45b556e6407b90
https://github.com/deep-compute/basescript/blob/f7233963c5291530fcb2444a7f45b556e6407b90/basescript/basescript.py#L123-L151
train
elbow-jason/Uno-deprecated
uno/parser/source_coder.py
SourceCoder.cleanup_payload
def cleanup_payload(self, payload): """ Basically, turns payload that looks like ' \\n ' to ''. In the calling function, if this function returns '' no object is added for that payload. """ p = payload.replace('\n', '') p = p.rstrip() p = p.lstrip() return p
python
def cleanup_payload(self, payload): """ Basically, turns payload that looks like ' \\n ' to ''. In the calling function, if this function returns '' no object is added for that payload. """ p = payload.replace('\n', '') p = p.rstrip() p = p.lstrip() return p
[ "def", "cleanup_payload", "(", "self", ",", "payload", ")", ":", "p", "=", "payload", ".", "replace", "(", "'\\n'", ",", "''", ")", "p", "=", "p", ".", "rstrip", "(", ")", "p", "=", "p", ".", "lstrip", "(", ")", "return", "p" ]
Basically, turns payload that looks like ' \\n ' to ''. In the calling function, if this function returns '' no object is added for that payload.
[ "Basically", "turns", "payload", "that", "looks", "like", "\\\\", "n", "to", ".", "In", "the", "calling", "function", "if", "this", "function", "returns", "no", "object", "is", "added", "for", "that", "payload", "." ]
4ad07d7b84e5b6e3e2b2c89db69448906f24b4e4
https://github.com/elbow-jason/Uno-deprecated/blob/4ad07d7b84e5b6e3e2b2c89db69448906f24b4e4/uno/parser/source_coder.py#L73-L82
train
consbio/gis-metadata-parser
gis_metadata/utils.py
get_default_for
def get_default_for(prop, value): """ Ensures complex property types have the correct default values """ prop = prop.strip('_') # Handle alternate props (leading underscores) val = reduce_value(value) # Filtering of value happens here if prop in _COMPLEX_LISTS: return wrap_value(val) elif prop in _COMPLEX_STRUCTS: return val or {} else: return u'' if val is None else val
python
def get_default_for(prop, value): """ Ensures complex property types have the correct default values """ prop = prop.strip('_') # Handle alternate props (leading underscores) val = reduce_value(value) # Filtering of value happens here if prop in _COMPLEX_LISTS: return wrap_value(val) elif prop in _COMPLEX_STRUCTS: return val or {} else: return u'' if val is None else val
[ "def", "get_default_for", "(", "prop", ",", "value", ")", ":", "prop", "=", "prop", ".", "strip", "(", "'_'", ")", "# Handle alternate props (leading underscores)", "val", "=", "reduce_value", "(", "value", ")", "# Filtering of value happens here", "if", "prop", "in", "_COMPLEX_LISTS", ":", "return", "wrap_value", "(", "val", ")", "elif", "prop", "in", "_COMPLEX_STRUCTS", ":", "return", "val", "or", "{", "}", "else", ":", "return", "u''", "if", "val", "is", "None", "else", "val" ]
Ensures complex property types have the correct default values
[ "Ensures", "complex", "property", "types", "have", "the", "correct", "default", "values" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L223-L234
train
consbio/gis-metadata-parser
gis_metadata/utils.py
update_property
def update_property(tree_to_update, xpath_root, xpaths, prop, values, supported=None): """ Either update the tree the default way, or call the custom updater Default Way: Existing values in the tree are overwritten. If xpaths contains a single path, then each value is written to the tree at that path. If xpaths contains a list of xpaths, then the values corresponding to each xpath index are written to their respective locations. In either case, empty values are ignored. :param tree_to_update: the XML tree compatible with element_utils to be updated :param xpath_root: the XPATH location shared by all the xpaths passed in :param xpaths: a string or a list of strings representing the XPATH location(s) to which to write values :param prop: the name of the property of the parser containing the value(s) with which to update the tree :param values: a single value, or a list of values to write to the specified XPATHs :see: ParserProperty for more on custom updaters :return: a list of all elements updated by this operation """ if supported and prop.startswith('_') and prop.strip('_') in supported: values = u'' # Remove alternate elements: write values only to primary location else: values = get_default_for(prop, values) # Enforce defaults as required per property if not xpaths: return [] elif not isinstance(xpaths, ParserProperty): return _update_property(tree_to_update, xpath_root, xpaths, values) else: # Call ParserProperty.set_prop without xpath_root (managed internally) return xpaths.set_prop(tree_to_update=tree_to_update, prop=prop, values=values)
python
def update_property(tree_to_update, xpath_root, xpaths, prop, values, supported=None): """ Either update the tree the default way, or call the custom updater Default Way: Existing values in the tree are overwritten. If xpaths contains a single path, then each value is written to the tree at that path. If xpaths contains a list of xpaths, then the values corresponding to each xpath index are written to their respective locations. In either case, empty values are ignored. :param tree_to_update: the XML tree compatible with element_utils to be updated :param xpath_root: the XPATH location shared by all the xpaths passed in :param xpaths: a string or a list of strings representing the XPATH location(s) to which to write values :param prop: the name of the property of the parser containing the value(s) with which to update the tree :param values: a single value, or a list of values to write to the specified XPATHs :see: ParserProperty for more on custom updaters :return: a list of all elements updated by this operation """ if supported and prop.startswith('_') and prop.strip('_') in supported: values = u'' # Remove alternate elements: write values only to primary location else: values = get_default_for(prop, values) # Enforce defaults as required per property if not xpaths: return [] elif not isinstance(xpaths, ParserProperty): return _update_property(tree_to_update, xpath_root, xpaths, values) else: # Call ParserProperty.set_prop without xpath_root (managed internally) return xpaths.set_prop(tree_to_update=tree_to_update, prop=prop, values=values)
[ "def", "update_property", "(", "tree_to_update", ",", "xpath_root", ",", "xpaths", ",", "prop", ",", "values", ",", "supported", "=", "None", ")", ":", "if", "supported", "and", "prop", ".", "startswith", "(", "'_'", ")", "and", "prop", ".", "strip", "(", "'_'", ")", "in", "supported", ":", "values", "=", "u''", "# Remove alternate elements: write values only to primary location", "else", ":", "values", "=", "get_default_for", "(", "prop", ",", "values", ")", "# Enforce defaults as required per property", "if", "not", "xpaths", ":", "return", "[", "]", "elif", "not", "isinstance", "(", "xpaths", ",", "ParserProperty", ")", ":", "return", "_update_property", "(", "tree_to_update", ",", "xpath_root", ",", "xpaths", ",", "values", ")", "else", ":", "# Call ParserProperty.set_prop without xpath_root (managed internally)", "return", "xpaths", ".", "set_prop", "(", "tree_to_update", "=", "tree_to_update", ",", "prop", "=", "prop", ",", "values", "=", "values", ")" ]
Either update the tree the default way, or call the custom updater Default Way: Existing values in the tree are overwritten. If xpaths contains a single path, then each value is written to the tree at that path. If xpaths contains a list of xpaths, then the values corresponding to each xpath index are written to their respective locations. In either case, empty values are ignored. :param tree_to_update: the XML tree compatible with element_utils to be updated :param xpath_root: the XPATH location shared by all the xpaths passed in :param xpaths: a string or a list of strings representing the XPATH location(s) to which to write values :param prop: the name of the property of the parser containing the value(s) with which to update the tree :param values: a single value, or a list of values to write to the specified XPATHs :see: ParserProperty for more on custom updaters :return: a list of all elements updated by this operation
[ "Either", "update", "the", "tree", "the", "default", "way", "or", "call", "the", "custom", "updater" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L392-L423
train
consbio/gis-metadata-parser
gis_metadata/utils.py
_update_property
def _update_property(tree_to_update, xpath_root, xpaths, values): """ Default update operation for a single parser property. If xpaths contains one xpath, then one element per value will be inserted at that location in the tree_to_update; otherwise, the number of values must match the number of xpaths. """ # Inner function to update a specific XPATH with the values provided def update_element(elem, idx, root, path, vals): """ Internal helper function to encapsulate single item update """ has_root = bool(root and len(path) > len(root) and path.startswith(root)) path, attr = get_xpath_tuple(path) # 'path/@attr' to ('path', 'attr') if attr: removed = [get_element(elem, path)] remove_element_attributes(removed[0], attr) elif not has_root: removed = wrap_value(remove_element(elem, path)) else: path = get_xpath_branch(root, path) removed = [] if idx != 0 else [remove_element(e, path, True) for e in get_elements(elem, root)] if not vals: return removed items = [] for i, val in enumerate(wrap_value(vals)): elem_to_update = elem if has_root: elem_to_update = insert_element(elem, (i + idx), root) val = val.decode('utf-8') if not isinstance(val, string_types) else val if not attr: items.append(insert_element(elem_to_update, i, path, val)) elif path: items.append(insert_element(elem_to_update, i, path, **{attr: val})) else: set_element_attributes(elem_to_update, **{attr: val}) items.append(elem_to_update) return items # Code to update each of the XPATHs with each of the values xpaths = reduce_value(xpaths) values = filter_empty(values) if isinstance(xpaths, string_types): return update_element(tree_to_update, 0, xpath_root, xpaths, values) else: each = [] for index, xpath in enumerate(xpaths): value = values[index] if values else None each.extend(update_element(tree_to_update, index, xpath_root, xpath, value)) return each
python
def _update_property(tree_to_update, xpath_root, xpaths, values): """ Default update operation for a single parser property. If xpaths contains one xpath, then one element per value will be inserted at that location in the tree_to_update; otherwise, the number of values must match the number of xpaths. """ # Inner function to update a specific XPATH with the values provided def update_element(elem, idx, root, path, vals): """ Internal helper function to encapsulate single item update """ has_root = bool(root and len(path) > len(root) and path.startswith(root)) path, attr = get_xpath_tuple(path) # 'path/@attr' to ('path', 'attr') if attr: removed = [get_element(elem, path)] remove_element_attributes(removed[0], attr) elif not has_root: removed = wrap_value(remove_element(elem, path)) else: path = get_xpath_branch(root, path) removed = [] if idx != 0 else [remove_element(e, path, True) for e in get_elements(elem, root)] if not vals: return removed items = [] for i, val in enumerate(wrap_value(vals)): elem_to_update = elem if has_root: elem_to_update = insert_element(elem, (i + idx), root) val = val.decode('utf-8') if not isinstance(val, string_types) else val if not attr: items.append(insert_element(elem_to_update, i, path, val)) elif path: items.append(insert_element(elem_to_update, i, path, **{attr: val})) else: set_element_attributes(elem_to_update, **{attr: val}) items.append(elem_to_update) return items # Code to update each of the XPATHs with each of the values xpaths = reduce_value(xpaths) values = filter_empty(values) if isinstance(xpaths, string_types): return update_element(tree_to_update, 0, xpath_root, xpaths, values) else: each = [] for index, xpath in enumerate(xpaths): value = values[index] if values else None each.extend(update_element(tree_to_update, index, xpath_root, xpath, value)) return each
[ "def", "_update_property", "(", "tree_to_update", ",", "xpath_root", ",", "xpaths", ",", "values", ")", ":", "# Inner function to update a specific XPATH with the values provided", "def", "update_element", "(", "elem", ",", "idx", ",", "root", ",", "path", ",", "vals", ")", ":", "\"\"\" Internal helper function to encapsulate single item update \"\"\"", "has_root", "=", "bool", "(", "root", "and", "len", "(", "path", ")", ">", "len", "(", "root", ")", "and", "path", ".", "startswith", "(", "root", ")", ")", "path", ",", "attr", "=", "get_xpath_tuple", "(", "path", ")", "# 'path/@attr' to ('path', 'attr')", "if", "attr", ":", "removed", "=", "[", "get_element", "(", "elem", ",", "path", ")", "]", "remove_element_attributes", "(", "removed", "[", "0", "]", ",", "attr", ")", "elif", "not", "has_root", ":", "removed", "=", "wrap_value", "(", "remove_element", "(", "elem", ",", "path", ")", ")", "else", ":", "path", "=", "get_xpath_branch", "(", "root", ",", "path", ")", "removed", "=", "[", "]", "if", "idx", "!=", "0", "else", "[", "remove_element", "(", "e", ",", "path", ",", "True", ")", "for", "e", "in", "get_elements", "(", "elem", ",", "root", ")", "]", "if", "not", "vals", ":", "return", "removed", "items", "=", "[", "]", "for", "i", ",", "val", "in", "enumerate", "(", "wrap_value", "(", "vals", ")", ")", ":", "elem_to_update", "=", "elem", "if", "has_root", ":", "elem_to_update", "=", "insert_element", "(", "elem", ",", "(", "i", "+", "idx", ")", ",", "root", ")", "val", "=", "val", ".", "decode", "(", "'utf-8'", ")", "if", "not", "isinstance", "(", "val", ",", "string_types", ")", "else", "val", "if", "not", "attr", ":", "items", ".", "append", "(", "insert_element", "(", "elem_to_update", ",", "i", ",", "path", ",", "val", ")", ")", "elif", "path", ":", "items", ".", "append", "(", "insert_element", "(", "elem_to_update", ",", "i", ",", "path", ",", "*", "*", "{", "attr", ":", "val", "}", ")", ")", "else", ":", "set_element_attributes", "(", "elem_to_update", ",", "*", "*", "{", "attr", ":", "val", "}", ")", "items", ".", "append", "(", "elem_to_update", ")", "return", "items", "# Code to update each of the XPATHs with each of the values", "xpaths", "=", "reduce_value", "(", "xpaths", ")", "values", "=", "filter_empty", "(", "values", ")", "if", "isinstance", "(", "xpaths", ",", "string_types", ")", ":", "return", "update_element", "(", "tree_to_update", ",", "0", ",", "xpath_root", ",", "xpaths", ",", "values", ")", "else", ":", "each", "=", "[", "]", "for", "index", ",", "xpath", "in", "enumerate", "(", "xpaths", ")", ":", "value", "=", "values", "[", "index", "]", "if", "values", "else", "None", "each", ".", "extend", "(", "update_element", "(", "tree_to_update", ",", "index", ",", "xpath_root", ",", "xpath", ",", "value", ")", ")", "return", "each" ]
Default update operation for a single parser property. If xpaths contains one xpath, then one element per value will be inserted at that location in the tree_to_update; otherwise, the number of values must match the number of xpaths.
[ "Default", "update", "operation", "for", "a", "single", "parser", "property", ".", "If", "xpaths", "contains", "one", "xpath", "then", "one", "element", "per", "value", "will", "be", "inserted", "at", "that", "location", "in", "the", "tree_to_update", ";", "otherwise", "the", "number", "of", "values", "must", "match", "the", "number", "of", "xpaths", "." ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L426-L486
train
consbio/gis-metadata-parser
gis_metadata/utils.py
validate_complex
def validate_complex(prop, value, xpath_map=None): """ Default validation for single complex data structure """ if value is not None: validate_type(prop, value, dict) if prop in _complex_definitions: complex_keys = _complex_definitions[prop] else: complex_keys = {} if xpath_map is None else xpath_map for complex_prop, complex_val in iteritems(value): complex_key = '.'.join((prop, complex_prop)) if complex_prop not in complex_keys: _validation_error(prop, None, value, ('keys: {0}'.format(','.join(complex_keys)))) validate_type(complex_key, complex_val, (string_types, list))
python
def validate_complex(prop, value, xpath_map=None): """ Default validation for single complex data structure """ if value is not None: validate_type(prop, value, dict) if prop in _complex_definitions: complex_keys = _complex_definitions[prop] else: complex_keys = {} if xpath_map is None else xpath_map for complex_prop, complex_val in iteritems(value): complex_key = '.'.join((prop, complex_prop)) if complex_prop not in complex_keys: _validation_error(prop, None, value, ('keys: {0}'.format(','.join(complex_keys)))) validate_type(complex_key, complex_val, (string_types, list))
[ "def", "validate_complex", "(", "prop", ",", "value", ",", "xpath_map", "=", "None", ")", ":", "if", "value", "is", "not", "None", ":", "validate_type", "(", "prop", ",", "value", ",", "dict", ")", "if", "prop", "in", "_complex_definitions", ":", "complex_keys", "=", "_complex_definitions", "[", "prop", "]", "else", ":", "complex_keys", "=", "{", "}", "if", "xpath_map", "is", "None", "else", "xpath_map", "for", "complex_prop", ",", "complex_val", "in", "iteritems", "(", "value", ")", ":", "complex_key", "=", "'.'", ".", "join", "(", "(", "prop", ",", "complex_prop", ")", ")", "if", "complex_prop", "not", "in", "complex_keys", ":", "_validation_error", "(", "prop", ",", "None", ",", "value", ",", "(", "'keys: {0}'", ".", "format", "(", "','", ".", "join", "(", "complex_keys", ")", ")", ")", ")", "validate_type", "(", "complex_key", ",", "complex_val", ",", "(", "string_types", ",", "list", ")", ")" ]
Default validation for single complex data structure
[ "Default", "validation", "for", "single", "complex", "data", "structure" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L572-L589
train
consbio/gis-metadata-parser
gis_metadata/utils.py
validate_complex_list
def validate_complex_list(prop, value, xpath_map=None): """ Default validation for Attribute Details data structure """ if value is not None: validate_type(prop, value, (dict, list)) if prop in _complex_definitions: complex_keys = _complex_definitions[prop] else: complex_keys = {} if xpath_map is None else xpath_map for idx, complex_struct in enumerate(wrap_value(value)): cs_idx = prop + '[' + str(idx) + ']' validate_type(cs_idx, complex_struct, dict) for cs_prop, cs_val in iteritems(complex_struct): cs_key = '.'.join((cs_idx, cs_prop)) if cs_prop not in complex_keys: _validation_error(prop, None, value, ('keys: {0}'.format(','.join(complex_keys)))) if not isinstance(cs_val, list): validate_type(cs_key, cs_val, (string_types, list)) else: for list_idx, list_val in enumerate(cs_val): list_prop = cs_key + '[' + str(list_idx) + ']' validate_type(list_prop, list_val, string_types)
python
def validate_complex_list(prop, value, xpath_map=None): """ Default validation for Attribute Details data structure """ if value is not None: validate_type(prop, value, (dict, list)) if prop in _complex_definitions: complex_keys = _complex_definitions[prop] else: complex_keys = {} if xpath_map is None else xpath_map for idx, complex_struct in enumerate(wrap_value(value)): cs_idx = prop + '[' + str(idx) + ']' validate_type(cs_idx, complex_struct, dict) for cs_prop, cs_val in iteritems(complex_struct): cs_key = '.'.join((cs_idx, cs_prop)) if cs_prop not in complex_keys: _validation_error(prop, None, value, ('keys: {0}'.format(','.join(complex_keys)))) if not isinstance(cs_val, list): validate_type(cs_key, cs_val, (string_types, list)) else: for list_idx, list_val in enumerate(cs_val): list_prop = cs_key + '[' + str(list_idx) + ']' validate_type(list_prop, list_val, string_types)
[ "def", "validate_complex_list", "(", "prop", ",", "value", ",", "xpath_map", "=", "None", ")", ":", "if", "value", "is", "not", "None", ":", "validate_type", "(", "prop", ",", "value", ",", "(", "dict", ",", "list", ")", ")", "if", "prop", "in", "_complex_definitions", ":", "complex_keys", "=", "_complex_definitions", "[", "prop", "]", "else", ":", "complex_keys", "=", "{", "}", "if", "xpath_map", "is", "None", "else", "xpath_map", "for", "idx", ",", "complex_struct", "in", "enumerate", "(", "wrap_value", "(", "value", ")", ")", ":", "cs_idx", "=", "prop", "+", "'['", "+", "str", "(", "idx", ")", "+", "']'", "validate_type", "(", "cs_idx", ",", "complex_struct", ",", "dict", ")", "for", "cs_prop", ",", "cs_val", "in", "iteritems", "(", "complex_struct", ")", ":", "cs_key", "=", "'.'", ".", "join", "(", "(", "cs_idx", ",", "cs_prop", ")", ")", "if", "cs_prop", "not", "in", "complex_keys", ":", "_validation_error", "(", "prop", ",", "None", ",", "value", ",", "(", "'keys: {0}'", ".", "format", "(", "','", ".", "join", "(", "complex_keys", ")", ")", ")", ")", "if", "not", "isinstance", "(", "cs_val", ",", "list", ")", ":", "validate_type", "(", "cs_key", ",", "cs_val", ",", "(", "string_types", ",", "list", ")", ")", "else", ":", "for", "list_idx", ",", "list_val", "in", "enumerate", "(", "cs_val", ")", ":", "list_prop", "=", "cs_key", "+", "'['", "+", "str", "(", "list_idx", ")", "+", "']'", "validate_type", "(", "list_prop", ",", "list_val", ",", "string_types", ")" ]
Default validation for Attribute Details data structure
[ "Default", "validation", "for", "Attribute", "Details", "data", "structure" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L592-L618
train
consbio/gis-metadata-parser
gis_metadata/utils.py
validate_dates
def validate_dates(prop, value, xpath_map=None): """ Default validation for Date Types data structure """ if value is not None: validate_type(prop, value, dict) date_keys = set(value) if date_keys: if DATE_TYPE not in date_keys or DATE_VALUES not in date_keys: if prop in _complex_definitions: complex_keys = _complex_definitions[prop] else: complex_keys = _complex_definitions[DATES] if xpath_map is None else xpath_map _validation_error(prop, None, value, ('keys: {0}'.format(','.join(complex_keys)))) date_type = value[DATE_TYPE] if date_type not in DATE_TYPES: _validation_error('dates.type', None, date_type, DATE_TYPES) date_vals = value[DATE_VALUES] validate_type('dates.values', date_vals, list) dates_len = len(date_vals) if date_type == DATE_TYPE_MISSING and dates_len != 0: _validation_error('len(dates.values)', None, dates_len, 0) if date_type == DATE_TYPE_SINGLE and dates_len != 1: _validation_error('len(dates.values)', None, dates_len, 1) if date_type == DATE_TYPE_RANGE and dates_len != 2: _validation_error('len(dates.values)', None, dates_len, 2) if date_type == DATE_TYPE_MULTIPLE and dates_len < 2: _validation_error('len(dates.values)', None, dates_len, 'at least two') for idx, date in enumerate(date_vals): date_key = 'dates.value[' + str(idx) + ']' validate_type(date_key, date, string_types)
python
def validate_dates(prop, value, xpath_map=None): """ Default validation for Date Types data structure """ if value is not None: validate_type(prop, value, dict) date_keys = set(value) if date_keys: if DATE_TYPE not in date_keys or DATE_VALUES not in date_keys: if prop in _complex_definitions: complex_keys = _complex_definitions[prop] else: complex_keys = _complex_definitions[DATES] if xpath_map is None else xpath_map _validation_error(prop, None, value, ('keys: {0}'.format(','.join(complex_keys)))) date_type = value[DATE_TYPE] if date_type not in DATE_TYPES: _validation_error('dates.type', None, date_type, DATE_TYPES) date_vals = value[DATE_VALUES] validate_type('dates.values', date_vals, list) dates_len = len(date_vals) if date_type == DATE_TYPE_MISSING and dates_len != 0: _validation_error('len(dates.values)', None, dates_len, 0) if date_type == DATE_TYPE_SINGLE and dates_len != 1: _validation_error('len(dates.values)', None, dates_len, 1) if date_type == DATE_TYPE_RANGE and dates_len != 2: _validation_error('len(dates.values)', None, dates_len, 2) if date_type == DATE_TYPE_MULTIPLE and dates_len < 2: _validation_error('len(dates.values)', None, dates_len, 'at least two') for idx, date in enumerate(date_vals): date_key = 'dates.value[' + str(idx) + ']' validate_type(date_key, date, string_types)
[ "def", "validate_dates", "(", "prop", ",", "value", ",", "xpath_map", "=", "None", ")", ":", "if", "value", "is", "not", "None", ":", "validate_type", "(", "prop", ",", "value", ",", "dict", ")", "date_keys", "=", "set", "(", "value", ")", "if", "date_keys", ":", "if", "DATE_TYPE", "not", "in", "date_keys", "or", "DATE_VALUES", "not", "in", "date_keys", ":", "if", "prop", "in", "_complex_definitions", ":", "complex_keys", "=", "_complex_definitions", "[", "prop", "]", "else", ":", "complex_keys", "=", "_complex_definitions", "[", "DATES", "]", "if", "xpath_map", "is", "None", "else", "xpath_map", "_validation_error", "(", "prop", ",", "None", ",", "value", ",", "(", "'keys: {0}'", ".", "format", "(", "','", ".", "join", "(", "complex_keys", ")", ")", ")", ")", "date_type", "=", "value", "[", "DATE_TYPE", "]", "if", "date_type", "not", "in", "DATE_TYPES", ":", "_validation_error", "(", "'dates.type'", ",", "None", ",", "date_type", ",", "DATE_TYPES", ")", "date_vals", "=", "value", "[", "DATE_VALUES", "]", "validate_type", "(", "'dates.values'", ",", "date_vals", ",", "list", ")", "dates_len", "=", "len", "(", "date_vals", ")", "if", "date_type", "==", "DATE_TYPE_MISSING", "and", "dates_len", "!=", "0", ":", "_validation_error", "(", "'len(dates.values)'", ",", "None", ",", "dates_len", ",", "0", ")", "if", "date_type", "==", "DATE_TYPE_SINGLE", "and", "dates_len", "!=", "1", ":", "_validation_error", "(", "'len(dates.values)'", ",", "None", ",", "dates_len", ",", "1", ")", "if", "date_type", "==", "DATE_TYPE_RANGE", "and", "dates_len", "!=", "2", ":", "_validation_error", "(", "'len(dates.values)'", ",", "None", ",", "dates_len", ",", "2", ")", "if", "date_type", "==", "DATE_TYPE_MULTIPLE", "and", "dates_len", "<", "2", ":", "_validation_error", "(", "'len(dates.values)'", ",", "None", ",", "dates_len", ",", "'at least two'", ")", "for", "idx", ",", "date", "in", "enumerate", "(", "date_vals", ")", ":", "date_key", "=", "'dates.value['", "+", "str", "(", "idx", ")", "+", "']'", "validate_type", "(", "date_key", ",", "date", ",", "string_types", ")" ]
Default validation for Date Types data structure
[ "Default", "validation", "for", "Date", "Types", "data", "structure" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L621-L663
train
consbio/gis-metadata-parser
gis_metadata/utils.py
validate_process_steps
def validate_process_steps(prop, value): """ Default validation for Process Steps data structure """ if value is not None: validate_type(prop, value, (dict, list)) procstep_keys = set(_complex_definitions[prop]) for idx, procstep in enumerate(wrap_value(value)): ps_idx = prop + '[' + str(idx) + ']' validate_type(ps_idx, procstep, dict) for ps_prop, ps_val in iteritems(procstep): ps_key = '.'.join((ps_idx, ps_prop)) if ps_prop not in procstep_keys: _validation_error(prop, None, value, ('keys: {0}'.format(','.join(procstep_keys)))) if ps_prop != 'sources': validate_type(ps_key, ps_val, string_types) else: validate_type(ps_key, ps_val, (string_types, list)) for src_idx, src_val in enumerate(wrap_value(ps_val)): src_key = ps_key + '[' + str(src_idx) + ']' validate_type(src_key, src_val, string_types)
python
def validate_process_steps(prop, value): """ Default validation for Process Steps data structure """ if value is not None: validate_type(prop, value, (dict, list)) procstep_keys = set(_complex_definitions[prop]) for idx, procstep in enumerate(wrap_value(value)): ps_idx = prop + '[' + str(idx) + ']' validate_type(ps_idx, procstep, dict) for ps_prop, ps_val in iteritems(procstep): ps_key = '.'.join((ps_idx, ps_prop)) if ps_prop not in procstep_keys: _validation_error(prop, None, value, ('keys: {0}'.format(','.join(procstep_keys)))) if ps_prop != 'sources': validate_type(ps_key, ps_val, string_types) else: validate_type(ps_key, ps_val, (string_types, list)) for src_idx, src_val in enumerate(wrap_value(ps_val)): src_key = ps_key + '[' + str(src_idx) + ']' validate_type(src_key, src_val, string_types)
[ "def", "validate_process_steps", "(", "prop", ",", "value", ")", ":", "if", "value", "is", "not", "None", ":", "validate_type", "(", "prop", ",", "value", ",", "(", "dict", ",", "list", ")", ")", "procstep_keys", "=", "set", "(", "_complex_definitions", "[", "prop", "]", ")", "for", "idx", ",", "procstep", "in", "enumerate", "(", "wrap_value", "(", "value", ")", ")", ":", "ps_idx", "=", "prop", "+", "'['", "+", "str", "(", "idx", ")", "+", "']'", "validate_type", "(", "ps_idx", ",", "procstep", ",", "dict", ")", "for", "ps_prop", ",", "ps_val", "in", "iteritems", "(", "procstep", ")", ":", "ps_key", "=", "'.'", ".", "join", "(", "(", "ps_idx", ",", "ps_prop", ")", ")", "if", "ps_prop", "not", "in", "procstep_keys", ":", "_validation_error", "(", "prop", ",", "None", ",", "value", ",", "(", "'keys: {0}'", ".", "format", "(", "','", ".", "join", "(", "procstep_keys", ")", ")", ")", ")", "if", "ps_prop", "!=", "'sources'", ":", "validate_type", "(", "ps_key", ",", "ps_val", ",", "string_types", ")", "else", ":", "validate_type", "(", "ps_key", ",", "ps_val", ",", "(", "string_types", ",", "list", ")", ")", "for", "src_idx", ",", "src_val", "in", "enumerate", "(", "wrap_value", "(", "ps_val", ")", ")", ":", "src_key", "=", "ps_key", "+", "'['", "+", "str", "(", "src_idx", ")", "+", "']'", "validate_type", "(", "src_key", ",", "src_val", ",", "string_types", ")" ]
Default validation for Process Steps data structure
[ "Default", "validation", "for", "Process", "Steps", "data", "structure" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L666-L691
train
consbio/gis-metadata-parser
gis_metadata/utils.py
validate_type
def validate_type(prop, value, expected): """ Default validation for all types """ # Validate on expected type(s), but ignore None: defaults handled elsewhere if value is not None and not isinstance(value, expected): _validation_error(prop, type(value).__name__, None, expected)
python
def validate_type(prop, value, expected): """ Default validation for all types """ # Validate on expected type(s), but ignore None: defaults handled elsewhere if value is not None and not isinstance(value, expected): _validation_error(prop, type(value).__name__, None, expected)
[ "def", "validate_type", "(", "prop", ",", "value", ",", "expected", ")", ":", "# Validate on expected type(s), but ignore None: defaults handled elsewhere", "if", "value", "is", "not", "None", "and", "not", "isinstance", "(", "value", ",", "expected", ")", ":", "_validation_error", "(", "prop", ",", "type", "(", "value", ")", ".", "__name__", ",", "None", ",", "expected", ")" ]
Default validation for all types
[ "Default", "validation", "for", "all", "types" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L710-L715
train
consbio/gis-metadata-parser
gis_metadata/utils.py
_validation_error
def _validation_error(prop, prop_type, prop_value, expected): """ Default validation for updated properties """ if prop_type is None: attrib = 'value' assigned = prop_value else: attrib = 'type' assigned = prop_type raise ValidationError( 'Invalid property {attrib} for {prop}:\n\t{attrib}: {assigned}\n\texpected: {expected}', attrib=attrib, prop=prop, assigned=assigned, expected=expected, invalid={prop: prop_value} if attrib == 'value' else {} )
python
def _validation_error(prop, prop_type, prop_value, expected): """ Default validation for updated properties """ if prop_type is None: attrib = 'value' assigned = prop_value else: attrib = 'type' assigned = prop_type raise ValidationError( 'Invalid property {attrib} for {prop}:\n\t{attrib}: {assigned}\n\texpected: {expected}', attrib=attrib, prop=prop, assigned=assigned, expected=expected, invalid={prop: prop_value} if attrib == 'value' else {} )
[ "def", "_validation_error", "(", "prop", ",", "prop_type", ",", "prop_value", ",", "expected", ")", ":", "if", "prop_type", "is", "None", ":", "attrib", "=", "'value'", "assigned", "=", "prop_value", "else", ":", "attrib", "=", "'type'", "assigned", "=", "prop_type", "raise", "ValidationError", "(", "'Invalid property {attrib} for {prop}:\\n\\t{attrib}: {assigned}\\n\\texpected: {expected}'", ",", "attrib", "=", "attrib", ",", "prop", "=", "prop", ",", "assigned", "=", "assigned", ",", "expected", "=", "expected", ",", "invalid", "=", "{", "prop", ":", "prop_value", "}", "if", "attrib", "==", "'value'", "else", "{", "}", ")" ]
Default validation for updated properties
[ "Default", "validation", "for", "updated", "properties" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L718-L732
train
consbio/gis-metadata-parser
gis_metadata/utils.py
ParserProperty.get_prop
def get_prop(self, prop): """ Calls the getter with no arguments and returns its value """ if self._parser is None: raise ConfigurationError('Cannot call ParserProperty."get_prop" with no parser configured') return self._parser(prop) if prop else self._parser()
python
def get_prop(self, prop): """ Calls the getter with no arguments and returns its value """ if self._parser is None: raise ConfigurationError('Cannot call ParserProperty."get_prop" with no parser configured') return self._parser(prop) if prop else self._parser()
[ "def", "get_prop", "(", "self", ",", "prop", ")", ":", "if", "self", ".", "_parser", "is", "None", ":", "raise", "ConfigurationError", "(", "'Cannot call ParserProperty.\"get_prop\" with no parser configured'", ")", "return", "self", ".", "_parser", "(", "prop", ")", "if", "prop", "else", "self", ".", "_parser", "(", ")" ]
Calls the getter with no arguments and returns its value
[ "Calls", "the", "getter", "with", "no", "arguments", "and", "returns", "its", "value" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/utils.py#L765-L771
train
disqus/nydus
nydus/db/backends/memcache.py
can_group_commands
def can_group_commands(command, next_command): """ Returns a boolean representing whether these commands can be grouped together or not. A few things are taken into account for this decision: For ``set`` commands: - Are all arguments other than the key/value the same? For ``delete`` and ``get`` commands: - Are all arguments other than the key the same? """ multi_capable_commands = ('get', 'set', 'delete') if next_command is None: return False name = command.get_name() # TODO: support multi commands if name not in multi_capable_commands: return False if name != next_command.get_name(): return False # if the shared args (key, or key/value) do not match, we cannot group if grouped_args_for_command(command) != grouped_args_for_command(next_command): return False # If the keyword arguments do not much (e.g. key_prefix, or timeout on set) # then we cannot group if command.get_kwargs() != next_command.get_kwargs(): return False return True
python
def can_group_commands(command, next_command): """ Returns a boolean representing whether these commands can be grouped together or not. A few things are taken into account for this decision: For ``set`` commands: - Are all arguments other than the key/value the same? For ``delete`` and ``get`` commands: - Are all arguments other than the key the same? """ multi_capable_commands = ('get', 'set', 'delete') if next_command is None: return False name = command.get_name() # TODO: support multi commands if name not in multi_capable_commands: return False if name != next_command.get_name(): return False # if the shared args (key, or key/value) do not match, we cannot group if grouped_args_for_command(command) != grouped_args_for_command(next_command): return False # If the keyword arguments do not much (e.g. key_prefix, or timeout on set) # then we cannot group if command.get_kwargs() != next_command.get_kwargs(): return False return True
[ "def", "can_group_commands", "(", "command", ",", "next_command", ")", ":", "multi_capable_commands", "=", "(", "'get'", ",", "'set'", ",", "'delete'", ")", "if", "next_command", "is", "None", ":", "return", "False", "name", "=", "command", ".", "get_name", "(", ")", "# TODO: support multi commands", "if", "name", "not", "in", "multi_capable_commands", ":", "return", "False", "if", "name", "!=", "next_command", ".", "get_name", "(", ")", ":", "return", "False", "# if the shared args (key, or key/value) do not match, we cannot group", "if", "grouped_args_for_command", "(", "command", ")", "!=", "grouped_args_for_command", "(", "next_command", ")", ":", "return", "False", "# If the keyword arguments do not much (e.g. key_prefix, or timeout on set)", "# then we cannot group", "if", "command", ".", "get_kwargs", "(", ")", "!=", "next_command", ".", "get_kwargs", "(", ")", ":", "return", "False", "return", "True" ]
Returns a boolean representing whether these commands can be grouped together or not. A few things are taken into account for this decision: For ``set`` commands: - Are all arguments other than the key/value the same? For ``delete`` and ``get`` commands: - Are all arguments other than the key the same?
[ "Returns", "a", "boolean", "representing", "whether", "these", "commands", "can", "be", "grouped", "together", "or", "not", "." ]
9b505840da47a34f758a830c3992fa5dcb7bb7ad
https://github.com/disqus/nydus/blob/9b505840da47a34f758a830c3992fa5dcb7bb7ad/nydus/db/backends/memcache.py#L97-L135
train
christophertbrown/bioscripts
ctbBio/rp16.py
find_databases
def find_databases(databases): """ define ribosomal proteins and location of curated databases """ # 16 ribosomal proteins in their expected order proteins = ['L15', 'L18', 'L6', 'S8', 'L5', 'L24', 'L14', 'S17', 'L16', 'S3', 'L22', 'S19', 'L2', 'L4', 'L3', 'S10'] # curated databases protein_databases = { 'L14': 'rpL14_JGI_MDM.filtered.faa', 'L15': 'rpL15_JGI_MDM.filtered.faa', 'L16': 'rpL16_JGI_MDM.filtered.faa', 'L18': 'rpL18_JGI_MDM.filtered.faa', 'L22': 'rpL22_JGI_MDM.filtered.faa', 'L24': 'rpL24_JGI_MDM.filtered.faa', 'L2': 'rpL2_JGI_MDM.filtered.faa', 'L3': 'rpL3_JGI_MDM.filtered.faa', 'L4': 'rpL4_JGI_MDM.filtered.faa', 'L5': 'rpL5_JGI_MDM.filtered.faa', 'L6': 'rpL6_JGI_MDM.filtered.faa', 'S10': 'rpS10_JGI_MDM.filtered.faa', 'S17': 'rpS17_JGI_MDM.filtered.faa', 'S19': 'rpS19_JGI_MDM.filtered.faa', 'S3': 'rpS3_JGI_MDM.filtered.faa', 'S8': 'rpS8_JGI_MDM.filtered.faa'} protein_databases = {key: '%s/%s' % (databases, database) \ for key, database in list(protein_databases.items())} return proteins, protein_databases
python
def find_databases(databases): """ define ribosomal proteins and location of curated databases """ # 16 ribosomal proteins in their expected order proteins = ['L15', 'L18', 'L6', 'S8', 'L5', 'L24', 'L14', 'S17', 'L16', 'S3', 'L22', 'S19', 'L2', 'L4', 'L3', 'S10'] # curated databases protein_databases = { 'L14': 'rpL14_JGI_MDM.filtered.faa', 'L15': 'rpL15_JGI_MDM.filtered.faa', 'L16': 'rpL16_JGI_MDM.filtered.faa', 'L18': 'rpL18_JGI_MDM.filtered.faa', 'L22': 'rpL22_JGI_MDM.filtered.faa', 'L24': 'rpL24_JGI_MDM.filtered.faa', 'L2': 'rpL2_JGI_MDM.filtered.faa', 'L3': 'rpL3_JGI_MDM.filtered.faa', 'L4': 'rpL4_JGI_MDM.filtered.faa', 'L5': 'rpL5_JGI_MDM.filtered.faa', 'L6': 'rpL6_JGI_MDM.filtered.faa', 'S10': 'rpS10_JGI_MDM.filtered.faa', 'S17': 'rpS17_JGI_MDM.filtered.faa', 'S19': 'rpS19_JGI_MDM.filtered.faa', 'S3': 'rpS3_JGI_MDM.filtered.faa', 'S8': 'rpS8_JGI_MDM.filtered.faa'} protein_databases = {key: '%s/%s' % (databases, database) \ for key, database in list(protein_databases.items())} return proteins, protein_databases
[ "def", "find_databases", "(", "databases", ")", ":", "# 16 ribosomal proteins in their expected order", "proteins", "=", "[", "'L15'", ",", "'L18'", ",", "'L6'", ",", "'S8'", ",", "'L5'", ",", "'L24'", ",", "'L14'", ",", "'S17'", ",", "'L16'", ",", "'S3'", ",", "'L22'", ",", "'S19'", ",", "'L2'", ",", "'L4'", ",", "'L3'", ",", "'S10'", "]", "# curated databases", "protein_databases", "=", "{", "'L14'", ":", "'rpL14_JGI_MDM.filtered.faa'", ",", "'L15'", ":", "'rpL15_JGI_MDM.filtered.faa'", ",", "'L16'", ":", "'rpL16_JGI_MDM.filtered.faa'", ",", "'L18'", ":", "'rpL18_JGI_MDM.filtered.faa'", ",", "'L22'", ":", "'rpL22_JGI_MDM.filtered.faa'", ",", "'L24'", ":", "'rpL24_JGI_MDM.filtered.faa'", ",", "'L2'", ":", "'rpL2_JGI_MDM.filtered.faa'", ",", "'L3'", ":", "'rpL3_JGI_MDM.filtered.faa'", ",", "'L4'", ":", "'rpL4_JGI_MDM.filtered.faa'", ",", "'L5'", ":", "'rpL5_JGI_MDM.filtered.faa'", ",", "'L6'", ":", "'rpL6_JGI_MDM.filtered.faa'", ",", "'S10'", ":", "'rpS10_JGI_MDM.filtered.faa'", ",", "'S17'", ":", "'rpS17_JGI_MDM.filtered.faa'", ",", "'S19'", ":", "'rpS19_JGI_MDM.filtered.faa'", ",", "'S3'", ":", "'rpS3_JGI_MDM.filtered.faa'", ",", "'S8'", ":", "'rpS8_JGI_MDM.filtered.faa'", "}", "protein_databases", "=", "{", "key", ":", "'%s/%s'", "%", "(", "databases", ",", "database", ")", "for", "key", ",", "database", "in", "list", "(", "protein_databases", ".", "items", "(", ")", ")", "}", "return", "proteins", ",", "protein_databases" ]
define ribosomal proteins and location of curated databases
[ "define", "ribosomal", "proteins", "and", "location", "of", "curated", "databases" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rp16.py#L21-L48
train
christophertbrown/bioscripts
ctbBio/rp16.py
find_next
def find_next(start, stop, i2hits): """ which protein has the best hit, the one to the 'right' or to the 'left?' """ if start not in i2hits and stop in i2hits: index = stop elif stop not in i2hits and start in i2hits: index = start elif start not in i2hits and stop not in i2hits: index = choice([start, stop]) i2hits[index] = [[False]] else: A, B = i2hits[start][0], i2hits[stop][0] if B[10] <= A[10]: index = stop else: index = start if index == start: nstart = start - 1 nstop = stop else: nstop = stop + 1 nstart = start match = i2hits[index][0] rp = match[-1] return index, nstart, nstop, rp, match
python
def find_next(start, stop, i2hits): """ which protein has the best hit, the one to the 'right' or to the 'left?' """ if start not in i2hits and stop in i2hits: index = stop elif stop not in i2hits and start in i2hits: index = start elif start not in i2hits and stop not in i2hits: index = choice([start, stop]) i2hits[index] = [[False]] else: A, B = i2hits[start][0], i2hits[stop][0] if B[10] <= A[10]: index = stop else: index = start if index == start: nstart = start - 1 nstop = stop else: nstop = stop + 1 nstart = start match = i2hits[index][0] rp = match[-1] return index, nstart, nstop, rp, match
[ "def", "find_next", "(", "start", ",", "stop", ",", "i2hits", ")", ":", "if", "start", "not", "in", "i2hits", "and", "stop", "in", "i2hits", ":", "index", "=", "stop", "elif", "stop", "not", "in", "i2hits", "and", "start", "in", "i2hits", ":", "index", "=", "start", "elif", "start", "not", "in", "i2hits", "and", "stop", "not", "in", "i2hits", ":", "index", "=", "choice", "(", "[", "start", ",", "stop", "]", ")", "i2hits", "[", "index", "]", "=", "[", "[", "False", "]", "]", "else", ":", "A", ",", "B", "=", "i2hits", "[", "start", "]", "[", "0", "]", ",", "i2hits", "[", "stop", "]", "[", "0", "]", "if", "B", "[", "10", "]", "<=", "A", "[", "10", "]", ":", "index", "=", "stop", "else", ":", "index", "=", "start", "if", "index", "==", "start", ":", "nstart", "=", "start", "-", "1", "nstop", "=", "stop", "else", ":", "nstop", "=", "stop", "+", "1", "nstart", "=", "start", "match", "=", "i2hits", "[", "index", "]", "[", "0", "]", "rp", "=", "match", "[", "-", "1", "]", "return", "index", ",", "nstart", ",", "nstop", ",", "rp", ",", "match" ]
which protein has the best hit, the one to the 'right' or to the 'left?'
[ "which", "protein", "has", "the", "best", "hit", "the", "one", "to", "the", "right", "or", "to", "the", "left?" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rp16.py#L77-L102
train
christophertbrown/bioscripts
ctbBio/rp16.py
find_ribosomal
def find_ribosomal(rps, scaffolds, s2rp, min_hits, max_hits_rp, max_errors): """ determine which hits represent real ribosomal proteins, identify each in syntenic block max_hits_rp = maximum number of hits to consider per ribosomal protein per scaffold """ for scaffold, proteins in list(s2rp.items()): # for each scaffold, get best hits for each rp hits = {p: [i for i in sorted(hits, key = itemgetter(10))][0:max_hits_rp] for p, hits in list(proteins.items()) if len(hits) > 0} # skip if fewer than min_hits RPs are identified if len(hits) < min_hits: continue best = sorted([hit[0] + [p] for p, hit in list(hits.items())], key = itemgetter(10))[0] block = find_block(rps, scaffolds[scaffold], hits, best, max_errors) if (len(block) - 1) >= min_hits: yield scaffold, block
python
def find_ribosomal(rps, scaffolds, s2rp, min_hits, max_hits_rp, max_errors): """ determine which hits represent real ribosomal proteins, identify each in syntenic block max_hits_rp = maximum number of hits to consider per ribosomal protein per scaffold """ for scaffold, proteins in list(s2rp.items()): # for each scaffold, get best hits for each rp hits = {p: [i for i in sorted(hits, key = itemgetter(10))][0:max_hits_rp] for p, hits in list(proteins.items()) if len(hits) > 0} # skip if fewer than min_hits RPs are identified if len(hits) < min_hits: continue best = sorted([hit[0] + [p] for p, hit in list(hits.items())], key = itemgetter(10))[0] block = find_block(rps, scaffolds[scaffold], hits, best, max_errors) if (len(block) - 1) >= min_hits: yield scaffold, block
[ "def", "find_ribosomal", "(", "rps", ",", "scaffolds", ",", "s2rp", ",", "min_hits", ",", "max_hits_rp", ",", "max_errors", ")", ":", "for", "scaffold", ",", "proteins", "in", "list", "(", "s2rp", ".", "items", "(", ")", ")", ":", "# for each scaffold, get best hits for each rp", "hits", "=", "{", "p", ":", "[", "i", "for", "i", "in", "sorted", "(", "hits", ",", "key", "=", "itemgetter", "(", "10", ")", ")", "]", "[", "0", ":", "max_hits_rp", "]", "for", "p", ",", "hits", "in", "list", "(", "proteins", ".", "items", "(", ")", ")", "if", "len", "(", "hits", ")", ">", "0", "}", "# skip if fewer than min_hits RPs are identified", "if", "len", "(", "hits", ")", "<", "min_hits", ":", "continue", "best", "=", "sorted", "(", "[", "hit", "[", "0", "]", "+", "[", "p", "]", "for", "p", ",", "hit", "in", "list", "(", "hits", ".", "items", "(", ")", ")", "]", ",", "key", "=", "itemgetter", "(", "10", ")", ")", "[", "0", "]", "block", "=", "find_block", "(", "rps", ",", "scaffolds", "[", "scaffold", "]", ",", "hits", ",", "best", ",", "max_errors", ")", "if", "(", "len", "(", "block", ")", "-", "1", ")", ">=", "min_hits", ":", "yield", "scaffold", ",", "block" ]
determine which hits represent real ribosomal proteins, identify each in syntenic block max_hits_rp = maximum number of hits to consider per ribosomal protein per scaffold
[ "determine", "which", "hits", "represent", "real", "ribosomal", "proteins", "identify", "each", "in", "syntenic", "block", "max_hits_rp", "=", "maximum", "number", "of", "hits", "to", "consider", "per", "ribosomal", "protein", "per", "scaffold" ]
83b2566b3a5745437ec651cd6cafddd056846240
https://github.com/christophertbrown/bioscripts/blob/83b2566b3a5745437ec651cd6cafddd056846240/ctbBio/rp16.py#L134-L150
train
smdabdoub/phylotoast
bin/filter_rep_set.py
filter_rep_set
def filter_rep_set(inF, otuSet): """ Parse the rep set file and remove all sequences not associated with unique OTUs. :@type inF: file :@param inF: The representative sequence set :@rtype: list :@return: The set of sequences associated with unique OTUs """ seqs = [] for record in SeqIO.parse(inF, "fasta"): if record.id in otuSet: seqs.append(record) return seqs
python
def filter_rep_set(inF, otuSet): """ Parse the rep set file and remove all sequences not associated with unique OTUs. :@type inF: file :@param inF: The representative sequence set :@rtype: list :@return: The set of sequences associated with unique OTUs """ seqs = [] for record in SeqIO.parse(inF, "fasta"): if record.id in otuSet: seqs.append(record) return seqs
[ "def", "filter_rep_set", "(", "inF", ",", "otuSet", ")", ":", "seqs", "=", "[", "]", "for", "record", "in", "SeqIO", ".", "parse", "(", "inF", ",", "\"fasta\"", ")", ":", "if", "record", ".", "id", "in", "otuSet", ":", "seqs", ".", "append", "(", "record", ")", "return", "seqs" ]
Parse the rep set file and remove all sequences not associated with unique OTUs. :@type inF: file :@param inF: The representative sequence set :@rtype: list :@return: The set of sequences associated with unique OTUs
[ "Parse", "the", "rep", "set", "file", "and", "remove", "all", "sequences", "not", "associated", "with", "unique", "OTUs", "." ]
0b74ef171e6a84761710548501dfac71285a58a3
https://github.com/smdabdoub/phylotoast/blob/0b74ef171e6a84761710548501dfac71285a58a3/bin/filter_rep_set.py#L32-L47
train
consbio/gis-metadata-parser
gis_metadata/arcgis_metadata_parser.py
ArcGISParser._update_report_item
def _update_report_item(self, **update_props): """ Update the text for each element at the configured path if attribute matches """ tree_to_update = update_props['tree_to_update'] prop = update_props['prop'] values = wrap_value(update_props['values']) xroot = self._get_xroot_for(prop) attr_key = 'type' attr_val = u'' if prop == 'attribute_accuracy': attr_val = 'DQQuanAttAcc' elif prop == 'dataset_completeness': attr_val = 'DQCompOm' # Clear (make empty) all elements of the appropriate type for elem in get_elements(tree_to_update, xroot): if get_element_attributes(elem).get(attr_key) == attr_val: clear_element(elem) # Remove all empty elements, including those previously cleared remove_empty_element(tree_to_update, xroot) # Insert elements with correct attributes for each new value attrs = {attr_key: attr_val} updated = [] for idx, value in enumerate(values): elem = insert_element(tree_to_update, idx, xroot, **attrs) updated.append(insert_element(elem, idx, 'measDesc', value)) return updated
python
def _update_report_item(self, **update_props): """ Update the text for each element at the configured path if attribute matches """ tree_to_update = update_props['tree_to_update'] prop = update_props['prop'] values = wrap_value(update_props['values']) xroot = self._get_xroot_for(prop) attr_key = 'type' attr_val = u'' if prop == 'attribute_accuracy': attr_val = 'DQQuanAttAcc' elif prop == 'dataset_completeness': attr_val = 'DQCompOm' # Clear (make empty) all elements of the appropriate type for elem in get_elements(tree_to_update, xroot): if get_element_attributes(elem).get(attr_key) == attr_val: clear_element(elem) # Remove all empty elements, including those previously cleared remove_empty_element(tree_to_update, xroot) # Insert elements with correct attributes for each new value attrs = {attr_key: attr_val} updated = [] for idx, value in enumerate(values): elem = insert_element(tree_to_update, idx, xroot, **attrs) updated.append(insert_element(elem, idx, 'measDesc', value)) return updated
[ "def", "_update_report_item", "(", "self", ",", "*", "*", "update_props", ")", ":", "tree_to_update", "=", "update_props", "[", "'tree_to_update'", "]", "prop", "=", "update_props", "[", "'prop'", "]", "values", "=", "wrap_value", "(", "update_props", "[", "'values'", "]", ")", "xroot", "=", "self", ".", "_get_xroot_for", "(", "prop", ")", "attr_key", "=", "'type'", "attr_val", "=", "u''", "if", "prop", "==", "'attribute_accuracy'", ":", "attr_val", "=", "'DQQuanAttAcc'", "elif", "prop", "==", "'dataset_completeness'", ":", "attr_val", "=", "'DQCompOm'", "# Clear (make empty) all elements of the appropriate type", "for", "elem", "in", "get_elements", "(", "tree_to_update", ",", "xroot", ")", ":", "if", "get_element_attributes", "(", "elem", ")", ".", "get", "(", "attr_key", ")", "==", "attr_val", ":", "clear_element", "(", "elem", ")", "# Remove all empty elements, including those previously cleared", "remove_empty_element", "(", "tree_to_update", ",", "xroot", ")", "# Insert elements with correct attributes for each new value", "attrs", "=", "{", "attr_key", ":", "attr_val", "}", "updated", "=", "[", "]", "for", "idx", ",", "value", "in", "enumerate", "(", "values", ")", ":", "elem", "=", "insert_element", "(", "tree_to_update", ",", "idx", ",", "xroot", ",", "*", "*", "attrs", ")", "updated", ".", "append", "(", "insert_element", "(", "elem", ",", "idx", ",", "'measDesc'", ",", "value", ")", ")", "return", "updated" ]
Update the text for each element at the configured path if attribute matches
[ "Update", "the", "text", "for", "each", "element", "at", "the", "configured", "path", "if", "attribute", "matches" ]
59eefb2e51cd4d8cc3e94623a2167499ca9ef70f
https://github.com/consbio/gis-metadata-parser/blob/59eefb2e51cd4d8cc3e94623a2167499ca9ef70f/gis_metadata/arcgis_metadata_parser.py#L407-L440
train
adafruit/Adafruit_Python_VCNL40xx
Adafruit_VCNL40xx/VCNL40xx.py
VCNL4010._clear_interrupt
def _clear_interrupt(self, intbit): """Clear the specified interrupt bit in the interrupt status register. """ int_status = self._device.readU8(VCNL4010_INTSTAT); int_status &= ~intbit; self._device.write8(VCNL4010_INTSTAT, int_status);
python
def _clear_interrupt(self, intbit): """Clear the specified interrupt bit in the interrupt status register. """ int_status = self._device.readU8(VCNL4010_INTSTAT); int_status &= ~intbit; self._device.write8(VCNL4010_INTSTAT, int_status);
[ "def", "_clear_interrupt", "(", "self", ",", "intbit", ")", ":", "int_status", "=", "self", ".", "_device", ".", "readU8", "(", "VCNL4010_INTSTAT", ")", "int_status", "&=", "~", "intbit", "self", ".", "_device", ".", "write8", "(", "VCNL4010_INTSTAT", ",", "int_status", ")" ]
Clear the specified interrupt bit in the interrupt status register.
[ "Clear", "the", "specified", "interrupt", "bit", "in", "the", "interrupt", "status", "register", "." ]
f88ec755fd23017028b6dec1be0607ff4a018e10
https://github.com/adafruit/Adafruit_Python_VCNL40xx/blob/f88ec755fd23017028b6dec1be0607ff4a018e10/Adafruit_VCNL40xx/VCNL40xx.py#L123-L128
train
skojaku/core-periphery-detection
cpalgorithm/Rombach.py
SimAlg.move
def move(self): """Swaps two nodes""" a = random.randint(0, len(self.state) - 1) b = random.randint(0, len(self.state) - 1) self.state[[a,b]] = self.state[[b,a]]
python
def move(self): """Swaps two nodes""" a = random.randint(0, len(self.state) - 1) b = random.randint(0, len(self.state) - 1) self.state[[a,b]] = self.state[[b,a]]
[ "def", "move", "(", "self", ")", ":", "a", "=", "random", ".", "randint", "(", "0", ",", "len", "(", "self", ".", "state", ")", "-", "1", ")", "b", "=", "random", ".", "randint", "(", "0", ",", "len", "(", "self", ".", "state", ")", "-", "1", ")", "self", ".", "state", "[", "[", "a", ",", "b", "]", "]", "=", "self", ".", "state", "[", "[", "b", ",", "a", "]", "]" ]
Swaps two nodes
[ "Swaps", "two", "nodes" ]
d724e6441066622506ddb54d81ee9a1cfd15f766
https://github.com/skojaku/core-periphery-detection/blob/d724e6441066622506ddb54d81ee9a1cfd15f766/cpalgorithm/Rombach.py#L19-L23
train
wbond/certbuilder
certbuilder/__init__.py
CertificateBuilder.self_signed
def self_signed(self, value): """ A bool - if the certificate should be self-signed. """ self._self_signed = bool(value) if self._self_signed: self._issuer = None
python
def self_signed(self, value): """ A bool - if the certificate should be self-signed. """ self._self_signed = bool(value) if self._self_signed: self._issuer = None
[ "def", "self_signed", "(", "self", ",", "value", ")", ":", "self", ".", "_self_signed", "=", "bool", "(", "value", ")", "if", "self", ".", "_self_signed", ":", "self", ".", "_issuer", "=", "None" ]
A bool - if the certificate should be self-signed.
[ "A", "bool", "-", "if", "the", "certificate", "should", "be", "self", "-", "signed", "." ]
969dae884fa7f73988bbf1dcbec4fb51e234a3c5
https://github.com/wbond/certbuilder/blob/969dae884fa7f73988bbf1dcbec4fb51e234a3c5/certbuilder/__init__.py#L122-L130
train
wbond/certbuilder
certbuilder/__init__.py
CertificateBuilder._get_crl_url
def _get_crl_url(self, distribution_points): """ Grabs the first URL out of a asn1crypto.x509.CRLDistributionPoints object :param distribution_points: The x509.CRLDistributionPoints object to pull the URL out of :return: A unicode string or None """ if distribution_points is None: return None for distribution_point in distribution_points: name = distribution_point['distribution_point'] if name.name == 'full_name' and name.chosen[0].name == 'uniform_resource_identifier': return name.chosen[0].chosen.native return None
python
def _get_crl_url(self, distribution_points): """ Grabs the first URL out of a asn1crypto.x509.CRLDistributionPoints object :param distribution_points: The x509.CRLDistributionPoints object to pull the URL out of :return: A unicode string or None """ if distribution_points is None: return None for distribution_point in distribution_points: name = distribution_point['distribution_point'] if name.name == 'full_name' and name.chosen[0].name == 'uniform_resource_identifier': return name.chosen[0].chosen.native return None
[ "def", "_get_crl_url", "(", "self", ",", "distribution_points", ")", ":", "if", "distribution_points", "is", "None", ":", "return", "None", "for", "distribution_point", "in", "distribution_points", ":", "name", "=", "distribution_point", "[", "'distribution_point'", "]", "if", "name", ".", "name", "==", "'full_name'", "and", "name", ".", "chosen", "[", "0", "]", ".", "name", "==", "'uniform_resource_identifier'", ":", "return", "name", ".", "chosen", "[", "0", "]", ".", "chosen", ".", "native", "return", "None" ]
Grabs the first URL out of a asn1crypto.x509.CRLDistributionPoints object :param distribution_points: The x509.CRLDistributionPoints object to pull the URL out of :return: A unicode string or None
[ "Grabs", "the", "first", "URL", "out", "of", "a", "asn1crypto", ".", "x509", ".", "CRLDistributionPoints", "object" ]
969dae884fa7f73988bbf1dcbec4fb51e234a3c5
https://github.com/wbond/certbuilder/blob/969dae884fa7f73988bbf1dcbec4fb51e234a3c5/certbuilder/__init__.py#L544-L564
train
wbond/certbuilder
certbuilder/__init__.py
CertificateBuilder.ocsp_no_check
def ocsp_no_check(self, value): """ A bool - if the certificate should have the OCSP no check extension. Only applicable to certificates created for signing OCSP responses. Such certificates should normally be issued for a very short period of time since they are effectively whitelisted by clients. """ if value is None: self._ocsp_no_check = None else: self._ocsp_no_check = bool(value)
python
def ocsp_no_check(self, value): """ A bool - if the certificate should have the OCSP no check extension. Only applicable to certificates created for signing OCSP responses. Such certificates should normally be issued for a very short period of time since they are effectively whitelisted by clients. """ if value is None: self._ocsp_no_check = None else: self._ocsp_no_check = bool(value)
[ "def", "ocsp_no_check", "(", "self", ",", "value", ")", ":", "if", "value", "is", "None", ":", "self", ".", "_ocsp_no_check", "=", "None", "else", ":", "self", ".", "_ocsp_no_check", "=", "bool", "(", "value", ")" ]
A bool - if the certificate should have the OCSP no check extension. Only applicable to certificates created for signing OCSP responses. Such certificates should normally be issued for a very short period of time since they are effectively whitelisted by clients.
[ "A", "bool", "-", "if", "the", "certificate", "should", "have", "the", "OCSP", "no", "check", "extension", ".", "Only", "applicable", "to", "certificates", "created", "for", "signing", "OCSP", "responses", ".", "Such", "certificates", "should", "normally", "be", "issued", "for", "a", "very", "short", "period", "of", "time", "since", "they", "are", "effectively", "whitelisted", "by", "clients", "." ]
969dae884fa7f73988bbf1dcbec4fb51e234a3c5
https://github.com/wbond/certbuilder/blob/969dae884fa7f73988bbf1dcbec4fb51e234a3c5/certbuilder/__init__.py#L688-L699
train
tell-k/django-modelsdoc
modelsdoc/templatetags/modelsdoc_tags.py
emptylineless
def emptylineless(parser, token): """ Removes empty line. Example usage:: {% emptylineless %} test1 test2 test3 {% endemptylineless %} This example would return this HTML:: test1 test2 test3 """ nodelist = parser.parse(('endemptylineless',)) parser.delete_first_token() return EmptylinelessNode(nodelist)
python
def emptylineless(parser, token): """ Removes empty line. Example usage:: {% emptylineless %} test1 test2 test3 {% endemptylineless %} This example would return this HTML:: test1 test2 test3 """ nodelist = parser.parse(('endemptylineless',)) parser.delete_first_token() return EmptylinelessNode(nodelist)
[ "def", "emptylineless", "(", "parser", ",", "token", ")", ":", "nodelist", "=", "parser", ".", "parse", "(", "(", "'endemptylineless'", ",", ")", ")", "parser", ".", "delete_first_token", "(", ")", "return", "EmptylinelessNode", "(", "nodelist", ")" ]
Removes empty line. Example usage:: {% emptylineless %} test1 test2 test3 {% endemptylineless %} This example would return this HTML:: test1 test2 test3
[ "Removes", "empty", "line", "." ]
c9d336e76251feb142347b3a41365430d3365436
https://github.com/tell-k/django-modelsdoc/blob/c9d336e76251feb142347b3a41365430d3365436/modelsdoc/templatetags/modelsdoc_tags.py#L31-L54
train
justquick/python-varnish
varnish.py
http_purge_url
def http_purge_url(url): """ Do an HTTP PURGE of the given asset. The URL is run through urlparse and must point to the varnish instance not the varnishadm """ url = urlparse(url) connection = HTTPConnection(url.hostname, url.port or 80) path = url.path or '/' connection.request('PURGE', '%s?%s' % (path, url.query) if url.query else path, '', {'Host': '%s:%s' % (url.hostname, url.port) if url.port else url.hostname}) response = connection.getresponse() if response.status != 200: logging.error('Purge failed with status: %s' % response.status) return response
python
def http_purge_url(url): """ Do an HTTP PURGE of the given asset. The URL is run through urlparse and must point to the varnish instance not the varnishadm """ url = urlparse(url) connection = HTTPConnection(url.hostname, url.port or 80) path = url.path or '/' connection.request('PURGE', '%s?%s' % (path, url.query) if url.query else path, '', {'Host': '%s:%s' % (url.hostname, url.port) if url.port else url.hostname}) response = connection.getresponse() if response.status != 200: logging.error('Purge failed with status: %s' % response.status) return response
[ "def", "http_purge_url", "(", "url", ")", ":", "url", "=", "urlparse", "(", "url", ")", "connection", "=", "HTTPConnection", "(", "url", ".", "hostname", ",", "url", ".", "port", "or", "80", ")", "path", "=", "url", ".", "path", "or", "'/'", "connection", ".", "request", "(", "'PURGE'", ",", "'%s?%s'", "%", "(", "path", ",", "url", ".", "query", ")", "if", "url", ".", "query", "else", "path", ",", "''", ",", "{", "'Host'", ":", "'%s:%s'", "%", "(", "url", ".", "hostname", ",", "url", ".", "port", ")", "if", "url", ".", "port", "else", "url", ".", "hostname", "}", ")", "response", "=", "connection", ".", "getresponse", "(", ")", "if", "response", ".", "status", "!=", "200", ":", "logging", ".", "error", "(", "'Purge failed with status: %s'", "%", "response", ".", "status", ")", "return", "response" ]
Do an HTTP PURGE of the given asset. The URL is run through urlparse and must point to the varnish instance not the varnishadm
[ "Do", "an", "HTTP", "PURGE", "of", "the", "given", "asset", ".", "The", "URL", "is", "run", "through", "urlparse", "and", "must", "point", "to", "the", "varnish", "instance", "not", "the", "varnishadm" ]
8f114c74898e6c5ade2ce49c8b595040bd150465
https://github.com/justquick/python-varnish/blob/8f114c74898e6c5ade2ce49c8b595040bd150465/varnish.py#L47-L60
train
justquick/python-varnish
varnish.py
run
def run(addr, *commands, **kwargs): """ Non-threaded batch command runner returning output results """ results = [] handler = VarnishHandler(addr, **kwargs) for cmd in commands: if isinstance(cmd, tuple) and len(cmd)>1: results.extend([getattr(handler, c[0].replace('.','_'))(*c[1:]) for c in cmd]) else: results.append(getattr(handler, cmd.replace('.','_'))(*commands[1:])) break handler.close() return results
python
def run(addr, *commands, **kwargs): """ Non-threaded batch command runner returning output results """ results = [] handler = VarnishHandler(addr, **kwargs) for cmd in commands: if isinstance(cmd, tuple) and len(cmd)>1: results.extend([getattr(handler, c[0].replace('.','_'))(*c[1:]) for c in cmd]) else: results.append(getattr(handler, cmd.replace('.','_'))(*commands[1:])) break handler.close() return results
[ "def", "run", "(", "addr", ",", "*", "commands", ",", "*", "*", "kwargs", ")", ":", "results", "=", "[", "]", "handler", "=", "VarnishHandler", "(", "addr", ",", "*", "*", "kwargs", ")", "for", "cmd", "in", "commands", ":", "if", "isinstance", "(", "cmd", ",", "tuple", ")", "and", "len", "(", "cmd", ")", ">", "1", ":", "results", ".", "extend", "(", "[", "getattr", "(", "handler", ",", "c", "[", "0", "]", ".", "replace", "(", "'.'", ",", "'_'", ")", ")", "(", "*", "c", "[", "1", ":", "]", ")", "for", "c", "in", "cmd", "]", ")", "else", ":", "results", ".", "append", "(", "getattr", "(", "handler", ",", "cmd", ".", "replace", "(", "'.'", ",", "'_'", ")", ")", "(", "*", "commands", "[", "1", ":", "]", ")", ")", "break", "handler", ".", "close", "(", ")", "return", "results" ]
Non-threaded batch command runner returning output results
[ "Non", "-", "threaded", "batch", "command", "runner", "returning", "output", "results" ]
8f114c74898e6c5ade2ce49c8b595040bd150465
https://github.com/justquick/python-varnish/blob/8f114c74898e6c5ade2ce49c8b595040bd150465/varnish.py#L289-L302
train
madeindjs/Super-Markdown
SuperMarkdown/SuperMarkdown.py
SuperMarkdown.add_stylesheets
def add_stylesheets(self, *css_files): """add stylesheet files in HTML head""" for css_file in css_files: self.main_soup.style.append(self._text_file(css_file))
python
def add_stylesheets(self, *css_files): """add stylesheet files in HTML head""" for css_file in css_files: self.main_soup.style.append(self._text_file(css_file))
[ "def", "add_stylesheets", "(", "self", ",", "*", "css_files", ")", ":", "for", "css_file", "in", "css_files", ":", "self", ".", "main_soup", ".", "style", ".", "append", "(", "self", ".", "_text_file", "(", "css_file", ")", ")" ]
add stylesheet files in HTML head
[ "add", "stylesheet", "files", "in", "HTML", "head" ]
fe2da746afa6a27aaaad27a2db1dca234f802eb0
https://github.com/madeindjs/Super-Markdown/blob/fe2da746afa6a27aaaad27a2db1dca234f802eb0/SuperMarkdown/SuperMarkdown.py#L43-L46
train
madeindjs/Super-Markdown
SuperMarkdown/SuperMarkdown.py
SuperMarkdown.add_javascripts
def add_javascripts(self, *js_files): """add javascripts files in HTML body""" # create the script tag if don't exists if self.main_soup.script is None: script_tag = self.main_soup.new_tag('script') self.main_soup.body.append(script_tag) for js_file in js_files: self.main_soup.script.append(self._text_file(js_file))
python
def add_javascripts(self, *js_files): """add javascripts files in HTML body""" # create the script tag if don't exists if self.main_soup.script is None: script_tag = self.main_soup.new_tag('script') self.main_soup.body.append(script_tag) for js_file in js_files: self.main_soup.script.append(self._text_file(js_file))
[ "def", "add_javascripts", "(", "self", ",", "*", "js_files", ")", ":", "# create the script tag if don't exists", "if", "self", ".", "main_soup", ".", "script", "is", "None", ":", "script_tag", "=", "self", ".", "main_soup", ".", "new_tag", "(", "'script'", ")", "self", ".", "main_soup", ".", "body", ".", "append", "(", "script_tag", ")", "for", "js_file", "in", "js_files", ":", "self", ".", "main_soup", ".", "script", ".", "append", "(", "self", ".", "_text_file", "(", "js_file", ")", ")" ]
add javascripts files in HTML body
[ "add", "javascripts", "files", "in", "HTML", "body" ]
fe2da746afa6a27aaaad27a2db1dca234f802eb0
https://github.com/madeindjs/Super-Markdown/blob/fe2da746afa6a27aaaad27a2db1dca234f802eb0/SuperMarkdown/SuperMarkdown.py#L48-L56
train
madeindjs/Super-Markdown
SuperMarkdown/SuperMarkdown.py
SuperMarkdown.export
def export(self): """return the object in a file""" with open(self.export_url, 'w', encoding='utf-8') as file: file.write(self.build()) if self.open_browser: webbrowser.open_new_tab(self.export_url)
python
def export(self): """return the object in a file""" with open(self.export_url, 'w', encoding='utf-8') as file: file.write(self.build()) if self.open_browser: webbrowser.open_new_tab(self.export_url)
[ "def", "export", "(", "self", ")", ":", "with", "open", "(", "self", ".", "export_url", ",", "'w'", ",", "encoding", "=", "'utf-8'", ")", "as", "file", ":", "file", ".", "write", "(", "self", ".", "build", "(", ")", ")", "if", "self", ".", "open_browser", ":", "webbrowser", ".", "open_new_tab", "(", "self", ".", "export_url", ")" ]
return the object in a file
[ "return", "the", "object", "in", "a", "file" ]
fe2da746afa6a27aaaad27a2db1dca234f802eb0
https://github.com/madeindjs/Super-Markdown/blob/fe2da746afa6a27aaaad27a2db1dca234f802eb0/SuperMarkdown/SuperMarkdown.py#L58-L64
train
madeindjs/Super-Markdown
SuperMarkdown/SuperMarkdown.py
SuperMarkdown.build
def build(self): """convert Markdown text as html. return the html file as string""" markdown_html = markdown.markdown(self.markdown_text, extensions=[ TocExtension(), 'fenced_code', 'markdown_checklist.extension', 'markdown.extensions.tables']) markdown_soup = BeautifulSoup(markdown_html, 'html.parser') # include jquery & mermaid.js only if there are Mermaid graph if markdown_soup.find('code', attrs={'class': 'mermaid'}): self._add_mermaid_js() # search in markdown html if there are Dot Graph & replace it with .svg result for dot_tag in markdown_soup.find_all('code', attrs={'class': 'dotgraph'}): grap_svg = self._text_to_graphiz(dot_tag.string) graph_soup = BeautifulSoup(grap_svg, 'html.parser') dot_tag.parent.replaceWith(graph_soup) self.main_soup.body.append(markdown_soup) return self.main_soup.prettify()
python
def build(self): """convert Markdown text as html. return the html file as string""" markdown_html = markdown.markdown(self.markdown_text, extensions=[ TocExtension(), 'fenced_code', 'markdown_checklist.extension', 'markdown.extensions.tables']) markdown_soup = BeautifulSoup(markdown_html, 'html.parser') # include jquery & mermaid.js only if there are Mermaid graph if markdown_soup.find('code', attrs={'class': 'mermaid'}): self._add_mermaid_js() # search in markdown html if there are Dot Graph & replace it with .svg result for dot_tag in markdown_soup.find_all('code', attrs={'class': 'dotgraph'}): grap_svg = self._text_to_graphiz(dot_tag.string) graph_soup = BeautifulSoup(grap_svg, 'html.parser') dot_tag.parent.replaceWith(graph_soup) self.main_soup.body.append(markdown_soup) return self.main_soup.prettify()
[ "def", "build", "(", "self", ")", ":", "markdown_html", "=", "markdown", ".", "markdown", "(", "self", ".", "markdown_text", ",", "extensions", "=", "[", "TocExtension", "(", ")", ",", "'fenced_code'", ",", "'markdown_checklist.extension'", ",", "'markdown.extensions.tables'", "]", ")", "markdown_soup", "=", "BeautifulSoup", "(", "markdown_html", ",", "'html.parser'", ")", "# include jquery & mermaid.js only if there are Mermaid graph", "if", "markdown_soup", ".", "find", "(", "'code'", ",", "attrs", "=", "{", "'class'", ":", "'mermaid'", "}", ")", ":", "self", ".", "_add_mermaid_js", "(", ")", "# search in markdown html if there are Dot Graph & replace it with .svg result", "for", "dot_tag", "in", "markdown_soup", ".", "find_all", "(", "'code'", ",", "attrs", "=", "{", "'class'", ":", "'dotgraph'", "}", ")", ":", "grap_svg", "=", "self", ".", "_text_to_graphiz", "(", "dot_tag", ".", "string", ")", "graph_soup", "=", "BeautifulSoup", "(", "grap_svg", ",", "'html.parser'", ")", "dot_tag", ".", "parent", ".", "replaceWith", "(", "graph_soup", ")", "self", ".", "main_soup", ".", "body", ".", "append", "(", "markdown_soup", ")", "return", "self", ".", "main_soup", ".", "prettify", "(", ")" ]
convert Markdown text as html. return the html file as string
[ "convert", "Markdown", "text", "as", "html", ".", "return", "the", "html", "file", "as", "string" ]
fe2da746afa6a27aaaad27a2db1dca234f802eb0
https://github.com/madeindjs/Super-Markdown/blob/fe2da746afa6a27aaaad27a2db1dca234f802eb0/SuperMarkdown/SuperMarkdown.py#L66-L84
train
madeindjs/Super-Markdown
SuperMarkdown/SuperMarkdown.py
SuperMarkdown._text_file
def _text_file(self, url): """return the content of a file""" try: with open(url, 'r', encoding='utf-8') as file: return file.read() except FileNotFoundError: print('File `{}` not found'.format(url)) sys.exit(0)
python
def _text_file(self, url): """return the content of a file""" try: with open(url, 'r', encoding='utf-8') as file: return file.read() except FileNotFoundError: print('File `{}` not found'.format(url)) sys.exit(0)
[ "def", "_text_file", "(", "self", ",", "url", ")", ":", "try", ":", "with", "open", "(", "url", ",", "'r'", ",", "encoding", "=", "'utf-8'", ")", "as", "file", ":", "return", "file", ".", "read", "(", ")", "except", "FileNotFoundError", ":", "print", "(", "'File `{}` not found'", ".", "format", "(", "url", ")", ")", "sys", ".", "exit", "(", "0", ")" ]
return the content of a file
[ "return", "the", "content", "of", "a", "file" ]
fe2da746afa6a27aaaad27a2db1dca234f802eb0
https://github.com/madeindjs/Super-Markdown/blob/fe2da746afa6a27aaaad27a2db1dca234f802eb0/SuperMarkdown/SuperMarkdown.py#L86-L93
train
madeindjs/Super-Markdown
SuperMarkdown/SuperMarkdown.py
SuperMarkdown._text_to_graphiz
def _text_to_graphiz(self, text): """create a graphviz graph from text""" dot = Source(text, format='svg') return dot.pipe().decode('utf-8')
python
def _text_to_graphiz(self, text): """create a graphviz graph from text""" dot = Source(text, format='svg') return dot.pipe().decode('utf-8')
[ "def", "_text_to_graphiz", "(", "self", ",", "text", ")", ":", "dot", "=", "Source", "(", "text", ",", "format", "=", "'svg'", ")", "return", "dot", ".", "pipe", "(", ")", ".", "decode", "(", "'utf-8'", ")" ]
create a graphviz graph from text
[ "create", "a", "graphviz", "graph", "from", "text" ]
fe2da746afa6a27aaaad27a2db1dca234f802eb0
https://github.com/madeindjs/Super-Markdown/blob/fe2da746afa6a27aaaad27a2db1dca234f802eb0/SuperMarkdown/SuperMarkdown.py#L95-L98
train
madeindjs/Super-Markdown
SuperMarkdown/SuperMarkdown.py
SuperMarkdown._add_mermaid_js
def _add_mermaid_js(self): """add js libraries and css files of mermaid js_file""" self.add_javascripts('{}/js/jquery-1.11.3.min.js'.format(self.resources_path)) self.add_javascripts('{}/js/mermaid.min.js'.format(self.resources_path)) self.add_stylesheets('{}/css/mermaid.css'.format(self.resources_path)) self.main_soup.script.append('mermaid.initialize({startOnLoad:true });')
python
def _add_mermaid_js(self): """add js libraries and css files of mermaid js_file""" self.add_javascripts('{}/js/jquery-1.11.3.min.js'.format(self.resources_path)) self.add_javascripts('{}/js/mermaid.min.js'.format(self.resources_path)) self.add_stylesheets('{}/css/mermaid.css'.format(self.resources_path)) self.main_soup.script.append('mermaid.initialize({startOnLoad:true });')
[ "def", "_add_mermaid_js", "(", "self", ")", ":", "self", ".", "add_javascripts", "(", "'{}/js/jquery-1.11.3.min.js'", ".", "format", "(", "self", ".", "resources_path", ")", ")", "self", ".", "add_javascripts", "(", "'{}/js/mermaid.min.js'", ".", "format", "(", "self", ".", "resources_path", ")", ")", "self", ".", "add_stylesheets", "(", "'{}/css/mermaid.css'", ".", "format", "(", "self", ".", "resources_path", ")", ")", "self", ".", "main_soup", ".", "script", ".", "append", "(", "'mermaid.initialize({startOnLoad:true });'", ")" ]
add js libraries and css files of mermaid js_file
[ "add", "js", "libraries", "and", "css", "files", "of", "mermaid", "js_file" ]
fe2da746afa6a27aaaad27a2db1dca234f802eb0
https://github.com/madeindjs/Super-Markdown/blob/fe2da746afa6a27aaaad27a2db1dca234f802eb0/SuperMarkdown/SuperMarkdown.py#L100-L105
train
paul-wolf/strgen
strgen/__init__.py
StringGenerator.getCharacterSet
def getCharacterSet(self): '''Get a character set with individual members or ranges. Current index is on '[', the start of the character set. ''' chars = u'' c = None cnt = 1 start = 0 while True: escaped_slash = False c = self.next() # print "pattern : ", self.pattern # print "C : ", c # print "Slash : ", c == u'\\' # print 'chars : ', chars # print 'index : ', self.index # print 'last : ', self.last() # print 'lookahead : ', self.lookahead() if self.lookahead() == u'-' and not c == u'\\': f = c self.next() # skip hyphen c = self.next() # get far range if not c or (c in self.meta_chars): raise StringGenerator.SyntaxError(u"unexpected end of class range") chars += self.getCharacterRange(f, c) elif c == u'\\': if self.lookahead() in self.meta_chars: c = self.next() chars += c continue elif self.lookahead() in self.string_code: c = self.next() chars += self.string_code[c] elif c and c not in self.meta_chars: chars += c if c == u']': if self.lookahead() == u'{': [start, cnt] = self.getQuantifier() else: start = -1 cnt = 1 break if c and c in self.meta_chars and not self.last() == u"\\": raise StringGenerator.SyntaxError(u"Un-escaped character in class definition: %s" % c) if not c: break return StringGenerator.CharacterSet(chars, start, cnt)
python
def getCharacterSet(self): '''Get a character set with individual members or ranges. Current index is on '[', the start of the character set. ''' chars = u'' c = None cnt = 1 start = 0 while True: escaped_slash = False c = self.next() # print "pattern : ", self.pattern # print "C : ", c # print "Slash : ", c == u'\\' # print 'chars : ', chars # print 'index : ', self.index # print 'last : ', self.last() # print 'lookahead : ', self.lookahead() if self.lookahead() == u'-' and not c == u'\\': f = c self.next() # skip hyphen c = self.next() # get far range if not c or (c in self.meta_chars): raise StringGenerator.SyntaxError(u"unexpected end of class range") chars += self.getCharacterRange(f, c) elif c == u'\\': if self.lookahead() in self.meta_chars: c = self.next() chars += c continue elif self.lookahead() in self.string_code: c = self.next() chars += self.string_code[c] elif c and c not in self.meta_chars: chars += c if c == u']': if self.lookahead() == u'{': [start, cnt] = self.getQuantifier() else: start = -1 cnt = 1 break if c and c in self.meta_chars and not self.last() == u"\\": raise StringGenerator.SyntaxError(u"Un-escaped character in class definition: %s" % c) if not c: break return StringGenerator.CharacterSet(chars, start, cnt)
[ "def", "getCharacterSet", "(", "self", ")", ":", "chars", "=", "u''", "c", "=", "None", "cnt", "=", "1", "start", "=", "0", "while", "True", ":", "escaped_slash", "=", "False", "c", "=", "self", ".", "next", "(", ")", "# print \"pattern : \", self.pattern", "# print \"C : \", c", "# print \"Slash : \", c == u'\\\\'", "# print 'chars : ', chars", "# print 'index : ', self.index", "# print 'last : ', self.last()", "# print 'lookahead : ', self.lookahead()", "if", "self", ".", "lookahead", "(", ")", "==", "u'-'", "and", "not", "c", "==", "u'\\\\'", ":", "f", "=", "c", "self", ".", "next", "(", ")", "# skip hyphen", "c", "=", "self", ".", "next", "(", ")", "# get far range", "if", "not", "c", "or", "(", "c", "in", "self", ".", "meta_chars", ")", ":", "raise", "StringGenerator", ".", "SyntaxError", "(", "u\"unexpected end of class range\"", ")", "chars", "+=", "self", ".", "getCharacterRange", "(", "f", ",", "c", ")", "elif", "c", "==", "u'\\\\'", ":", "if", "self", ".", "lookahead", "(", ")", "in", "self", ".", "meta_chars", ":", "c", "=", "self", ".", "next", "(", ")", "chars", "+=", "c", "continue", "elif", "self", ".", "lookahead", "(", ")", "in", "self", ".", "string_code", ":", "c", "=", "self", ".", "next", "(", ")", "chars", "+=", "self", ".", "string_code", "[", "c", "]", "elif", "c", "and", "c", "not", "in", "self", ".", "meta_chars", ":", "chars", "+=", "c", "if", "c", "==", "u']'", ":", "if", "self", ".", "lookahead", "(", ")", "==", "u'{'", ":", "[", "start", ",", "cnt", "]", "=", "self", ".", "getQuantifier", "(", ")", "else", ":", "start", "=", "-", "1", "cnt", "=", "1", "break", "if", "c", "and", "c", "in", "self", ".", "meta_chars", "and", "not", "self", ".", "last", "(", ")", "==", "u\"\\\\\"", ":", "raise", "StringGenerator", ".", "SyntaxError", "(", "u\"Un-escaped character in class definition: %s\"", "%", "c", ")", "if", "not", "c", ":", "break", "return", "StringGenerator", ".", "CharacterSet", "(", "chars", ",", "start", ",", "cnt", ")" ]
Get a character set with individual members or ranges. Current index is on '[', the start of the character set.
[ "Get", "a", "character", "set", "with", "individual", "members", "or", "ranges", "." ]
ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f
https://github.com/paul-wolf/strgen/blob/ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f/strgen/__init__.py#L368-L419
train
paul-wolf/strgen
strgen/__init__.py
StringGenerator.getLiteral
def getLiteral(self): '''Get a sequence of non-special characters.''' # we are on the first non-special character chars = u'' c = self.current() while True: if c and c == u"\\": c = self.next() if c: chars += c continue elif not c or (c in self.meta_chars): break else: chars += c if self.lookahead() and self.lookahead() in self.meta_chars: break c = self.next() return StringGenerator.Literal(chars)
python
def getLiteral(self): '''Get a sequence of non-special characters.''' # we are on the first non-special character chars = u'' c = self.current() while True: if c and c == u"\\": c = self.next() if c: chars += c continue elif not c or (c in self.meta_chars): break else: chars += c if self.lookahead() and self.lookahead() in self.meta_chars: break c = self.next() return StringGenerator.Literal(chars)
[ "def", "getLiteral", "(", "self", ")", ":", "# we are on the first non-special character", "chars", "=", "u''", "c", "=", "self", ".", "current", "(", ")", "while", "True", ":", "if", "c", "and", "c", "==", "u\"\\\\\"", ":", "c", "=", "self", ".", "next", "(", ")", "if", "c", ":", "chars", "+=", "c", "continue", "elif", "not", "c", "or", "(", "c", "in", "self", ".", "meta_chars", ")", ":", "break", "else", ":", "chars", "+=", "c", "if", "self", ".", "lookahead", "(", ")", "and", "self", ".", "lookahead", "(", ")", "in", "self", ".", "meta_chars", ":", "break", "c", "=", "self", ".", "next", "(", ")", "return", "StringGenerator", ".", "Literal", "(", "chars", ")" ]
Get a sequence of non-special characters.
[ "Get", "a", "sequence", "of", "non", "-", "special", "characters", "." ]
ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f
https://github.com/paul-wolf/strgen/blob/ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f/strgen/__init__.py#L421-L439
train
paul-wolf/strgen
strgen/__init__.py
StringGenerator.getSequence
def getSequence(self, level=0): '''Get a sequence of nodes.''' seq = [] op = '' left_operand = None right_operand = None sequence_closed = False while True: c = self.next() if not c: break if c and c not in self.meta_chars: seq.append(self.getLiteral()) elif c and c == u'$' and self.lookahead() == u'{': seq.append(self.getSource()) elif c == u'[' and not self.last() == u'\\': seq.append(self.getCharacterSet()) elif c == u'(' and not self.last() == u'\\': seq.append(self.getSequence(level + 1)) elif c == u')' and not self.last() == u'\\': # end of this sequence if level == 0: # there should be no parens here raise StringGenerator.SyntaxError(u"Extra closing parenthesis") sequence_closed = True break elif c == u'|' and not self.last() == u'\\': op = c elif c == u'&' and not self.last() == u'\\': op = c else: if c in self.meta_chars and not self.last() == u"\\": raise StringGenerator.SyntaxError(u"Un-escaped special character: %s" % c) #print( op,len(seq) ) if op and not left_operand: if not seq or len(seq) < 1: raise StringGenerator.SyntaxError(u"Operator: %s with no left operand" % op) left_operand = seq.pop() elif op and len(seq) >= 1 and left_operand: right_operand = seq.pop() #print( "popped: [%s] %s:%s"%( op, left_operand, right_operand) ) if op == u'|': seq.append(StringGenerator.SequenceOR([left_operand, right_operand])) elif op == u'&': seq.append(StringGenerator.SequenceAND([left_operand, right_operand])) op = u'' left_operand = None right_operand = None # check for syntax errors if op: raise StringGenerator.SyntaxError(u"Operator: %s with no right operand" % op) if level > 0 and not sequence_closed: # it means we are finishing a non-first-level sequence without closing parens raise StringGenerator.SyntaxError(u"Missing closing parenthesis") return StringGenerator.Sequence(seq)
python
def getSequence(self, level=0): '''Get a sequence of nodes.''' seq = [] op = '' left_operand = None right_operand = None sequence_closed = False while True: c = self.next() if not c: break if c and c not in self.meta_chars: seq.append(self.getLiteral()) elif c and c == u'$' and self.lookahead() == u'{': seq.append(self.getSource()) elif c == u'[' and not self.last() == u'\\': seq.append(self.getCharacterSet()) elif c == u'(' and not self.last() == u'\\': seq.append(self.getSequence(level + 1)) elif c == u')' and not self.last() == u'\\': # end of this sequence if level == 0: # there should be no parens here raise StringGenerator.SyntaxError(u"Extra closing parenthesis") sequence_closed = True break elif c == u'|' and not self.last() == u'\\': op = c elif c == u'&' and not self.last() == u'\\': op = c else: if c in self.meta_chars and not self.last() == u"\\": raise StringGenerator.SyntaxError(u"Un-escaped special character: %s" % c) #print( op,len(seq) ) if op and not left_operand: if not seq or len(seq) < 1: raise StringGenerator.SyntaxError(u"Operator: %s with no left operand" % op) left_operand = seq.pop() elif op and len(seq) >= 1 and left_operand: right_operand = seq.pop() #print( "popped: [%s] %s:%s"%( op, left_operand, right_operand) ) if op == u'|': seq.append(StringGenerator.SequenceOR([left_operand, right_operand])) elif op == u'&': seq.append(StringGenerator.SequenceAND([left_operand, right_operand])) op = u'' left_operand = None right_operand = None # check for syntax errors if op: raise StringGenerator.SyntaxError(u"Operator: %s with no right operand" % op) if level > 0 and not sequence_closed: # it means we are finishing a non-first-level sequence without closing parens raise StringGenerator.SyntaxError(u"Missing closing parenthesis") return StringGenerator.Sequence(seq)
[ "def", "getSequence", "(", "self", ",", "level", "=", "0", ")", ":", "seq", "=", "[", "]", "op", "=", "''", "left_operand", "=", "None", "right_operand", "=", "None", "sequence_closed", "=", "False", "while", "True", ":", "c", "=", "self", ".", "next", "(", ")", "if", "not", "c", ":", "break", "if", "c", "and", "c", "not", "in", "self", ".", "meta_chars", ":", "seq", ".", "append", "(", "self", ".", "getLiteral", "(", ")", ")", "elif", "c", "and", "c", "==", "u'$'", "and", "self", ".", "lookahead", "(", ")", "==", "u'{'", ":", "seq", ".", "append", "(", "self", ".", "getSource", "(", ")", ")", "elif", "c", "==", "u'['", "and", "not", "self", ".", "last", "(", ")", "==", "u'\\\\'", ":", "seq", ".", "append", "(", "self", ".", "getCharacterSet", "(", ")", ")", "elif", "c", "==", "u'('", "and", "not", "self", ".", "last", "(", ")", "==", "u'\\\\'", ":", "seq", ".", "append", "(", "self", ".", "getSequence", "(", "level", "+", "1", ")", ")", "elif", "c", "==", "u')'", "and", "not", "self", ".", "last", "(", ")", "==", "u'\\\\'", ":", "# end of this sequence", "if", "level", "==", "0", ":", "# there should be no parens here", "raise", "StringGenerator", ".", "SyntaxError", "(", "u\"Extra closing parenthesis\"", ")", "sequence_closed", "=", "True", "break", "elif", "c", "==", "u'|'", "and", "not", "self", ".", "last", "(", ")", "==", "u'\\\\'", ":", "op", "=", "c", "elif", "c", "==", "u'&'", "and", "not", "self", ".", "last", "(", ")", "==", "u'\\\\'", ":", "op", "=", "c", "else", ":", "if", "c", "in", "self", ".", "meta_chars", "and", "not", "self", ".", "last", "(", ")", "==", "u\"\\\\\"", ":", "raise", "StringGenerator", ".", "SyntaxError", "(", "u\"Un-escaped special character: %s\"", "%", "c", ")", "#print( op,len(seq) )", "if", "op", "and", "not", "left_operand", ":", "if", "not", "seq", "or", "len", "(", "seq", ")", "<", "1", ":", "raise", "StringGenerator", ".", "SyntaxError", "(", "u\"Operator: %s with no left operand\"", "%", "op", ")", "left_operand", "=", "seq", ".", "pop", "(", ")", "elif", "op", "and", "len", "(", "seq", ")", ">=", "1", "and", "left_operand", ":", "right_operand", "=", "seq", ".", "pop", "(", ")", "#print( \"popped: [%s] %s:%s\"%( op, left_operand, right_operand) )", "if", "op", "==", "u'|'", ":", "seq", ".", "append", "(", "StringGenerator", ".", "SequenceOR", "(", "[", "left_operand", ",", "right_operand", "]", ")", ")", "elif", "op", "==", "u'&'", ":", "seq", ".", "append", "(", "StringGenerator", ".", "SequenceAND", "(", "[", "left_operand", ",", "right_operand", "]", ")", ")", "op", "=", "u''", "left_operand", "=", "None", "right_operand", "=", "None", "# check for syntax errors", "if", "op", ":", "raise", "StringGenerator", ".", "SyntaxError", "(", "u\"Operator: %s with no right operand\"", "%", "op", ")", "if", "level", ">", "0", "and", "not", "sequence_closed", ":", "# it means we are finishing a non-first-level sequence without closing parens", "raise", "StringGenerator", ".", "SyntaxError", "(", "u\"Missing closing parenthesis\"", ")", "return", "StringGenerator", ".", "Sequence", "(", "seq", ")" ]
Get a sequence of nodes.
[ "Get", "a", "sequence", "of", "nodes", "." ]
ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f
https://github.com/paul-wolf/strgen/blob/ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f/strgen/__init__.py#L441-L501
train
paul-wolf/strgen
strgen/__init__.py
StringGenerator.dump
def dump(self, **kwargs): import sys '''Print the parse tree and then call render for an example.''' if not self.seq: self.seq = self.getSequence() print("StringGenerator version: %s" % (__version__)) print("Python version: %s" % sys.version) # this doesn't work anymore in p3 # print("Random method provider class: %s" % randint.im_class.__name__) self.seq.dump() return self.render(**kwargs)
python
def dump(self, **kwargs): import sys '''Print the parse tree and then call render for an example.''' if not self.seq: self.seq = self.getSequence() print("StringGenerator version: %s" % (__version__)) print("Python version: %s" % sys.version) # this doesn't work anymore in p3 # print("Random method provider class: %s" % randint.im_class.__name__) self.seq.dump() return self.render(**kwargs)
[ "def", "dump", "(", "self", ",", "*", "*", "kwargs", ")", ":", "import", "sys", "if", "not", "self", ".", "seq", ":", "self", ".", "seq", "=", "self", ".", "getSequence", "(", ")", "print", "(", "\"StringGenerator version: %s\"", "%", "(", "__version__", ")", ")", "print", "(", "\"Python version: %s\"", "%", "sys", ".", "version", ")", "# this doesn't work anymore in p3", "# print(\"Random method provider class: %s\" % randint.im_class.__name__)", "self", ".", "seq", ".", "dump", "(", ")", "return", "self", ".", "render", "(", "*", "*", "kwargs", ")" ]
Print the parse tree and then call render for an example.
[ "Print", "the", "parse", "tree", "and", "then", "call", "render", "for", "an", "example", "." ]
ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f
https://github.com/paul-wolf/strgen/blob/ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f/strgen/__init__.py#L521-L531
train
paul-wolf/strgen
strgen/__init__.py
StringGenerator.render_list
def render_list(self, cnt, unique=False, progress_callback=None, **kwargs): '''Return a list of generated strings. Args: cnt (int): length of list unique (bool): whether to make entries unique Returns: list. We keep track of total attempts because a template may specify something impossible to attain, like [1-9]{} with cnt==1000 ''' rendered_list = [] i = 0 total_attempts = 0 while True: if i >= cnt: break if total_attempts > cnt * self.unique_attempts_factor: raise StringGenerator.UniquenessError(u"couldn't satisfy uniqueness") s = self.render(**kwargs) if unique: if not s in rendered_list: rendered_list.append(s) i += 1 else: rendered_list.append(s) i += 1 total_attempts += 1 # Optionally trigger the progress indicator to inform others about our progress if progress_callback and callable(progress_callback): progress_callback(i, cnt) return rendered_list
python
def render_list(self, cnt, unique=False, progress_callback=None, **kwargs): '''Return a list of generated strings. Args: cnt (int): length of list unique (bool): whether to make entries unique Returns: list. We keep track of total attempts because a template may specify something impossible to attain, like [1-9]{} with cnt==1000 ''' rendered_list = [] i = 0 total_attempts = 0 while True: if i >= cnt: break if total_attempts > cnt * self.unique_attempts_factor: raise StringGenerator.UniquenessError(u"couldn't satisfy uniqueness") s = self.render(**kwargs) if unique: if not s in rendered_list: rendered_list.append(s) i += 1 else: rendered_list.append(s) i += 1 total_attempts += 1 # Optionally trigger the progress indicator to inform others about our progress if progress_callback and callable(progress_callback): progress_callback(i, cnt) return rendered_list
[ "def", "render_list", "(", "self", ",", "cnt", ",", "unique", "=", "False", ",", "progress_callback", "=", "None", ",", "*", "*", "kwargs", ")", ":", "rendered_list", "=", "[", "]", "i", "=", "0", "total_attempts", "=", "0", "while", "True", ":", "if", "i", ">=", "cnt", ":", "break", "if", "total_attempts", ">", "cnt", "*", "self", ".", "unique_attempts_factor", ":", "raise", "StringGenerator", ".", "UniquenessError", "(", "u\"couldn't satisfy uniqueness\"", ")", "s", "=", "self", ".", "render", "(", "*", "*", "kwargs", ")", "if", "unique", ":", "if", "not", "s", "in", "rendered_list", ":", "rendered_list", ".", "append", "(", "s", ")", "i", "+=", "1", "else", ":", "rendered_list", ".", "append", "(", "s", ")", "i", "+=", "1", "total_attempts", "+=", "1", "# Optionally trigger the progress indicator to inform others about our progress", "if", "progress_callback", "and", "callable", "(", "progress_callback", ")", ":", "progress_callback", "(", "i", ",", "cnt", ")", "return", "rendered_list" ]
Return a list of generated strings. Args: cnt (int): length of list unique (bool): whether to make entries unique Returns: list. We keep track of total attempts because a template may specify something impossible to attain, like [1-9]{} with cnt==1000
[ "Return", "a", "list", "of", "generated", "strings", "." ]
ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f
https://github.com/paul-wolf/strgen/blob/ca1a1484bed5a31dc9ceaef1ab62dd5582cc0d9f/strgen/__init__.py#L533-L570
train
seperman/s3utils
s3utils/s3utils.py
S3utils.connect
def connect(self): """ Establish the connection. This is done automatically for you. If you lose the connection, you can manually run this to be re-connected. """ self.conn = boto.connect_s3(self.AWS_ACCESS_KEY_ID, self.AWS_SECRET_ACCESS_KEY, debug=self.S3UTILS_DEBUG_LEVEL) self.bucket = self.conn.get_bucket(self.AWS_STORAGE_BUCKET_NAME) self.k = Key(self.bucket)
python
def connect(self): """ Establish the connection. This is done automatically for you. If you lose the connection, you can manually run this to be re-connected. """ self.conn = boto.connect_s3(self.AWS_ACCESS_KEY_ID, self.AWS_SECRET_ACCESS_KEY, debug=self.S3UTILS_DEBUG_LEVEL) self.bucket = self.conn.get_bucket(self.AWS_STORAGE_BUCKET_NAME) self.k = Key(self.bucket)
[ "def", "connect", "(", "self", ")", ":", "self", ".", "conn", "=", "boto", ".", "connect_s3", "(", "self", ".", "AWS_ACCESS_KEY_ID", ",", "self", ".", "AWS_SECRET_ACCESS_KEY", ",", "debug", "=", "self", ".", "S3UTILS_DEBUG_LEVEL", ")", "self", ".", "bucket", "=", "self", ".", "conn", ".", "get_bucket", "(", "self", ".", "AWS_STORAGE_BUCKET_NAME", ")", "self", ".", "k", "=", "Key", "(", "self", ".", "bucket", ")" ]
Establish the connection. This is done automatically for you. If you lose the connection, you can manually run this to be re-connected.
[ "Establish", "the", "connection", ".", "This", "is", "done", "automatically", "for", "you", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L145-L155
train
seperman/s3utils
s3utils/s3utils.py
S3utils.connect_cloudfront
def connect_cloudfront(self): "Connect to Cloud Front. This is done automatically for you when needed." self.conn_cloudfront = connect_cloudfront(self.AWS_ACCESS_KEY_ID, self.AWS_SECRET_ACCESS_KEY, debug=self.S3UTILS_DEBUG_LEVEL)
python
def connect_cloudfront(self): "Connect to Cloud Front. This is done automatically for you when needed." self.conn_cloudfront = connect_cloudfront(self.AWS_ACCESS_KEY_ID, self.AWS_SECRET_ACCESS_KEY, debug=self.S3UTILS_DEBUG_LEVEL)
[ "def", "connect_cloudfront", "(", "self", ")", ":", "self", ".", "conn_cloudfront", "=", "connect_cloudfront", "(", "self", ".", "AWS_ACCESS_KEY_ID", ",", "self", ".", "AWS_SECRET_ACCESS_KEY", ",", "debug", "=", "self", ".", "S3UTILS_DEBUG_LEVEL", ")" ]
Connect to Cloud Front. This is done automatically for you when needed.
[ "Connect", "to", "Cloud", "Front", ".", "This", "is", "done", "automatically", "for", "you", "when", "needed", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L166-L168
train
seperman/s3utils
s3utils/s3utils.py
S3utils.mkdir
def mkdir(self, target_folder): """ Create a folder on S3. Examples -------- >>> s3utils.mkdir("path/to/my_folder") Making directory: path/to/my_folder """ self.printv("Making directory: %s" % target_folder) self.k.key = re.sub(r"^/|/$", "", target_folder) + "/" self.k.set_contents_from_string('') self.k.close()
python
def mkdir(self, target_folder): """ Create a folder on S3. Examples -------- >>> s3utils.mkdir("path/to/my_folder") Making directory: path/to/my_folder """ self.printv("Making directory: %s" % target_folder) self.k.key = re.sub(r"^/|/$", "", target_folder) + "/" self.k.set_contents_from_string('') self.k.close()
[ "def", "mkdir", "(", "self", ",", "target_folder", ")", ":", "self", ".", "printv", "(", "\"Making directory: %s\"", "%", "target_folder", ")", "self", ".", "k", ".", "key", "=", "re", ".", "sub", "(", "r\"^/|/$\"", ",", "\"\"", ",", "target_folder", ")", "+", "\"/\"", "self", ".", "k", ".", "set_contents_from_string", "(", "''", ")", "self", ".", "k", ".", "close", "(", ")" ]
Create a folder on S3. Examples -------- >>> s3utils.mkdir("path/to/my_folder") Making directory: path/to/my_folder
[ "Create", "a", "folder", "on", "S3", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L171-L183
train
seperman/s3utils
s3utils/s3utils.py
S3utils.rm
def rm(self, path): """ Delete the path and anything under the path. Example ------- >>> s3utils.rm("path/to/file_or_folder") """ list_of_files = list(self.ls(path)) if list_of_files: if len(list_of_files) == 1: self.bucket.delete_key(list_of_files[0]) else: self.bucket.delete_keys(list_of_files) self.printv("Deleted: %s" % list_of_files) else: logger.error("There was nothing to remove under %s", path)
python
def rm(self, path): """ Delete the path and anything under the path. Example ------- >>> s3utils.rm("path/to/file_or_folder") """ list_of_files = list(self.ls(path)) if list_of_files: if len(list_of_files) == 1: self.bucket.delete_key(list_of_files[0]) else: self.bucket.delete_keys(list_of_files) self.printv("Deleted: %s" % list_of_files) else: logger.error("There was nothing to remove under %s", path)
[ "def", "rm", "(", "self", ",", "path", ")", ":", "list_of_files", "=", "list", "(", "self", ".", "ls", "(", "path", ")", ")", "if", "list_of_files", ":", "if", "len", "(", "list_of_files", ")", "==", "1", ":", "self", ".", "bucket", ".", "delete_key", "(", "list_of_files", "[", "0", "]", ")", "else", ":", "self", ".", "bucket", ".", "delete_keys", "(", "list_of_files", ")", "self", ".", "printv", "(", "\"Deleted: %s\"", "%", "list_of_files", ")", "else", ":", "logger", ".", "error", "(", "\"There was nothing to remove under %s\"", ",", "path", ")" ]
Delete the path and anything under the path. Example ------- >>> s3utils.rm("path/to/file_or_folder")
[ "Delete", "the", "path", "and", "anything", "under", "the", "path", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L186-L204
train
seperman/s3utils
s3utils/s3utils.py
S3utils.__put_key
def __put_key(self, local_file, target_file, acl='public-read', del_after_upload=False, overwrite=True, source="filename"): """Copy a file to s3.""" action_word = "moving" if del_after_upload else "copying" try: self.k.key = target_file # setting the path (key) of file in the container if source == "filename": # grabs the contents from local_file address. Note that it loads the whole file into memory self.k.set_contents_from_filename(local_file, self.AWS_HEADERS) elif source == "fileobj": self.k.set_contents_from_file(local_file, self.AWS_HEADERS) elif source == "string": self.k.set_contents_from_string(local_file, self.AWS_HEADERS) else: raise Exception("%s is not implemented as a source." % source) self.k.set_acl(acl) # setting the file permissions self.k.close() # not sure if it is needed. Somewhere I read it is recommended. self.printv("%s %s to %s" % (action_word, local_file, target_file)) # if it is supposed to delete the local file after uploading if del_after_upload and source == "filename": try: os.remove(local_file) except: logger.error("Unable to delete the file: ", local_file, exc_info=True) return True except: logger.error("Error in writing to %s", target_file, exc_info=True) return False
python
def __put_key(self, local_file, target_file, acl='public-read', del_after_upload=False, overwrite=True, source="filename"): """Copy a file to s3.""" action_word = "moving" if del_after_upload else "copying" try: self.k.key = target_file # setting the path (key) of file in the container if source == "filename": # grabs the contents from local_file address. Note that it loads the whole file into memory self.k.set_contents_from_filename(local_file, self.AWS_HEADERS) elif source == "fileobj": self.k.set_contents_from_file(local_file, self.AWS_HEADERS) elif source == "string": self.k.set_contents_from_string(local_file, self.AWS_HEADERS) else: raise Exception("%s is not implemented as a source." % source) self.k.set_acl(acl) # setting the file permissions self.k.close() # not sure if it is needed. Somewhere I read it is recommended. self.printv("%s %s to %s" % (action_word, local_file, target_file)) # if it is supposed to delete the local file after uploading if del_after_upload and source == "filename": try: os.remove(local_file) except: logger.error("Unable to delete the file: ", local_file, exc_info=True) return True except: logger.error("Error in writing to %s", target_file, exc_info=True) return False
[ "def", "__put_key", "(", "self", ",", "local_file", ",", "target_file", ",", "acl", "=", "'public-read'", ",", "del_after_upload", "=", "False", ",", "overwrite", "=", "True", ",", "source", "=", "\"filename\"", ")", ":", "action_word", "=", "\"moving\"", "if", "del_after_upload", "else", "\"copying\"", "try", ":", "self", ".", "k", ".", "key", "=", "target_file", "# setting the path (key) of file in the container", "if", "source", "==", "\"filename\"", ":", "# grabs the contents from local_file address. Note that it loads the whole file into memory", "self", ".", "k", ".", "set_contents_from_filename", "(", "local_file", ",", "self", ".", "AWS_HEADERS", ")", "elif", "source", "==", "\"fileobj\"", ":", "self", ".", "k", ".", "set_contents_from_file", "(", "local_file", ",", "self", ".", "AWS_HEADERS", ")", "elif", "source", "==", "\"string\"", ":", "self", ".", "k", ".", "set_contents_from_string", "(", "local_file", ",", "self", ".", "AWS_HEADERS", ")", "else", ":", "raise", "Exception", "(", "\"%s is not implemented as a source.\"", "%", "source", ")", "self", ".", "k", ".", "set_acl", "(", "acl", ")", "# setting the file permissions", "self", ".", "k", ".", "close", "(", ")", "# not sure if it is needed. Somewhere I read it is recommended.", "self", ".", "printv", "(", "\"%s %s to %s\"", "%", "(", "action_word", ",", "local_file", ",", "target_file", ")", ")", "# if it is supposed to delete the local file after uploading", "if", "del_after_upload", "and", "source", "==", "\"filename\"", ":", "try", ":", "os", ".", "remove", "(", "local_file", ")", "except", ":", "logger", ".", "error", "(", "\"Unable to delete the file: \"", ",", "local_file", ",", "exc_info", "=", "True", ")", "return", "True", "except", ":", "logger", ".", "error", "(", "\"Error in writing to %s\"", ",", "target_file", ",", "exc_info", "=", "True", ")", "return", "False" ]
Copy a file to s3.
[ "Copy", "a", "file", "to", "s3", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L207-L238
train
seperman/s3utils
s3utils/s3utils.py
S3utils.cp
def cp(self, local_path, target_path, acl='public-read', del_after_upload=False, overwrite=True, invalidate=False): """ Copy a file or folder from local to s3. Parameters ---------- local_path : string Path to file or folder. Or if you want to copy only the contents of folder, add /* at the end of folder name target_path : string Target path on S3 bucket. acl : string, optional File permissions on S3. Default is public-read options: - private: Owner gets FULL_CONTROL. No one else has any access rights. - public-read: Owners gets FULL_CONTROL and the anonymous principal is granted READ access. - public-read-write: Owner gets FULL_CONTROL and the anonymous principal is granted READ and WRITE access. - authenticated-read: Owner gets FULL_CONTROL and any principal authenticated as a registered Amazon S3 user is granted READ access del_after_upload : boolean, optional delete the local file after uploading. This is effectively like moving the file. You can use s3utils.mv instead of s3utils.cp to move files from local to S3. It basically sets this flag to True. default = False overwrite : boolean, optional overwrites files on S3 if set to True. Default is True invalidate : boolean, optional invalidates the CDN (a.k.a Distribution) cache if the file already exists on S3 default = False Note that invalidation might take up to 15 minutes to take place. It is easier and faster to use cache buster to grab lastest version of your file on CDN than invalidation. **Returns** Nothing on success but it will return what went wrong if something fails. Examples -------- >>> s3utils.cp("path/to/folder","/test/") copying /path/to/myfolder/test2.txt to test/myfolder/test2.txt copying /path/to/myfolder/test.txt to test/myfolder/test.txt copying /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG copying /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff >>> # When overwrite is set to False, it returns the file(s) that were already existing on s3 and were not overwritten. >>> s3utils.cp("/tmp/test3.txt", "test3.txt", overwrite=False) ERROR:root:test3.txt already exist. Not overwriting. >>> {'existing_files': {'test3.txt'}} >>> # To overwrite the files on S3 and invalidate the CDN (cloudfront) cache so the new file goes on CDN: >>> s3utils.cp("path/to/folder","/test/", invalidate=True) copying /path/to/myfolder/test2.txt to test/myfolder/test2.txt copying /path/to/myfolder/test.txt to test/myfolder/test.txt copying /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG copying /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff >>> # When file does not exist, it returns a dictionary of what went wrong. >>> s3utils.cp("/tmp/does_not_exist", "somewhere") ERROR:root:trying to upload to s3 but file doesn't exist: /tmp/does_not_exist >>> {'file_does_not_exist': '/tmp/does_not_exist'} """ result = None if overwrite: list_of_files = [] else: list_of_files = self.ls(folder=target_path, begin_from_file="", num=-1, get_grants=False, all_grant_data=False) # copying the contents of the folder and not folder itself if local_path.endswith("/*"): local_path = local_path[:-2] target_path = re.sub(r"^/|/$", "", target_path) # Amazon S3 doesn't let the name to begin with / # copying folder too else: local_base_name = os.path.basename(local_path) local_path = re.sub(r"/$", "", local_path) target_path = re.sub(r"^/", "", target_path) if not target_path.endswith(local_base_name): target_path = os.path.join(target_path, local_base_name) if os.path.exists(local_path): result = self.__find_files_and_copy(local_path, target_path, acl, del_after_upload, overwrite, invalidate, list_of_files) else: result = {'file_does_not_exist': local_path} logger.error("trying to upload to s3 but file doesn't exist: %s" % local_path) return result
python
def cp(self, local_path, target_path, acl='public-read', del_after_upload=False, overwrite=True, invalidate=False): """ Copy a file or folder from local to s3. Parameters ---------- local_path : string Path to file or folder. Or if you want to copy only the contents of folder, add /* at the end of folder name target_path : string Target path on S3 bucket. acl : string, optional File permissions on S3. Default is public-read options: - private: Owner gets FULL_CONTROL. No one else has any access rights. - public-read: Owners gets FULL_CONTROL and the anonymous principal is granted READ access. - public-read-write: Owner gets FULL_CONTROL and the anonymous principal is granted READ and WRITE access. - authenticated-read: Owner gets FULL_CONTROL and any principal authenticated as a registered Amazon S3 user is granted READ access del_after_upload : boolean, optional delete the local file after uploading. This is effectively like moving the file. You can use s3utils.mv instead of s3utils.cp to move files from local to S3. It basically sets this flag to True. default = False overwrite : boolean, optional overwrites files on S3 if set to True. Default is True invalidate : boolean, optional invalidates the CDN (a.k.a Distribution) cache if the file already exists on S3 default = False Note that invalidation might take up to 15 minutes to take place. It is easier and faster to use cache buster to grab lastest version of your file on CDN than invalidation. **Returns** Nothing on success but it will return what went wrong if something fails. Examples -------- >>> s3utils.cp("path/to/folder","/test/") copying /path/to/myfolder/test2.txt to test/myfolder/test2.txt copying /path/to/myfolder/test.txt to test/myfolder/test.txt copying /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG copying /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff >>> # When overwrite is set to False, it returns the file(s) that were already existing on s3 and were not overwritten. >>> s3utils.cp("/tmp/test3.txt", "test3.txt", overwrite=False) ERROR:root:test3.txt already exist. Not overwriting. >>> {'existing_files': {'test3.txt'}} >>> # To overwrite the files on S3 and invalidate the CDN (cloudfront) cache so the new file goes on CDN: >>> s3utils.cp("path/to/folder","/test/", invalidate=True) copying /path/to/myfolder/test2.txt to test/myfolder/test2.txt copying /path/to/myfolder/test.txt to test/myfolder/test.txt copying /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG copying /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff >>> # When file does not exist, it returns a dictionary of what went wrong. >>> s3utils.cp("/tmp/does_not_exist", "somewhere") ERROR:root:trying to upload to s3 but file doesn't exist: /tmp/does_not_exist >>> {'file_does_not_exist': '/tmp/does_not_exist'} """ result = None if overwrite: list_of_files = [] else: list_of_files = self.ls(folder=target_path, begin_from_file="", num=-1, get_grants=False, all_grant_data=False) # copying the contents of the folder and not folder itself if local_path.endswith("/*"): local_path = local_path[:-2] target_path = re.sub(r"^/|/$", "", target_path) # Amazon S3 doesn't let the name to begin with / # copying folder too else: local_base_name = os.path.basename(local_path) local_path = re.sub(r"/$", "", local_path) target_path = re.sub(r"^/", "", target_path) if not target_path.endswith(local_base_name): target_path = os.path.join(target_path, local_base_name) if os.path.exists(local_path): result = self.__find_files_and_copy(local_path, target_path, acl, del_after_upload, overwrite, invalidate, list_of_files) else: result = {'file_does_not_exist': local_path} logger.error("trying to upload to s3 but file doesn't exist: %s" % local_path) return result
[ "def", "cp", "(", "self", ",", "local_path", ",", "target_path", ",", "acl", "=", "'public-read'", ",", "del_after_upload", "=", "False", ",", "overwrite", "=", "True", ",", "invalidate", "=", "False", ")", ":", "result", "=", "None", "if", "overwrite", ":", "list_of_files", "=", "[", "]", "else", ":", "list_of_files", "=", "self", ".", "ls", "(", "folder", "=", "target_path", ",", "begin_from_file", "=", "\"\"", ",", "num", "=", "-", "1", ",", "get_grants", "=", "False", ",", "all_grant_data", "=", "False", ")", "# copying the contents of the folder and not folder itself", "if", "local_path", ".", "endswith", "(", "\"/*\"", ")", ":", "local_path", "=", "local_path", "[", ":", "-", "2", "]", "target_path", "=", "re", ".", "sub", "(", "r\"^/|/$\"", ",", "\"\"", ",", "target_path", ")", "# Amazon S3 doesn't let the name to begin with /", "# copying folder too", "else", ":", "local_base_name", "=", "os", ".", "path", ".", "basename", "(", "local_path", ")", "local_path", "=", "re", ".", "sub", "(", "r\"/$\"", ",", "\"\"", ",", "local_path", ")", "target_path", "=", "re", ".", "sub", "(", "r\"^/\"", ",", "\"\"", ",", "target_path", ")", "if", "not", "target_path", ".", "endswith", "(", "local_base_name", ")", ":", "target_path", "=", "os", ".", "path", ".", "join", "(", "target_path", ",", "local_base_name", ")", "if", "os", ".", "path", ".", "exists", "(", "local_path", ")", ":", "result", "=", "self", ".", "__find_files_and_copy", "(", "local_path", ",", "target_path", ",", "acl", ",", "del_after_upload", ",", "overwrite", ",", "invalidate", ",", "list_of_files", ")", "else", ":", "result", "=", "{", "'file_does_not_exist'", ":", "local_path", "}", "logger", ".", "error", "(", "\"trying to upload to s3 but file doesn't exist: %s\"", "%", "local_path", ")", "return", "result" ]
Copy a file or folder from local to s3. Parameters ---------- local_path : string Path to file or folder. Or if you want to copy only the contents of folder, add /* at the end of folder name target_path : string Target path on S3 bucket. acl : string, optional File permissions on S3. Default is public-read options: - private: Owner gets FULL_CONTROL. No one else has any access rights. - public-read: Owners gets FULL_CONTROL and the anonymous principal is granted READ access. - public-read-write: Owner gets FULL_CONTROL and the anonymous principal is granted READ and WRITE access. - authenticated-read: Owner gets FULL_CONTROL and any principal authenticated as a registered Amazon S3 user is granted READ access del_after_upload : boolean, optional delete the local file after uploading. This is effectively like moving the file. You can use s3utils.mv instead of s3utils.cp to move files from local to S3. It basically sets this flag to True. default = False overwrite : boolean, optional overwrites files on S3 if set to True. Default is True invalidate : boolean, optional invalidates the CDN (a.k.a Distribution) cache if the file already exists on S3 default = False Note that invalidation might take up to 15 minutes to take place. It is easier and faster to use cache buster to grab lastest version of your file on CDN than invalidation. **Returns** Nothing on success but it will return what went wrong if something fails. Examples -------- >>> s3utils.cp("path/to/folder","/test/") copying /path/to/myfolder/test2.txt to test/myfolder/test2.txt copying /path/to/myfolder/test.txt to test/myfolder/test.txt copying /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG copying /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff >>> # When overwrite is set to False, it returns the file(s) that were already existing on s3 and were not overwritten. >>> s3utils.cp("/tmp/test3.txt", "test3.txt", overwrite=False) ERROR:root:test3.txt already exist. Not overwriting. >>> {'existing_files': {'test3.txt'}} >>> # To overwrite the files on S3 and invalidate the CDN (cloudfront) cache so the new file goes on CDN: >>> s3utils.cp("path/to/folder","/test/", invalidate=True) copying /path/to/myfolder/test2.txt to test/myfolder/test2.txt copying /path/to/myfolder/test.txt to test/myfolder/test.txt copying /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG copying /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff >>> # When file does not exist, it returns a dictionary of what went wrong. >>> s3utils.cp("/tmp/does_not_exist", "somewhere") ERROR:root:trying to upload to s3 but file doesn't exist: /tmp/does_not_exist >>> {'file_does_not_exist': '/tmp/does_not_exist'}
[ "Copy", "a", "file", "or", "folder", "from", "local", "to", "s3", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L240-L336
train
seperman/s3utils
s3utils/s3utils.py
S3utils.mv
def mv(self, local_file, target_file, acl='public-read', overwrite=True, invalidate=False): """ Similar to Linux mv command. Move the file to the S3 and deletes the local copy It is basically s3utils.cp that has del_after_upload=True Examples -------- >>> s3utils.mv("path/to/folder","/test/") moving /path/to/myfolder/test2.txt to test/myfolder/test2.txt moving /path/to/myfolder/test.txt to test/myfolder/test.txt moving /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG moving /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff **Returns:** Nothing on success, otherwise what went wrong. Return type: dict """ self.cp(local_file, target_file, acl=acl, del_after_upload=True, overwrite=overwrite, invalidate=invalidate)
python
def mv(self, local_file, target_file, acl='public-read', overwrite=True, invalidate=False): """ Similar to Linux mv command. Move the file to the S3 and deletes the local copy It is basically s3utils.cp that has del_after_upload=True Examples -------- >>> s3utils.mv("path/to/folder","/test/") moving /path/to/myfolder/test2.txt to test/myfolder/test2.txt moving /path/to/myfolder/test.txt to test/myfolder/test.txt moving /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG moving /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff **Returns:** Nothing on success, otherwise what went wrong. Return type: dict """ self.cp(local_file, target_file, acl=acl, del_after_upload=True, overwrite=overwrite, invalidate=invalidate)
[ "def", "mv", "(", "self", ",", "local_file", ",", "target_file", ",", "acl", "=", "'public-read'", ",", "overwrite", "=", "True", ",", "invalidate", "=", "False", ")", ":", "self", ".", "cp", "(", "local_file", ",", "target_file", ",", "acl", "=", "acl", ",", "del_after_upload", "=", "True", ",", "overwrite", "=", "overwrite", ",", "invalidate", "=", "invalidate", ")" ]
Similar to Linux mv command. Move the file to the S3 and deletes the local copy It is basically s3utils.cp that has del_after_upload=True Examples -------- >>> s3utils.mv("path/to/folder","/test/") moving /path/to/myfolder/test2.txt to test/myfolder/test2.txt moving /path/to/myfolder/test.txt to test/myfolder/test.txt moving /path/to/myfolder/hoho/photo.JPG to test/myfolder/hoho/photo.JPG moving /path/to/myfolder/hoho/haha/ff to test/myfolder/hoho/haha/ff **Returns:** Nothing on success, otherwise what went wrong. Return type: dict
[ "Similar", "to", "Linux", "mv", "command", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L483-L507
train
seperman/s3utils
s3utils/s3utils.py
S3utils.cp_cropduster_image
def cp_cropduster_image(self, the_image_path, del_after_upload=False, overwrite=False, invalidate=False): """ Deal with saving cropduster images to S3. Cropduster is a Django library for resizing editorial images. S3utils was originally written to put cropduster images on S3 bucket. Extra Items in your Django Settings ----------------------------------- MEDIA_ROOT : string Django media root. Currently it is ONLY used in cp_cropduster_image method. NOT any other method as this library was originally made to put Django cropduster images on s3 bucket. S3_ROOT_BASE : string S3 media root base. This will be the root folder in S3. Currently it is ONLY used in cp_cropduster_image method. NOT any other method as this library was originally made to put Django cropduster images on s3 bucket. """ local_file = os.path.join(settings.MEDIA_ROOT, the_image_path) # only try to upload things if the origin cropduster file exists (so it is not already uploaded to the CDN) if os.path.exists(local_file): the_image_crops_path = os.path.splitext(the_image_path)[0] the_image_crops_path_full_path = os.path.join(settings.MEDIA_ROOT, the_image_crops_path) self.cp(local_path=local_file, target_path=os.path.join(settings.S3_ROOT_BASE, the_image_path), del_after_upload=del_after_upload, overwrite=overwrite, invalidate=invalidate, ) self.cp(local_path=the_image_crops_path_full_path + "/*", target_path=os.path.join(settings.S3_ROOT_BASE, the_image_crops_path), del_after_upload=del_after_upload, overwrite=overwrite, invalidate=invalidate, )
python
def cp_cropduster_image(self, the_image_path, del_after_upload=False, overwrite=False, invalidate=False): """ Deal with saving cropduster images to S3. Cropduster is a Django library for resizing editorial images. S3utils was originally written to put cropduster images on S3 bucket. Extra Items in your Django Settings ----------------------------------- MEDIA_ROOT : string Django media root. Currently it is ONLY used in cp_cropduster_image method. NOT any other method as this library was originally made to put Django cropduster images on s3 bucket. S3_ROOT_BASE : string S3 media root base. This will be the root folder in S3. Currently it is ONLY used in cp_cropduster_image method. NOT any other method as this library was originally made to put Django cropduster images on s3 bucket. """ local_file = os.path.join(settings.MEDIA_ROOT, the_image_path) # only try to upload things if the origin cropduster file exists (so it is not already uploaded to the CDN) if os.path.exists(local_file): the_image_crops_path = os.path.splitext(the_image_path)[0] the_image_crops_path_full_path = os.path.join(settings.MEDIA_ROOT, the_image_crops_path) self.cp(local_path=local_file, target_path=os.path.join(settings.S3_ROOT_BASE, the_image_path), del_after_upload=del_after_upload, overwrite=overwrite, invalidate=invalidate, ) self.cp(local_path=the_image_crops_path_full_path + "/*", target_path=os.path.join(settings.S3_ROOT_BASE, the_image_crops_path), del_after_upload=del_after_upload, overwrite=overwrite, invalidate=invalidate, )
[ "def", "cp_cropduster_image", "(", "self", ",", "the_image_path", ",", "del_after_upload", "=", "False", ",", "overwrite", "=", "False", ",", "invalidate", "=", "False", ")", ":", "local_file", "=", "os", ".", "path", ".", "join", "(", "settings", ".", "MEDIA_ROOT", ",", "the_image_path", ")", "# only try to upload things if the origin cropduster file exists (so it is not already uploaded to the CDN)", "if", "os", ".", "path", ".", "exists", "(", "local_file", ")", ":", "the_image_crops_path", "=", "os", ".", "path", ".", "splitext", "(", "the_image_path", ")", "[", "0", "]", "the_image_crops_path_full_path", "=", "os", ".", "path", ".", "join", "(", "settings", ".", "MEDIA_ROOT", ",", "the_image_crops_path", ")", "self", ".", "cp", "(", "local_path", "=", "local_file", ",", "target_path", "=", "os", ".", "path", ".", "join", "(", "settings", ".", "S3_ROOT_BASE", ",", "the_image_path", ")", ",", "del_after_upload", "=", "del_after_upload", ",", "overwrite", "=", "overwrite", ",", "invalidate", "=", "invalidate", ",", ")", "self", ".", "cp", "(", "local_path", "=", "the_image_crops_path_full_path", "+", "\"/*\"", ",", "target_path", "=", "os", ".", "path", ".", "join", "(", "settings", ".", "S3_ROOT_BASE", ",", "the_image_crops_path", ")", ",", "del_after_upload", "=", "del_after_upload", ",", "overwrite", "=", "overwrite", ",", "invalidate", "=", "invalidate", ",", ")" ]
Deal with saving cropduster images to S3. Cropduster is a Django library for resizing editorial images. S3utils was originally written to put cropduster images on S3 bucket. Extra Items in your Django Settings ----------------------------------- MEDIA_ROOT : string Django media root. Currently it is ONLY used in cp_cropduster_image method. NOT any other method as this library was originally made to put Django cropduster images on s3 bucket. S3_ROOT_BASE : string S3 media root base. This will be the root folder in S3. Currently it is ONLY used in cp_cropduster_image method. NOT any other method as this library was originally made to put Django cropduster images on s3 bucket.
[ "Deal", "with", "saving", "cropduster", "images", "to", "S3", ".", "Cropduster", "is", "a", "Django", "library", "for", "resizing", "editorial", "images", ".", "S3utils", "was", "originally", "written", "to", "put", "cropduster", "images", "on", "S3", "bucket", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L510-L551
train
seperman/s3utils
s3utils/s3utils.py
S3utils.chmod
def chmod(self, target_file, acl='public-read'): """ sets permissions for a file on S3 Parameters ---------- target_file : string Path to file on S3 acl : string, optional File permissions on S3. Default is public-read options: - private: Owner gets FULL_CONTROL. No one else has any access rights. - public-read: Owners gets FULL_CONTROL and the anonymous principal is granted READ access. - public-read-write: Owner gets FULL_CONTROL and the anonymous principal is granted READ and WRITE access. - authenticated-read: Owner gets FULL_CONTROL and any principal authenticated as a registered Amazon S3 user is granted READ access Examples -------- >>> s3utils.chmod("path/to/file","private") """ self.k.key = target_file # setting the path (key) of file in the container self.k.set_acl(acl) # setting the file permissions self.k.close()
python
def chmod(self, target_file, acl='public-read'): """ sets permissions for a file on S3 Parameters ---------- target_file : string Path to file on S3 acl : string, optional File permissions on S3. Default is public-read options: - private: Owner gets FULL_CONTROL. No one else has any access rights. - public-read: Owners gets FULL_CONTROL and the anonymous principal is granted READ access. - public-read-write: Owner gets FULL_CONTROL and the anonymous principal is granted READ and WRITE access. - authenticated-read: Owner gets FULL_CONTROL and any principal authenticated as a registered Amazon S3 user is granted READ access Examples -------- >>> s3utils.chmod("path/to/file","private") """ self.k.key = target_file # setting the path (key) of file in the container self.k.set_acl(acl) # setting the file permissions self.k.close()
[ "def", "chmod", "(", "self", ",", "target_file", ",", "acl", "=", "'public-read'", ")", ":", "self", ".", "k", ".", "key", "=", "target_file", "# setting the path (key) of file in the container", "self", ".", "k", ".", "set_acl", "(", "acl", ")", "# setting the file permissions", "self", ".", "k", ".", "close", "(", ")" ]
sets permissions for a file on S3 Parameters ---------- target_file : string Path to file on S3 acl : string, optional File permissions on S3. Default is public-read options: - private: Owner gets FULL_CONTROL. No one else has any access rights. - public-read: Owners gets FULL_CONTROL and the anonymous principal is granted READ access. - public-read-write: Owner gets FULL_CONTROL and the anonymous principal is granted READ and WRITE access. - authenticated-read: Owner gets FULL_CONTROL and any principal authenticated as a registered Amazon S3 user is granted READ access Examples -------- >>> s3utils.chmod("path/to/file","private")
[ "sets", "permissions", "for", "a", "file", "on", "S3" ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L582-L610
train
seperman/s3utils
s3utils/s3utils.py
S3utils.ll
def ll(self, folder="", begin_from_file="", num=-1, all_grant_data=False): """ Get the list of files and permissions from S3. This is similar to LL (ls -lah) in Linux: List of files with permissions. Parameters ---------- folder : string Path to file on S3 num: integer, optional number of results to return, by default it returns all results. begin_from_file : string, optional which file to start from on S3. This is usedful in case you are iterating over lists of files and you need to page the result by starting listing from a certain file and fetching certain num (number) of files. all_grant_data : Boolean, optional More detailed file permission data will be returned. Examples -------- >>> from s3utils import S3utils >>> s3utils = S3utils( ... AWS_ACCESS_KEY_ID = 'your access key', ... AWS_SECRET_ACCESS_KEY = 'your secret key', ... AWS_STORAGE_BUCKET_NAME = 'your bucket name', ... S3UTILS_DEBUG_LEVEL = 1, #change it to 0 for less verbose ... ) >>> import json >>> # We use json.dumps to print the results more readable: >>> my_folder_stuff = s3utils.ll("/test/") >>> print(json.dumps(my_folder_stuff, indent=2)) { "test/myfolder/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/em/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/.DS_Store": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], "test/myfolder/hoho/haha/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/haha/ff": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], "test/myfolder/hoho/photo.JPG": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], } """ return self.ls(folder=folder, begin_from_file=begin_from_file, num=num, get_grants=True, all_grant_data=all_grant_data)
python
def ll(self, folder="", begin_from_file="", num=-1, all_grant_data=False): """ Get the list of files and permissions from S3. This is similar to LL (ls -lah) in Linux: List of files with permissions. Parameters ---------- folder : string Path to file on S3 num: integer, optional number of results to return, by default it returns all results. begin_from_file : string, optional which file to start from on S3. This is usedful in case you are iterating over lists of files and you need to page the result by starting listing from a certain file and fetching certain num (number) of files. all_grant_data : Boolean, optional More detailed file permission data will be returned. Examples -------- >>> from s3utils import S3utils >>> s3utils = S3utils( ... AWS_ACCESS_KEY_ID = 'your access key', ... AWS_SECRET_ACCESS_KEY = 'your secret key', ... AWS_STORAGE_BUCKET_NAME = 'your bucket name', ... S3UTILS_DEBUG_LEVEL = 1, #change it to 0 for less verbose ... ) >>> import json >>> # We use json.dumps to print the results more readable: >>> my_folder_stuff = s3utils.ll("/test/") >>> print(json.dumps(my_folder_stuff, indent=2)) { "test/myfolder/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/em/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/.DS_Store": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], "test/myfolder/hoho/haha/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/haha/ff": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], "test/myfolder/hoho/photo.JPG": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], } """ return self.ls(folder=folder, begin_from_file=begin_from_file, num=num, get_grants=True, all_grant_data=all_grant_data)
[ "def", "ll", "(", "self", ",", "folder", "=", "\"\"", ",", "begin_from_file", "=", "\"\"", ",", "num", "=", "-", "1", ",", "all_grant_data", "=", "False", ")", ":", "return", "self", ".", "ls", "(", "folder", "=", "folder", ",", "begin_from_file", "=", "begin_from_file", ",", "num", "=", "num", ",", "get_grants", "=", "True", ",", "all_grant_data", "=", "all_grant_data", ")" ]
Get the list of files and permissions from S3. This is similar to LL (ls -lah) in Linux: List of files with permissions. Parameters ---------- folder : string Path to file on S3 num: integer, optional number of results to return, by default it returns all results. begin_from_file : string, optional which file to start from on S3. This is usedful in case you are iterating over lists of files and you need to page the result by starting listing from a certain file and fetching certain num (number) of files. all_grant_data : Boolean, optional More detailed file permission data will be returned. Examples -------- >>> from s3utils import S3utils >>> s3utils = S3utils( ... AWS_ACCESS_KEY_ID = 'your access key', ... AWS_SECRET_ACCESS_KEY = 'your secret key', ... AWS_STORAGE_BUCKET_NAME = 'your bucket name', ... S3UTILS_DEBUG_LEVEL = 1, #change it to 0 for less verbose ... ) >>> import json >>> # We use json.dumps to print the results more readable: >>> my_folder_stuff = s3utils.ll("/test/") >>> print(json.dumps(my_folder_stuff, indent=2)) { "test/myfolder/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/em/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/.DS_Store": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], "test/myfolder/hoho/haha/": [ { "name": "owner's name", "permission": "FULL_CONTROL" } ], "test/myfolder/hoho/haha/ff": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], "test/myfolder/hoho/photo.JPG": [ { "name": "owner's name", "permission": "FULL_CONTROL" }, { "name": null, "permission": "READ" } ], }
[ "Get", "the", "list", "of", "files", "and", "permissions", "from", "S3", "." ]
aea41388a023dcf1e95588402077e31097514cf1
https://github.com/seperman/s3utils/blob/aea41388a023dcf1e95588402077e31097514cf1/s3utils/s3utils.py#L669-L764
train
uktrade/directory-signature-auth
sigauth/helpers.py
get_path
def get_path(url): """ Get the path from a given url, including the querystring. Args: url (str) Returns: str """ url = urlsplit(url) path = url.path if url.query: path += "?{}".format(url.query) return path
python
def get_path(url): """ Get the path from a given url, including the querystring. Args: url (str) Returns: str """ url = urlsplit(url) path = url.path if url.query: path += "?{}".format(url.query) return path
[ "def", "get_path", "(", "url", ")", ":", "url", "=", "urlsplit", "(", "url", ")", "path", "=", "url", ".", "path", "if", "url", ".", "query", ":", "path", "+=", "\"?{}\"", ".", "format", "(", "url", ".", "query", ")", "return", "path" ]
Get the path from a given url, including the querystring. Args: url (str) Returns: str
[ "Get", "the", "path", "from", "a", "given", "url", "including", "the", "querystring", "." ]
1a1b1e887b25a938133d7bcc146d3fecf1079313
https://github.com/uktrade/directory-signature-auth/blob/1a1b1e887b25a938133d7bcc146d3fecf1079313/sigauth/helpers.py#L79-L94
train
vkruoso/receita-tools
receita/tools/build.py
Build.run
def run(self): """Reads data from disk and generates CSV files.""" # Try to create the directory if not os.path.exists(self.output): try: os.mkdir(self.output) except: print 'failed to create output directory %s' % self.output # Be sure it is a directory if not os.path.isdir(self.output): print 'invalid output directory %s' % self.output sys.exit(1) # Create the CSV handlers visitors = [ _CompaniesCSV(self.output), _ActivitiesCSV(self.output), _ActivitiesSeenCSV(self.output), _QSACSV(self.output), ] # Run by each company populating the CSV files for path in glob.glob(os.path.join(self.input, '*.json')): with open(path, 'r') as f: try: data = json.load(f, encoding='utf-8') except ValueError: continue for visitor in visitors: visitor.visit(data)
python
def run(self): """Reads data from disk and generates CSV files.""" # Try to create the directory if not os.path.exists(self.output): try: os.mkdir(self.output) except: print 'failed to create output directory %s' % self.output # Be sure it is a directory if not os.path.isdir(self.output): print 'invalid output directory %s' % self.output sys.exit(1) # Create the CSV handlers visitors = [ _CompaniesCSV(self.output), _ActivitiesCSV(self.output), _ActivitiesSeenCSV(self.output), _QSACSV(self.output), ] # Run by each company populating the CSV files for path in glob.glob(os.path.join(self.input, '*.json')): with open(path, 'r') as f: try: data = json.load(f, encoding='utf-8') except ValueError: continue for visitor in visitors: visitor.visit(data)
[ "def", "run", "(", "self", ")", ":", "# Try to create the directory", "if", "not", "os", ".", "path", ".", "exists", "(", "self", ".", "output", ")", ":", "try", ":", "os", ".", "mkdir", "(", "self", ".", "output", ")", "except", ":", "print", "'failed to create output directory %s'", "%", "self", ".", "output", "# Be sure it is a directory", "if", "not", "os", ".", "path", ".", "isdir", "(", "self", ".", "output", ")", ":", "print", "'invalid output directory %s'", "%", "self", ".", "output", "sys", ".", "exit", "(", "1", ")", "# Create the CSV handlers", "visitors", "=", "[", "_CompaniesCSV", "(", "self", ".", "output", ")", ",", "_ActivitiesCSV", "(", "self", ".", "output", ")", ",", "_ActivitiesSeenCSV", "(", "self", ".", "output", ")", ",", "_QSACSV", "(", "self", ".", "output", ")", ",", "]", "# Run by each company populating the CSV files", "for", "path", "in", "glob", ".", "glob", "(", "os", ".", "path", ".", "join", "(", "self", ".", "input", ",", "'*.json'", ")", ")", ":", "with", "open", "(", "path", ",", "'r'", ")", "as", "f", ":", "try", ":", "data", "=", "json", ".", "load", "(", "f", ",", "encoding", "=", "'utf-8'", ")", "except", "ValueError", ":", "continue", "for", "visitor", "in", "visitors", ":", "visitor", ".", "visit", "(", "data", ")" ]
Reads data from disk and generates CSV files.
[ "Reads", "data", "from", "disk", "and", "generates", "CSV", "files", "." ]
fd62a252c76541c9feac6470b9048b31348ffe86
https://github.com/vkruoso/receita-tools/blob/fd62a252c76541c9feac6470b9048b31348ffe86/receita/tools/build.py#L144-L175
train
marrow/mongo
marrow/mongo/core/index.py
Index.process_fields
def process_fields(self, fields): """Process a list of simple string field definitions and assign their order based on prefix.""" result = [] strip = ''.join(self.PREFIX_MAP) for field in fields: direction = self.PREFIX_MAP[''] if field[0] in self.PREFIX_MAP: direction = self.PREFIX_MAP[field[0]] field = field.lstrip(strip) result.append((field, direction)) return result
python
def process_fields(self, fields): """Process a list of simple string field definitions and assign their order based on prefix.""" result = [] strip = ''.join(self.PREFIX_MAP) for field in fields: direction = self.PREFIX_MAP[''] if field[0] in self.PREFIX_MAP: direction = self.PREFIX_MAP[field[0]] field = field.lstrip(strip) result.append((field, direction)) return result
[ "def", "process_fields", "(", "self", ",", "fields", ")", ":", "result", "=", "[", "]", "strip", "=", "''", ".", "join", "(", "self", ".", "PREFIX_MAP", ")", "for", "field", "in", "fields", ":", "direction", "=", "self", ".", "PREFIX_MAP", "[", "''", "]", "if", "field", "[", "0", "]", "in", "self", ".", "PREFIX_MAP", ":", "direction", "=", "self", ".", "PREFIX_MAP", "[", "field", "[", "0", "]", "]", "field", "=", "field", ".", "lstrip", "(", "strip", ")", "result", ".", "append", "(", "(", "field", ",", "direction", ")", ")", "return", "result" ]
Process a list of simple string field definitions and assign their order based on prefix.
[ "Process", "a", "list", "of", "simple", "string", "field", "definitions", "and", "assign", "their", "order", "based", "on", "prefix", "." ]
2066dc73e281b8a46cb5fc965267d6b8e1b18467
https://github.com/marrow/mongo/blob/2066dc73e281b8a46cb5fc965267d6b8e1b18467/marrow/mongo/core/index.py#L60-L75
train
svartalf/python-2gis
dgis/__init__.py
API.search_in_rubric
def search_in_rubric(self, **kwargs): """Firms search in rubric http://api.2gis.ru/doc/firms/searches/searchinrubric/ """ point = kwargs.pop('point', False) if point: kwargs['point'] = '%s,%s' % point bound = kwargs.pop('bound', False) if bound: kwargs['bound[point1]'] = bound[0] kwargs['bound[point2]'] = bound[1] filters = kwargs.pop('filters', False) if filters: for k, v in filters.items(): kwargs['filters[%s]' % k] = v return self._search_in_rubric(**kwargs)
python
def search_in_rubric(self, **kwargs): """Firms search in rubric http://api.2gis.ru/doc/firms/searches/searchinrubric/ """ point = kwargs.pop('point', False) if point: kwargs['point'] = '%s,%s' % point bound = kwargs.pop('bound', False) if bound: kwargs['bound[point1]'] = bound[0] kwargs['bound[point2]'] = bound[1] filters = kwargs.pop('filters', False) if filters: for k, v in filters.items(): kwargs['filters[%s]' % k] = v return self._search_in_rubric(**kwargs)
[ "def", "search_in_rubric", "(", "self", ",", "*", "*", "kwargs", ")", ":", "point", "=", "kwargs", ".", "pop", "(", "'point'", ",", "False", ")", "if", "point", ":", "kwargs", "[", "'point'", "]", "=", "'%s,%s'", "%", "point", "bound", "=", "kwargs", ".", "pop", "(", "'bound'", ",", "False", ")", "if", "bound", ":", "kwargs", "[", "'bound[point1]'", "]", "=", "bound", "[", "0", "]", "kwargs", "[", "'bound[point2]'", "]", "=", "bound", "[", "1", "]", "filters", "=", "kwargs", ".", "pop", "(", "'filters'", ",", "False", ")", "if", "filters", ":", "for", "k", ",", "v", "in", "filters", ".", "items", "(", ")", ":", "kwargs", "[", "'filters[%s]'", "%", "k", "]", "=", "v", "return", "self", ".", "_search_in_rubric", "(", "*", "*", "kwargs", ")" ]
Firms search in rubric http://api.2gis.ru/doc/firms/searches/searchinrubric/
[ "Firms", "search", "in", "rubric" ]
6eccd6073c99494b7abf20b38a5455cbd55d6420
https://github.com/svartalf/python-2gis/blob/6eccd6073c99494b7abf20b38a5455cbd55d6420/dgis/__init__.py#L89-L109
train
tonybaloney/retox
retox/ui.py
RetoxRefreshMixin.refresh
def refresh(self): ''' Refresh the list and the screen ''' self._screen.force_update() self._screen.refresh() self._update(1)
python
def refresh(self): ''' Refresh the list and the screen ''' self._screen.force_update() self._screen.refresh() self._update(1)
[ "def", "refresh", "(", "self", ")", ":", "self", ".", "_screen", ".", "force_update", "(", ")", "self", ".", "_screen", ".", "refresh", "(", ")", "self", ".", "_update", "(", "1", ")" ]
Refresh the list and the screen
[ "Refresh", "the", "list", "and", "the", "screen" ]
4635e31001d2ac083423f46766249ac8daca7c9c
https://github.com/tonybaloney/retox/blob/4635e31001d2ac083423f46766249ac8daca7c9c/retox/ui.py#L54-L60
train
tonybaloney/retox
retox/ui.py
VirtualEnvironmentFrame.start
def start(self, activity, action): ''' Mark an action as started :param activity: The virtualenv activity name :type activity: ``str`` :param action: The virtualenv action :type action: :class:`tox.session.Action` ''' try: self._start_action(activity, action) except ValueError: retox_log.debug("Could not find action %s in env %s" % (activity, self.name)) self.refresh()
python
def start(self, activity, action): ''' Mark an action as started :param activity: The virtualenv activity name :type activity: ``str`` :param action: The virtualenv action :type action: :class:`tox.session.Action` ''' try: self._start_action(activity, action) except ValueError: retox_log.debug("Could not find action %s in env %s" % (activity, self.name)) self.refresh()
[ "def", "start", "(", "self", ",", "activity", ",", "action", ")", ":", "try", ":", "self", ".", "_start_action", "(", "activity", ",", "action", ")", "except", "ValueError", ":", "retox_log", ".", "debug", "(", "\"Could not find action %s in env %s\"", "%", "(", "activity", ",", "self", ".", "name", ")", ")", "self", ".", "refresh", "(", ")" ]
Mark an action as started :param activity: The virtualenv activity name :type activity: ``str`` :param action: The virtualenv action :type action: :class:`tox.session.Action`
[ "Mark", "an", "action", "as", "started" ]
4635e31001d2ac083423f46766249ac8daca7c9c
https://github.com/tonybaloney/retox/blob/4635e31001d2ac083423f46766249ac8daca7c9c/retox/ui.py#L233-L247
train
tonybaloney/retox
retox/ui.py
VirtualEnvironmentFrame.stop
def stop(self, activity, action): ''' Mark a task as completed :param activity: The virtualenv activity name :type activity: ``str`` :param action: The virtualenv action :type action: :class:`tox.session.Action` ''' try: self._remove_running_action(activity, action) except ValueError: retox_log.debug("Could not find action %s in env %s" % (activity, self.name)) self._mark_action_completed(activity, action) self.refresh()
python
def stop(self, activity, action): ''' Mark a task as completed :param activity: The virtualenv activity name :type activity: ``str`` :param action: The virtualenv action :type action: :class:`tox.session.Action` ''' try: self._remove_running_action(activity, action) except ValueError: retox_log.debug("Could not find action %s in env %s" % (activity, self.name)) self._mark_action_completed(activity, action) self.refresh()
[ "def", "stop", "(", "self", ",", "activity", ",", "action", ")", ":", "try", ":", "self", ".", "_remove_running_action", "(", "activity", ",", "action", ")", "except", "ValueError", ":", "retox_log", ".", "debug", "(", "\"Could not find action %s in env %s\"", "%", "(", "activity", ",", "self", ".", "name", ")", ")", "self", ".", "_mark_action_completed", "(", "activity", ",", "action", ")", "self", ".", "refresh", "(", ")" ]
Mark a task as completed :param activity: The virtualenv activity name :type activity: ``str`` :param action: The virtualenv action :type action: :class:`tox.session.Action`
[ "Mark", "a", "task", "as", "completed" ]
4635e31001d2ac083423f46766249ac8daca7c9c
https://github.com/tonybaloney/retox/blob/4635e31001d2ac083423f46766249ac8daca7c9c/retox/ui.py#L249-L264
train
tonybaloney/retox
retox/ui.py
VirtualEnvironmentFrame.finish
def finish(self, status): ''' Move laggard tasks over :param activity: The virtualenv status :type activity: ``str`` ''' retox_log.info("Completing %s with status %s" % (self.name, status)) result = Screen.COLOUR_GREEN if not status else Screen.COLOUR_RED self.palette['title'] = (Screen.COLOUR_WHITE, Screen.A_BOLD, result) for item in list(self._task_view.options): self._task_view.options.remove(item) self._completed_view.options.append(item) self.refresh()
python
def finish(self, status): ''' Move laggard tasks over :param activity: The virtualenv status :type activity: ``str`` ''' retox_log.info("Completing %s with status %s" % (self.name, status)) result = Screen.COLOUR_GREEN if not status else Screen.COLOUR_RED self.palette['title'] = (Screen.COLOUR_WHITE, Screen.A_BOLD, result) for item in list(self._task_view.options): self._task_view.options.remove(item) self._completed_view.options.append(item) self.refresh()
[ "def", "finish", "(", "self", ",", "status", ")", ":", "retox_log", ".", "info", "(", "\"Completing %s with status %s\"", "%", "(", "self", ".", "name", ",", "status", ")", ")", "result", "=", "Screen", ".", "COLOUR_GREEN", "if", "not", "status", "else", "Screen", ".", "COLOUR_RED", "self", ".", "palette", "[", "'title'", "]", "=", "(", "Screen", ".", "COLOUR_WHITE", ",", "Screen", ".", "A_BOLD", ",", "result", ")", "for", "item", "in", "list", "(", "self", ".", "_task_view", ".", "options", ")", ":", "self", ".", "_task_view", ".", "options", ".", "remove", "(", "item", ")", "self", ".", "_completed_view", ".", "options", ".", "append", "(", "item", ")", "self", ".", "refresh", "(", ")" ]
Move laggard tasks over :param activity: The virtualenv status :type activity: ``str``
[ "Move", "laggard", "tasks", "over" ]
4635e31001d2ac083423f46766249ac8daca7c9c
https://github.com/tonybaloney/retox/blob/4635e31001d2ac083423f46766249ac8daca7c9c/retox/ui.py#L266-L279
train
tonybaloney/retox
retox/ui.py
VirtualEnvironmentFrame.reset
def reset(self): ''' Reset the frame between jobs ''' self.palette['title'] = (Screen.COLOUR_WHITE, Screen.A_BOLD, Screen.COLOUR_BLUE) self._completed_view.options = [] self._task_view.options = [] self.refresh()
python
def reset(self): ''' Reset the frame between jobs ''' self.palette['title'] = (Screen.COLOUR_WHITE, Screen.A_BOLD, Screen.COLOUR_BLUE) self._completed_view.options = [] self._task_view.options = [] self.refresh()
[ "def", "reset", "(", "self", ")", ":", "self", ".", "palette", "[", "'title'", "]", "=", "(", "Screen", ".", "COLOUR_WHITE", ",", "Screen", ".", "A_BOLD", ",", "Screen", ".", "COLOUR_BLUE", ")", "self", ".", "_completed_view", ".", "options", "=", "[", "]", "self", ".", "_task_view", ".", "options", "=", "[", "]", "self", ".", "refresh", "(", ")" ]
Reset the frame between jobs
[ "Reset", "the", "frame", "between", "jobs" ]
4635e31001d2ac083423f46766249ac8daca7c9c
https://github.com/tonybaloney/retox/blob/4635e31001d2ac083423f46766249ac8daca7c9c/retox/ui.py#L281-L288
train
ryukinix/decorating
decorating/decorator.py
Decorator.default_arguments
def default_arguments(cls): """Returns the available kwargs of the called class""" func = cls.__init__ args = func.__code__.co_varnames defaults = func.__defaults__ index = -len(defaults) return {k: v for k, v in zip(args[index:], defaults)}
python
def default_arguments(cls): """Returns the available kwargs of the called class""" func = cls.__init__ args = func.__code__.co_varnames defaults = func.__defaults__ index = -len(defaults) return {k: v for k, v in zip(args[index:], defaults)}
[ "def", "default_arguments", "(", "cls", ")", ":", "func", "=", "cls", ".", "__init__", "args", "=", "func", ".", "__code__", ".", "co_varnames", "defaults", "=", "func", ".", "__defaults__", "index", "=", "-", "len", "(", "defaults", ")", "return", "{", "k", ":", "v", "for", "k", ",", "v", "in", "zip", "(", "args", "[", "index", ":", "]", ",", "defaults", ")", "}" ]
Returns the available kwargs of the called class
[ "Returns", "the", "available", "kwargs", "of", "the", "called", "class" ]
df78c3f87800205701704c0bc0fb9b6bb908ba7e
https://github.com/ryukinix/decorating/blob/df78c3f87800205701704c0bc0fb9b6bb908ba7e/decorating/decorator.py#L134-L140
train
ryukinix/decorating
decorating/decorator.py
Decorator.recreate
def recreate(cls, *args, **kwargs): """Recreate the class based in your args, multiple uses""" cls.check_arguments(kwargs) first_is_callable = True if any(args) and callable(args[0]) else False signature = cls.default_arguments() allowed_arguments = {k: v for k, v in kwargs.items() if k in signature} if (any(allowed_arguments) or any(args)) and not first_is_callable: if any(args) and not first_is_callable: return cls(args[0], **allowed_arguments) elif any(allowed_arguments): return cls(**allowed_arguments) return cls.instances[-1] if any(cls.instances) else cls()
python
def recreate(cls, *args, **kwargs): """Recreate the class based in your args, multiple uses""" cls.check_arguments(kwargs) first_is_callable = True if any(args) and callable(args[0]) else False signature = cls.default_arguments() allowed_arguments = {k: v for k, v in kwargs.items() if k in signature} if (any(allowed_arguments) or any(args)) and not first_is_callable: if any(args) and not first_is_callable: return cls(args[0], **allowed_arguments) elif any(allowed_arguments): return cls(**allowed_arguments) return cls.instances[-1] if any(cls.instances) else cls()
[ "def", "recreate", "(", "cls", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "cls", ".", "check_arguments", "(", "kwargs", ")", "first_is_callable", "=", "True", "if", "any", "(", "args", ")", "and", "callable", "(", "args", "[", "0", "]", ")", "else", "False", "signature", "=", "cls", ".", "default_arguments", "(", ")", "allowed_arguments", "=", "{", "k", ":", "v", "for", "k", ",", "v", "in", "kwargs", ".", "items", "(", ")", "if", "k", "in", "signature", "}", "if", "(", "any", "(", "allowed_arguments", ")", "or", "any", "(", "args", ")", ")", "and", "not", "first_is_callable", ":", "if", "any", "(", "args", ")", "and", "not", "first_is_callable", ":", "return", "cls", "(", "args", "[", "0", "]", ",", "*", "*", "allowed_arguments", ")", "elif", "any", "(", "allowed_arguments", ")", ":", "return", "cls", "(", "*", "*", "allowed_arguments", ")", "return", "cls", ".", "instances", "[", "-", "1", "]", "if", "any", "(", "cls", ".", "instances", ")", "else", "cls", "(", ")" ]
Recreate the class based in your args, multiple uses
[ "Recreate", "the", "class", "based", "in", "your", "args", "multiple", "uses" ]
df78c3f87800205701704c0bc0fb9b6bb908ba7e
https://github.com/ryukinix/decorating/blob/df78c3f87800205701704c0bc0fb9b6bb908ba7e/decorating/decorator.py#L143-L155
train
ryukinix/decorating
decorating/decorator.py
Decorator.check_arguments
def check_arguments(cls, passed): """Put warnings of arguments whose can't be handle by the class""" defaults = list(cls.default_arguments().keys()) template = ("Pass arg {argument:!r} in {cname:!r}, can be a typo? " "Supported key arguments: {defaults}") fails = [] for arg in passed: if arg not in defaults: warn(template.format(argument=arg, cname=cls.__name__, defaults=defaults)) fails.append(arg) return any(fails)
python
def check_arguments(cls, passed): """Put warnings of arguments whose can't be handle by the class""" defaults = list(cls.default_arguments().keys()) template = ("Pass arg {argument:!r} in {cname:!r}, can be a typo? " "Supported key arguments: {defaults}") fails = [] for arg in passed: if arg not in defaults: warn(template.format(argument=arg, cname=cls.__name__, defaults=defaults)) fails.append(arg) return any(fails)
[ "def", "check_arguments", "(", "cls", ",", "passed", ")", ":", "defaults", "=", "list", "(", "cls", ".", "default_arguments", "(", ")", ".", "keys", "(", ")", ")", "template", "=", "(", "\"Pass arg {argument:!r} in {cname:!r}, can be a typo? \"", "\"Supported key arguments: {defaults}\"", ")", "fails", "=", "[", "]", "for", "arg", "in", "passed", ":", "if", "arg", "not", "in", "defaults", ":", "warn", "(", "template", ".", "format", "(", "argument", "=", "arg", ",", "cname", "=", "cls", ".", "__name__", ",", "defaults", "=", "defaults", ")", ")", "fails", ".", "append", "(", "arg", ")", "return", "any", "(", "fails", ")" ]
Put warnings of arguments whose can't be handle by the class
[ "Put", "warnings", "of", "arguments", "whose", "can", "t", "be", "handle", "by", "the", "class" ]
df78c3f87800205701704c0bc0fb9b6bb908ba7e
https://github.com/ryukinix/decorating/blob/df78c3f87800205701704c0bc0fb9b6bb908ba7e/decorating/decorator.py#L158-L171
train
ovnicraft/suds2
suds/builder.py
Builder.process
def process(self, data, type, history): """ process the specified type then process its children """ if type in history: return if type.enum(): return history.append(type) resolved = type.resolve() value = None if type.multi_occurrence(): value = [] else: if len(resolved) > 0: if resolved.mixed(): value = Factory.property(resolved.name) md = value.__metadata__ md.sxtype = resolved else: value = Factory.object(resolved.name) md = value.__metadata__ md.sxtype = resolved md.ordering = self.ordering(resolved) setattr(data, type.name, value) if value is not None: data = value if not isinstance(data, list): self.add_attributes(data, resolved) for child, ancestry in resolved.children(): if self.skip_child(child, ancestry): continue self.process(data, child, history[:])
python
def process(self, data, type, history): """ process the specified type then process its children """ if type in history: return if type.enum(): return history.append(type) resolved = type.resolve() value = None if type.multi_occurrence(): value = [] else: if len(resolved) > 0: if resolved.mixed(): value = Factory.property(resolved.name) md = value.__metadata__ md.sxtype = resolved else: value = Factory.object(resolved.name) md = value.__metadata__ md.sxtype = resolved md.ordering = self.ordering(resolved) setattr(data, type.name, value) if value is not None: data = value if not isinstance(data, list): self.add_attributes(data, resolved) for child, ancestry in resolved.children(): if self.skip_child(child, ancestry): continue self.process(data, child, history[:])
[ "def", "process", "(", "self", ",", "data", ",", "type", ",", "history", ")", ":", "if", "type", "in", "history", ":", "return", "if", "type", ".", "enum", "(", ")", ":", "return", "history", ".", "append", "(", "type", ")", "resolved", "=", "type", ".", "resolve", "(", ")", "value", "=", "None", "if", "type", ".", "multi_occurrence", "(", ")", ":", "value", "=", "[", "]", "else", ":", "if", "len", "(", "resolved", ")", ">", "0", ":", "if", "resolved", ".", "mixed", "(", ")", ":", "value", "=", "Factory", ".", "property", "(", "resolved", ".", "name", ")", "md", "=", "value", ".", "__metadata__", "md", ".", "sxtype", "=", "resolved", "else", ":", "value", "=", "Factory", ".", "object", "(", "resolved", ".", "name", ")", "md", "=", "value", ".", "__metadata__", "md", ".", "sxtype", "=", "resolved", "md", ".", "ordering", "=", "self", ".", "ordering", "(", "resolved", ")", "setattr", "(", "data", ",", "type", ".", "name", ",", "value", ")", "if", "value", "is", "not", "None", ":", "data", "=", "value", "if", "not", "isinstance", "(", "data", ",", "list", ")", ":", "self", ".", "add_attributes", "(", "data", ",", "resolved", ")", "for", "child", ",", "ancestry", "in", "resolved", ".", "children", "(", ")", ":", "if", "self", ".", "skip_child", "(", "child", ",", "ancestry", ")", ":", "continue", "self", ".", "process", "(", "data", ",", "child", ",", "history", "[", ":", "]", ")" ]
process the specified type then process its children
[ "process", "the", "specified", "type", "then", "process", "its", "children" ]
e5b540792206a41efc22f5d5b9cfac2dbe7a7992
https://github.com/ovnicraft/suds2/blob/e5b540792206a41efc22f5d5b9cfac2dbe7a7992/suds/builder.py#L60-L90
train
ovnicraft/suds2
suds/builder.py
Builder.skip_child
def skip_child(self, child, ancestry): """ get whether or not to skip the specified child """ if child.any(): return True for x in ancestry: if x.choice(): return True return False
python
def skip_child(self, child, ancestry): """ get whether or not to skip the specified child """ if child.any(): return True for x in ancestry: if x.choice(): return True return False
[ "def", "skip_child", "(", "self", ",", "child", ",", "ancestry", ")", ":", "if", "child", ".", "any", "(", ")", ":", "return", "True", "for", "x", "in", "ancestry", ":", "if", "x", ".", "choice", "(", ")", ":", "return", "True", "return", "False" ]
get whether or not to skip the specified child
[ "get", "whether", "or", "not", "to", "skip", "the", "specified", "child" ]
e5b540792206a41efc22f5d5b9cfac2dbe7a7992
https://github.com/ovnicraft/suds2/blob/e5b540792206a41efc22f5d5b9cfac2dbe7a7992/suds/builder.py#L99-L105
train
nephila/django-knocker
knocker/signals.py
active_knocks
def active_knocks(obj): """ Checks whether knocks are enabled for the model given as argument :param obj: model instance :return True if knocks are active """ if not hasattr(_thread_locals, 'knock_enabled'): return True return _thread_locals.knock_enabled.get(obj.__class__, True)
python
def active_knocks(obj): """ Checks whether knocks are enabled for the model given as argument :param obj: model instance :return True if knocks are active """ if not hasattr(_thread_locals, 'knock_enabled'): return True return _thread_locals.knock_enabled.get(obj.__class__, True)
[ "def", "active_knocks", "(", "obj", ")", ":", "if", "not", "hasattr", "(", "_thread_locals", ",", "'knock_enabled'", ")", ":", "return", "True", "return", "_thread_locals", ".", "knock_enabled", ".", "get", "(", "obj", ".", "__class__", ",", "True", ")" ]
Checks whether knocks are enabled for the model given as argument :param obj: model instance :return True if knocks are active
[ "Checks", "whether", "knocks", "are", "enabled", "for", "the", "model", "given", "as", "argument" ]
d25380d43a1f91285f1581dcf9db8510fe87f354
https://github.com/nephila/django-knocker/blob/d25380d43a1f91285f1581dcf9db8510fe87f354/knocker/signals.py#L34-L43
train
nephila/django-knocker
knocker/signals.py
pause_knocks
def pause_knocks(obj): """ Context manager to suspend sending knocks for the given model :param obj: model instance """ if not hasattr(_thread_locals, 'knock_enabled'): _thread_locals.knock_enabled = {} obj.__class__._disconnect() _thread_locals.knock_enabled[obj.__class__] = False yield _thread_locals.knock_enabled[obj.__class__] = True obj.__class__._connect()
python
def pause_knocks(obj): """ Context manager to suspend sending knocks for the given model :param obj: model instance """ if not hasattr(_thread_locals, 'knock_enabled'): _thread_locals.knock_enabled = {} obj.__class__._disconnect() _thread_locals.knock_enabled[obj.__class__] = False yield _thread_locals.knock_enabled[obj.__class__] = True obj.__class__._connect()
[ "def", "pause_knocks", "(", "obj", ")", ":", "if", "not", "hasattr", "(", "_thread_locals", ",", "'knock_enabled'", ")", ":", "_thread_locals", ".", "knock_enabled", "=", "{", "}", "obj", ".", "__class__", ".", "_disconnect", "(", ")", "_thread_locals", ".", "knock_enabled", "[", "obj", ".", "__class__", "]", "=", "False", "yield", "_thread_locals", ".", "knock_enabled", "[", "obj", ".", "__class__", "]", "=", "True", "obj", ".", "__class__", ".", "_connect", "(", ")" ]
Context manager to suspend sending knocks for the given model :param obj: model instance
[ "Context", "manager", "to", "suspend", "sending", "knocks", "for", "the", "given", "model" ]
d25380d43a1f91285f1581dcf9db8510fe87f354
https://github.com/nephila/django-knocker/blob/d25380d43a1f91285f1581dcf9db8510fe87f354/knocker/signals.py#L47-L59
train
tonybaloney/retox
retox/reporter.py
RetoxReporter._loopreport
def _loopreport(self): ''' Loop over the report progress ''' while 1: eventlet.sleep(0.2) ac2popenlist = {} for action in self.session._actions: for popen in action._popenlist: if popen.poll() is None: lst = ac2popenlist.setdefault(action.activity, []) lst.append(popen) if not action._popenlist and action in self._actionmayfinish: super(RetoxReporter, self).logaction_finish(action) self._actionmayfinish.remove(action) self.screen.draw_next_frame(repeat=False)
python
def _loopreport(self): ''' Loop over the report progress ''' while 1: eventlet.sleep(0.2) ac2popenlist = {} for action in self.session._actions: for popen in action._popenlist: if popen.poll() is None: lst = ac2popenlist.setdefault(action.activity, []) lst.append(popen) if not action._popenlist and action in self._actionmayfinish: super(RetoxReporter, self).logaction_finish(action) self._actionmayfinish.remove(action) self.screen.draw_next_frame(repeat=False)
[ "def", "_loopreport", "(", "self", ")", ":", "while", "1", ":", "eventlet", ".", "sleep", "(", "0.2", ")", "ac2popenlist", "=", "{", "}", "for", "action", "in", "self", ".", "session", ".", "_actions", ":", "for", "popen", "in", "action", ".", "_popenlist", ":", "if", "popen", ".", "poll", "(", ")", "is", "None", ":", "lst", "=", "ac2popenlist", ".", "setdefault", "(", "action", ".", "activity", ",", "[", "]", ")", "lst", ".", "append", "(", "popen", ")", "if", "not", "action", ".", "_popenlist", "and", "action", "in", "self", ".", "_actionmayfinish", ":", "super", "(", "RetoxReporter", ",", "self", ")", ".", "logaction_finish", "(", "action", ")", "self", ".", "_actionmayfinish", ".", "remove", "(", "action", ")", "self", ".", "screen", ".", "draw_next_frame", "(", "repeat", "=", "False", ")" ]
Loop over the report progress
[ "Loop", "over", "the", "report", "progress" ]
4635e31001d2ac083423f46766249ac8daca7c9c
https://github.com/tonybaloney/retox/blob/4635e31001d2ac083423f46766249ac8daca7c9c/retox/reporter.py#L49-L65
train
yejianye/mdmail
mdmail/api.py
send
def send(email, subject=None, from_email=None, to_email=None, cc=None, bcc=None, reply_to=None, smtp=None): """Send markdown email Args: email (str/obj): A markdown string or EmailContent object subject (str): subject line from_email (str): sender email address to_email (str/list): recipient email addresses cc (str/list): CC email addresses (string or a list) bcc (str/list): BCC email addresses (string or a list) reply_to (str): Reply-to email address smtp (dict): SMTP configuration (dict) Schema of smtp dict: host (str): SMTP server host. Default: localhost port (int): SMTP server port. Default: 25 tls (bool): Use TLS. Default: False ssl (bool): Use SSL. Default: False user (bool): SMTP login user. Default empty password (bool): SMTP login password. Default empty """ if is_string(email): email = EmailContent(email) from_email = sanitize_email_address(from_email or email.headers.get('from')) to_email = sanitize_email_address(to_email or email.headers.get('to')) cc = sanitize_email_address(cc or email.headers.get('cc')) bcc = sanitize_email_address(bcc or email.headers.get('bcc')) reply_to = sanitize_email_address(reply_to or email.headers.get('reply-to')) message_args = { 'html': email.html, 'text': email.text, 'subject': (subject or email.headers.get('subject', '')), 'mail_from': from_email, 'mail_to': to_email } if cc: message_args['cc'] = cc if bcc: message_args['bcc'] = bcc if reply_to: message_args['headers'] = {'reply-to': reply_to} message = emails.Message(**message_args) for filename, data in email.inline_images: message.attach(filename=filename, content_disposition='inline', data=data) message.send(smtp=smtp)
python
def send(email, subject=None, from_email=None, to_email=None, cc=None, bcc=None, reply_to=None, smtp=None): """Send markdown email Args: email (str/obj): A markdown string or EmailContent object subject (str): subject line from_email (str): sender email address to_email (str/list): recipient email addresses cc (str/list): CC email addresses (string or a list) bcc (str/list): BCC email addresses (string or a list) reply_to (str): Reply-to email address smtp (dict): SMTP configuration (dict) Schema of smtp dict: host (str): SMTP server host. Default: localhost port (int): SMTP server port. Default: 25 tls (bool): Use TLS. Default: False ssl (bool): Use SSL. Default: False user (bool): SMTP login user. Default empty password (bool): SMTP login password. Default empty """ if is_string(email): email = EmailContent(email) from_email = sanitize_email_address(from_email or email.headers.get('from')) to_email = sanitize_email_address(to_email or email.headers.get('to')) cc = sanitize_email_address(cc or email.headers.get('cc')) bcc = sanitize_email_address(bcc or email.headers.get('bcc')) reply_to = sanitize_email_address(reply_to or email.headers.get('reply-to')) message_args = { 'html': email.html, 'text': email.text, 'subject': (subject or email.headers.get('subject', '')), 'mail_from': from_email, 'mail_to': to_email } if cc: message_args['cc'] = cc if bcc: message_args['bcc'] = bcc if reply_to: message_args['headers'] = {'reply-to': reply_to} message = emails.Message(**message_args) for filename, data in email.inline_images: message.attach(filename=filename, content_disposition='inline', data=data) message.send(smtp=smtp)
[ "def", "send", "(", "email", ",", "subject", "=", "None", ",", "from_email", "=", "None", ",", "to_email", "=", "None", ",", "cc", "=", "None", ",", "bcc", "=", "None", ",", "reply_to", "=", "None", ",", "smtp", "=", "None", ")", ":", "if", "is_string", "(", "email", ")", ":", "email", "=", "EmailContent", "(", "email", ")", "from_email", "=", "sanitize_email_address", "(", "from_email", "or", "email", ".", "headers", ".", "get", "(", "'from'", ")", ")", "to_email", "=", "sanitize_email_address", "(", "to_email", "or", "email", ".", "headers", ".", "get", "(", "'to'", ")", ")", "cc", "=", "sanitize_email_address", "(", "cc", "or", "email", ".", "headers", ".", "get", "(", "'cc'", ")", ")", "bcc", "=", "sanitize_email_address", "(", "bcc", "or", "email", ".", "headers", ".", "get", "(", "'bcc'", ")", ")", "reply_to", "=", "sanitize_email_address", "(", "reply_to", "or", "email", ".", "headers", ".", "get", "(", "'reply-to'", ")", ")", "message_args", "=", "{", "'html'", ":", "email", ".", "html", ",", "'text'", ":", "email", ".", "text", ",", "'subject'", ":", "(", "subject", "or", "email", ".", "headers", ".", "get", "(", "'subject'", ",", "''", ")", ")", ",", "'mail_from'", ":", "from_email", ",", "'mail_to'", ":", "to_email", "}", "if", "cc", ":", "message_args", "[", "'cc'", "]", "=", "cc", "if", "bcc", ":", "message_args", "[", "'bcc'", "]", "=", "bcc", "if", "reply_to", ":", "message_args", "[", "'headers'", "]", "=", "{", "'reply-to'", ":", "reply_to", "}", "message", "=", "emails", ".", "Message", "(", "*", "*", "message_args", ")", "for", "filename", ",", "data", "in", "email", ".", "inline_images", ":", "message", ".", "attach", "(", "filename", "=", "filename", ",", "content_disposition", "=", "'inline'", ",", "data", "=", "data", ")", "message", ".", "send", "(", "smtp", "=", "smtp", ")" ]
Send markdown email Args: email (str/obj): A markdown string or EmailContent object subject (str): subject line from_email (str): sender email address to_email (str/list): recipient email addresses cc (str/list): CC email addresses (string or a list) bcc (str/list): BCC email addresses (string or a list) reply_to (str): Reply-to email address smtp (dict): SMTP configuration (dict) Schema of smtp dict: host (str): SMTP server host. Default: localhost port (int): SMTP server port. Default: 25 tls (bool): Use TLS. Default: False ssl (bool): Use SSL. Default: False user (bool): SMTP login user. Default empty password (bool): SMTP login password. Default empty
[ "Send", "markdown", "email" ]
ef03da8d5836b5ae0a4ad8c44f2fe4936a896644
https://github.com/yejianye/mdmail/blob/ef03da8d5836b5ae0a4ad8c44f2fe4936a896644/mdmail/api.py#L11-L63
train
marrow/mongo
marrow/mongo/core/field/date.py
Date._process_tz
def _process_tz(self, dt, naive, tz): """Process timezone casting and conversion.""" def _tz(t): if t in (None, 'naive'): return t if t == 'local': if __debug__ and not localtz: raise ValueError("Requested conversion to local timezone, but `localtz` not installed.") t = localtz if not isinstance(t, tzinfo): if __debug__ and not localtz: raise ValueError("The `pytz` package must be installed to look up timezone: " + repr(t)) t = get_tz(t) if not hasattr(t, 'normalize') and get_tz: # Attempt to handle non-pytz tzinfo. t = get_tz(t.tzname(dt)) return t naive = _tz(naive) tz = _tz(tz) if not dt.tzinfo and naive: if hasattr(naive, 'localize'): dt = naive.localize(dt) else: dt = dt.replace(tzinfo=naive) if not tz: return dt if hasattr(tz, 'normalize'): dt = tz.normalize(dt.astimezone(tz)) elif tz == 'naive': dt = dt.replace(tzinfo=None) else: dt = dt.astimezone(tz) # Warning: this might not always be entirely correct! return dt
python
def _process_tz(self, dt, naive, tz): """Process timezone casting and conversion.""" def _tz(t): if t in (None, 'naive'): return t if t == 'local': if __debug__ and not localtz: raise ValueError("Requested conversion to local timezone, but `localtz` not installed.") t = localtz if not isinstance(t, tzinfo): if __debug__ and not localtz: raise ValueError("The `pytz` package must be installed to look up timezone: " + repr(t)) t = get_tz(t) if not hasattr(t, 'normalize') and get_tz: # Attempt to handle non-pytz tzinfo. t = get_tz(t.tzname(dt)) return t naive = _tz(naive) tz = _tz(tz) if not dt.tzinfo and naive: if hasattr(naive, 'localize'): dt = naive.localize(dt) else: dt = dt.replace(tzinfo=naive) if not tz: return dt if hasattr(tz, 'normalize'): dt = tz.normalize(dt.astimezone(tz)) elif tz == 'naive': dt = dt.replace(tzinfo=None) else: dt = dt.astimezone(tz) # Warning: this might not always be entirely correct! return dt
[ "def", "_process_tz", "(", "self", ",", "dt", ",", "naive", ",", "tz", ")", ":", "def", "_tz", "(", "t", ")", ":", "if", "t", "in", "(", "None", ",", "'naive'", ")", ":", "return", "t", "if", "t", "==", "'local'", ":", "if", "__debug__", "and", "not", "localtz", ":", "raise", "ValueError", "(", "\"Requested conversion to local timezone, but `localtz` not installed.\"", ")", "t", "=", "localtz", "if", "not", "isinstance", "(", "t", ",", "tzinfo", ")", ":", "if", "__debug__", "and", "not", "localtz", ":", "raise", "ValueError", "(", "\"The `pytz` package must be installed to look up timezone: \"", "+", "repr", "(", "t", ")", ")", "t", "=", "get_tz", "(", "t", ")", "if", "not", "hasattr", "(", "t", ",", "'normalize'", ")", "and", "get_tz", ":", "# Attempt to handle non-pytz tzinfo.", "t", "=", "get_tz", "(", "t", ".", "tzname", "(", "dt", ")", ")", "return", "t", "naive", "=", "_tz", "(", "naive", ")", "tz", "=", "_tz", "(", "tz", ")", "if", "not", "dt", ".", "tzinfo", "and", "naive", ":", "if", "hasattr", "(", "naive", ",", "'localize'", ")", ":", "dt", "=", "naive", ".", "localize", "(", "dt", ")", "else", ":", "dt", "=", "dt", ".", "replace", "(", "tzinfo", "=", "naive", ")", "if", "not", "tz", ":", "return", "dt", "if", "hasattr", "(", "tz", ",", "'normalize'", ")", ":", "dt", "=", "tz", ".", "normalize", "(", "dt", ".", "astimezone", "(", "tz", ")", ")", "elif", "tz", "==", "'naive'", ":", "dt", "=", "dt", ".", "replace", "(", "tzinfo", "=", "None", ")", "else", ":", "dt", "=", "dt", ".", "astimezone", "(", "tz", ")", "# Warning: this might not always be entirely correct!", "return", "dt" ]
Process timezone casting and conversion.
[ "Process", "timezone", "casting", "and", "conversion", "." ]
2066dc73e281b8a46cb5fc965267d6b8e1b18467
https://github.com/marrow/mongo/blob/2066dc73e281b8a46cb5fc965267d6b8e1b18467/marrow/mongo/core/field/date.py#L59-L102
train
marrow/mongo
marrow/mongo/core/document.py
Document._prepare_defaults
def _prepare_defaults(self): """Trigger assignment of default values.""" for name, field in self.__fields__.items(): if field.assign: getattr(self, name)
python
def _prepare_defaults(self): """Trigger assignment of default values.""" for name, field in self.__fields__.items(): if field.assign: getattr(self, name)
[ "def", "_prepare_defaults", "(", "self", ")", ":", "for", "name", ",", "field", "in", "self", ".", "__fields__", ".", "items", "(", ")", ":", "if", "field", ".", "assign", ":", "getattr", "(", "self", ",", "name", ")" ]
Trigger assignment of default values.
[ "Trigger", "assignment", "of", "default", "values", "." ]
2066dc73e281b8a46cb5fc965267d6b8e1b18467
https://github.com/marrow/mongo/blob/2066dc73e281b8a46cb5fc965267d6b8e1b18467/marrow/mongo/core/document.py#L71-L76
train
marrow/mongo
marrow/mongo/core/document.py
Document.from_mongo
def from_mongo(cls, doc): """Convert data coming in from the MongoDB wire driver into a Document instance.""" if doc is None: # To support simplified iterative use, None should return None. return None if isinstance(doc, Document): # No need to perform processing on existing Document instances. return doc if cls.__type_store__ and cls.__type_store__ in doc: # Instantiate specific class mentioned in the data. cls = load(doc[cls.__type_store__], 'marrow.mongo.document') # Prepare a new instance in such a way that changes to the instance will be reflected in the originating doc. instance = cls(_prepare_defaults=False) # Construct an instance, but delay default value processing. instance.__data__ = doc # I am Popeye of Borg (pattern); you will be askimilgrated. instance._prepare_defaults() # pylint:disable=protected-access -- deferred default value processing. return instance
python
def from_mongo(cls, doc): """Convert data coming in from the MongoDB wire driver into a Document instance.""" if doc is None: # To support simplified iterative use, None should return None. return None if isinstance(doc, Document): # No need to perform processing on existing Document instances. return doc if cls.__type_store__ and cls.__type_store__ in doc: # Instantiate specific class mentioned in the data. cls = load(doc[cls.__type_store__], 'marrow.mongo.document') # Prepare a new instance in such a way that changes to the instance will be reflected in the originating doc. instance = cls(_prepare_defaults=False) # Construct an instance, but delay default value processing. instance.__data__ = doc # I am Popeye of Borg (pattern); you will be askimilgrated. instance._prepare_defaults() # pylint:disable=protected-access -- deferred default value processing. return instance
[ "def", "from_mongo", "(", "cls", ",", "doc", ")", ":", "if", "doc", "is", "None", ":", "# To support simplified iterative use, None should return None.", "return", "None", "if", "isinstance", "(", "doc", ",", "Document", ")", ":", "# No need to perform processing on existing Document instances.", "return", "doc", "if", "cls", ".", "__type_store__", "and", "cls", ".", "__type_store__", "in", "doc", ":", "# Instantiate specific class mentioned in the data.", "cls", "=", "load", "(", "doc", "[", "cls", ".", "__type_store__", "]", ",", "'marrow.mongo.document'", ")", "# Prepare a new instance in such a way that changes to the instance will be reflected in the originating doc.", "instance", "=", "cls", "(", "_prepare_defaults", "=", "False", ")", "# Construct an instance, but delay default value processing.", "instance", ".", "__data__", "=", "doc", "# I am Popeye of Borg (pattern); you will be askimilgrated.", "instance", ".", "_prepare_defaults", "(", ")", "# pylint:disable=protected-access -- deferred default value processing.", "return", "instance" ]
Convert data coming in from the MongoDB wire driver into a Document instance.
[ "Convert", "data", "coming", "in", "from", "the", "MongoDB", "wire", "driver", "into", "a", "Document", "instance", "." ]
2066dc73e281b8a46cb5fc965267d6b8e1b18467
https://github.com/marrow/mongo/blob/2066dc73e281b8a46cb5fc965267d6b8e1b18467/marrow/mongo/core/document.py#L81-L98
train
marrow/mongo
marrow/mongo/core/document.py
Document.pop
def pop(self, name, default=SENTINEL): """Retrieve and remove a value from the backing store, optionally with a default.""" if default is SENTINEL: return self.__data__.pop(name) return self.__data__.pop(name, default)
python
def pop(self, name, default=SENTINEL): """Retrieve and remove a value from the backing store, optionally with a default.""" if default is SENTINEL: return self.__data__.pop(name) return self.__data__.pop(name, default)
[ "def", "pop", "(", "self", ",", "name", ",", "default", "=", "SENTINEL", ")", ":", "if", "default", "is", "SENTINEL", ":", "return", "self", ".", "__data__", ".", "pop", "(", "name", ")", "return", "self", ".", "__data__", ".", "pop", "(", "name", ",", "default", ")" ]
Retrieve and remove a value from the backing store, optionally with a default.
[ "Retrieve", "and", "remove", "a", "value", "from", "the", "backing", "store", "optionally", "with", "a", "default", "." ]
2066dc73e281b8a46cb5fc965267d6b8e1b18467
https://github.com/marrow/mongo/blob/2066dc73e281b8a46cb5fc965267d6b8e1b18467/marrow/mongo/core/document.py#L246-L252
train
marrow/mongo
marrow/mongo/query/query.py
Q._op
def _op(self, operation, other, *allowed): """A basic operation operating on a single value.""" f = self._field if self._combining: # We are a field-compound query fragment, e.g. (Foo.bar & Foo.baz). return reduce(self._combining, (q._op(operation, other, *allowed) for q in f)) # pylint:disable=protected-access # Optimize this away in production; diagnosic aide. if __debug__ and _complex_safety_check(f, {operation} | set(allowed)): # pragma: no cover raise NotImplementedError("{self!r} does not allow {op} comparison.".format(self=self, op=operation)) if other is not None: other = f.transformer.foreign(other, (f, self._document)) return Filter({self._name: {operation: other}})
python
def _op(self, operation, other, *allowed): """A basic operation operating on a single value.""" f = self._field if self._combining: # We are a field-compound query fragment, e.g. (Foo.bar & Foo.baz). return reduce(self._combining, (q._op(operation, other, *allowed) for q in f)) # pylint:disable=protected-access # Optimize this away in production; diagnosic aide. if __debug__ and _complex_safety_check(f, {operation} | set(allowed)): # pragma: no cover raise NotImplementedError("{self!r} does not allow {op} comparison.".format(self=self, op=operation)) if other is not None: other = f.transformer.foreign(other, (f, self._document)) return Filter({self._name: {operation: other}})
[ "def", "_op", "(", "self", ",", "operation", ",", "other", ",", "*", "allowed", ")", ":", "f", "=", "self", ".", "_field", "if", "self", ".", "_combining", ":", "# We are a field-compound query fragment, e.g. (Foo.bar & Foo.baz).", "return", "reduce", "(", "self", ".", "_combining", ",", "(", "q", ".", "_op", "(", "operation", ",", "other", ",", "*", "allowed", ")", "for", "q", "in", "f", ")", ")", "# pylint:disable=protected-access", "# Optimize this away in production; diagnosic aide.", "if", "__debug__", "and", "_complex_safety_check", "(", "f", ",", "{", "operation", "}", "|", "set", "(", "allowed", ")", ")", ":", "# pragma: no cover", "raise", "NotImplementedError", "(", "\"{self!r} does not allow {op} comparison.\"", ".", "format", "(", "self", "=", "self", ",", "op", "=", "operation", ")", ")", "if", "other", "is", "not", "None", ":", "other", "=", "f", ".", "transformer", ".", "foreign", "(", "other", ",", "(", "f", ",", "self", ".", "_document", ")", ")", "return", "Filter", "(", "{", "self", ".", "_name", ":", "{", "operation", ":", "other", "}", "}", ")" ]
A basic operation operating on a single value.
[ "A", "basic", "operation", "operating", "on", "a", "single", "value", "." ]
2066dc73e281b8a46cb5fc965267d6b8e1b18467
https://github.com/marrow/mongo/blob/2066dc73e281b8a46cb5fc965267d6b8e1b18467/marrow/mongo/query/query.py#L154-L170
train
marrow/mongo
marrow/mongo/query/query.py
Q._iop
def _iop(self, operation, other, *allowed): """An iterative operation operating on multiple values. Consumes iterators to construct a concrete list at time of execution. """ f = self._field if self._combining: # We are a field-compound query fragment, e.g. (Foo.bar & Foo.baz). return reduce(self._combining, (q._iop(operation, other, *allowed) for q in f)) # pylint:disable=protected-access # Optimize this away in production; diagnosic aide. if __debug__ and _complex_safety_check(f, {operation} | set(allowed)): # pragma: no cover raise NotImplementedError("{self!r} does not allow {op} comparison.".format( self=self, op=operation)) def _t(o): for value in o: yield None if value is None else f.transformer.foreign(value, (f, self._document)) other = other if len(other) > 1 else other[0] values = list(_t(other)) return Filter({self._name: {operation: values}})
python
def _iop(self, operation, other, *allowed): """An iterative operation operating on multiple values. Consumes iterators to construct a concrete list at time of execution. """ f = self._field if self._combining: # We are a field-compound query fragment, e.g. (Foo.bar & Foo.baz). return reduce(self._combining, (q._iop(operation, other, *allowed) for q in f)) # pylint:disable=protected-access # Optimize this away in production; diagnosic aide. if __debug__ and _complex_safety_check(f, {operation} | set(allowed)): # pragma: no cover raise NotImplementedError("{self!r} does not allow {op} comparison.".format( self=self, op=operation)) def _t(o): for value in o: yield None if value is None else f.transformer.foreign(value, (f, self._document)) other = other if len(other) > 1 else other[0] values = list(_t(other)) return Filter({self._name: {operation: values}})
[ "def", "_iop", "(", "self", ",", "operation", ",", "other", ",", "*", "allowed", ")", ":", "f", "=", "self", ".", "_field", "if", "self", ".", "_combining", ":", "# We are a field-compound query fragment, e.g. (Foo.bar & Foo.baz).", "return", "reduce", "(", "self", ".", "_combining", ",", "(", "q", ".", "_iop", "(", "operation", ",", "other", ",", "*", "allowed", ")", "for", "q", "in", "f", ")", ")", "# pylint:disable=protected-access", "# Optimize this away in production; diagnosic aide.", "if", "__debug__", "and", "_complex_safety_check", "(", "f", ",", "{", "operation", "}", "|", "set", "(", "allowed", ")", ")", ":", "# pragma: no cover", "raise", "NotImplementedError", "(", "\"{self!r} does not allow {op} comparison.\"", ".", "format", "(", "self", "=", "self", ",", "op", "=", "operation", ")", ")", "def", "_t", "(", "o", ")", ":", "for", "value", "in", "o", ":", "yield", "None", "if", "value", "is", "None", "else", "f", ".", "transformer", ".", "foreign", "(", "value", ",", "(", "f", ",", "self", ".", "_document", ")", ")", "other", "=", "other", "if", "len", "(", "other", ")", ">", "1", "else", "other", "[", "0", "]", "values", "=", "list", "(", "_t", "(", "other", ")", ")", "return", "Filter", "(", "{", "self", ".", "_name", ":", "{", "operation", ":", "values", "}", "}", ")" ]
An iterative operation operating on multiple values. Consumes iterators to construct a concrete list at time of execution.
[ "An", "iterative", "operation", "operating", "on", "multiple", "values", ".", "Consumes", "iterators", "to", "construct", "a", "concrete", "list", "at", "time", "of", "execution", "." ]
2066dc73e281b8a46cb5fc965267d6b8e1b18467
https://github.com/marrow/mongo/blob/2066dc73e281b8a46cb5fc965267d6b8e1b18467/marrow/mongo/query/query.py#L172-L196
train