identifier
stringlengths
1
155
parameters
stringlengths
2
6.09k
docstring
stringlengths
11
63.4k
docstring_summary
stringlengths
0
63.4k
function
stringlengths
29
99.8k
function_tokens
sequence
start_point
sequence
end_point
sequence
language
stringclasses
1 value
docstring_language
stringlengths
2
7
docstring_language_predictions
stringlengths
18
23
is_langid_reliable
stringclasses
2 values
user_in_user_groups
(user_id, **options)
Get all user groups a user belongs to :param user_id: The id of user :param user_id: str :param options: Generic advanced options dict, see online documentation :type options: dict, optional :return: List of groups user is in :rtype: dict
Get all user groups a user belongs to :param user_id: The id of user :param user_id: str :param options: Generic advanced options dict, see online documentation :type options: dict, optional :return: List of groups user is in :rtype: dict
def user_in_user_groups(user_id, **options): """ Get all user groups a user belongs to :param user_id: The id of user :param user_id: str :param options: Generic advanced options dict, see online documentation :type options: dict, optional :return: List of groups user is in :rtype: dict """ uri = [USER_GROUPS_SUB_PATH, user_id] return _call_account_api("get", uri, {}, **options)
[ "def", "user_in_user_groups", "(", "user_id", ",", "*", "*", "options", ")", ":", "uri", "=", "[", "USER_GROUPS_SUB_PATH", ",", "user_id", "]", "return", "_call_account_api", "(", "\"get\"", ",", "uri", ",", "{", "}", ",", "*", "*", "options", ")" ]
[ 354, 0 ]
[ 365, 55 ]
python
en
['en', 'error', 'th']
False
scaled_dot_product_attention
(q, k, v)
Calculate the attention weights. q, k, v must have matching leading dimensions. k, v must have matching penultimate dimension, i.e.: seq_len_k = seq_len_v. The mask has different shapes depending on its type(padding or look ahead) but it must be broadcastable for addition. Args: q: query shape == (..., seq_len_q, depth) k: key shape == (..., seq_len_k, depth) v: value shape == (..., seq_len_v, depth_v) Returns: output, attention_weights from : https://www.tensorflow.org/tutorials/text/transformer
Calculate the attention weights. q, k, v must have matching leading dimensions. k, v must have matching penultimate dimension, i.e.: seq_len_k = seq_len_v. The mask has different shapes depending on its type(padding or look ahead) but it must be broadcastable for addition.
def scaled_dot_product_attention(q, k, v): """Calculate the attention weights. q, k, v must have matching leading dimensions. k, v must have matching penultimate dimension, i.e.: seq_len_k = seq_len_v. The mask has different shapes depending on its type(padding or look ahead) but it must be broadcastable for addition. Args: q: query shape == (..., seq_len_q, depth) k: key shape == (..., seq_len_k, depth) v: value shape == (..., seq_len_v, depth_v) Returns: output, attention_weights from : https://www.tensorflow.org/tutorials/text/transformer """ matmul_qk = tf.matmul(q, k, transpose_b=True) # (..., seq_len_q, seq_len_k) # scale matmul_qk dk = tf.cast(tf.shape(k)[-1], tf.float32) scaled_attention_logits = matmul_qk / tf.math.sqrt(dk) # softmax is normalized on the last axis (seq_len_k) so that the scores # add up to 1. attention_weights = tf.nn.softmax(scaled_attention_logits, axis=-1) # (..., seq_len_q, seq_len_k) output = tf.matmul(attention_weights, v) # (..., seq_len_q, depth_v) return output, attention_weights
[ "def", "scaled_dot_product_attention", "(", "q", ",", "k", ",", "v", ")", ":", "matmul_qk", "=", "tf", ".", "matmul", "(", "q", ",", "k", ",", "transpose_b", "=", "True", ")", "# (..., seq_len_q, seq_len_k)", "# scale matmul_qk", "dk", "=", "tf", ".", "cast", "(", "tf", ".", "shape", "(", "k", ")", "[", "-", "1", "]", ",", "tf", ".", "float32", ")", "scaled_attention_logits", "=", "matmul_qk", "/", "tf", ".", "math", ".", "sqrt", "(", "dk", ")", "# softmax is normalized on the last axis (seq_len_k) so that the scores", "# add up to 1.", "attention_weights", "=", "tf", ".", "nn", ".", "softmax", "(", "scaled_attention_logits", ",", "axis", "=", "-", "1", ")", "# (..., seq_len_q, seq_len_k)", "output", "=", "tf", ".", "matmul", "(", "attention_weights", ",", "v", ")", "# (..., seq_len_q, depth_v)", "return", "output", ",", "attention_weights" ]
[ 60, 0 ]
[ 90, 36 ]
python
en
['en', 'en', 'en']
True
Attention.__init__
(self, d_model, spatial_dims, positional_encoding=True, name="self_attention")
d_model : number of output channels spatial_dim : spatial dimensions of input tensor (x , y) if positional_encoding: depth must correspond to input channel number adapted from: https://www.tensorflow.org/tutorials/text/transformer
d_model : number of output channels spatial_dim : spatial dimensions of input tensor (x , y) if positional_encoding: depth must correspond to input channel number adapted from: https://www.tensorflow.org/tutorials/text/transformer
def __init__(self, d_model, spatial_dims, positional_encoding=True, name="self_attention"): ''' d_model : number of output channels spatial_dim : spatial dimensions of input tensor (x , y) if positional_encoding: depth must correspond to input channel number adapted from: https://www.tensorflow.org/tutorials/text/transformer ''' super().__init__(name=name) self.d_model = d_model self.spatial_dims=spatial_dims self.spatial_dim = np.prod(spatial_dims) self.wq = Dense(self.d_model, name=name+"_q") self.wk = Dense(self.d_model, name=name+"_k") self.wv = Dense(self.d_model, name=name+"_w") self.positional_encoding=positional_encoding if positional_encoding: self.pos_embedding = Embedding(self.spatial_dim, d_model, name=name+"pos_enc")
[ "def", "__init__", "(", "self", ",", "d_model", ",", "spatial_dims", ",", "positional_encoding", "=", "True", ",", "name", "=", "\"self_attention\"", ")", ":", "super", "(", ")", ".", "__init__", "(", "name", "=", "name", ")", "self", ".", "d_model", "=", "d_model", "self", ".", "spatial_dims", "=", "spatial_dims", "self", ".", "spatial_dim", "=", "np", ".", "prod", "(", "spatial_dims", ")", "self", ".", "wq", "=", "Dense", "(", "self", ".", "d_model", ",", "name", "=", "name", "+", "\"_q\"", ")", "self", ".", "wk", "=", "Dense", "(", "self", ".", "d_model", ",", "name", "=", "name", "+", "\"_k\"", ")", "self", ".", "wv", "=", "Dense", "(", "self", ".", "d_model", ",", "name", "=", "name", "+", "\"_w\"", ")", "self", ".", "positional_encoding", "=", "positional_encoding", "if", "positional_encoding", ":", "self", ".", "pos_embedding", "=", "Embedding", "(", "self", ".", "spatial_dim", ",", "d_model", ",", "name", "=", "name", "+", "\"pos_enc\"", ")" ]
[ 6, 4 ]
[ 22, 90 ]
python
en
['en', 'error', 'th']
False
Attention.call
(self, x)
x : list of 2 tensor with shape (batch_size, y, x, channels)
x : list of 2 tensor with shape (batch_size, y, x, channels)
def call(self, x): ''' x : list of 2 tensor with shape (batch_size, y, x, channels) ''' [input, output] = x shape = tf.shape(output) batch_size = shape[0] #spatial_dims = shape[1:-1] #spatial_dim = tf.reduce_prod(spatial_dims) depth_dim = shape[3] if self.positional_encoding: x_index = tf.range(self.spatial_dim, dtype=tf.int32) pos_emb = self.pos_embedding(x_index) # (spa_dim, d_model) pos_emb = tf.reshape(pos_emb, (self.spatial_dims[0], self.spatial_dims[1], self.d_model)) #for broadcasting purpose input = input + pos_emb # broadcast output = output + pos_emb # broadcast q = self.wq(output) # (batch_size, *spa_dims, d_model) k = self.wk(input) # (batch_size, *spa_dims, d_model) v = self.wv(input) # (batch_size, *spa_dims, d_model) q = tf.reshape(q, (batch_size, -1, depth_dim)) # (batch_size, spa_dim, d_model) k = tf.reshape(k, (batch_size, -1, depth_dim)) v = tf.reshape(v, (batch_size, -1, depth_dim)) # scaled_attention.shape == (batch_size, spa_dims, depth) # attention_weights.shape == (batch_size, spa_dims, spa_dims) scaled_attention, attention_weights = scaled_dot_product_attention(q, k, v) output = tf.reshape(scaled_attention, (batch_size, self.spatial_dims[0], self.spatial_dims[1], self.d_model)) tf.identity(attention_weights, name=self.name+"_attention_weights") return output, attention_weights
[ "def", "call", "(", "self", ",", "x", ")", ":", "[", "input", ",", "output", "]", "=", "x", "shape", "=", "tf", ".", "shape", "(", "output", ")", "batch_size", "=", "shape", "[", "0", "]", "#spatial_dims = shape[1:-1]", "#spatial_dim = tf.reduce_prod(spatial_dims)", "depth_dim", "=", "shape", "[", "3", "]", "if", "self", ".", "positional_encoding", ":", "x_index", "=", "tf", ".", "range", "(", "self", ".", "spatial_dim", ",", "dtype", "=", "tf", ".", "int32", ")", "pos_emb", "=", "self", ".", "pos_embedding", "(", "x_index", ")", "# (spa_dim, d_model)", "pos_emb", "=", "tf", ".", "reshape", "(", "pos_emb", ",", "(", "self", ".", "spatial_dims", "[", "0", "]", ",", "self", ".", "spatial_dims", "[", "1", "]", ",", "self", ".", "d_model", ")", ")", "#for broadcasting purpose", "input", "=", "input", "+", "pos_emb", "# broadcast", "output", "=", "output", "+", "pos_emb", "# broadcast", "q", "=", "self", ".", "wq", "(", "output", ")", "# (batch_size, *spa_dims, d_model)", "k", "=", "self", ".", "wk", "(", "input", ")", "# (batch_size, *spa_dims, d_model)", "v", "=", "self", ".", "wv", "(", "input", ")", "# (batch_size, *spa_dims, d_model)", "q", "=", "tf", ".", "reshape", "(", "q", ",", "(", "batch_size", ",", "-", "1", ",", "depth_dim", ")", ")", "# (batch_size, spa_dim, d_model)", "k", "=", "tf", ".", "reshape", "(", "k", ",", "(", "batch_size", ",", "-", "1", ",", "depth_dim", ")", ")", "v", "=", "tf", ".", "reshape", "(", "v", ",", "(", "batch_size", ",", "-", "1", ",", "depth_dim", ")", ")", "# scaled_attention.shape == (batch_size, spa_dims, depth)", "# attention_weights.shape == (batch_size, spa_dims, spa_dims)", "scaled_attention", ",", "attention_weights", "=", "scaled_dot_product_attention", "(", "q", ",", "k", ",", "v", ")", "output", "=", "tf", ".", "reshape", "(", "scaled_attention", ",", "(", "batch_size", ",", "self", ".", "spatial_dims", "[", "0", "]", ",", "self", ".", "spatial_dims", "[", "1", "]", ",", "self", ".", "d_model", ")", ")", "tf", ".", "identity", "(", "attention_weights", ",", "name", "=", "self", ".", "name", "+", "\"_attention_weights\"", ")", "return", "output", ",", "attention_weights" ]
[ 24, 4 ]
[ 55, 40 ]
python
en
['en', 'error', 'th']
False
all_frames
(im, func=None)
Applies a given function to all frames in an image or a list of images. The frames are returned as a list of separate images. :param im: An image, or a list of images. :param func: The function to apply to all of the image frames. :returns: A list of images.
Applies a given function to all frames in an image or a list of images. The frames are returned as a list of separate images.
def all_frames(im, func=None): """ Applies a given function to all frames in an image or a list of images. The frames are returned as a list of separate images. :param im: An image, or a list of images. :param func: The function to apply to all of the image frames. :returns: A list of images. """ if not isinstance(im, list): im = [im] ims = [] for imSequence in im: current = imSequence.tell() ims += [im_frame.copy() for im_frame in Iterator(imSequence)] imSequence.seek(current) return [func(im) for im in ims] if func else ims
[ "def", "all_frames", "(", "im", ",", "func", "=", "None", ")", ":", "if", "not", "isinstance", "(", "im", ",", "list", ")", ":", "im", "=", "[", "im", "]", "ims", "=", "[", "]", "for", "imSequence", "in", "im", ":", "current", "=", "imSequence", ".", "tell", "(", ")", "ims", "+=", "[", "im_frame", ".", "copy", "(", ")", "for", "im_frame", "in", "Iterator", "(", "imSequence", ")", "]", "imSequence", ".", "seek", "(", "current", ")", "return", "[", "func", "(", "im", ")", "for", "im", "in", "ims", "]", "if", "func", "else", "ims" ]
[ 55, 0 ]
[ 74, 52 ]
python
en
['en', 'error', 'th']
False
command_swanson
(bot, user, channel, args)
Prints Ron Swanson quotes.
Prints Ron Swanson quotes.
def command_swanson(bot, user, channel, args): "Prints Ron Swanson quotes." quotefile = os.path.join(bot.factory.moduledir, "swanson.txt") try: with open(quotefile) as f: quotes = f.readlines() except IOError: log.error("Could not read {}".format(quotefile)) else: quote = random.choice(quotes).strip() return bot.say(channel, "{} - Ron Swanson".format(quote))
[ "def", "command_swanson", "(", "bot", ",", "user", ",", "channel", ",", "args", ")", ":", "quotefile", "=", "os", ".", "path", ".", "join", "(", "bot", ".", "factory", ".", "moduledir", ",", "\"swanson.txt\"", ")", "try", ":", "with", "open", "(", "quotefile", ")", "as", "f", ":", "quotes", "=", "f", ".", "readlines", "(", ")", "except", "IOError", ":", "log", ".", "error", "(", "\"Could not read {}\"", ".", "format", "(", "quotefile", ")", ")", "else", ":", "quote", "=", "random", ".", "choice", "(", "quotes", ")", ".", "strip", "(", ")", "return", "bot", ".", "say", "(", "channel", ",", "\"{} - Ron Swanson\"", ".", "format", "(", "quote", ")", ")" ]
[ 8, 0 ]
[ 20, 61 ]
python
en
['fr', 'sr', 'en']
False
command_whatshesaid
(bot, user, channel, args)
Prints quotes of accomplished women.
Prints quotes of accomplished women.
def command_whatshesaid(bot, user, channel, args): "Prints quotes of accomplished women." quotefile = os.path.join(bot.factory.moduledir, "whatshesaid.txt") try: with open(quotefile) as f: quotes = f.readlines() except IOError: log.error("Could not read {}".format(quotefile)) else: quote = random.choice(quotes).strip() return bot.say(channel, quote)
[ "def", "command_whatshesaid", "(", "bot", ",", "user", ",", "channel", ",", "args", ")", ":", "quotefile", "=", "os", ".", "path", ".", "join", "(", "bot", ".", "factory", ".", "moduledir", ",", "\"whatshesaid.txt\"", ")", "try", ":", "with", "open", "(", "quotefile", ")", "as", "f", ":", "quotes", "=", "f", ".", "readlines", "(", ")", "except", "IOError", ":", "log", ".", "error", "(", "\"Could not read {}\"", ".", "format", "(", "quotefile", ")", ")", "else", ":", "quote", "=", "random", ".", "choice", "(", "quotes", ")", ".", "strip", "(", ")", "return", "bot", ".", "say", "(", "channel", ",", "quote", ")" ]
[ 23, 0 ]
[ 35, 34 ]
python
en
['en', 'en', 'en']
True
do_cache
(parser, token)
This will cache the contents of a template fragment for a given amount of time. Usage:: {% load cache %} {% cache [expire_time] [fragment_name] %} .. some expensive processing .. {% endcache %} This tag also supports varying by a list of arguments:: {% load cache %} {% cache [expire_time] [fragment_name] [var1] [var2] .. %} .. some expensive processing .. {% endcache %} Optionally the cache to use may be specified thus:: {% cache .... using="cachename" %} Each unique set of arguments will result in a unique cache entry.
This will cache the contents of a template fragment for a given amount of time.
def do_cache(parser, token): """ This will cache the contents of a template fragment for a given amount of time. Usage:: {% load cache %} {% cache [expire_time] [fragment_name] %} .. some expensive processing .. {% endcache %} This tag also supports varying by a list of arguments:: {% load cache %} {% cache [expire_time] [fragment_name] [var1] [var2] .. %} .. some expensive processing .. {% endcache %} Optionally the cache to use may be specified thus:: {% cache .... using="cachename" %} Each unique set of arguments will result in a unique cache entry. """ nodelist = parser.parse(('endcache',)) parser.delete_first_token() tokens = token.split_contents() if len(tokens) < 3: raise TemplateSyntaxError("'%r' tag requires at least 2 arguments." % tokens[0]) if len(tokens) > 3 and tokens[-1].startswith('using='): cache_name = parser.compile_filter(tokens[-1][len('using='):]) tokens = tokens[:-1] else: cache_name = None return CacheNode( nodelist, parser.compile_filter(tokens[1]), tokens[2], # fragment_name can't be a variable. [parser.compile_filter(t) for t in tokens[3:]], cache_name, )
[ "def", "do_cache", "(", "parser", ",", "token", ")", ":", "nodelist", "=", "parser", ".", "parse", "(", "(", "'endcache'", ",", ")", ")", "parser", ".", "delete_first_token", "(", ")", "tokens", "=", "token", ".", "split_contents", "(", ")", "if", "len", "(", "tokens", ")", "<", "3", ":", "raise", "TemplateSyntaxError", "(", "\"'%r' tag requires at least 2 arguments.\"", "%", "tokens", "[", "0", "]", ")", "if", "len", "(", "tokens", ")", ">", "3", "and", "tokens", "[", "-", "1", "]", ".", "startswith", "(", "'using='", ")", ":", "cache_name", "=", "parser", ".", "compile_filter", "(", "tokens", "[", "-", "1", "]", "[", "len", "(", "'using='", ")", ":", "]", ")", "tokens", "=", "tokens", "[", ":", "-", "1", "]", "else", ":", "cache_name", "=", "None", "return", "CacheNode", "(", "nodelist", ",", "parser", ".", "compile_filter", "(", "tokens", "[", "1", "]", ")", ",", "tokens", "[", "2", "]", ",", "# fragment_name can't be a variable.", "[", "parser", ".", "compile_filter", "(", "t", ")", "for", "t", "in", "tokens", "[", "3", ":", "]", "]", ",", "cache_name", ",", ")" ]
[ 52, 0 ]
[ 92, 5 ]
python
en
['en', 'error', 'th']
False
MigrationQuestioner.ask_initial
(self, app_label)
Should we create an initial migration for the app?
Should we create an initial migration for the app?
def ask_initial(self, app_label): """Should we create an initial migration for the app?""" # If it was specified on the command line, definitely true if app_label in self.specified_apps: return True # Otherwise, we look to see if it has a migrations module # without any Python files in it, apart from __init__.py. # Apps from the new app template will have these; the Python # file check will ensure we skip South ones. try: app_config = apps.get_app_config(app_label) except LookupError: # It's a fake app. return self.defaults.get("ask_initial", False) migrations_import_path, _ = MigrationLoader.migrations_module(app_config.label) if migrations_import_path is None: # It's an application with migrations disabled. return self.defaults.get("ask_initial", False) try: migrations_module = importlib.import_module(migrations_import_path) except ImportError: return self.defaults.get("ask_initial", False) else: # getattr() needed on PY36 and older (replace with attribute access). if getattr(migrations_module, "__file__", None): filenames = os.listdir(os.path.dirname(migrations_module.__file__)) elif hasattr(migrations_module, "__path__"): if len(migrations_module.__path__) > 1: return False filenames = os.listdir(list(migrations_module.__path__)[0]) return not any(x.endswith(".py") for x in filenames if x != "__init__.py")
[ "def", "ask_initial", "(", "self", ",", "app_label", ")", ":", "# If it was specified on the command line, definitely true", "if", "app_label", "in", "self", ".", "specified_apps", ":", "return", "True", "# Otherwise, we look to see if it has a migrations module", "# without any Python files in it, apart from __init__.py.", "# Apps from the new app template will have these; the Python", "# file check will ensure we skip South ones.", "try", ":", "app_config", "=", "apps", ".", "get_app_config", "(", "app_label", ")", "except", "LookupError", ":", "# It's a fake app.", "return", "self", ".", "defaults", ".", "get", "(", "\"ask_initial\"", ",", "False", ")", "migrations_import_path", ",", "_", "=", "MigrationLoader", ".", "migrations_module", "(", "app_config", ".", "label", ")", "if", "migrations_import_path", "is", "None", ":", "# It's an application with migrations disabled.", "return", "self", ".", "defaults", ".", "get", "(", "\"ask_initial\"", ",", "False", ")", "try", ":", "migrations_module", "=", "importlib", ".", "import_module", "(", "migrations_import_path", ")", "except", "ImportError", ":", "return", "self", ".", "defaults", ".", "get", "(", "\"ask_initial\"", ",", "False", ")", "else", ":", "# getattr() needed on PY36 and older (replace with attribute access).", "if", "getattr", "(", "migrations_module", ",", "\"__file__\"", ",", "None", ")", ":", "filenames", "=", "os", ".", "listdir", "(", "os", ".", "path", ".", "dirname", "(", "migrations_module", ".", "__file__", ")", ")", "elif", "hasattr", "(", "migrations_module", ",", "\"__path__\"", ")", ":", "if", "len", "(", "migrations_module", ".", "__path__", ")", ">", "1", ":", "return", "False", "filenames", "=", "os", ".", "listdir", "(", "list", "(", "migrations_module", ".", "__path__", ")", "[", "0", "]", ")", "return", "not", "any", "(", "x", ".", "endswith", "(", "\".py\"", ")", "for", "x", "in", "filenames", "if", "x", "!=", "\"__init__.py\"", ")" ]
[ 24, 4 ]
[ 53, 86 ]
python
en
['en', 'en', 'en']
True
MigrationQuestioner.ask_not_null_addition
(self, field_name, model_name)
Adding a NOT NULL field to a model.
Adding a NOT NULL field to a model.
def ask_not_null_addition(self, field_name, model_name): """Adding a NOT NULL field to a model.""" # None means quit return None
[ "def", "ask_not_null_addition", "(", "self", ",", "field_name", ",", "model_name", ")", ":", "# None means quit", "return", "None" ]
[ 55, 4 ]
[ 58, 19 ]
python
en
['en', 'en', 'en']
True
MigrationQuestioner.ask_not_null_alteration
(self, field_name, model_name)
Changing a NULL field to NOT NULL.
Changing a NULL field to NOT NULL.
def ask_not_null_alteration(self, field_name, model_name): """Changing a NULL field to NOT NULL.""" # None means quit return None
[ "def", "ask_not_null_alteration", "(", "self", ",", "field_name", ",", "model_name", ")", ":", "# None means quit", "return", "None" ]
[ 60, 4 ]
[ 63, 19 ]
python
en
['en', 'en', 'en']
True
MigrationQuestioner.ask_rename
(self, model_name, old_name, new_name, field_instance)
Was this field really renamed?
Was this field really renamed?
def ask_rename(self, model_name, old_name, new_name, field_instance): """Was this field really renamed?""" return self.defaults.get("ask_rename", False)
[ "def", "ask_rename", "(", "self", ",", "model_name", ",", "old_name", ",", "new_name", ",", "field_instance", ")", ":", "return", "self", ".", "defaults", ".", "get", "(", "\"ask_rename\"", ",", "False", ")" ]
[ 65, 4 ]
[ 67, 53 ]
python
en
['en', 'en', 'en']
True
MigrationQuestioner.ask_rename_model
(self, old_model_state, new_model_state)
Was this model really renamed?
Was this model really renamed?
def ask_rename_model(self, old_model_state, new_model_state): """Was this model really renamed?""" return self.defaults.get("ask_rename_model", False)
[ "def", "ask_rename_model", "(", "self", ",", "old_model_state", ",", "new_model_state", ")", ":", "return", "self", ".", "defaults", ".", "get", "(", "\"ask_rename_model\"", ",", "False", ")" ]
[ 69, 4 ]
[ 71, 59 ]
python
en
['en', 'en', 'en']
True
MigrationQuestioner.ask_merge
(self, app_label)
Do you really want to merge these migrations?
Do you really want to merge these migrations?
def ask_merge(self, app_label): """Do you really want to merge these migrations?""" return self.defaults.get("ask_merge", False)
[ "def", "ask_merge", "(", "self", ",", "app_label", ")", ":", "return", "self", ".", "defaults", ".", "get", "(", "\"ask_merge\"", ",", "False", ")" ]
[ 73, 4 ]
[ 75, 52 ]
python
en
['en', 'en', 'en']
True
MigrationQuestioner.ask_auto_now_add_addition
(self, field_name, model_name)
Adding an auto_now_add field to a model.
Adding an auto_now_add field to a model.
def ask_auto_now_add_addition(self, field_name, model_name): """Adding an auto_now_add field to a model.""" # None means quit return None
[ "def", "ask_auto_now_add_addition", "(", "self", ",", "field_name", ",", "model_name", ")", ":", "# None means quit", "return", "None" ]
[ 77, 4 ]
[ 80, 19 ]
python
en
['en', 'en', 'en']
True
InteractiveMigrationQuestioner._ask_default
(self, default='')
Prompt for a default value. The ``default`` argument allows providing a custom default value (as a string) which will be shown to the user and used as the return value if the user doesn't provide any other input.
Prompt for a default value.
def _ask_default(self, default=''): """ Prompt for a default value. The ``default`` argument allows providing a custom default value (as a string) which will be shown to the user and used as the return value if the user doesn't provide any other input. """ print("Please enter the default value now, as valid Python") if default: print( "You can accept the default '{}' by pressing 'Enter' or you " "can provide another value.".format(default) ) print("The datetime and django.utils.timezone modules are available, so you can do e.g. timezone.now") print("Type 'exit' to exit this prompt") while True: if default: prompt = "[default: {}] >>> ".format(default) else: prompt = ">>> " code = input(prompt) if not code and default: code = default if not code: print("Please enter some code, or 'exit' (with no quotes) to exit.") elif code == "exit": sys.exit(1) else: try: return eval(code, {}, {'datetime': datetime, 'timezone': timezone}) except (SyntaxError, NameError) as e: print("Invalid input: %s" % e)
[ "def", "_ask_default", "(", "self", ",", "default", "=", "''", ")", ":", "print", "(", "\"Please enter the default value now, as valid Python\"", ")", "if", "default", ":", "print", "(", "\"You can accept the default '{}' by pressing 'Enter' or you \"", "\"can provide another value.\"", ".", "format", "(", "default", ")", ")", "print", "(", "\"The datetime and django.utils.timezone modules are available, so you can do e.g. timezone.now\"", ")", "print", "(", "\"Type 'exit' to exit this prompt\"", ")", "while", "True", ":", "if", "default", ":", "prompt", "=", "\"[default: {}] >>> \"", ".", "format", "(", "default", ")", "else", ":", "prompt", "=", "\">>> \"", "code", "=", "input", "(", "prompt", ")", "if", "not", "code", "and", "default", ":", "code", "=", "default", "if", "not", "code", ":", "print", "(", "\"Please enter some code, or 'exit' (with no quotes) to exit.\"", ")", "elif", "code", "==", "\"exit\"", ":", "sys", ".", "exit", "(", "1", ")", "else", ":", "try", ":", "return", "eval", "(", "code", ",", "{", "}", ",", "{", "'datetime'", ":", "datetime", ",", "'timezone'", ":", "timezone", "}", ")", "except", "(", "SyntaxError", ",", "NameError", ")", "as", "e", ":", "print", "(", "\"Invalid input: %s\"", "%", "e", ")" ]
[ 108, 4 ]
[ 140, 50 ]
python
en
['en', 'error', 'th']
False
InteractiveMigrationQuestioner.ask_not_null_addition
(self, field_name, model_name)
Adding a NOT NULL field to a model.
Adding a NOT NULL field to a model.
def ask_not_null_addition(self, field_name, model_name): """Adding a NOT NULL field to a model.""" if not self.dry_run: choice = self._choice_input( "You are trying to add a non-nullable field '%s' to %s without a default; " "we can't do that (the database needs something to populate existing rows).\n" "Please select a fix:" % (field_name, model_name), [ ("Provide a one-off default now (will be set on all existing " "rows with a null value for this column)"), "Quit, and let me add a default in models.py", ] ) if choice == 2: sys.exit(3) else: return self._ask_default() return None
[ "def", "ask_not_null_addition", "(", "self", ",", "field_name", ",", "model_name", ")", ":", "if", "not", "self", ".", "dry_run", ":", "choice", "=", "self", ".", "_choice_input", "(", "\"You are trying to add a non-nullable field '%s' to %s without a default; \"", "\"we can't do that (the database needs something to populate existing rows).\\n\"", "\"Please select a fix:\"", "%", "(", "field_name", ",", "model_name", ")", ",", "[", "(", "\"Provide a one-off default now (will be set on all existing \"", "\"rows with a null value for this column)\"", ")", ",", "\"Quit, and let me add a default in models.py\"", ",", "]", ")", "if", "choice", "==", "2", ":", "sys", ".", "exit", "(", "3", ")", "else", ":", "return", "self", ".", "_ask_default", "(", ")", "return", "None" ]
[ 142, 4 ]
[ 159, 19 ]
python
en
['en', 'en', 'en']
True
InteractiveMigrationQuestioner.ask_not_null_alteration
(self, field_name, model_name)
Changing a NULL field to NOT NULL.
Changing a NULL field to NOT NULL.
def ask_not_null_alteration(self, field_name, model_name): """Changing a NULL field to NOT NULL.""" if not self.dry_run: choice = self._choice_input( "You are trying to change the nullable field '%s' on %s to non-nullable " "without a default; we can't do that (the database needs something to " "populate existing rows).\n" "Please select a fix:" % (field_name, model_name), [ ("Provide a one-off default now (will be set on all existing " "rows with a null value for this column)"), ("Ignore for now, and let me handle existing rows with NULL myself " "(e.g. because you added a RunPython or RunSQL operation to handle " "NULL values in a previous data migration)"), "Quit, and let me add a default in models.py", ] ) if choice == 2: return NOT_PROVIDED elif choice == 3: sys.exit(3) else: return self._ask_default() return None
[ "def", "ask_not_null_alteration", "(", "self", ",", "field_name", ",", "model_name", ")", ":", "if", "not", "self", ".", "dry_run", ":", "choice", "=", "self", ".", "_choice_input", "(", "\"You are trying to change the nullable field '%s' on %s to non-nullable \"", "\"without a default; we can't do that (the database needs something to \"", "\"populate existing rows).\\n\"", "\"Please select a fix:\"", "%", "(", "field_name", ",", "model_name", ")", ",", "[", "(", "\"Provide a one-off default now (will be set on all existing \"", "\"rows with a null value for this column)\"", ")", ",", "(", "\"Ignore for now, and let me handle existing rows with NULL myself \"", "\"(e.g. because you added a RunPython or RunSQL operation to handle \"", "\"NULL values in a previous data migration)\"", ")", ",", "\"Quit, and let me add a default in models.py\"", ",", "]", ")", "if", "choice", "==", "2", ":", "return", "NOT_PROVIDED", "elif", "choice", "==", "3", ":", "sys", ".", "exit", "(", "3", ")", "else", ":", "return", "self", ".", "_ask_default", "(", ")", "return", "None" ]
[ 161, 4 ]
[ 184, 19 ]
python
en
['en', 'en', 'en']
True
InteractiveMigrationQuestioner.ask_rename
(self, model_name, old_name, new_name, field_instance)
Was this field really renamed?
Was this field really renamed?
def ask_rename(self, model_name, old_name, new_name, field_instance): """Was this field really renamed?""" msg = "Did you rename %s.%s to %s.%s (a %s)? [y/N]" return self._boolean_input(msg % (model_name, old_name, model_name, new_name, field_instance.__class__.__name__), False)
[ "def", "ask_rename", "(", "self", ",", "model_name", ",", "old_name", ",", "new_name", ",", "field_instance", ")", ":", "msg", "=", "\"Did you rename %s.%s to %s.%s (a %s)? [y/N]\"", "return", "self", ".", "_boolean_input", "(", "msg", "%", "(", "model_name", ",", "old_name", ",", "model_name", ",", "new_name", ",", "field_instance", ".", "__class__", ".", "__name__", ")", ",", "False", ")" ]
[ 186, 4 ]
[ 190, 84 ]
python
en
['en', 'en', 'en']
True
InteractiveMigrationQuestioner.ask_rename_model
(self, old_model_state, new_model_state)
Was this model really renamed?
Was this model really renamed?
def ask_rename_model(self, old_model_state, new_model_state): """Was this model really renamed?""" msg = "Did you rename the %s.%s model to %s? [y/N]" return self._boolean_input(msg % (old_model_state.app_label, old_model_state.name, new_model_state.name), False)
[ "def", "ask_rename_model", "(", "self", ",", "old_model_state", ",", "new_model_state", ")", ":", "msg", "=", "\"Did you rename the %s.%s model to %s? [y/N]\"", "return", "self", ".", "_boolean_input", "(", "msg", "%", "(", "old_model_state", ".", "app_label", ",", "old_model_state", ".", "name", ",", "new_model_state", ".", "name", ")", ",", "False", ")" ]
[ 192, 4 ]
[ 196, 71 ]
python
en
['en', 'en', 'en']
True
InteractiveMigrationQuestioner.ask_auto_now_add_addition
(self, field_name, model_name)
Adding an auto_now_add field to a model.
Adding an auto_now_add field to a model.
def ask_auto_now_add_addition(self, field_name, model_name): """Adding an auto_now_add field to a model.""" if not self.dry_run: choice = self._choice_input( "You are trying to add the field '{}' with 'auto_now_add=True' " "to {} without a default; the database needs something to " "populate existing rows.\n".format(field_name, model_name), [ "Provide a one-off default now (will be set on all " "existing rows)", "Quit, and let me add a default in models.py", ] ) if choice == 2: sys.exit(3) else: return self._ask_default(default='timezone.now') return None
[ "def", "ask_auto_now_add_addition", "(", "self", ",", "field_name", ",", "model_name", ")", ":", "if", "not", "self", ".", "dry_run", ":", "choice", "=", "self", ".", "_choice_input", "(", "\"You are trying to add the field '{}' with 'auto_now_add=True' \"", "\"to {} without a default; the database needs something to \"", "\"populate existing rows.\\n\"", ".", "format", "(", "field_name", ",", "model_name", ")", ",", "[", "\"Provide a one-off default now (will be set on all \"", "\"existing rows)\"", ",", "\"Quit, and let me add a default in models.py\"", ",", "]", ")", "if", "choice", "==", "2", ":", "sys", ".", "exit", "(", "3", ")", "else", ":", "return", "self", ".", "_ask_default", "(", "default", "=", "'timezone.now'", ")", "return", "None" ]
[ 206, 4 ]
[ 223, 19 ]
python
en
['en', 'en', 'en']
True
upload_large
(file, **options)
Upload large files.
Upload large files.
def upload_large(file, **options): """ Upload large files. """ if utils.is_remote_url(file): return upload(file, **options) if hasattr(file, 'read') and callable(file.read): file_io = file else: file_io = open(file, 'rb') upload_result = None with file_io: upload_id = utils.random_public_id() current_loc = 0 chunk_size = options.get("chunk_size", UPLOAD_LARGE_CHUNK_SIZE) file_size = utils.file_io_size(file_io) file_name = options.get( "filename", file_io.name if hasattr(file_io, 'name') and isinstance(file_io.name, str) else "stream") chunk = file_io.read(chunk_size) while chunk: content_range = "bytes {0}-{1}/{2}".format(current_loc, current_loc + len(chunk) - 1, file_size) current_loc += len(chunk) http_headers = {"Content-Range": content_range, "X-Unique-Upload-Id": upload_id} upload_result = upload_large_part((file_name, chunk), http_headers=http_headers, **options) options["public_id"] = upload_result.get("public_id") chunk = file_io.read(chunk_size) return upload_result
[ "def", "upload_large", "(", "file", ",", "*", "*", "options", ")", ":", "if", "utils", ".", "is_remote_url", "(", "file", ")", ":", "return", "upload", "(", "file", ",", "*", "*", "options", ")", "if", "hasattr", "(", "file", ",", "'read'", ")", "and", "callable", "(", "file", ".", "read", ")", ":", "file_io", "=", "file", "else", ":", "file_io", "=", "open", "(", "file", ",", "'rb'", ")", "upload_result", "=", "None", "with", "file_io", ":", "upload_id", "=", "utils", ".", "random_public_id", "(", ")", "current_loc", "=", "0", "chunk_size", "=", "options", ".", "get", "(", "\"chunk_size\"", ",", "UPLOAD_LARGE_CHUNK_SIZE", ")", "file_size", "=", "utils", ".", "file_io_size", "(", "file_io", ")", "file_name", "=", "options", ".", "get", "(", "\"filename\"", ",", "file_io", ".", "name", "if", "hasattr", "(", "file_io", ",", "'name'", ")", "and", "isinstance", "(", "file_io", ".", "name", ",", "str", ")", "else", "\"stream\"", ")", "chunk", "=", "file_io", ".", "read", "(", "chunk_size", ")", "while", "chunk", ":", "content_range", "=", "\"bytes {0}-{1}/{2}\"", ".", "format", "(", "current_loc", ",", "current_loc", "+", "len", "(", "chunk", ")", "-", "1", ",", "file_size", ")", "current_loc", "+=", "len", "(", "chunk", ")", "http_headers", "=", "{", "\"Content-Range\"", ":", "content_range", ",", "\"X-Unique-Upload-Id\"", ":", "upload_id", "}", "upload_result", "=", "upload_large_part", "(", "(", "file_name", ",", "chunk", ")", ",", "http_headers", "=", "http_headers", ",", "*", "*", "options", ")", "options", "[", "\"public_id\"", "]", "=", "upload_result", ".", "get", "(", "\"public_id\"", ")", "chunk", "=", "file_io", ".", "read", "(", "chunk_size", ")", "return", "upload_result" ]
[ 67, 0 ]
[ 102, 24 ]
python
en
['en', 'da', 'en']
True
upload_large_part
(file, **options)
Upload large files.
Upload large files.
def upload_large_part(file, **options): """ Upload large files. """ params = utils.build_upload_params(**options) if 'resource_type' not in options: options['resource_type'] = "raw" return call_cacheable_api("upload", params, file=file, **options)
[ "def", "upload_large_part", "(", "file", ",", "*", "*", "options", ")", ":", "params", "=", "utils", ".", "build_upload_params", "(", "*", "*", "options", ")", "if", "'resource_type'", "not", "in", "options", ":", "options", "[", "'resource_type'", "]", "=", "\"raw\"", "return", "call_cacheable_api", "(", "\"upload\"", ",", "params", ",", "file", "=", "file", ",", "*", "*", "options", ")" ]
[ 105, 0 ]
[ 112, 69 ]
python
en
['en', 'da', 'en']
True
update_metadata
(metadata, public_ids, **options)
Populates metadata fields with the given values. Existing values will be overwritten. Any metadata-value pairs given are merged with any existing metadata-value pairs (an empty value for an existing metadata field clears the value) :param metadata: A list of custom metadata fields (by external_id) and the values to assign to each of them. :param public_ids: An array of Public IDs of assets uploaded to Cloudinary. :param options: Options such as *resource_type* (the type of file. Default: image. Valid values: image, raw, or video) and *type* (The storage type. Default: upload. Valid values: upload, private, or authenticated.) :return: A list of public IDs that were updated :rtype: mixed
Populates metadata fields with the given values. Existing values will be overwritten.
def update_metadata(metadata, public_ids, **options): """ Populates metadata fields with the given values. Existing values will be overwritten. Any metadata-value pairs given are merged with any existing metadata-value pairs (an empty value for an existing metadata field clears the value) :param metadata: A list of custom metadata fields (by external_id) and the values to assign to each of them. :param public_ids: An array of Public IDs of assets uploaded to Cloudinary. :param options: Options such as *resource_type* (the type of file. Default: image. Valid values: image, raw, or video) and *type* (The storage type. Default: upload. Valid values: upload, private, or authenticated.) :return: A list of public IDs that were updated :rtype: mixed """ params = { "timestamp": utils.now(), "metadata": utils.encode_context(metadata), "public_ids": utils.build_array(public_ids), "type": options.get("type") } return call_api("metadata", params, **options)
[ "def", "update_metadata", "(", "metadata", ",", "public_ids", ",", "*", "*", "options", ")", ":", "params", "=", "{", "\"timestamp\"", ":", "utils", ".", "now", "(", ")", ",", "\"metadata\"", ":", "utils", ".", "encode_context", "(", "metadata", ")", ",", "\"public_ids\"", ":", "utils", ".", "build_array", "(", "public_ids", ")", ",", "\"type\"", ":", "options", ".", "get", "(", "\"type\"", ")", "}", "return", "call_api", "(", "\"metadata\"", ",", "params", ",", "*", "*", "options", ")" ]
[ 140, 0 ]
[ 164, 50 ]
python
en
['en', 'error', 'th']
False
generate_sprite
(tag=None, urls=None, **options)
Generates sprites by merging multiple images into a single large image. See: `Sprite method API reference <https://cloudinary.com/documentation/image_upload_api_reference#sprite_method>`_ :param tag: The sprite is created from all images with this tag. If not set - `urls` parameter is required :type tag: str :param urls: List of URLs to create a sprite from. Can only be used if `tag` is not set :type urls: list :param options: Additional options :type options: dict, optional :return: Dictionary with meta information URLs of generated sprite resources :rtype: dict
Generates sprites by merging multiple images into a single large image.
def generate_sprite(tag=None, urls=None, **options): """ Generates sprites by merging multiple images into a single large image. See: `Sprite method API reference <https://cloudinary.com/documentation/image_upload_api_reference#sprite_method>`_ :param tag: The sprite is created from all images with this tag. If not set - `urls` parameter is required :type tag: str :param urls: List of URLs to create a sprite from. Can only be used if `tag` is not set :type urls: list :param options: Additional options :type options: dict, optional :return: Dictionary with meta information URLs of generated sprite resources :rtype: dict """ params = utils.build_multi_and_sprite_params(tag=tag, urls=urls, **options) return call_api("sprite", params, **options)
[ "def", "generate_sprite", "(", "tag", "=", "None", ",", "urls", "=", "None", ",", "*", "*", "options", ")", ":", "params", "=", "utils", ".", "build_multi_and_sprite_params", "(", "tag", "=", "tag", ",", "urls", "=", "urls", ",", "*", "*", "options", ")", "return", "call_api", "(", "\"sprite\"", ",", "params", ",", "*", "*", "options", ")" ]
[ 184, 0 ]
[ 201, 48 ]
python
en
['en', 'error', 'th']
False
download_generated_sprite
(tag=None, urls=None, **options)
Returns signed URL for the sprite endpoint with `mode=download` :param tag: The sprite is created from all images with this tag. If not set - `urls` parameter is required :type tag: str :param urls: List of URLs to create a sprite from. Can only be used if `tag` is not set :type urls: list :param options: Additional options :type options: dict, optional :return: The signed URL to download sprite :rtype: str
Returns signed URL for the sprite endpoint with `mode=download`
def download_generated_sprite(tag=None, urls=None, **options): """ Returns signed URL for the sprite endpoint with `mode=download` :param tag: The sprite is created from all images with this tag. If not set - `urls` parameter is required :type tag: str :param urls: List of URLs to create a sprite from. Can only be used if `tag` is not set :type urls: list :param options: Additional options :type options: dict, optional :return: The signed URL to download sprite :rtype: str """ params = utils.build_multi_and_sprite_params(tag=tag, urls=urls, **options) return utils.cloudinary_api_download_url(action="sprite", params=params, **options)
[ "def", "download_generated_sprite", "(", "tag", "=", "None", ",", "urls", "=", "None", ",", "*", "*", "options", ")", ":", "params", "=", "utils", ".", "build_multi_and_sprite_params", "(", "tag", "=", "tag", ",", "urls", "=", "urls", ",", "*", "*", "options", ")", "return", "utils", ".", "cloudinary_api_download_url", "(", "action", "=", "\"sprite\"", ",", "params", "=", "params", ",", "*", "*", "options", ")" ]
[ 204, 0 ]
[ 218, 87 ]
python
en
['en', 'error', 'th']
False
multi
(tag=None, urls=None, **options)
Creates either a single animated image, video or a PDF. See: `Upload method API reference <https://cloudinary.com/documentation/image_upload_api_reference#multi_method>`_ :param tag: The animated image, video or PDF is created from all images with this tag. If not set - `urls` parameter is required :type tag: str :param urls: List of URLs to create an animated image, video or PDF from. Can only be used if `tag` is not set :type urls: list :param options: Additional options :type options: dict, optional :return: Dictionary with meta information URLs of the generated file :rtype: dict
Creates either a single animated image, video or a PDF.
def multi(tag=None, urls=None, **options): """ Creates either a single animated image, video or a PDF. See: `Upload method API reference <https://cloudinary.com/documentation/image_upload_api_reference#multi_method>`_ :param tag: The animated image, video or PDF is created from all images with this tag. If not set - `urls` parameter is required :type tag: str :param urls: List of URLs to create an animated image, video or PDF from. Can only be used if `tag` is not set :type urls: list :param options: Additional options :type options: dict, optional :return: Dictionary with meta information URLs of the generated file :rtype: dict """ params = utils.build_multi_and_sprite_params(tag=tag, urls=urls, **options) return call_api("multi", params, **options)
[ "def", "multi", "(", "tag", "=", "None", ",", "urls", "=", "None", ",", "*", "*", "options", ")", ":", "params", "=", "utils", ".", "build_multi_and_sprite_params", "(", "tag", "=", "tag", ",", "urls", "=", "urls", ",", "*", "*", "options", ")", "return", "call_api", "(", "\"multi\"", ",", "params", ",", "*", "*", "options", ")" ]
[ 221, 0 ]
[ 239, 47 ]
python
en
['en', 'error', 'th']
False
download_multi
(tag=None, urls=None, **options)
Returns signed URL for the multi endpoint with `mode=download` :param tag: The sprite is created from all images with this tag. If not set - `urls` parameter is required :type tag: str :param urls: List of URLs to create a sprite from. Can only be used if `tag` is not set :type urls: list :param options: Additional options :type options: dict, optional :return: The signed URL to download multi :rtype: str
Returns signed URL for the multi endpoint with `mode=download`
def download_multi(tag=None, urls=None, **options): """ Returns signed URL for the multi endpoint with `mode=download` :param tag: The sprite is created from all images with this tag. If not set - `urls` parameter is required :type tag: str :param urls: List of URLs to create a sprite from. Can only be used if `tag` is not set :type urls: list :param options: Additional options :type options: dict, optional :return: The signed URL to download multi :rtype: str """ params = utils.build_multi_and_sprite_params(tag=tag, urls=urls, **options) return utils.cloudinary_api_download_url(action="multi", params=params, **options)
[ "def", "download_multi", "(", "tag", "=", "None", ",", "urls", "=", "None", ",", "*", "*", "options", ")", ":", "params", "=", "utils", ".", "build_multi_and_sprite_params", "(", "tag", "=", "tag", ",", "urls", "=", "urls", ",", "*", "*", "options", ")", "return", "utils", ".", "cloudinary_api_download_url", "(", "action", "=", "\"multi\"", ",", "params", "=", "params", ",", "*", "*", "options", ")" ]
[ 242, 0 ]
[ 256, 86 ]
python
en
['en', 'error', 'th']
False
remove_all_tags
(public_ids, **options)
Remove all tags from the specified public IDs. :param public_ids: the public IDs of the resources to update :param options: additional options passed to the request :return: dictionary with a list of public IDs that were updated
Remove all tags from the specified public IDs.
def remove_all_tags(public_ids, **options): """ Remove all tags from the specified public IDs. :param public_ids: the public IDs of the resources to update :param options: additional options passed to the request :return: dictionary with a list of public IDs that were updated """ return call_tags_api(None, "remove_all", public_ids, **options)
[ "def", "remove_all_tags", "(", "public_ids", ",", "*", "*", "options", ")", ":", "return", "call_tags_api", "(", "None", ",", "\"remove_all\"", ",", "public_ids", ",", "*", "*", "options", ")" ]
[ 285, 0 ]
[ 294, 67 ]
python
en
['en', 'error', 'th']
False
add_context
(context, public_ids, **options)
Add a context keys and values. If a particular key already exists, the value associated with the key is updated. :param context: dictionary of context :param public_ids: the public IDs of the resources to update :param options: additional options passed to the request :return: dictionary with a list of public IDs that were updated
Add a context keys and values. If a particular key already exists, the value associated with the key is updated.
def add_context(context, public_ids, **options): """ Add a context keys and values. If a particular key already exists, the value associated with the key is updated. :param context: dictionary of context :param public_ids: the public IDs of the resources to update :param options: additional options passed to the request :return: dictionary with a list of public IDs that were updated """ return call_context_api(context, "add", public_ids, **options)
[ "def", "add_context", "(", "context", ",", "public_ids", ",", "*", "*", "options", ")", ":", "return", "call_context_api", "(", "context", ",", "\"add\"", ",", "public_ids", ",", "*", "*", "options", ")" ]
[ 297, 0 ]
[ 307, 66 ]
python
en
['en', 'error', 'th']
False
remove_all_context
(public_ids, **options)
Remove all custom context from the specified public IDs. :param public_ids: the public IDs of the resources to update :param options: additional options passed to the request :return: dictionary with a list of public IDs that were updated
Remove all custom context from the specified public IDs.
def remove_all_context(public_ids, **options): """ Remove all custom context from the specified public IDs. :param public_ids: the public IDs of the resources to update :param options: additional options passed to the request :return: dictionary with a list of public IDs that were updated """ return call_context_api(None, "remove_all", public_ids, **options)
[ "def", "remove_all_context", "(", "public_ids", ",", "*", "*", "options", ")", ":", "return", "call_context_api", "(", "None", ",", "\"remove_all\"", ",", "public_ids", ",", "*", "*", "options", ")" ]
[ 310, 0 ]
[ 319, 70 ]
python
en
['en', 'error', 'th']
False
_save_responsive_breakpoints_to_cache
(result)
Saves responsive breakpoints parsed from upload result to cache :param result: Upload result
Saves responsive breakpoints parsed from upload result to cache
def _save_responsive_breakpoints_to_cache(result): """ Saves responsive breakpoints parsed from upload result to cache :param result: Upload result """ if "responsive_breakpoints" not in result: return if "public_id" not in result: # We have some faulty result, nothing to cache return options = dict((k, result[k]) for k in ["type", "resource_type"] if k in result) for transformation in result.get("responsive_breakpoints", []): options["raw_transformation"] = transformation.get("transformation", "") options["format"] = os.path.splitext(transformation["breakpoints"][0]["url"])[1][1:] breakpoints = [bp["width"] for bp in transformation["breakpoints"]] responsive_breakpoints_cache_instance.set(result["public_id"], breakpoints, **options)
[ "def", "_save_responsive_breakpoints_to_cache", "(", "result", ")", ":", "if", "\"responsive_breakpoints\"", "not", "in", "result", ":", "return", "if", "\"public_id\"", "not", "in", "result", ":", "# We have some faulty result, nothing to cache", "return", "options", "=", "dict", "(", "(", "k", ",", "result", "[", "k", "]", ")", "for", "k", "in", "[", "\"type\"", ",", "\"resource_type\"", "]", "if", "k", "in", "result", ")", "for", "transformation", "in", "result", ".", "get", "(", "\"responsive_breakpoints\"", ",", "[", "]", ")", ":", "options", "[", "\"raw_transformation\"", "]", "=", "transformation", ".", "get", "(", "\"transformation\"", ",", "\"\"", ")", "options", "[", "\"format\"", "]", "=", "os", ".", "path", ".", "splitext", "(", "transformation", "[", "\"breakpoints\"", "]", "[", "0", "]", "[", "\"url\"", "]", ")", "[", "1", "]", "[", "1", ":", "]", "breakpoints", "=", "[", "bp", "[", "\"width\"", "]", "for", "bp", "in", "transformation", "[", "\"breakpoints\"", "]", "]", "responsive_breakpoints_cache_instance", ".", "set", "(", "result", "[", "\"public_id\"", "]", ",", "breakpoints", ",", "*", "*", "options", ")" ]
[ 365, 0 ]
[ 384, 94 ]
python
en
['en', 'error', 'th']
False
call_cacheable_api
(action, params, http_headers=None, return_error=False, unsigned=False, file=None, timeout=None, **options)
Calls Upload API and saves results to cache (if enabled)
Calls Upload API and saves results to cache (if enabled)
def call_cacheable_api(action, params, http_headers=None, return_error=False, unsigned=False, file=None, timeout=None, **options): """ Calls Upload API and saves results to cache (if enabled) """ result = call_api(action, params, http_headers, return_error, unsigned, file, timeout, **options) if "use_cache" in options or cloudinary.config().use_cache: _save_responsive_breakpoints_to_cache(result) return result
[ "def", "call_cacheable_api", "(", "action", ",", "params", ",", "http_headers", "=", "None", ",", "return_error", "=", "False", ",", "unsigned", "=", "False", ",", "file", "=", "None", ",", "timeout", "=", "None", ",", "*", "*", "options", ")", ":", "result", "=", "call_api", "(", "action", ",", "params", ",", "http_headers", ",", "return_error", ",", "unsigned", ",", "file", ",", "timeout", ",", "*", "*", "options", ")", "if", "\"use_cache\"", "in", "options", "or", "cloudinary", ".", "config", "(", ")", ".", "use_cache", ":", "_save_responsive_breakpoints_to_cache", "(", "result", ")", "return", "result" ]
[ 387, 0 ]
[ 398, 17 ]
python
en
['en', 'error', 'th']
False
Link.__init__
( self, url: str, comes_from: Optional[Union[str, "HTMLPage"]] = None, requires_python: Optional[str] = None, yanked_reason: Optional[str] = None, cache_link_parsing: bool = True, )
:param url: url of the resource pointed to (href of the link) :param comes_from: instance of HTMLPage where the link was found, or string. :param requires_python: String containing the `Requires-Python` metadata field, specified in PEP 345. This may be specified by a data-requires-python attribute in the HTML link tag, as described in PEP 503. :param yanked_reason: the reason the file has been yanked, if the file has been yanked, or None if the file hasn't been yanked. This is the value of the "data-yanked" attribute, if present, in a simple repository HTML link. If the file has been yanked but no reason was provided, this should be the empty string. See PEP 592 for more information and the specification. :param cache_link_parsing: A flag that is used elsewhere to determine whether resources retrieved from this link should be cached. PyPI index urls should generally have this set to False, for example.
:param url: url of the resource pointed to (href of the link) :param comes_from: instance of HTMLPage where the link was found, or string. :param requires_python: String containing the `Requires-Python` metadata field, specified in PEP 345. This may be specified by a data-requires-python attribute in the HTML link tag, as described in PEP 503. :param yanked_reason: the reason the file has been yanked, if the file has been yanked, or None if the file hasn't been yanked. This is the value of the "data-yanked" attribute, if present, in a simple repository HTML link. If the file has been yanked but no reason was provided, this should be the empty string. See PEP 592 for more information and the specification. :param cache_link_parsing: A flag that is used elsewhere to determine whether resources retrieved from this link should be cached. PyPI index urls should generally have this set to False, for example.
def __init__( self, url: str, comes_from: Optional[Union[str, "HTMLPage"]] = None, requires_python: Optional[str] = None, yanked_reason: Optional[str] = None, cache_link_parsing: bool = True, ) -> None: """ :param url: url of the resource pointed to (href of the link) :param comes_from: instance of HTMLPage where the link was found, or string. :param requires_python: String containing the `Requires-Python` metadata field, specified in PEP 345. This may be specified by a data-requires-python attribute in the HTML link tag, as described in PEP 503. :param yanked_reason: the reason the file has been yanked, if the file has been yanked, or None if the file hasn't been yanked. This is the value of the "data-yanked" attribute, if present, in a simple repository HTML link. If the file has been yanked but no reason was provided, this should be the empty string. See PEP 592 for more information and the specification. :param cache_link_parsing: A flag that is used elsewhere to determine whether resources retrieved from this link should be cached. PyPI index urls should generally have this set to False, for example. """ # url can be a UNC windows share if url.startswith('\\\\'): url = path_to_url(url) self._parsed_url = urllib.parse.urlsplit(url) # Store the url as a private attribute to prevent accidentally # trying to set a new value. self._url = url self.comes_from = comes_from self.requires_python = requires_python if requires_python else None self.yanked_reason = yanked_reason super().__init__(key=url, defining_class=Link) self.cache_link_parsing = cache_link_parsing
[ "def", "__init__", "(", "self", ",", "url", ":", "str", ",", "comes_from", ":", "Optional", "[", "Union", "[", "str", ",", "\"HTMLPage\"", "]", "]", "=", "None", ",", "requires_python", ":", "Optional", "[", "str", "]", "=", "None", ",", "yanked_reason", ":", "Optional", "[", "str", "]", "=", "None", ",", "cache_link_parsing", ":", "bool", "=", "True", ",", ")", "->", "None", ":", "# url can be a UNC windows share", "if", "url", ".", "startswith", "(", "'\\\\\\\\'", ")", ":", "url", "=", "path_to_url", "(", "url", ")", "self", ".", "_parsed_url", "=", "urllib", ".", "parse", ".", "urlsplit", "(", "url", ")", "# Store the url as a private attribute to prevent accidentally", "# trying to set a new value.", "self", ".", "_url", "=", "url", "self", ".", "comes_from", "=", "comes_from", "self", ".", "requires_python", "=", "requires_python", "if", "requires_python", "else", "None", "self", ".", "yanked_reason", "=", "yanked_reason", "super", "(", ")", ".", "__init__", "(", "key", "=", "url", ",", "defining_class", "=", "Link", ")", "self", ".", "cache_link_parsing", "=", "cache_link_parsing" ]
[ 40, 4 ]
[ 84, 52 ]
python
en
['en', 'error', 'th']
False
Link.netloc
(self)
This can contain auth information.
This can contain auth information.
def netloc(self) -> str: """ This can contain auth information. """ return self._parsed_url.netloc
[ "def", "netloc", "(", "self", ")", "->", "str", ":", "return", "self", ".", "_parsed_url", ".", "netloc" ]
[ 127, 4 ]
[ 131, 38 ]
python
en
['en', 'error', 'th']
False
Link.is_hash_allowed
(self, hashes: Optional[Hashes])
Return True if the link has a hash and it is allowed.
Return True if the link has a hash and it is allowed.
def is_hash_allowed(self, hashes: Optional[Hashes]) -> bool: """ Return True if the link has a hash and it is allowed. """ if hashes is None or not self.has_hash: return False # Assert non-None so mypy knows self.hash_name and self.hash are str. assert self.hash_name is not None assert self.hash is not None return hashes.is_hash_allowed(self.hash_name, hex_digest=self.hash)
[ "def", "is_hash_allowed", "(", "self", ",", "hashes", ":", "Optional", "[", "Hashes", "]", ")", "->", "bool", ":", "if", "hashes", "is", "None", "or", "not", "self", ".", "has_hash", ":", "return", "False", "# Assert non-None so mypy knows self.hash_name and self.hash are str.", "assert", "self", ".", "hash_name", "is", "not", "None", "assert", "self", ".", "hash", "is", "not", "None", "return", "hashes", ".", "is_hash_allowed", "(", "self", ".", "hash_name", ",", "hex_digest", "=", "self", ".", "hash", ")" ]
[ 214, 4 ]
[ 224, 75 ]
python
en
['en', 'error', 'th']
False
SimpleTemplateResponse.__getstate__
(self)
Raise an exception if trying to pickle an unrendered response. Pickle only rendered data, not the data used to construct the response.
Raise an exception if trying to pickle an unrendered response. Pickle only rendered data, not the data used to construct the response.
def __getstate__(self): """ Raise an exception if trying to pickle an unrendered response. Pickle only rendered data, not the data used to construct the response. """ obj_dict = self.__dict__.copy() if not self._is_rendered: raise ContentNotRenderedError('The response content must be ' 'rendered before it can be pickled.') for attr in self.rendering_attrs: if attr in obj_dict: del obj_dict[attr] return obj_dict
[ "def", "__getstate__", "(", "self", ")", ":", "obj_dict", "=", "self", ".", "__dict__", ".", "copy", "(", ")", "if", "not", "self", ".", "_is_rendered", ":", "raise", "ContentNotRenderedError", "(", "'The response content must be '", "'rendered before it can be pickled.'", ")", "for", "attr", "in", "self", ".", "rendering_attrs", ":", "if", "attr", "in", "obj_dict", ":", "del", "obj_dict", "[", "attr", "]", "return", "obj_dict" ]
[ 44, 4 ]
[ 57, 23 ]
python
en
['en', 'error', 'th']
False
SimpleTemplateResponse.resolve_template
(self, template)
Accept a template object, path-to-template, or list of paths.
Accept a template object, path-to-template, or list of paths.
def resolve_template(self, template): """Accept a template object, path-to-template, or list of paths.""" if isinstance(template, (list, tuple)): return select_template(template, using=self.using) elif isinstance(template, str): return get_template(template, using=self.using) else: return template
[ "def", "resolve_template", "(", "self", ",", "template", ")", ":", "if", "isinstance", "(", "template", ",", "(", "list", ",", "tuple", ")", ")", ":", "return", "select_template", "(", "template", ",", "using", "=", "self", ".", "using", ")", "elif", "isinstance", "(", "template", ",", "str", ")", ":", "return", "get_template", "(", "template", ",", "using", "=", "self", ".", "using", ")", "else", ":", "return", "template" ]
[ 59, 4 ]
[ 66, 27 ]
python
en
['en', 'en', 'en']
True
SimpleTemplateResponse.rendered_content
(self)
Return the freshly rendered content for the template and context described by the TemplateResponse. This *does not* set the final content of the response. To set the response content, you must either call render(), or set the content explicitly using the value of this property.
Return the freshly rendered content for the template and context described by the TemplateResponse.
def rendered_content(self): """Return the freshly rendered content for the template and context described by the TemplateResponse. This *does not* set the final content of the response. To set the response content, you must either call render(), or set the content explicitly using the value of this property. """ template = self.resolve_template(self.template_name) context = self.resolve_context(self.context_data) return template.render(context, self._request)
[ "def", "rendered_content", "(", "self", ")", ":", "template", "=", "self", ".", "resolve_template", "(", "self", ".", "template_name", ")", "context", "=", "self", ".", "resolve_context", "(", "self", ".", "context_data", ")", "return", "template", ".", "render", "(", "context", ",", "self", ".", "_request", ")" ]
[ 72, 4 ]
[ 82, 54 ]
python
en
['en', 'en', 'en']
True
SimpleTemplateResponse.add_post_render_callback
(self, callback)
Add a new post-rendering callback. If the response has already been rendered, invoke the callback immediately.
Add a new post-rendering callback.
def add_post_render_callback(self, callback): """Add a new post-rendering callback. If the response has already been rendered, invoke the callback immediately. """ if self._is_rendered: callback(self) else: self._post_render_callbacks.append(callback)
[ "def", "add_post_render_callback", "(", "self", ",", "callback", ")", ":", "if", "self", ".", "_is_rendered", ":", "callback", "(", "self", ")", "else", ":", "self", ".", "_post_render_callbacks", ".", "append", "(", "callback", ")" ]
[ 84, 4 ]
[ 93, 56 ]
python
en
['en', 'en', 'en']
True
SimpleTemplateResponse.render
(self)
Render (thereby finalizing) the content of the response. If the content has already been rendered, this is a no-op. Return the baked response instance.
Render (thereby finalizing) the content of the response.
def render(self): """Render (thereby finalizing) the content of the response. If the content has already been rendered, this is a no-op. Return the baked response instance. """ retval = self if not self._is_rendered: self.content = self.rendered_content for post_callback in self._post_render_callbacks: newretval = post_callback(retval) if newretval is not None: retval = newretval return retval
[ "def", "render", "(", "self", ")", ":", "retval", "=", "self", "if", "not", "self", ".", "_is_rendered", ":", "self", ".", "content", "=", "self", ".", "rendered_content", "for", "post_callback", "in", "self", ".", "_post_render_callbacks", ":", "newretval", "=", "post_callback", "(", "retval", ")", "if", "newretval", "is", "not", "None", ":", "retval", "=", "newretval", "return", "retval" ]
[ 95, 4 ]
[ 109, 21 ]
python
en
['en', 'en', 'en']
True
SimpleTemplateResponse.content
(self, value)
Set the content for the response.
Set the content for the response.
def content(self, value): """Set the content for the response.""" HttpResponse.content.fset(self, value) self._is_rendered = True
[ "def", "content", "(", "self", ",", "value", ")", ":", "HttpResponse", ".", "content", ".", "fset", "(", "self", ",", "value", ")", "self", ".", "_is_rendered", "=", "True" ]
[ 131, 4 ]
[ 134, 32 ]
python
en
['en', 'en', 'en']
True
RequestException.__init__
(self, *args, **kwargs)
Initialize RequestException with `request` and `response` objects.
Initialize RequestException with `request` and `response` objects.
def __init__(self, *args, **kwargs): """Initialize RequestException with `request` and `response` objects.""" response = kwargs.pop('response', None) self.response = response self.request = kwargs.pop('request', None) if (response is not None and not self.request and hasattr(response, 'request')): self.request = self.response.request super(RequestException, self).__init__(*args, **kwargs)
[ "def", "__init__", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "response", "=", "kwargs", ".", "pop", "(", "'response'", ",", "None", ")", "self", ".", "response", "=", "response", "self", ".", "request", "=", "kwargs", ".", "pop", "(", "'request'", ",", "None", ")", "if", "(", "response", "is", "not", "None", "and", "not", "self", ".", "request", "and", "hasattr", "(", "response", ",", "'request'", ")", ")", ":", "self", ".", "request", "=", "self", ".", "response", ".", "request", "super", "(", "RequestException", ",", "self", ")", ".", "__init__", "(", "*", "args", ",", "*", "*", "kwargs", ")" ]
[ 16, 4 ]
[ 24, 63 ]
python
en
['en', 'en', 'en']
True
certs
()
Returns a dictionary of current Google public key certificates for validating Google-signed JWTs. Since these change rarely, the result is cached on first request for faster subsequent responses.
Returns a dictionary of current Google public key certificates for validating Google-signed JWTs. Since these change rarely, the result is cached on first request for faster subsequent responses.
def certs(): """Returns a dictionary of current Google public key certificates for validating Google-signed JWTs. Since these change rarely, the result is cached on first request for faster subsequent responses. """ import requests global CERTS if CERTS is None: response = requests.get( 'https://www.gstatic.com/iap/verify/public_key' ) CERTS = response.json() return CERTS
[ "def", "certs", "(", ")", ":", "import", "requests", "global", "CERTS", "if", "CERTS", "is", "None", ":", "response", "=", "requests", ".", "get", "(", "'https://www.gstatic.com/iap/verify/public_key'", ")", "CERTS", "=", "response", ".", "json", "(", ")", "return", "CERTS" ]
[ 25, 0 ]
[ 38, 16 ]
python
en
['en', 'en', 'en']
True
get_metadata
(item_name)
Returns a string with the project metadata value for the item_name. See https://cloud.google.com/compute/docs/storing-retrieving-metadata for possible item_name values.
Returns a string with the project metadata value for the item_name. See https://cloud.google.com/compute/docs/storing-retrieving-metadata for possible item_name values.
def get_metadata(item_name): """Returns a string with the project metadata value for the item_name. See https://cloud.google.com/compute/docs/storing-retrieving-metadata for possible item_name values. """ import requests endpoint = 'http://metadata.google.internal' path = '/computeMetadata/v1/project/' path += item_name response = requests.get( '{}{}'.format(endpoint, path), headers={'Metadata-Flavor': 'Google'} ) metadata = response.text return metadata
[ "def", "get_metadata", "(", "item_name", ")", ":", "import", "requests", "endpoint", "=", "'http://metadata.google.internal'", "path", "=", "'/computeMetadata/v1/project/'", "path", "+=", "item_name", "response", "=", "requests", ".", "get", "(", "'{}{}'", ".", "format", "(", "endpoint", ",", "path", ")", ",", "headers", "=", "{", "'Metadata-Flavor'", ":", "'Google'", "}", ")", "metadata", "=", "response", ".", "text", "return", "metadata" ]
[ 43, 0 ]
[ 58, 19 ]
python
en
['en', 'en', 'en']
True
audience
()
Returns the audience value (the JWT 'aud' property) for the current running instance. Since this involves a metadata lookup, the result is cached when first requested for faster future responses.
Returns the audience value (the JWT 'aud' property) for the current running instance. Since this involves a metadata lookup, the result is cached when first requested for faster future responses.
def audience(): """Returns the audience value (the JWT 'aud' property) for the current running instance. Since this involves a metadata lookup, the result is cached when first requested for faster future responses. """ global AUDIENCE if AUDIENCE is None: project_number = get_metadata('numeric-project-id') project_id = get_metadata('project-id') AUDIENCE = '/projects/{}/apps/{}'.format( project_number, project_id ) return AUDIENCE
[ "def", "audience", "(", ")", ":", "global", "AUDIENCE", "if", "AUDIENCE", "is", "None", ":", "project_number", "=", "get_metadata", "(", "'numeric-project-id'", ")", "project_id", "=", "get_metadata", "(", "'project-id'", ")", "AUDIENCE", "=", "'/projects/{}/apps/{}'", ".", "format", "(", "project_number", ",", "project_id", ")", "return", "AUDIENCE" ]
[ 63, 0 ]
[ 75, 19 ]
python
en
['en', 'en', 'en']
True
validate_assertion
(assertion)
Checks that the JWT assertion is valid (properly signed, for the correct audience) and if so, returns strings for the requesting user's email and a persistent user ID. If not valid, returns None for each field.
Checks that the JWT assertion is valid (properly signed, for the correct audience) and if so, returns strings for the requesting user's email and a persistent user ID. If not valid, returns None for each field.
def validate_assertion(assertion): """Checks that the JWT assertion is valid (properly signed, for the correct audience) and if so, returns strings for the requesting user's email and a persistent user ID. If not valid, returns None for each field. """ from jose import jwt try: info = jwt.decode( assertion, certs(), algorithms=['ES256'], audience=audience() ) return info['email'], info['sub'] except Exception as e: print('Failed to validate assertion: {}'.format(e), file=sys.stderr) return None, None
[ "def", "validate_assertion", "(", "assertion", ")", ":", "from", "jose", "import", "jwt", "try", ":", "info", "=", "jwt", ".", "decode", "(", "assertion", ",", "certs", "(", ")", ",", "algorithms", "=", "[", "'ES256'", "]", ",", "audience", "=", "audience", "(", ")", ")", "return", "info", "[", "'email'", "]", ",", "info", "[", "'sub'", "]", "except", "Exception", "as", "e", ":", "print", "(", "'Failed to validate assertion: {}'", ".", "format", "(", "e", ")", ",", "file", "=", "sys", ".", "stderr", ")", "return", "None", ",", "None" ]
[ 80, 0 ]
[ 97, 25 ]
python
en
['en', 'en', 'en']
True
UserProfileSerializer.create
(self, validated_data)
Create and return a new user
Create and return a new user
def create(self, validated_data): """Create and return a new user""" user = models.UserProfile.objects.create_user( email=validated_data['email'], name=validated_data['name'], password=validated_data['password'] ) return user
[ "def", "create", "(", "self", ",", "validated_data", ")", ":", "user", "=", "models", ".", "UserProfile", ".", "objects", ".", "create_user", "(", "email", "=", "validated_data", "[", "'email'", "]", ",", "name", "=", "validated_data", "[", "'name'", "]", ",", "password", "=", "validated_data", "[", "'password'", "]", ")", "return", "user" ]
[ 23, 4 ]
[ 31, 19 ]
python
en
['en', 'en', 'en']
True
parse_docstring
(docstring)
Parse out the parts of a docstring. Return (title, body, metadata).
Parse out the parts of a docstring. Return (title, body, metadata).
def parse_docstring(docstring): """ Parse out the parts of a docstring. Return (title, body, metadata). """ if not docstring: return '', '', {} docstring = cleandoc(docstring) parts = re.split(r'\n{2,}', docstring) title = parts[0] if len(parts) == 1: body = '' metadata = {} else: parser = HeaderParser() try: metadata = parser.parsestr(parts[-1]) except HeaderParseError: metadata = {} body = "\n\n".join(parts[1:]) else: metadata = dict(metadata.items()) if metadata: body = "\n\n".join(parts[1:-1]) else: body = "\n\n".join(parts[1:]) return title, body, metadata
[ "def", "parse_docstring", "(", "docstring", ")", ":", "if", "not", "docstring", ":", "return", "''", ",", "''", ",", "{", "}", "docstring", "=", "cleandoc", "(", "docstring", ")", "parts", "=", "re", ".", "split", "(", "r'\\n{2,}'", ",", "docstring", ")", "title", "=", "parts", "[", "0", "]", "if", "len", "(", "parts", ")", "==", "1", ":", "body", "=", "''", "metadata", "=", "{", "}", "else", ":", "parser", "=", "HeaderParser", "(", ")", "try", ":", "metadata", "=", "parser", ".", "parsestr", "(", "parts", "[", "-", "1", "]", ")", "except", "HeaderParseError", ":", "metadata", "=", "{", "}", "body", "=", "\"\\n\\n\"", ".", "join", "(", "parts", "[", "1", ":", "]", ")", "else", ":", "metadata", "=", "dict", "(", "metadata", ".", "items", "(", ")", ")", "if", "metadata", ":", "body", "=", "\"\\n\\n\"", ".", "join", "(", "parts", "[", "1", ":", "-", "1", "]", ")", "else", ":", "body", "=", "\"\\n\\n\"", ".", "join", "(", "parts", "[", "1", ":", "]", ")", "return", "title", ",", "body", ",", "metadata" ]
[ 27, 0 ]
[ 52, 32 ]
python
en
['en', 'error', 'th']
False
parse_rst
(text, default_reference_context, thing_being_parsed=None)
Convert the string from reST to an XHTML fragment.
Convert the string from reST to an XHTML fragment.
def parse_rst(text, default_reference_context, thing_being_parsed=None): """ Convert the string from reST to an XHTML fragment. """ overrides = { 'doctitle_xform': True, 'initial_header_level': 3, "default_reference_context": default_reference_context, "link_base": reverse('django-admindocs-docroot').rstrip('/'), 'raw_enabled': False, 'file_insertion_enabled': False, } thing_being_parsed = thing_being_parsed and '<%s>' % thing_being_parsed # Wrap ``text`` in some reST that sets the default role to ``cmsreference``, # then restores it. source = """ .. default-role:: cmsreference %s .. default-role:: """ parts = docutils.core.publish_parts( source % text, source_path=thing_being_parsed, destination_path=None, writer_name='html', settings_overrides=overrides, ) return mark_safe(parts['fragment'])
[ "def", "parse_rst", "(", "text", ",", "default_reference_context", ",", "thing_being_parsed", "=", "None", ")", ":", "overrides", "=", "{", "'doctitle_xform'", ":", "True", ",", "'initial_header_level'", ":", "3", ",", "\"default_reference_context\"", ":", "default_reference_context", ",", "\"link_base\"", ":", "reverse", "(", "'django-admindocs-docroot'", ")", ".", "rstrip", "(", "'/'", ")", ",", "'raw_enabled'", ":", "False", ",", "'file_insertion_enabled'", ":", "False", ",", "}", "thing_being_parsed", "=", "thing_being_parsed", "and", "'<%s>'", "%", "thing_being_parsed", "# Wrap ``text`` in some reST that sets the default role to ``cmsreference``,", "# then restores it.", "source", "=", "\"\"\"\n.. default-role:: cmsreference\n\n%s\n\n.. default-role::\n\"\"\"", "parts", "=", "docutils", ".", "core", ".", "publish_parts", "(", "source", "%", "text", ",", "source_path", "=", "thing_being_parsed", ",", "destination_path", "=", "None", ",", "writer_name", "=", "'html'", ",", "settings_overrides", "=", "overrides", ",", ")", "return", "mark_safe", "(", "parts", "[", "'fragment'", "]", ")" ]
[ 55, 0 ]
[ 82, 39 ]
python
en
['en', 'error', 'th']
False
replace_named_groups
(pattern)
r""" Find named groups in `pattern` and replace them with the group name. E.g., 1. ^(?P<a>\w+)/b/(\w+)$ ==> ^<a>/b/(\w+)$ 2. ^(?P<a>\w+)/b/(?P<c>\w+)/$ ==> ^<a>/b/<c>/$ 3. ^(?P<a>\w+)/b/(\w+) ==> ^<a>/b/(\w+) 4. ^(?P<a>\w+)/b/(?P<c>\w+) ==> ^<a>/b/<c>
r""" Find named groups in `pattern` and replace them with the group name. E.g., 1. ^(?P<a>\w+)/b/(\w+)$ ==> ^<a>/b/(\w+)$ 2. ^(?P<a>\w+)/b/(?P<c>\w+)/$ ==> ^<a>/b/<c>/$ 3. ^(?P<a>\w+)/b/(\w+) ==> ^<a>/b/(\w+) 4. ^(?P<a>\w+)/b/(?P<c>\w+) ==> ^<a>/b/<c>
def replace_named_groups(pattern): r""" Find named groups in `pattern` and replace them with the group name. E.g., 1. ^(?P<a>\w+)/b/(\w+)$ ==> ^<a>/b/(\w+)$ 2. ^(?P<a>\w+)/b/(?P<c>\w+)/$ ==> ^<a>/b/<c>/$ 3. ^(?P<a>\w+)/b/(\w+) ==> ^<a>/b/(\w+) 4. ^(?P<a>\w+)/b/(?P<c>\w+) ==> ^<a>/b/<c> """ named_group_indices = [ (m.start(0), m.end(0), m[1]) for m in named_group_matcher.finditer(pattern) ] # Tuples of (named capture group pattern, group name). group_pattern_and_name = [] # Loop over the groups and their start and end indices. for start, end, group_name in named_group_indices: # Handle nested parentheses, e.g. '^(?P<a>(x|y))/b'. unmatched_open_brackets, prev_char = 1, None for idx, val in enumerate(pattern[end:]): # Check for unescaped `(` and `)`. They mark the start and end of a # nested group. if val == '(' and prev_char != '\\': unmatched_open_brackets += 1 elif val == ')' and prev_char != '\\': unmatched_open_brackets -= 1 prev_char = val # If brackets are balanced, the end of the string for the current # named capture group pattern has been reached. if unmatched_open_brackets == 0: group_pattern_and_name.append((pattern[start:end + idx + 1], group_name)) break # Replace the string for named capture groups with their group names. for group_pattern, group_name in group_pattern_and_name: pattern = pattern.replace(group_pattern, group_name) return pattern
[ "def", "replace_named_groups", "(", "pattern", ")", ":", "named_group_indices", "=", "[", "(", "m", ".", "start", "(", "0", ")", ",", "m", ".", "end", "(", "0", ")", ",", "m", "[", "1", "]", ")", "for", "m", "in", "named_group_matcher", ".", "finditer", "(", "pattern", ")", "]", "# Tuples of (named capture group pattern, group name).", "group_pattern_and_name", "=", "[", "]", "# Loop over the groups and their start and end indices.", "for", "start", ",", "end", ",", "group_name", "in", "named_group_indices", ":", "# Handle nested parentheses, e.g. '^(?P<a>(x|y))/b'.", "unmatched_open_brackets", ",", "prev_char", "=", "1", ",", "None", "for", "idx", ",", "val", "in", "enumerate", "(", "pattern", "[", "end", ":", "]", ")", ":", "# Check for unescaped `(` and `)`. They mark the start and end of a", "# nested group.", "if", "val", "==", "'('", "and", "prev_char", "!=", "'\\\\'", ":", "unmatched_open_brackets", "+=", "1", "elif", "val", "==", "')'", "and", "prev_char", "!=", "'\\\\'", ":", "unmatched_open_brackets", "-=", "1", "prev_char", "=", "val", "# If brackets are balanced, the end of the string for the current", "# named capture group pattern has been reached.", "if", "unmatched_open_brackets", "==", "0", ":", "group_pattern_and_name", ".", "append", "(", "(", "pattern", "[", "start", ":", "end", "+", "idx", "+", "1", "]", ",", "group_name", ")", ")", "break", "# Replace the string for named capture groups with their group names.", "for", "group_pattern", ",", "group_name", "in", "group_pattern_and_name", ":", "pattern", "=", "pattern", ".", "replace", "(", "group_pattern", ",", "group_name", ")", "return", "pattern" ]
[ 141, 0 ]
[ 176, 18 ]
python
cy
['en', 'cy', 'hi']
False
replace_unnamed_groups
(pattern)
r""" Find unnamed groups in `pattern` and replace them with '<var>'. E.g., 1. ^(?P<a>\w+)/b/(\w+)$ ==> ^(?P<a>\w+)/b/<var>$ 2. ^(?P<a>\w+)/b/((x|y)\w+)$ ==> ^(?P<a>\w+)/b/<var>$ 3. ^(?P<a>\w+)/b/(\w+) ==> ^(?P<a>\w+)/b/<var> 4. ^(?P<a>\w+)/b/((x|y)\w+) ==> ^(?P<a>\w+)/b/<var>
r""" Find unnamed groups in `pattern` and replace them with '<var>'. E.g., 1. ^(?P<a>\w+)/b/(\w+)$ ==> ^(?P<a>\w+)/b/<var>$ 2. ^(?P<a>\w+)/b/((x|y)\w+)$ ==> ^(?P<a>\w+)/b/<var>$ 3. ^(?P<a>\w+)/b/(\w+) ==> ^(?P<a>\w+)/b/<var> 4. ^(?P<a>\w+)/b/((x|y)\w+) ==> ^(?P<a>\w+)/b/<var>
def replace_unnamed_groups(pattern): r""" Find unnamed groups in `pattern` and replace them with '<var>'. E.g., 1. ^(?P<a>\w+)/b/(\w+)$ ==> ^(?P<a>\w+)/b/<var>$ 2. ^(?P<a>\w+)/b/((x|y)\w+)$ ==> ^(?P<a>\w+)/b/<var>$ 3. ^(?P<a>\w+)/b/(\w+) ==> ^(?P<a>\w+)/b/<var> 4. ^(?P<a>\w+)/b/((x|y)\w+) ==> ^(?P<a>\w+)/b/<var> """ unnamed_group_indices = [m.start(0) for m in unnamed_group_matcher.finditer(pattern)] # Indices of the start of unnamed capture groups. group_indices = [] # Loop over the start indices of the groups. for start in unnamed_group_indices: # Handle nested parentheses, e.g. '^b/((x|y)\w+)$'. unmatched_open_brackets, prev_char = 1, None for idx, val in enumerate(pattern[start + 1:]): # Check for unescaped `(` and `)`. They mark the start and end of # a nested group. if val == '(' and prev_char != '\\': unmatched_open_brackets += 1 elif val == ')' and prev_char != '\\': unmatched_open_brackets -= 1 prev_char = val if unmatched_open_brackets == 0: group_indices.append((start, start + 2 + idx)) break # Remove unnamed group matches inside other unnamed capture groups. group_start_end_indices = [] prev_end = None for start, end in group_indices: if prev_end and start > prev_end or not prev_end: group_start_end_indices.append((start, end)) prev_end = end if group_start_end_indices: # Replace unnamed groups with <var>. Handle the fact that replacing the # string between indices will change string length and thus indices # will point to the wrong substring if not corrected. final_pattern, prev_end = [], None for start, end in group_start_end_indices: if prev_end: final_pattern.append(pattern[prev_end:start]) final_pattern.append(pattern[:start] + '<var>') prev_end = end final_pattern.append(pattern[prev_end:]) return ''.join(final_pattern) else: return pattern
[ "def", "replace_unnamed_groups", "(", "pattern", ")", ":", "unnamed_group_indices", "=", "[", "m", ".", "start", "(", "0", ")", "for", "m", "in", "unnamed_group_matcher", ".", "finditer", "(", "pattern", ")", "]", "# Indices of the start of unnamed capture groups.", "group_indices", "=", "[", "]", "# Loop over the start indices of the groups.", "for", "start", "in", "unnamed_group_indices", ":", "# Handle nested parentheses, e.g. '^b/((x|y)\\w+)$'.", "unmatched_open_brackets", ",", "prev_char", "=", "1", ",", "None", "for", "idx", ",", "val", "in", "enumerate", "(", "pattern", "[", "start", "+", "1", ":", "]", ")", ":", "# Check for unescaped `(` and `)`. They mark the start and end of", "# a nested group.", "if", "val", "==", "'('", "and", "prev_char", "!=", "'\\\\'", ":", "unmatched_open_brackets", "+=", "1", "elif", "val", "==", "')'", "and", "prev_char", "!=", "'\\\\'", ":", "unmatched_open_brackets", "-=", "1", "prev_char", "=", "val", "if", "unmatched_open_brackets", "==", "0", ":", "group_indices", ".", "append", "(", "(", "start", ",", "start", "+", "2", "+", "idx", ")", ")", "break", "# Remove unnamed group matches inside other unnamed capture groups.", "group_start_end_indices", "=", "[", "]", "prev_end", "=", "None", "for", "start", ",", "end", "in", "group_indices", ":", "if", "prev_end", "and", "start", ">", "prev_end", "or", "not", "prev_end", ":", "group_start_end_indices", ".", "append", "(", "(", "start", ",", "end", ")", ")", "prev_end", "=", "end", "if", "group_start_end_indices", ":", "# Replace unnamed groups with <var>. Handle the fact that replacing the", "# string between indices will change string length and thus indices", "# will point to the wrong substring if not corrected.", "final_pattern", ",", "prev_end", "=", "[", "]", ",", "None", "for", "start", ",", "end", "in", "group_start_end_indices", ":", "if", "prev_end", ":", "final_pattern", ".", "append", "(", "pattern", "[", "prev_end", ":", "start", "]", ")", "final_pattern", ".", "append", "(", "pattern", "[", ":", "start", "]", "+", "'<var>'", ")", "prev_end", "=", "end", "final_pattern", ".", "append", "(", "pattern", "[", "prev_end", ":", "]", ")", "return", "''", ".", "join", "(", "final_pattern", ")", "else", ":", "return", "pattern" ]
[ 179, 0 ]
[ 227, 22 ]
python
cy
['en', 'cy', 'hi']
False
ReindentFilter._flatten_up_to_token
(self, token)
Yields all tokens up to token but excluding current.
Yields all tokens up to token but excluding current.
def _flatten_up_to_token(self, token): """Yields all tokens up to token but excluding current.""" if token.is_group: token = next(token.flatten()) for t in self._curr_stmt.flatten(): if t == token: break yield t
[ "def", "_flatten_up_to_token", "(", "self", ",", "token", ")", ":", "if", "token", ".", "is_group", ":", "token", "=", "next", "(", "token", ".", "flatten", "(", ")", ")", "for", "t", "in", "self", ".", "_curr_stmt", ".", "flatten", "(", ")", ":", "if", "t", "==", "token", ":", "break", "yield", "t" ]
[ 27, 4 ]
[ 35, 19 ]
python
en
['en', 'en', 'en']
True
BaseDistribution.location
(self)
Where the distribution is loaded from. A string value is not necessarily a filesystem path, since distributions can be loaded from other sources, e.g. arbitrary zip archives. ``None`` means the distribution is created in-memory. Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If this is a symbolic link, we want to preserve the relative path between it and files in the distribution.
Where the distribution is loaded from.
def location(self) -> Optional[str]: """Where the distribution is loaded from. A string value is not necessarily a filesystem path, since distributions can be loaded from other sources, e.g. arbitrary zip archives. ``None`` means the distribution is created in-memory. Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If this is a symbolic link, we want to preserve the relative path between it and files in the distribution. """ raise NotImplementedError()
[ "def", "location", "(", "self", ")", "->", "Optional", "[", "str", "]", ":", "raise", "NotImplementedError", "(", ")" ]
[ 53, 4 ]
[ 64, 35 ]
python
en
['en', 'en', 'en']
True
BaseDistribution.info_directory
(self)
Location of the .[egg|dist]-info directory. Similarly to ``location``, a string value is not necessarily a filesystem path. ``None`` means the distribution is created in-memory. For a modern .dist-info installation on disk, this should be something like ``{location}/{raw_name}-{version}.dist-info``. Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If this is a symbolic link, we want to preserve the relative path between it and other files in the distribution.
Location of the .[egg|dist]-info directory.
def info_directory(self) -> Optional[str]: """Location of the .[egg|dist]-info directory. Similarly to ``location``, a string value is not necessarily a filesystem path. ``None`` means the distribution is created in-memory. For a modern .dist-info installation on disk, this should be something like ``{location}/{raw_name}-{version}.dist-info``. Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If this is a symbolic link, we want to preserve the relative path between it and other files in the distribution. """ raise NotImplementedError()
[ "def", "info_directory", "(", "self", ")", "->", "Optional", "[", "str", "]", ":", "raise", "NotImplementedError", "(", ")" ]
[ 67, 4 ]
[ 80, 35 ]
python
en
['en', 'en', 'en']
True
BaseDistribution.direct_url
(self)
Obtain a DirectUrl from this distribution. Returns None if the distribution has no `direct_url.json` metadata, or if `direct_url.json` is invalid.
Obtain a DirectUrl from this distribution.
def direct_url(self) -> Optional[DirectUrl]: """Obtain a DirectUrl from this distribution. Returns None if the distribution has no `direct_url.json` metadata, or if `direct_url.json` is invalid. """ try: content = self.read_text(DIRECT_URL_METADATA_NAME) except FileNotFoundError: return None try: return DirectUrl.from_json(content) except ( UnicodeDecodeError, json.JSONDecodeError, DirectUrlValidationError, ) as e: logger.warning( "Error parsing %s for %s: %s", DIRECT_URL_METADATA_NAME, self.canonical_name, e, ) return None
[ "def", "direct_url", "(", "self", ")", "->", "Optional", "[", "DirectUrl", "]", ":", "try", ":", "content", "=", "self", ".", "read_text", "(", "DIRECT_URL_METADATA_NAME", ")", "except", "FileNotFoundError", ":", "return", "None", "try", ":", "return", "DirectUrl", ".", "from_json", "(", "content", ")", "except", "(", "UnicodeDecodeError", ",", "json", ".", "JSONDecodeError", ",", "DirectUrlValidationError", ",", ")", "as", "e", ":", "logger", ".", "warning", "(", "\"Error parsing %s for %s: %s\"", ",", "DIRECT_URL_METADATA_NAME", ",", "self", ".", "canonical_name", ",", "e", ",", ")", "return", "None" ]
[ 91, 4 ]
[ 114, 23 ]
python
en
['en', 'en', 'en']
True
BaseDistribution.read_text
(self, name: str)
Read a file in the .dist-info (or .egg-info) directory. Should raise ``FileNotFoundError`` if ``name`` does not exist in the metadata directory.
Read a file in the .dist-info (or .egg-info) directory.
def read_text(self, name: str) -> str: """Read a file in the .dist-info (or .egg-info) directory. Should raise ``FileNotFoundError`` if ``name`` does not exist in the metadata directory. """ raise NotImplementedError()
[ "def", "read_text", "(", "self", ",", "name", ":", "str", ")", "->", "str", ":", "raise", "NotImplementedError", "(", ")" ]
[ 136, 4 ]
[ 142, 35 ]
python
en
['en', 'en', 'en']
True
BaseDistribution.metadata
(self)
Metadata of distribution parsed from e.g. METADATA or PKG-INFO.
Metadata of distribution parsed from e.g. METADATA or PKG-INFO.
def metadata(self) -> email.message.Message: """Metadata of distribution parsed from e.g. METADATA or PKG-INFO.""" raise NotImplementedError()
[ "def", "metadata", "(", "self", ")", "->", "email", ".", "message", ".", "Message", ":", "raise", "NotImplementedError", "(", ")" ]
[ 148, 4 ]
[ 150, 35 ]
python
en
['en', 'en', 'en']
True
BaseDistribution.metadata_version
(self)
Value of "Metadata-Version:" in distribution metadata, if available.
Value of "Metadata-Version:" in distribution metadata, if available.
def metadata_version(self) -> Optional[str]: """Value of "Metadata-Version:" in distribution metadata, if available.""" return self.metadata.get("Metadata-Version")
[ "def", "metadata_version", "(", "self", ")", "->", "Optional", "[", "str", "]", ":", "return", "self", ".", "metadata", ".", "get", "(", "\"Metadata-Version\"", ")" ]
[ 153, 4 ]
[ 155, 52 ]
python
en
['en', 'en', 'en']
True
BaseDistribution.raw_name
(self)
Value of "Name:" in distribution metadata.
Value of "Name:" in distribution metadata.
def raw_name(self) -> str: """Value of "Name:" in distribution metadata.""" # The metadata should NEVER be missing the Name: key, but if it somehow # does not, fall back to the known canonical name. return self.metadata.get("Name", self.canonical_name)
[ "def", "raw_name", "(", "self", ")", "->", "str", ":", "# The metadata should NEVER be missing the Name: key, but if it somehow", "# does not, fall back to the known canonical name.", "return", "self", ".", "metadata", ".", "get", "(", "\"Name\"", ",", "self", ".", "canonical_name", ")" ]
[ 158, 4 ]
[ 162, 61 ]
python
en
['en', 'en', 'en']
True
BaseEnvironment.get_distribution
(self, name: str)
Given a requirement name, return the installed distributions.
Given a requirement name, return the installed distributions.
def get_distribution(self, name: str) -> Optional["BaseDistribution"]: """Given a requirement name, return the installed distributions.""" raise NotImplementedError()
[ "def", "get_distribution", "(", "self", ",", "name", ":", "str", ")", "->", "Optional", "[", "\"BaseDistribution\"", "]", ":", "raise", "NotImplementedError", "(", ")" ]
[ 179, 4 ]
[ 181, 35 ]
python
en
['en', 'en', 'en']
True
BaseEnvironment._iter_distributions
(self)
Iterate through installed distributions. This function should be implemented by subclass, but never called directly. Use the public ``iter_distribution()`` instead, which implements additional logic to make sure the distributions are valid.
Iterate through installed distributions.
def _iter_distributions(self) -> Iterator["BaseDistribution"]: """Iterate through installed distributions. This function should be implemented by subclass, but never called directly. Use the public ``iter_distribution()`` instead, which implements additional logic to make sure the distributions are valid. """ raise NotImplementedError()
[ "def", "_iter_distributions", "(", "self", ")", "->", "Iterator", "[", "\"BaseDistribution\"", "]", ":", "raise", "NotImplementedError", "(", ")" ]
[ 183, 4 ]
[ 190, 35 ]
python
en
['en', 'en', 'en']
True
BaseEnvironment.iter_distributions
(self)
Iterate through installed distributions.
Iterate through installed distributions.
def iter_distributions(self) -> Iterator["BaseDistribution"]: """Iterate through installed distributions.""" for dist in self._iter_distributions(): # Make sure the distribution actually comes from a valid Python # packaging distribution. Pip's AdjacentTempDirectory leaves folders # e.g. ``~atplotlib.dist-info`` if cleanup was interrupted. The # valid project name pattern is taken from PEP 508. project_name_valid = re.match( r"^([A-Z0-9]|[A-Z0-9][A-Z0-9._-]*[A-Z0-9])$", dist.canonical_name, flags=re.IGNORECASE, ) if not project_name_valid: logger.warning( "Ignoring invalid distribution %s (%s)", dist.canonical_name, dist.location, ) continue yield dist
[ "def", "iter_distributions", "(", "self", ")", "->", "Iterator", "[", "\"BaseDistribution\"", "]", ":", "for", "dist", "in", "self", ".", "_iter_distributions", "(", ")", ":", "# Make sure the distribution actually comes from a valid Python", "# packaging distribution. Pip's AdjacentTempDirectory leaves folders", "# e.g. ``~atplotlib.dist-info`` if cleanup was interrupted. The", "# valid project name pattern is taken from PEP 508.", "project_name_valid", "=", "re", ".", "match", "(", "r\"^([A-Z0-9]|[A-Z0-9][A-Z0-9._-]*[A-Z0-9])$\"", ",", "dist", ".", "canonical_name", ",", "flags", "=", "re", ".", "IGNORECASE", ",", ")", "if", "not", "project_name_valid", ":", "logger", ".", "warning", "(", "\"Ignoring invalid distribution %s (%s)\"", ",", "dist", ".", "canonical_name", ",", "dist", ".", "location", ",", ")", "continue", "yield", "dist" ]
[ 192, 4 ]
[ 211, 22 ]
python
en
['en', 'en', 'en']
True
BaseEnvironment.iter_installed_distributions
( self, local_only: bool = True, skip: Container[str] = stdlib_pkgs, include_editables: bool = True, editables_only: bool = False, user_only: bool = False, )
Return a list of installed distributions. :param local_only: If True (default), only return installations local to the current virtualenv, if in a virtualenv. :param skip: An iterable of canonicalized project names to ignore; defaults to ``stdlib_pkgs``. :param include_editables: If False, don't report editables. :param editables_only: If True, only report editables. :param user_only: If True, only report installations in the user site directory.
Return a list of installed distributions.
def iter_installed_distributions( self, local_only: bool = True, skip: Container[str] = stdlib_pkgs, include_editables: bool = True, editables_only: bool = False, user_only: bool = False, ) -> Iterator[BaseDistribution]: """Return a list of installed distributions. :param local_only: If True (default), only return installations local to the current virtualenv, if in a virtualenv. :param skip: An iterable of canonicalized project names to ignore; defaults to ``stdlib_pkgs``. :param include_editables: If False, don't report editables. :param editables_only: If True, only report editables. :param user_only: If True, only report installations in the user site directory. """ it = self.iter_distributions() if local_only: it = (d for d in it if d.local) if not include_editables: it = (d for d in it if not d.editable) if editables_only: it = (d for d in it if d.editable) if user_only: it = (d for d in it if d.in_usersite) return (d for d in it if d.canonical_name not in skip)
[ "def", "iter_installed_distributions", "(", "self", ",", "local_only", ":", "bool", "=", "True", ",", "skip", ":", "Container", "[", "str", "]", "=", "stdlib_pkgs", ",", "include_editables", ":", "bool", "=", "True", ",", "editables_only", ":", "bool", "=", "False", ",", "user_only", ":", "bool", "=", "False", ",", ")", "->", "Iterator", "[", "BaseDistribution", "]", ":", "it", "=", "self", ".", "iter_distributions", "(", ")", "if", "local_only", ":", "it", "=", "(", "d", "for", "d", "in", "it", "if", "d", ".", "local", ")", "if", "not", "include_editables", ":", "it", "=", "(", "d", "for", "d", "in", "it", "if", "not", "d", ".", "editable", ")", "if", "editables_only", ":", "it", "=", "(", "d", "for", "d", "in", "it", "if", "d", ".", "editable", ")", "if", "user_only", ":", "it", "=", "(", "d", "for", "d", "in", "it", "if", "d", ".", "in_usersite", ")", "return", "(", "d", "for", "d", "in", "it", "if", "d", ".", "canonical_name", "not", "in", "skip", ")" ]
[ 213, 4 ]
[ 241, 62 ]
python
en
['en', 'en', 'en']
True
unicode_to_ascii
(s)
Transforms an ascii string into unicode.
Transforms an ascii string into unicode.
def unicode_to_ascii(s): """Transforms an ascii string into unicode.""" normalized = unicodedata.normalize('NFD', s) return ''.join(c for c in normalized if unicodedata.category(c) != 'Mn')
[ "def", "unicode_to_ascii", "(", "s", ")", ":", "normalized", "=", "unicodedata", ".", "normalize", "(", "'NFD'", ",", "s", ")", "return", "''", ".", "join", "(", "c", "for", "c", "in", "normalized", "if", "unicodedata", ".", "category", "(", "c", ")", "!=", "'Mn'", ")" ]
[ 7, 0 ]
[ 10, 76 ]
python
en
['en', 'pt', 'en']
True
preprocess_sentence
(w)
Lowers, strips, and adds <start> and <end> tags to a sentence.
Lowers, strips, and adds <start> and <end> tags to a sentence.
def preprocess_sentence(w): """Lowers, strips, and adds <start> and <end> tags to a sentence. """ w = unicode_to_ascii(w.lower().strip()) # creating a space between a word and the punctuation following it # eg: "he is a boy." => "he is a boy ." w = re.sub(r"([?.!,¿])", r" \1 ", w) w = re.sub(r'[" "]+', " ", w) # replacing everything with space except (a-z, A-Z, ".", "?", "!", ",") w = re.sub(r"[^a-zA-Z?.!,¿]+", " ", w) w = w.rstrip().strip() # adding a start and an end token to the sentence # so that the model know when to start and stop predicting. w = '<start> ' + w + ' <end>' return w
[ "def", "preprocess_sentence", "(", "w", ")", ":", "w", "=", "unicode_to_ascii", "(", "w", ".", "lower", "(", ")", ".", "strip", "(", ")", ")", "# creating a space between a word and the punctuation following it", "# eg: \"he is a boy.\" => \"he is a boy .\"", "w", "=", "re", ".", "sub", "(", "r\"([?.!,¿])\",", " ", "\" \\1 \",", " ", ")", "", "w", "=", "re", ".", "sub", "(", "r'[\" \"]+'", ",", "\" \"", ",", "w", ")", "# replacing everything with space except (a-z, A-Z, \".\", \"?\", \"!\", \",\")", "w", "=", "re", ".", "sub", "(", "r\"[^a-zA-Z?.!,¿]+\",", " ", " \",", " ", ")", "", "w", "=", "w", ".", "rstrip", "(", ")", ".", "strip", "(", ")", "# adding a start and an end token to the sentence", "# so that the model know when to start and stop predicting.", "w", "=", "'<start> '", "+", "w", "+", "' <end>'", "return", "w" ]
[ 13, 0 ]
[ 32, 12 ]
python
en
['en', 'en', 'en']
True
tokenize
(lang, lang_tokenizer=None)
Given a list of sentences, return an integer representation Arguments: lang -- a python list of sentences lang_tokenizer -- keras_preprocessing.text.Tokenizer, if None this will be created for you Returns: tensor -- int tensor of shape (NUM_EXAMPLES,MAX_SENTENCE_LENGTH) lang_tokenizer -- keras_preprocessing.text.Tokenizer
Given a list of sentences, return an integer representation
def tokenize(lang, lang_tokenizer=None): """Given a list of sentences, return an integer representation Arguments: lang -- a python list of sentences lang_tokenizer -- keras_preprocessing.text.Tokenizer, if None this will be created for you Returns: tensor -- int tensor of shape (NUM_EXAMPLES,MAX_SENTENCE_LENGTH) lang_tokenizer -- keras_preprocessing.text.Tokenizer """ if lang_tokenizer is None: lang_tokenizer = tf.keras.preprocessing.text.Tokenizer( filters='') lang_tokenizer.fit_on_texts(lang) tensor = lang_tokenizer.texts_to_sequences(lang) tensor = tf.keras.preprocessing.sequence.pad_sequences( tensor, padding='post') return tensor, lang_tokenizer
[ "def", "tokenize", "(", "lang", ",", "lang_tokenizer", "=", "None", ")", ":", "if", "lang_tokenizer", "is", "None", ":", "lang_tokenizer", "=", "tf", ".", "keras", ".", "preprocessing", ".", "text", ".", "Tokenizer", "(", "filters", "=", "''", ")", "lang_tokenizer", ".", "fit_on_texts", "(", "lang", ")", "tensor", "=", "lang_tokenizer", ".", "texts_to_sequences", "(", "lang", ")", "tensor", "=", "tf", ".", "keras", ".", "preprocessing", ".", "sequence", ".", "pad_sequences", "(", "tensor", ",", "padding", "=", "'post'", ")", "return", "tensor", ",", "lang_tokenizer" ]
[ 35, 0 ]
[ 57, 33 ]
python
en
['en', 'lb', 'en']
True
preprocess
(sentences, tokenizer)
Preprocesses then tokenizes text Arguments: sentences -- a python list of of strings tokenizer -- Tokenizer for mapping words to integers Returns: tensor -- int tensor of shape (NUM_EXAMPLES,MAX_SENTENCE_LENGTH) lang_tokenizer -- keras_preprocessing.text.Tokenizer
Preprocesses then tokenizes text
def preprocess(sentences, tokenizer): """Preprocesses then tokenizes text Arguments: sentences -- a python list of of strings tokenizer -- Tokenizer for mapping words to integers Returns: tensor -- int tensor of shape (NUM_EXAMPLES,MAX_SENTENCE_LENGTH) lang_tokenizer -- keras_preprocessing.text.Tokenizer """ sentences = [preprocess_sentence(sentence) for sentence in sentences] tokens, _ = tokenize(sentences, tokenizer) return tokens
[ "def", "preprocess", "(", "sentences", ",", "tokenizer", ")", ":", "sentences", "=", "[", "preprocess_sentence", "(", "sentence", ")", "for", "sentence", "in", "sentences", "]", "tokens", ",", "_", "=", "tokenize", "(", "sentences", ",", "tokenizer", ")", "return", "tokens" ]
[ 60, 0 ]
[ 73, 17 ]
python
en
['en', 'el-Latn', 'en']
True
int2word
(tokenizer, int_sequence)
Converts integer representation to natural language representation Arguments: tokenizer -- keras_preprocessing.text.Tokenizer int_sequence -- an iterable or rank 1 tensor of integers Returns list of string tokens
Converts integer representation to natural language representation
def int2word(tokenizer, int_sequence): """Converts integer representation to natural language representation Arguments: tokenizer -- keras_preprocessing.text.Tokenizer int_sequence -- an iterable or rank 1 tensor of integers Returns list of string tokens """ return [tokenizer.index_word[t] if t != 0 else '' for t in int_sequence]
[ "def", "int2word", "(", "tokenizer", ",", "int_sequence", ")", ":", "return", "[", "tokenizer", ".", "index_word", "[", "t", "]", "if", "t", "!=", "0", "else", "''", "for", "t", "in", "int_sequence", "]" ]
[ 76, 0 ]
[ 85, 76 ]
python
en
['en', 'fil', 'en']
True
Schemas.__init__
(self, discovery)
Constructor. Args: discovery: object, Deserialized discovery document from which we pull out the named schema.
Constructor.
def __init__(self, discovery): """Constructor. Args: discovery: object, Deserialized discovery document from which we pull out the named schema. """ self.schemas = discovery.get('schemas', {}) # Cache of pretty printed schemas. self.pretty = {}
[ "def", "__init__", "(", "self", ",", "discovery", ")", ":", "self", ".", "schemas", "=", "discovery", ".", "get", "(", "'schemas'", ",", "{", "}", ")", "# Cache of pretty printed schemas.", "self", ".", "pretty", "=", "{", "}" ]
[ 78, 2 ]
[ 88, 20 ]
python
en
['en', 'en', 'en']
False
Schemas._prettyPrintByName
(self, name, seen=None, dent=0)
Get pretty printed object prototype from the schema name. Args: name: string, Name of schema in the discovery document. seen: list of string, Names of schema already seen. Used to handle recursive definitions. Returns: string, A string that contains a prototype object with comments that conforms to the given schema.
Get pretty printed object prototype from the schema name.
def _prettyPrintByName(self, name, seen=None, dent=0): """Get pretty printed object prototype from the schema name. Args: name: string, Name of schema in the discovery document. seen: list of string, Names of schema already seen. Used to handle recursive definitions. Returns: string, A string that contains a prototype object with comments that conforms to the given schema. """ if seen is None: seen = [] if name in seen: # Do not fall into an infinite loop over recursive definitions. return '# Object with schema name: %s' % name seen.append(name) if name not in self.pretty: self.pretty[name] = _SchemaToStruct(self.schemas[name], seen, dent=dent).to_str(self._prettyPrintByName) seen.pop() return self.pretty[name]
[ "def", "_prettyPrintByName", "(", "self", ",", "name", ",", "seen", "=", "None", ",", "dent", "=", "0", ")", ":", "if", "seen", "is", "None", ":", "seen", "=", "[", "]", "if", "name", "in", "seen", ":", "# Do not fall into an infinite loop over recursive definitions.", "return", "'# Object with schema name: %s'", "%", "name", "seen", ".", "append", "(", "name", ")", "if", "name", "not", "in", "self", ".", "pretty", ":", "self", ".", "pretty", "[", "name", "]", "=", "_SchemaToStruct", "(", "self", ".", "schemas", "[", "name", "]", ",", "seen", ",", "dent", "=", "dent", ")", ".", "to_str", "(", "self", ".", "_prettyPrintByName", ")", "seen", ".", "pop", "(", ")", "return", "self", ".", "pretty", "[", "name", "]" ]
[ 91, 2 ]
[ 117, 28 ]
python
en
['en', 'en', 'en']
True
Schemas.prettyPrintByName
(self, name)
Get pretty printed object prototype from the schema name. Args: name: string, Name of schema in the discovery document. Returns: string, A string that contains a prototype object with comments that conforms to the given schema.
Get pretty printed object prototype from the schema name.
def prettyPrintByName(self, name): """Get pretty printed object prototype from the schema name. Args: name: string, Name of schema in the discovery document. Returns: string, A string that contains a prototype object with comments that conforms to the given schema. """ # Return with trailing comma and newline removed. return self._prettyPrintByName(name, seen=[], dent=1)[:-2]
[ "def", "prettyPrintByName", "(", "self", ",", "name", ")", ":", "# Return with trailing comma and newline removed.", "return", "self", ".", "_prettyPrintByName", "(", "name", ",", "seen", "=", "[", "]", ",", "dent", "=", "1", ")", "[", ":", "-", "2", "]" ]
[ 119, 2 ]
[ 130, 62 ]
python
en
['en', 'en', 'en']
True
Schemas._prettyPrintSchema
(self, schema, seen=None, dent=0)
Get pretty printed object prototype of schema. Args: schema: object, Parsed JSON schema. seen: list of string, Names of schema already seen. Used to handle recursive definitions. Returns: string, A string that contains a prototype object with comments that conforms to the given schema.
Get pretty printed object prototype of schema.
def _prettyPrintSchema(self, schema, seen=None, dent=0): """Get pretty printed object prototype of schema. Args: schema: object, Parsed JSON schema. seen: list of string, Names of schema already seen. Used to handle recursive definitions. Returns: string, A string that contains a prototype object with comments that conforms to the given schema. """ if seen is None: seen = [] return _SchemaToStruct(schema, seen, dent=dent).to_str(self._prettyPrintByName)
[ "def", "_prettyPrintSchema", "(", "self", ",", "schema", ",", "seen", "=", "None", ",", "dent", "=", "0", ")", ":", "if", "seen", "is", "None", ":", "seen", "=", "[", "]", "return", "_SchemaToStruct", "(", "schema", ",", "seen", ",", "dent", "=", "dent", ")", ".", "to_str", "(", "self", ".", "_prettyPrintByName", ")" ]
[ 133, 2 ]
[ 148, 83 ]
python
en
['en', 'en', 'en']
True
Schemas.prettyPrintSchema
(self, schema)
Get pretty printed object prototype of schema. Args: schema: object, Parsed JSON schema. Returns: string, A string that contains a prototype object with comments that conforms to the given schema.
Get pretty printed object prototype of schema.
def prettyPrintSchema(self, schema): """Get pretty printed object prototype of schema. Args: schema: object, Parsed JSON schema. Returns: string, A string that contains a prototype object with comments that conforms to the given schema. """ # Return with trailing comma and newline removed. return self._prettyPrintSchema(schema, dent=1)[:-2]
[ "def", "prettyPrintSchema", "(", "self", ",", "schema", ")", ":", "# Return with trailing comma and newline removed.", "return", "self", ".", "_prettyPrintSchema", "(", "schema", ",", "dent", "=", "1", ")", "[", ":", "-", "2", "]" ]
[ 150, 2 ]
[ 161, 55 ]
python
en
['en', 'en', 'en']
True
Schemas.get
(self, name)
Get deserialized JSON schema from the schema name. Args: name: string, Schema name.
Get deserialized JSON schema from the schema name.
def get(self, name): """Get deserialized JSON schema from the schema name. Args: name: string, Schema name. """ return self.schemas[name]
[ "def", "get", "(", "self", ",", "name", ")", ":", "return", "self", ".", "schemas", "[", "name", "]" ]
[ 163, 2 ]
[ 169, 29 ]
python
en
['en', 'en', 'en']
True
_SchemaToStruct.__init__
(self, schema, seen, dent=0)
Constructor. Args: schema: object, Parsed JSON schema. seen: list, List of names of schema already seen while parsing. Used to handle recursive definitions. dent: int, Initial indentation depth.
Constructor.
def __init__(self, schema, seen, dent=0): """Constructor. Args: schema: object, Parsed JSON schema. seen: list, List of names of schema already seen while parsing. Used to handle recursive definitions. dent: int, Initial indentation depth. """ # The result of this parsing kept as list of strings. self.value = [] # The final value of the parsing. self.string = None # The parsed JSON schema. self.schema = schema # Indentation level. self.dent = dent # Method that when called returns a prototype object for the schema with # the given name. self.from_cache = None # List of names of schema already seen while parsing. self.seen = seen
[ "def", "__init__", "(", "self", ",", "schema", ",", "seen", ",", "dent", "=", "0", ")", ":", "# The result of this parsing kept as list of strings.", "self", ".", "value", "=", "[", "]", "# The final value of the parsing.", "self", ".", "string", "=", "None", "# The parsed JSON schema.", "self", ".", "schema", "=", "schema", "# Indentation level.", "self", ".", "dent", "=", "dent", "# Method that when called returns a prototype object for the schema with", "# the given name.", "self", ".", "from_cache", "=", "None", "# List of names of schema already seen while parsing.", "self", ".", "seen", "=", "seen" ]
[ 176, 2 ]
[ 202, 20 ]
python
en
['en', 'en', 'en']
False
_SchemaToStruct.emit
(self, text)
Add text as a line to the output. Args: text: string, Text to output.
Add text as a line to the output.
def emit(self, text): """Add text as a line to the output. Args: text: string, Text to output. """ self.value.extend([" " * self.dent, text, '\n'])
[ "def", "emit", "(", "self", ",", "text", ")", ":", "self", ".", "value", ".", "extend", "(", "[", "\" \"", "*", "self", ".", "dent", ",", "text", ",", "'\\n'", "]", ")" ]
[ 204, 2 ]
[ 210, 53 ]
python
en
['en', 'en', 'en']
True
_SchemaToStruct.emitBegin
(self, text)
Add text to the output, but with no line terminator. Args: text: string, Text to output.
Add text to the output, but with no line terminator.
def emitBegin(self, text): """Add text to the output, but with no line terminator. Args: text: string, Text to output. """ self.value.extend([" " * self.dent, text])
[ "def", "emitBegin", "(", "self", ",", "text", ")", ":", "self", ".", "value", ".", "extend", "(", "[", "\" \"", "*", "self", ".", "dent", ",", "text", "]", ")" ]
[ 212, 2 ]
[ 218, 47 ]
python
en
['en', 'en', 'en']
True
_SchemaToStruct.emitEnd
(self, text, comment)
Add text and comment to the output with line terminator. Args: text: string, Text to output. comment: string, Python comment.
Add text and comment to the output with line terminator.
def emitEnd(self, text, comment): """Add text and comment to the output with line terminator. Args: text: string, Text to output. comment: string, Python comment. """ if comment: divider = '\n' + ' ' * (self.dent + 2) + '# ' lines = comment.splitlines() lines = [x.rstrip() for x in lines] comment = divider.join(lines) self.value.extend([text, ' # ', comment, '\n']) else: self.value.extend([text, '\n'])
[ "def", "emitEnd", "(", "self", ",", "text", ",", "comment", ")", ":", "if", "comment", ":", "divider", "=", "'\\n'", "+", "' '", "*", "(", "self", ".", "dent", "+", "2", ")", "+", "'# '", "lines", "=", "comment", ".", "splitlines", "(", ")", "lines", "=", "[", "x", ".", "rstrip", "(", ")", "for", "x", "in", "lines", "]", "comment", "=", "divider", ".", "join", "(", "lines", ")", "self", ".", "value", ".", "extend", "(", "[", "text", ",", "' # '", ",", "comment", ",", "'\\n'", "]", ")", "else", ":", "self", ".", "value", ".", "extend", "(", "[", "text", ",", "'\\n'", "]", ")" ]
[ 220, 2 ]
[ 234, 37 ]
python
en
['en', 'en', 'en']
True
_SchemaToStruct.indent
(self)
Increase indentation level.
Increase indentation level.
def indent(self): """Increase indentation level.""" self.dent += 1
[ "def", "indent", "(", "self", ")", ":", "self", ".", "dent", "+=", "1" ]
[ 236, 2 ]
[ 238, 18 ]
python
en
['en', 'zu', 'en']
True
_SchemaToStruct.undent
(self)
Decrease indentation level.
Decrease indentation level.
def undent(self): """Decrease indentation level.""" self.dent -= 1
[ "def", "undent", "(", "self", ")", ":", "self", ".", "dent", "-=", "1" ]
[ 240, 2 ]
[ 242, 18 ]
python
en
['en', 'zu', 'en']
True
_SchemaToStruct._to_str_impl
(self, schema)
Prototype object based on the schema, in Python code with comments. Args: schema: object, Parsed JSON schema file. Returns: Prototype object based on the schema, in Python code with comments.
Prototype object based on the schema, in Python code with comments.
def _to_str_impl(self, schema): """Prototype object based on the schema, in Python code with comments. Args: schema: object, Parsed JSON schema file. Returns: Prototype object based on the schema, in Python code with comments. """ stype = schema.get('type') if stype == 'object': self.emitEnd('{', schema.get('description', '')) self.indent() if 'properties' in schema: for pname, pschema in six.iteritems(schema.get('properties', {})): self.emitBegin('"%s": ' % pname) self._to_str_impl(pschema) elif 'additionalProperties' in schema: self.emitBegin('"a_key": ') self._to_str_impl(schema['additionalProperties']) self.undent() self.emit('},') elif '$ref' in schema: schemaName = schema['$ref'] description = schema.get('description', '') s = self.from_cache(schemaName, seen=self.seen) parts = s.splitlines() self.emitEnd(parts[0], description) for line in parts[1:]: self.emit(line.rstrip()) elif stype == 'boolean': value = schema.get('default', 'True or False') self.emitEnd('%s,' % str(value), schema.get('description', '')) elif stype == 'string': value = schema.get('default', 'A String') self.emitEnd('"%s",' % str(value), schema.get('description', '')) elif stype == 'integer': value = schema.get('default', '42') self.emitEnd('%s,' % str(value), schema.get('description', '')) elif stype == 'number': value = schema.get('default', '3.14') self.emitEnd('%s,' % str(value), schema.get('description', '')) elif stype == 'null': self.emitEnd('None,', schema.get('description', '')) elif stype == 'any': self.emitEnd('"",', schema.get('description', '')) elif stype == 'array': self.emitEnd('[', schema.get('description')) self.indent() self.emitBegin('') self._to_str_impl(schema['items']) self.undent() self.emit('],') else: self.emit('Unknown type! %s' % stype) self.emitEnd('', '') self.string = ''.join(self.value) return self.string
[ "def", "_to_str_impl", "(", "self", ",", "schema", ")", ":", "stype", "=", "schema", ".", "get", "(", "'type'", ")", "if", "stype", "==", "'object'", ":", "self", ".", "emitEnd", "(", "'{'", ",", "schema", ".", "get", "(", "'description'", ",", "''", ")", ")", "self", ".", "indent", "(", ")", "if", "'properties'", "in", "schema", ":", "for", "pname", ",", "pschema", "in", "six", ".", "iteritems", "(", "schema", ".", "get", "(", "'properties'", ",", "{", "}", ")", ")", ":", "self", ".", "emitBegin", "(", "'\"%s\": '", "%", "pname", ")", "self", ".", "_to_str_impl", "(", "pschema", ")", "elif", "'additionalProperties'", "in", "schema", ":", "self", ".", "emitBegin", "(", "'\"a_key\": '", ")", "self", ".", "_to_str_impl", "(", "schema", "[", "'additionalProperties'", "]", ")", "self", ".", "undent", "(", ")", "self", ".", "emit", "(", "'},'", ")", "elif", "'$ref'", "in", "schema", ":", "schemaName", "=", "schema", "[", "'$ref'", "]", "description", "=", "schema", ".", "get", "(", "'description'", ",", "''", ")", "s", "=", "self", ".", "from_cache", "(", "schemaName", ",", "seen", "=", "self", ".", "seen", ")", "parts", "=", "s", ".", "splitlines", "(", ")", "self", ".", "emitEnd", "(", "parts", "[", "0", "]", ",", "description", ")", "for", "line", "in", "parts", "[", "1", ":", "]", ":", "self", ".", "emit", "(", "line", ".", "rstrip", "(", ")", ")", "elif", "stype", "==", "'boolean'", ":", "value", "=", "schema", ".", "get", "(", "'default'", ",", "'True or False'", ")", "self", ".", "emitEnd", "(", "'%s,'", "%", "str", "(", "value", ")", ",", "schema", ".", "get", "(", "'description'", ",", "''", ")", ")", "elif", "stype", "==", "'string'", ":", "value", "=", "schema", ".", "get", "(", "'default'", ",", "'A String'", ")", "self", ".", "emitEnd", "(", "'\"%s\",'", "%", "str", "(", "value", ")", ",", "schema", ".", "get", "(", "'description'", ",", "''", ")", ")", "elif", "stype", "==", "'integer'", ":", "value", "=", "schema", ".", "get", "(", "'default'", ",", "'42'", ")", "self", ".", "emitEnd", "(", "'%s,'", "%", "str", "(", "value", ")", ",", "schema", ".", "get", "(", "'description'", ",", "''", ")", ")", "elif", "stype", "==", "'number'", ":", "value", "=", "schema", ".", "get", "(", "'default'", ",", "'3.14'", ")", "self", ".", "emitEnd", "(", "'%s,'", "%", "str", "(", "value", ")", ",", "schema", ".", "get", "(", "'description'", ",", "''", ")", ")", "elif", "stype", "==", "'null'", ":", "self", ".", "emitEnd", "(", "'None,'", ",", "schema", ".", "get", "(", "'description'", ",", "''", ")", ")", "elif", "stype", "==", "'any'", ":", "self", ".", "emitEnd", "(", "'\"\",'", ",", "schema", ".", "get", "(", "'description'", ",", "''", ")", ")", "elif", "stype", "==", "'array'", ":", "self", ".", "emitEnd", "(", "'['", ",", "schema", ".", "get", "(", "'description'", ")", ")", "self", ".", "indent", "(", ")", "self", ".", "emitBegin", "(", "''", ")", "self", ".", "_to_str_impl", "(", "schema", "[", "'items'", "]", ")", "self", ".", "undent", "(", ")", "self", ".", "emit", "(", "'],'", ")", "else", ":", "self", ".", "emit", "(", "'Unknown type! %s'", "%", "stype", ")", "self", ".", "emitEnd", "(", "''", ",", "''", ")", "self", ".", "string", "=", "''", ".", "join", "(", "self", ".", "value", ")", "return", "self", ".", "string" ]
[ 244, 2 ]
[ 302, 22 ]
python
en
['en', 'en', 'en']
True
_SchemaToStruct.to_str
(self, from_cache)
Prototype object based on the schema, in Python code with comments. Args: from_cache: callable(name, seen), Callable that retrieves an object prototype for a schema with the given name. Seen is a list of schema names already seen as we recursively descend the schema definition. Returns: Prototype object based on the schema, in Python code with comments. The lines of the code will all be properly indented.
Prototype object based on the schema, in Python code with comments.
def to_str(self, from_cache): """Prototype object based on the schema, in Python code with comments. Args: from_cache: callable(name, seen), Callable that retrieves an object prototype for a schema with the given name. Seen is a list of schema names already seen as we recursively descend the schema definition. Returns: Prototype object based on the schema, in Python code with comments. The lines of the code will all be properly indented. """ self.from_cache = from_cache return self._to_str_impl(self.schema)
[ "def", "to_str", "(", "self", ",", "from_cache", ")", ":", "self", ".", "from_cache", "=", "from_cache", "return", "self", ".", "_to_str_impl", "(", "self", ".", "schema", ")" ]
[ 304, 2 ]
[ 317, 41 ]
python
en
['en', 'en', 'en']
True
GenerateConfig
(context)
Generate configuration.
Generate configuration.
def GenerateConfig(context): """Generate configuration.""" name_prefix = context.env['deployment'] + '-' + context.env['name'] instance = { 'zone': context.properties['zone'], 'machineType': ZonalComputeUrl( context.env['project'], context.properties['zone'], 'machineTypes', 'f1-micro'), 'metadata': { 'items': [{ 'key': 'startup-script', 'value': context.properties['startup-script'] }] }, 'disks': [{ 'deviceName': 'boot', 'type': 'PERSISTENT', 'autoDelete': True, 'boot': True, 'initializeParams': { 'diskName': name_prefix + '-disk', 'sourceImage': GlobalComputeUrl( 'debian-cloud', 'images', 'family/debian-8' ) }, }], 'networkInterfaces': [{ 'accessConfigs': [{ 'name': 'external-nat', 'type': 'ONE_TO_ONE_NAT' }], 'network': GlobalComputeUrl( context.env['project'], 'networks', 'default') }] } # Resources to return. resources = { 'resources': [{ 'name': name_prefix + '-vm', 'type': 'compute.v1.instance', 'properties': instance }] } return resources
[ "def", "GenerateConfig", "(", "context", ")", ":", "name_prefix", "=", "context", ".", "env", "[", "'deployment'", "]", "+", "'-'", "+", "context", ".", "env", "[", "'name'", "]", "instance", "=", "{", "'zone'", ":", "context", ".", "properties", "[", "'zone'", "]", ",", "'machineType'", ":", "ZonalComputeUrl", "(", "context", ".", "env", "[", "'project'", "]", ",", "context", ".", "properties", "[", "'zone'", "]", ",", "'machineTypes'", ",", "'f1-micro'", ")", ",", "'metadata'", ":", "{", "'items'", ":", "[", "{", "'key'", ":", "'startup-script'", ",", "'value'", ":", "context", ".", "properties", "[", "'startup-script'", "]", "}", "]", "}", ",", "'disks'", ":", "[", "{", "'deviceName'", ":", "'boot'", ",", "'type'", ":", "'PERSISTENT'", ",", "'autoDelete'", ":", "True", ",", "'boot'", ":", "True", ",", "'initializeParams'", ":", "{", "'diskName'", ":", "name_prefix", "+", "'-disk'", ",", "'sourceImage'", ":", "GlobalComputeUrl", "(", "'debian-cloud'", ",", "'images'", ",", "'family/debian-8'", ")", "}", ",", "}", "]", ",", "'networkInterfaces'", ":", "[", "{", "'accessConfigs'", ":", "[", "{", "'name'", ":", "'external-nat'", ",", "'type'", ":", "'ONE_TO_ONE_NAT'", "}", "]", ",", "'network'", ":", "GlobalComputeUrl", "(", "context", ".", "env", "[", "'project'", "]", ",", "'networks'", ",", "'default'", ")", "}", "]", "}", "# Resources to return.", "resources", "=", "{", "'resources'", ":", "[", "{", "'name'", ":", "name_prefix", "+", "'-vm'", ",", "'type'", ":", "'compute.v1.instance'", ",", "'properties'", ":", "instance", "}", "]", "}", "return", "resources" ]
[ 30, 0 ]
[ 77, 18 ]
python
en
['en', 'la', 'en']
False
eval_setup_to_action_tier
(eval_setup_name: str)
Gets a default action tier for an eval setup.
Gets a default action tier for an eval setup.
def eval_setup_to_action_tier(eval_setup_name: str) -> str: """Gets a default action tier for an eval setup.""" for tier in phyre.action_mappers.ACTION_MAPPERS: if eval_setup_name.startswith(tier): return tier raise ValueError('Failed to derive action tier for eval setup %s' % eval_setup_name)
[ "def", "eval_setup_to_action_tier", "(", "eval_setup_name", ":", "str", ")", "->", "str", ":", "for", "tier", "in", "phyre", ".", "action_mappers", ".", "ACTION_MAPPERS", ":", "if", "eval_setup_name", ".", "startswith", "(", "tier", ")", ":", "return", "tier", "raise", "ValueError", "(", "'Failed to derive action tier for eval setup %s'", "%", "eval_setup_name", ")" ]
[ 56, 0 ]
[ 62, 37 ]
python
en
['en', 'da', 'en']
True
list_eval_setups
()
Get a list of names for all known eval setups.
Get a list of names for all known eval setups.
def list_eval_setups() -> Tuple[str, ...]: """Get a list of names for all known eval setups.""" return tuple(sorted(EVAL_SETUP_BUILDERS))
[ "def", "list_eval_setups", "(", ")", "->", "Tuple", "[", "str", ",", "...", "]", ":", "return", "tuple", "(", "sorted", "(", "EVAL_SETUP_BUILDERS", ")", ")" ]
[ 65, 0 ]
[ 67, 45 ]
python
en
['en', 'en', 'en']
True
get_fold
(eval_setup: str, seed: int )
Get seed'th fold for specified evaluation setup. Args: eval_setup: The name of the evaluation setup to use. E.g., ball_cross_template. seed: The random seed to create the fold. Returns: Tuple (train_ids, dev_ids, test_ids) Contains task ids to use for each split. Raises: ValueError: Eval_setup is not valid evaluation setup.
Get seed'th fold for specified evaluation setup.
def get_fold(eval_setup: str, seed: int ) -> Tuple[Tuple[str, ...], Tuple[str, ...], Tuple[str, ...]]: """Get seed'th fold for specified evaluation setup. Args: eval_setup: The name of the evaluation setup to use. E.g., ball_cross_template. seed: The random seed to create the fold. Returns: Tuple (train_ids, dev_ids, test_ids) Contains task ids to use for each split. Raises: ValueError: Eval_setup is not valid evaluation setup. """ try: builder = EVAL_SETUP_BUILDERS[eval_setup] except KeyError: raise ValueError(f'Unknown eval setup: {eval_setup}. Chose one of' f' {",".join(EVAL_SETUP_BUILDERS)}') _, test_ids = _flatten_eval_setup(builder(seed)) train_ids, dev_ids = _flatten_eval_setup(builder(seed=seed, dev_seed=0)) return train_ids, dev_ids, test_ids
[ "def", "get_fold", "(", "eval_setup", ":", "str", ",", "seed", ":", "int", ")", "->", "Tuple", "[", "Tuple", "[", "str", ",", "...", "]", ",", "Tuple", "[", "str", ",", "...", "]", ",", "Tuple", "[", "str", ",", "...", "]", "]", ":", "try", ":", "builder", "=", "EVAL_SETUP_BUILDERS", "[", "eval_setup", "]", "except", "KeyError", ":", "raise", "ValueError", "(", "f'Unknown eval setup: {eval_setup}. Chose one of'", "f' {\",\".join(EVAL_SETUP_BUILDERS)}'", ")", "_", ",", "test_ids", "=", "_flatten_eval_setup", "(", "builder", "(", "seed", ")", ")", "train_ids", ",", "dev_ids", "=", "_flatten_eval_setup", "(", "builder", "(", "seed", "=", "seed", ",", "dev_seed", "=", "0", ")", ")", "return", "train_ids", ",", "dev_ids", ",", "test_ids" ]
[ 70, 0 ]
[ 93, 39 ]
python
en
['en', 'en', 'en']
True
_stable_rng
(task_id: str)
Map a string to a number in [0, 1).
Map a string to a number in [0, 1).
def _stable_rng(task_id: str) -> float: """Map a string to a number in [0, 1).""" divider = 100000 return (int(hashlib.md5(task_id.encode('utf8')).hexdigest(), 16) % divider) / divider
[ "def", "_stable_rng", "(", "task_id", ":", "str", ")", "->", "float", ":", "divider", "=", "100000", "return", "(", "int", "(", "hashlib", ".", "md5", "(", "task_id", ".", "encode", "(", "'utf8'", ")", ")", ".", "hexdigest", "(", ")", ",", "16", ")", "%", "divider", ")", "/", "divider" ]
[ 129, 0 ]
[ 133, 30 ]
python
en
['en', 'en', 'en']
True
get_task_ids_in_tier
(tier_name)
Returns a list of all task_ids in iter.
Returns a list of all task_ids in iter.
def get_task_ids_in_tier(tier_name): """Returns a list of all task_ids in iter.""" # dict of dicts: template_id -> task_id -> tier. template_task_tiers = collections.defaultdict(dict) for task_id, task in phyre.loader.load_compiled_task_dict().items(): template_id = task_id.split(':')[0] template_task_tiers[template_id][task_id] = task.tier selected_task_ids = set() tier_name = tier_name.upper() for template_id, task_to_tier in template_task_tiers.items(): tiers = frozenset(task_to_tier.values()) if len(tiers) == 1 and next(iter(tiers)) == tier_name: selected_task_ids.update(task_to_tier) return sorted(selected_task_ids)
[ "def", "get_task_ids_in_tier", "(", "tier_name", ")", ":", "# dict of dicts: template_id -> task_id -> tier.", "template_task_tiers", "=", "collections", ".", "defaultdict", "(", "dict", ")", "for", "task_id", ",", "task", "in", "phyre", ".", "loader", ".", "load_compiled_task_dict", "(", ")", ".", "items", "(", ")", ":", "template_id", "=", "task_id", ".", "split", "(", "':'", ")", "[", "0", "]", "template_task_tiers", "[", "template_id", "]", "[", "task_id", "]", "=", "task", ".", "tier", "selected_task_ids", "=", "set", "(", ")", "tier_name", "=", "tier_name", ".", "upper", "(", ")", "for", "template_id", ",", "task_to_tier", "in", "template_task_tiers", ".", "items", "(", ")", ":", "tiers", "=", "frozenset", "(", "task_to_tier", ".", "values", "(", ")", ")", "if", "len", "(", "tiers", ")", "==", "1", "and", "next", "(", "iter", "(", "tiers", ")", ")", "==", "tier_name", ":", "selected_task_ids", ".", "update", "(", "task_to_tier", ")", "return", "sorted", "(", "selected_task_ids", ")" ]
[ 136, 0 ]
[ 150, 36 ]
python
en
['en', 'en', 'en']
True
create_dev_set
(eval_setup, train_share=TRAIN_SHARE, seed=0)
Create a new train/test split from a train part of another eval setup.
Create a new train/test split from a train part of another eval setup.
def create_dev_set(eval_setup, train_share=TRAIN_SHARE, seed=0): """Create a new train/test split from a train part of another eval setup.""" dev_eval_setup = [] for train_task_ids, _ in eval_setup: train_task_ids = phyre.util.stable_shuffle(train_task_ids, f'make_dev{seed}') num_train = int(len(train_task_ids) * train_share) train, dev = train_task_ids[:num_train], train_task_ids[num_train:] dev_eval_setup.append((train, [dev])) return dev_eval_setup
[ "def", "create_dev_set", "(", "eval_setup", ",", "train_share", "=", "TRAIN_SHARE", ",", "seed", "=", "0", ")", ":", "dev_eval_setup", "=", "[", "]", "for", "train_task_ids", ",", "_", "in", "eval_setup", ":", "train_task_ids", "=", "phyre", ".", "util", ".", "stable_shuffle", "(", "train_task_ids", ",", "f'make_dev{seed}'", ")", "num_train", "=", "int", "(", "len", "(", "train_task_ids", ")", "*", "train_share", ")", "train", ",", "dev", "=", "train_task_ids", "[", ":", "num_train", "]", ",", "train_task_ids", "[", "num_train", ":", "]", "dev_eval_setup", ".", "append", "(", "(", "train", ",", "[", "dev", "]", ")", ")", "return", "dev_eval_setup" ]
[ 153, 0 ]
[ 162, 25 ]
python
en
['en', 'en', 'en']
True
ball_single_instance
(max_per_tpl=10)
Eval setup where each task instance is in separate eval group. The number of tasks that is randomly picked from each template is limited to 10.
Eval setup where each task instance is in separate eval group.
def ball_single_instance(max_per_tpl=10) -> EvalSetup: """Eval setup where each task instance is in separate eval group. The number of tasks that is randomly picked from each template is limited to 10. """ task_ids = get_task_ids_in_tier('ball') tasks_per_tpl = _get_task_per_tpl(task_ids) eval_setup = [] for _, task_ids_group in sorted(tasks_per_tpl.items()): eval_task_ids = phyre.util.stable_shuffle( task_ids_group, 'ball_online_ind_tasks')[:max_per_tpl] for task_id in eval_task_ids: eval_groups = ((task_id,),) train_set = () train_group = (train_set, eval_groups) eval_setup.append(train_group) return eval_setup
[ "def", "ball_single_instance", "(", "max_per_tpl", "=", "10", ")", "->", "EvalSetup", ":", "task_ids", "=", "get_task_ids_in_tier", "(", "'ball'", ")", "tasks_per_tpl", "=", "_get_task_per_tpl", "(", "task_ids", ")", "eval_setup", "=", "[", "]", "for", "_", ",", "task_ids_group", "in", "sorted", "(", "tasks_per_tpl", ".", "items", "(", ")", ")", ":", "eval_task_ids", "=", "phyre", ".", "util", ".", "stable_shuffle", "(", "task_ids_group", ",", "'ball_online_ind_tasks'", ")", "[", ":", "max_per_tpl", "]", "for", "task_id", "in", "eval_task_ids", ":", "eval_groups", "=", "(", "(", "task_id", ",", ")", ",", ")", "train_set", "=", "(", ")", "train_group", "=", "(", "train_set", ",", "eval_groups", ")", "eval_setup", ".", "append", "(", "train_group", ")", "return", "eval_setup" ]
[ 196, 0 ]
[ 214, 21 ]
python
en
['en', 'en', 'en']
True
ball_single_template
()
Eval setup where template tasks is a separate eval group.
Eval setup where template tasks is a separate eval group.
def ball_single_template() -> EvalSetup: """Eval setup where template tasks is a separate eval group.""" max_per_tpl = 10 task_ids = get_task_ids_in_tier('ball') tasks_per_tpl = _get_task_per_tpl(task_ids) eval_setup = [] for _, task_ids_group in sorted(tasks_per_tpl.items()): task_ids_group = phyre.util.stable_shuffle( task_ids_group, 'ball_online_ind_tasks')[:max_per_tpl] eval_groups = (task_ids_group,) train_set = () train_group = (train_set, eval_groups) eval_setup.append(train_group) return eval_setup
[ "def", "ball_single_template", "(", ")", "->", "EvalSetup", ":", "max_per_tpl", "=", "10", "task_ids", "=", "get_task_ids_in_tier", "(", "'ball'", ")", "tasks_per_tpl", "=", "_get_task_per_tpl", "(", "task_ids", ")", "eval_setup", "=", "[", "]", "for", "_", ",", "task_ids_group", "in", "sorted", "(", "tasks_per_tpl", ".", "items", "(", ")", ")", ":", "task_ids_group", "=", "phyre", ".", "util", ".", "stable_shuffle", "(", "task_ids_group", ",", "'ball_online_ind_tasks'", ")", "[", ":", "max_per_tpl", "]", "eval_groups", "=", "(", "task_ids_group", ",", ")", "train_set", "=", "(", ")", "train_group", "=", "(", "train_set", ",", "eval_groups", ")", "eval_setup", ".", "append", "(", "train_group", ")", "return", "eval_setup" ]
[ 233, 0 ]
[ 247, 21 ]
python
en
['en', 'en', 'en']
True
_cross_template
(tier, seed=1, dev_seed=None, train_share=TRAIN_SHARE)
A set of train groups with half templates in train and half in test.
A set of train groups with half templates in train and half in test.
def _cross_template(tier, seed=1, dev_seed=None, train_share=TRAIN_SHARE) -> EvalSetup: """A set of train groups with half templates in train and half in test.""" task_ids = get_task_ids_in_tier(tier) tasks_per_tpl = collections.defaultdict(list) for task_id in task_ids: tasks_per_tpl[task_id.split(':')[0]].append(task_id) key_order = phyre.util.stable_shuffle(list(tasks_per_tpl), f'ball_cross_template_half_{seed}') train_size = int(round(len(key_order) * train_share)) if dev_seed is not None: key_order = key_order[:train_size] key_order = phyre.util.stable_shuffle( key_order, f'dev_ball_cross_template_half_{dev_seed}') train_size = int(len(key_order) * train_share) tasks_per_tpl = [tasks_per_tpl[key] for key in key_order] train, test = tasks_per_tpl[:train_size], tasks_per_tpl[train_size:] eval_setup = [] train_ids = sum(train, []) eval_ids = sum(test, []) train_group = (tuple(train_ids), [tuple(eval_ids)]) eval_setup.append(train_group) return eval_setup
[ "def", "_cross_template", "(", "tier", ",", "seed", "=", "1", ",", "dev_seed", "=", "None", ",", "train_share", "=", "TRAIN_SHARE", ")", "->", "EvalSetup", ":", "task_ids", "=", "get_task_ids_in_tier", "(", "tier", ")", "tasks_per_tpl", "=", "collections", ".", "defaultdict", "(", "list", ")", "for", "task_id", "in", "task_ids", ":", "tasks_per_tpl", "[", "task_id", ".", "split", "(", "':'", ")", "[", "0", "]", "]", ".", "append", "(", "task_id", ")", "key_order", "=", "phyre", ".", "util", ".", "stable_shuffle", "(", "list", "(", "tasks_per_tpl", ")", ",", "f'ball_cross_template_half_{seed}'", ")", "train_size", "=", "int", "(", "round", "(", "len", "(", "key_order", ")", "*", "train_share", ")", ")", "if", "dev_seed", "is", "not", "None", ":", "key_order", "=", "key_order", "[", ":", "train_size", "]", "key_order", "=", "phyre", ".", "util", ".", "stable_shuffle", "(", "key_order", ",", "f'dev_ball_cross_template_half_{dev_seed}'", ")", "train_size", "=", "int", "(", "len", "(", "key_order", ")", "*", "train_share", ")", "tasks_per_tpl", "=", "[", "tasks_per_tpl", "[", "key", "]", "for", "key", "in", "key_order", "]", "train", ",", "test", "=", "tasks_per_tpl", "[", ":", "train_size", "]", ",", "tasks_per_tpl", "[", "train_size", ":", "]", "eval_setup", "=", "[", "]", "train_ids", "=", "sum", "(", "train", ",", "[", "]", ")", "eval_ids", "=", "sum", "(", "test", ",", "[", "]", ")", "train_group", "=", "(", "tuple", "(", "train_ids", ")", ",", "[", "tuple", "(", "eval_ids", ")", "]", ")", "eval_setup", ".", "append", "(", "train_group", ")", "return", "eval_setup" ]
[ 251, 0 ]
[ 273, 21 ]
python
en
['en', 'en', 'en']
True
_single_template
(tier, seed=1, dev_seed=None, train_share=TRAIN_SHARE)
Each template is a separate group.
Each template is a separate group.
def _single_template(tier, seed=1, dev_seed=None, train_share=TRAIN_SHARE) -> EvalSetup: """Each template is a separate group.""" task_ids = get_task_ids_in_tier(tier) tasks_per_tpl = _get_task_per_tpl(task_ids) eval_setup = [] for _, task_ids_group in sorted(tasks_per_tpl.items()): task_ids_group = phyre.util.stable_shuffle( task_ids_group, f'ball_online_ind_tasks{seed}') train_size = int(round(len(task_ids_group) * train_share)) if not train_size: continue train_ids, eval_ids = (task_ids_group[:train_size], task_ids_group[train_size:]) eval_groups = (tuple(eval_ids),) train_set = tuple(train_ids) train_group = (train_set, eval_groups) eval_setup.append(train_group) if dev_seed is not None: eval_setup = create_dev_set(eval_setup, train_share, seed=dev_seed) return eval_setup
[ "def", "_single_template", "(", "tier", ",", "seed", "=", "1", ",", "dev_seed", "=", "None", ",", "train_share", "=", "TRAIN_SHARE", ")", "->", "EvalSetup", ":", "task_ids", "=", "get_task_ids_in_tier", "(", "tier", ")", "tasks_per_tpl", "=", "_get_task_per_tpl", "(", "task_ids", ")", "eval_setup", "=", "[", "]", "for", "_", ",", "task_ids_group", "in", "sorted", "(", "tasks_per_tpl", ".", "items", "(", ")", ")", ":", "task_ids_group", "=", "phyre", ".", "util", ".", "stable_shuffle", "(", "task_ids_group", ",", "f'ball_online_ind_tasks{seed}'", ")", "train_size", "=", "int", "(", "round", "(", "len", "(", "task_ids_group", ")", "*", "train_share", ")", ")", "if", "not", "train_size", ":", "continue", "train_ids", ",", "eval_ids", "=", "(", "task_ids_group", "[", ":", "train_size", "]", ",", "task_ids_group", "[", "train_size", ":", "]", ")", "eval_groups", "=", "(", "tuple", "(", "eval_ids", ")", ",", ")", "train_set", "=", "tuple", "(", "train_ids", ")", "train_group", "=", "(", "train_set", ",", "eval_groups", ")", "eval_setup", ".", "append", "(", "train_group", ")", "if", "dev_seed", "is", "not", "None", ":", "eval_setup", "=", "create_dev_set", "(", "eval_setup", ",", "train_share", ",", "seed", "=", "dev_seed", ")", "return", "eval_setup" ]
[ 277, 0 ]
[ 299, 21 ]
python
en
['en', 'en', 'en']
True
Evaluator.maybe_log_attempt
(self, task_index: int, status: SimulationStatusLike, score: float = 0.0, action: Sequence = None)
Logs status of attempt on task iff status is for a valid action. Args: task_index: index into task_ids of task. status: simulation status of attempt on task. score: Normalized (typically via sigmoid) score of how likely the model thinks this action will solve the task. Added by rgirdhar, optional for previous parts of code. action: Actual action taken, to store in the results. Also added by rgirdhar. Returns: True if attempt was logged (valid action, less than MAX_TEST_ATTEMPTS made on task), else False. Raises: AssertionError: More than MAX_TEST_ATTEMPTS attempts were made on the task.
Logs status of attempt on task iff status is for a valid action.
def maybe_log_attempt(self, task_index: int, status: SimulationStatusLike, score: float = 0.0, action: Sequence = None) -> bool: """Logs status of attempt on task iff status is for a valid action. Args: task_index: index into task_ids of task. status: simulation status of attempt on task. score: Normalized (typically via sigmoid) score of how likely the model thinks this action will solve the task. Added by rgirdhar, optional for previous parts of code. action: Actual action taken, to store in the results. Also added by rgirdhar. Returns: True if attempt was logged (valid action, less than MAX_TEST_ATTEMPTS made on task), else False. Raises: AssertionError: More than MAX_TEST_ATTEMPTS attempts were made on the task. """ status = _normalize_sumulation_status(status) if status.is_invalid(): return False assert self.attempts_per_task_index[task_index] < MAX_TEST_ATTEMPTS, ( f'Task {self._task_ids[task_index]} of index {task_index} has ' f'{self.attempts_per_task_index[task_index]} attempts made, ' 'greater than maximum number of test attempts ' f'{MAX_TEST_ATTEMPTS}') task_id = self._task_ids[task_index] data_tuple = (task_id, status, score, action.tolist()) self._log.append(data_tuple) if task_id not in self._log_per_task_id: self._log_per_task_id[task_id] = [] self._log_per_task_id[task_id].append(data_tuple) self.attempts_per_task_index[task_index] += 1 return True
[ "def", "maybe_log_attempt", "(", "self", ",", "task_index", ":", "int", ",", "status", ":", "SimulationStatusLike", ",", "score", ":", "float", "=", "0.0", ",", "action", ":", "Sequence", "=", "None", ")", "->", "bool", ":", "status", "=", "_normalize_sumulation_status", "(", "status", ")", "if", "status", ".", "is_invalid", "(", ")", ":", "return", "False", "assert", "self", ".", "attempts_per_task_index", "[", "task_index", "]", "<", "MAX_TEST_ATTEMPTS", ",", "(", "f'Task {self._task_ids[task_index]} of index {task_index} has '", "f'{self.attempts_per_task_index[task_index]} attempts made, '", "'greater than maximum number of test attempts '", "f'{MAX_TEST_ATTEMPTS}'", ")", "task_id", "=", "self", ".", "_task_ids", "[", "task_index", "]", "data_tuple", "=", "(", "task_id", ",", "status", ",", "score", ",", "action", ".", "tolist", "(", ")", ")", "self", ".", "_log", ".", "append", "(", "data_tuple", ")", "if", "task_id", "not", "in", "self", ".", "_log_per_task_id", ":", "self", ".", "_log_per_task_id", "[", "task_id", "]", "=", "[", "]", "self", ".", "_log_per_task_id", "[", "task_id", "]", ".", "append", "(", "data_tuple", ")", "self", ".", "attempts_per_task_index", "[", "task_index", "]", "+=", "1", "return", "True" ]
[ 429, 4 ]
[ 469, 19 ]
python
en
['en', 'en', 'en']
True
Evaluator.compute_all_metrics
(self)
Computes metrics based on recorded log of simulation results. Returns: Dictionary mapping metric name to computed value.
Computes metrics based on recorded log of simulation results.
def compute_all_metrics(self) -> Metrics: """Computes metrics based on recorded log of simulation results. Returns: Dictionary mapping metric name to computed value. """ return compute_metrics_normalized(self._log, len(self._task_ids))
[ "def", "compute_all_metrics", "(", "self", ")", "->", "Metrics", ":", "return", "compute_metrics_normalized", "(", "self", ".", "_log", ",", "len", "(", "self", ".", "_task_ids", ")", ")" ]
[ 471, 4 ]
[ 477, 73 ]
python
en
['en', 'en', 'en']
True
Evaluator.compute_all_metrics_per_task
(self)
Computes metrics for each task separately. Returns: Dictionary of task to dictionary mapping metric name to computed value.
Computes metrics for each task separately.
def compute_all_metrics_per_task(self) -> Dict[str, Metrics]: """Computes metrics for each task separately. Returns: Dictionary of task to dictionary mapping metric name to computed value. """ res = {} for task_id in self._task_ids: if task_id in self._log_per_task_id: res[task_id] = compute_metrics_normalized( self._log_per_task_id[task_id], 1) res[task_id]['logs'] = self._log_per_task_id[task_id] else: res[task_id] = {} res[task_id]['logs'] = [] return res
[ "def", "compute_all_metrics_per_task", "(", "self", ")", "->", "Dict", "[", "str", ",", "Metrics", "]", ":", "res", "=", "{", "}", "for", "task_id", "in", "self", ".", "_task_ids", ":", "if", "task_id", "in", "self", ".", "_log_per_task_id", ":", "res", "[", "task_id", "]", "=", "compute_metrics_normalized", "(", "self", ".", "_log_per_task_id", "[", "task_id", "]", ",", "1", ")", "res", "[", "task_id", "]", "[", "'logs'", "]", "=", "self", ".", "_log_per_task_id", "[", "task_id", "]", "else", ":", "res", "[", "task_id", "]", "=", "{", "}", "res", "[", "task_id", "]", "[", "'logs'", "]", "=", "[", "]", "return", "res" ]
[ 479, 4 ]
[ 495, 18 ]
python
en
['en', 'en', 'en']
True
Evaluator.get_auccess
(self, attempts: int = MAX_TEST_ATTEMPTS)
Calculated AUCCESS metric. Starting in v0.0.1.1 renamed from get_aucess to get_auccess. Args: attempts: Number of attempts to use for calulation of auccess, default MAX_TEST_ATTEMPTS. Returns: Result of AUCCESS calculation.
Calculated AUCCESS metric.
def get_auccess(self, attempts: int = MAX_TEST_ATTEMPTS) -> float: """Calculated AUCCESS metric. Starting in v0.0.1.1 renamed from get_aucess to get_auccess. Args: attempts: Number of attempts to use for calulation of auccess, default MAX_TEST_ATTEMPTS. Returns: Result of AUCCESS calculation. """ metrics = self.compute_all_metrics() return metrics[AUCCESS_METRIC][attempts]
[ "def", "get_auccess", "(", "self", ",", "attempts", ":", "int", "=", "MAX_TEST_ATTEMPTS", ")", "->", "float", ":", "metrics", "=", "self", ".", "compute_all_metrics", "(", ")", "return", "metrics", "[", "AUCCESS_METRIC", "]", "[", "attempts", "]" ]
[ 501, 4 ]
[ 514, 48 ]
python
en
['en', 'en', 'en']
True
Evaluator.get_attempts_for_task
(self, task_index)
Args: task_index: index into task_ids of task. Returns: Number recorded attempts on task_index.
Args: task_index: index into task_ids of task.
def get_attempts_for_task(self, task_index): """ Args: task_index: index into task_ids of task. Returns: Number recorded attempts on task_index. """ return self.attempts_per_task_index[task_index]
[ "def", "get_attempts_for_task", "(", "self", ",", "task_index", ")", ":", "return", "self", ".", "attempts_per_task_index", "[", "task_index", "]" ]
[ 516, 4 ]
[ 524, 55 ]
python
en
['en', 'error', 'th']
False
Evaluator.task_ids
(self)
Returns ordered list of tasks ids.
Returns ordered list of tasks ids.
def task_ids(self) -> Tuple[str, ...]: """Returns ordered list of tasks ids.""" return self._task_ids
[ "def", "task_ids", "(", "self", ")", "->", "Tuple", "[", "str", ",", "...", "]", ":", "return", "self", ".", "_task_ids" ]
[ 527, 4 ]
[ 529, 29 ]
python
en
['en', 'et', 'en']
True