code stringlengths 26 79.6k | docstring stringlengths 1 46.9k |
|---|---|
def attention_lm_moe_memory_efficient():
hparams = attention_lm_moe_large()
hparams.diet_experts = True
hparams.layer_preprocess_sequence = "n"
hparams.layer_postprocess_sequence = "da"
hparams.layer_prepostprocess_dropout = 0.0
hparams.memory_efficient_ffn = True
hparams.attention_type = AttentionType... | Memory-efficient version. |
def figure_rst(figure_list, sources_dir):
figure_paths = [os.path.relpath(figure_path, sources_dir)
.replace(os.sep, ).lstrip()
for figure_path in figure_list]
images_rst = ""
if len(figure_paths) == 1:
figure_name = figure_paths[0]
images_rst = ... | Generate RST for a list of PNG filenames.
Depending on whether we have one or more figures, we use a
single rst call to 'image' or a horizontal list.
Parameters
----------
figure_list : list
List of strings of the figures' absolute paths.
sources_dir : str
absolute path of Sphi... |
def _stack_bands(one, other):
assert set(one.band_names).intersection(set(other.band_names)) == set()
if one.band_names == other.band_names:
raise ValueError("rasters have the same bands, use another merge strategy")
new_mask = np.ma.getmaskarray(one.image)[0] | ... | Merges two rasters with non overlapping bands by stacking the bands. |
def handle_abs(self):
x_raw = self.microbit.accelerometer.get_x()
y_raw = self.microbit.accelerometer.get_y()
x_abs = (, 0x00, x_raw)
y_abs = (, 0x01, y_raw)
return x_abs, y_abs | Gets the state as the raw abolute numbers. |
def walk_processes(top, topname=, topdown=True, ignoreFlag=False):
if not ignoreFlag:
flag = topdown
else:
flag = True
proc = top
level = 0
if flag:
yield topname, proc, level
if len(proc.subprocess) > 0:
level += 1
for name, subproc in proc.subproc... | Generator for recursive tree of climlab processes
Starts walking from climlab process ``top`` and generates a complete
list of all processes and sub-processes that are managed from ``top`` process.
``level`` indicades the rank of specific process in the process hierarchy:
.. note::
* level 0:... |
def get_audit_log(self, begin_time=None, end_time=None):
query_parms = self._time_query_parms(begin_time, end_time)
uri = self.uri + + query_parms
result = self.manager.session.post(uri)
return result | Return the console audit log entries, optionally filtered by their
creation time.
Authorization requirements:
* Task permission to the "Audit and Log Management" task.
Parameters:
begin_time (:class:`~py:datetime.datetime`):
Begin time for filtering. Log entries... |
def locks(self):
return self.execute(
sql.LOCKS.format(
pid_column=self.pid_column,
query_column=self.query_column
)
) | Display queries with active locks.
Record(
procpid=31776,
relname=None,
transactionid=None,
granted=True,
query_snippet='select * from hello;',
age=datetime.timedelta(0, 0, 288174),
)
:returns: list of Records |
def read_csr(csr):
*
csr = _get_request_obj(csr)
ret = {
: csr.get_version() + 1,
: _parse_subject(csr.get_subject()),
: _dec2hex(csr.get_subject().as_hash()),
: hashlib.sha1(csr.get_pubkey().get_modulus()).hexdigest(),
: _get_csr_e... | Returns a dict containing details of a certificate request.
:depends: - OpenSSL command line tool
csr:
A path or PEM encoded string containing the CSR to read.
CLI Example:
.. code-block:: bash
salt '*' x509.read_csr /etc/pki/mycert.csr |
def authenticate_credentials(self, request, access_token):
try:
token = oauth2_provider.oauth2.models.AccessToken.objects.select_related()
token = token.get(token=access_token, expires__gt=provider_now())
except oauth2_provider.oauth2.models.Ac... | Authenticate the request, given the access token. |
def encoded_datastream(self):
size = 0
if self.verify:
md5 = hashlib.md5()
leftover = None
while self.within_file:
content = self.get_next_section()
if content == BINARY_CONTENT_END:
if self.verify:
... | Generator for datastream content. Takes a list of sections
of data within the current chunk (split on binaryContent start and
end tags), runs a base64 decode, and yields the data. Computes
datastream size and MD5 as data is decoded for sanity-checking
purposes. If binary content is not... |
def set_transfer_spec(self):
_ret = False
try:
self._args.transfer_spec_func(self._args)
_ret = True
except Exception as ex:
self.notify_exception(AsperaTransferSpecError(ex), False)
return _ret | run the function to set the transfer spec on error set associated exception |
def _determine_timeout(default_timeout, specified_timeout, retry):
if specified_timeout is DEFAULT:
specified_timeout = default_timeout
if specified_timeout is default_timeout:
if (
retry
and retry is not DEFAULT
and isinstance... | Determines how timeout should be applied to a wrapped method.
Args:
default_timeout (Optional[Timeout]): The default timeout specified
at method creation time.
specified_timeout (Optional[Timeout]): The timeout specified at
invocation time. If :attr:`DEFAULT`, this will be s... |
def cli(ctx, board, fpga, pack, type, size, project_dir,
verbose, verbose_yosys, verbose_arachne):
exit_code = SCons(project_dir).time({
: board,
: fpga,
: size,
: type,
: pack,
: {
: verbose,
: verbose_yosys,
: v... | Bitstream timing analysis. |
def set_level(self, level):
for handler in self.__coloredlogs_handlers:
handler.setLevel(level=level)
self.logger.setLevel(level=level) | Set the logging level of this logger.
:param level: must be an int or a str. |
def _is_path(instance, attribute, s, exists=True):
"Validator for path-yness"
if not s:
return
if exists:
if os.path.exists(s):
return
else:
raise OSError("path does not exist")
else:
raise TypeError("Not a path?") | Validator for path-yness |
def nucmer_hits_to_ref_and_qry_coords(cls, nucmer_hits, contig=None):
if contig is None:
ctg_coords = {key: [] for key in nucmer_hits.keys()}
else:
ctg_coords = {contig: []}
ref_coords = {}
for key in ctg_coords:
hits = copy.copy(nucmer_hits... | Same as nucmer_hits_to_ref_coords, except removes containing hits first,
and returns ref and qry coords lists |
def _get_capabilities(self):
conn = self._get_connection()
capabilities = []
for c in conn.server_capabilities:
capabilities.append(c)
LOG.debug("Server capabilities: %s", capabilities)
return capabilities | Get the servers NETCONF capabilities.
:return: List of server capabilities. |
def set_loop_points(self, start_sample=-1, end_sample=0):
lib.SetVoiceLoopPoints(self._handle, start_sample, end_sample) | Set the loop points within the sound.
The sound must have been created with ``loop=True``. The default parameters cause the loop points to be set to
the entire sound duration.
:note: There is currently no API for converting sample numbers to times.
:param start_sample: sample number t... |
def _compute_all_features(self):
self._audio, _ = librosa.load(self.file_struct.audio_file,
sr=self.sr)
self.dur = len(self._audio) / float(self.sr)
self._framesync_features = self.compute_features()
se... | Computes all the features (beatsync, framesync) from the audio. |
def aggregate_by_index(self, function, level=0):
result = self._map_by_index(function, level=level)
return result.map(lambda v: array(v), index=result.index) | Aggregrate data in each record, grouping by index values.
For each unique value of the index, applies a function to the group
indexed by that value. Returns a Series indexed by those unique values.
For the result to be a valid Series object, the aggregating function should
return a simp... |
def association(self, group_xid):
association = {: group_xid}
self._indicator_data.setdefault(, []).append(association) | Add association using xid value.
Args:
group_xid (str): The external id of the Group to associate. |
def _get_timezone(self, root):
tz_str = root.xpath()[0].text
hours = int(self._tz_re.search(tz_str).group(1))
return tzoffset(tz_str, hours * 60) | Find timezone informatation on bottom of the page. |
def _erase_card(self, number):
with self._lock:
if number < (len(self.cards) - 1):
self._erase_card(number + 1)
if number > (len(self.cards) - 1):
return
max_cards_horiz = int(curses.COLS / 35)
obliterate = curses.newwin(
... | Destroy cards with this or higher number. |
def inspect_task(self, task):
url = self._url(, task)
return self._result(self._get(url), True) | Retrieve information about a task.
Args:
task (str): Task ID
Returns:
(dict): Information about the task.
Raises:
:py:class:`docker.errors.APIError`
If the server returns an error. |
def config_name_from_full_name(full_name):
projects, _, configs, result = full_name.split("/")
if projects != "projects" or configs != "configs":
raise ValueError(
"Unexpected format of resource",
full_name,
,
)
return result | Extract the config name from a full resource name.
>>> config_name_from_full_name('projects/my-proj/configs/my-config')
"my-config"
:type full_name: str
:param full_name:
The full resource name of a config. The full resource name looks like
``projects/project-name/configs/config-na... |
def golfclap(rest):
"Clap for something"
clapv = random.choice(phrases.clapvl)
adv = random.choice(phrases.advl)
adj = random.choice(phrases.adjl)
if rest:
clapee = rest.strip()
karma.Karma.store.change(clapee, 1)
return "/me claps %s for %s, %s %s." % (clapv, rest, adv, adj)
return "/me claps %s, %s %s." %... | Clap for something |
def exc_thrown_by_descriptor():
traceback = sys.exc_info()[2]
tb_locals = traceback.tb_frame.f_locals
if "self" in tb_locals:
if not isinstance(tb_locals["self"], Descriptor):
return False
return True
return False | Return True if the last exception was thrown by a
Descriptor instance. |
def AnnotateBED(bed, GTF, genome_file, bedcols=None, promoter=[1000,200]):
if type(bed) == type("string"):
bed=pd.read_table(bed,header=None)
bed.columns=bedcols.split(",")
print("Reading GTF file.")
sys.stdout.flush()
GTF=readGTF(GTF)
GTF["gene_name"]=retrieve_GTF_field("gene... | Annotates a bed file.
:param bed: either a /path/to/file.bed or a Pandas dataframe in bed format. /path/to/file.bed implies bedcols.
:param GTF: /path/to/file.gtf
:param genome_file: /path/to/file.genome - a tab separated values of chr name and size information
:param bedcols: a comma separated string ... |
def process_file(self, path, dryrun):
if dryrun:
return path
ret = []
with open(path, "r") as infile:
for line in infile:
if re.search(self.__exp, line):
ret.append(line)
return ret if len(r... | Print files path. |
def tunnel_to_kernel(connection_info, sshserver, sshkey=None):
if isinstance(connection_info, basestring):
if tunnel.try_passwordless_ssh(sshserver, sshkey):
password=False
else:
password = getpass("SSH Password for %s: "%sshserver)
for lp,rp in zip(lports, rports... | tunnel connections to a kernel via ssh
This will open four SSH tunnels from localhost on this machine to the
ports associated with the kernel. They can be either direct
localhost-localhost tunnels, or if an intermediate server is necessary,
the kernel must be listening on a public IP.
Par... |
def get_domain_config(self, domain):
domain_root = self.identify_domain_root(domain)
host =
if len(domain_root) != len(domain):
host = domain.replace( + domain_root, )
domain_connect_api = self._identify_domain_connect_api(domain_root)
ret = self._get_dom... | Makes a discovery of domain name and resolves configuration of DNS provider
:param domain: str
domain name
:return: DomainConnectConfig
domain connect config
:raises: NoDomainConnectRecordException
when no _domainconnect record found
:raises: NoDomain... |
def my_protocol_parser(out, buf):
while True:
tp = yield from buf.read(5)
if tp in (MSG_PING, MSG_PONG):
yield from buf.skipuntil(b)
out.feed_data(Message(tp, None))
elif tp == MSG_STOP:
out.feed_data(Message(tp, None))
elif tp ==... | Parser is used with StreamParser for incremental protocol parsing.
Parser is a generator function, but it is not a coroutine. Usually
parsers are implemented as a state machine.
more details in asyncio/parsers.py
existing parsers:
* HTTP protocol parsers asyncio/http/protocol.py
* websocket... |
def hash_data(data, hasher=NoParam, base=NoParam, types=False,
hashlen=NoParam, convert=False):
if convert and isinstance(data, six.string_types):
try:
data = json.dumps(data)
except TypeError as ex:
pass
base = _rectify_bas... | Get a unique hash depending on the state of the data.
Args:
data (object):
Any sort of loosely organized data
hasher (str or HASHER):
Hash algorithm from hashlib, defaults to `sha512`.
base (str or List[str]):
Shorthand key or a list of symbols. Valid ... |
def haiz(obj, chart):
objGender = obj.gender()
objFaction = obj.faction()
if obj.id == const.MERCURY:
sun = chart.getObject(const.SUN)
orientalityM = orientality(obj, sun)
if orientalityM == ORIENTAL:
objGender = const.MASCULINE
objFaction =... | Returns if an object is in Haiz. |
def scrape_wikinews(conn, project, articleset, query):
url = "http://en.wikinews.org/w/index.php?search={}&limit=50".format(query)
logging.info(url)
for page in get_pages(url):
urls = get_article_urls(page)
arts = list(get_articles(urls))
logging.info("Adding {} articles to set ... | Scrape wikinews articles from the given query
@param conn: The AmcatAPI object
@param articleset: The target articleset ID
@param category: The wikinews category name |
def _send_textmetrics(metrics):
data = [.join(map(six.text_type, metric)) for metric in metrics] + []
return .join(data) | Format metrics for the carbon plaintext protocol |
def get_citation_by_reference(self, type: str, reference: str) -> Optional[Citation]:
citation_hash = hash_citation(type=type, reference=reference)
return self.get_citation_by_hash(citation_hash) | Get a citation object by its type and reference. |
def entropy(args):
p = OptionParser(entropy.__doc__)
p.add_option("--threshold", default=0, type="int",
help="Complexity needs to be above")
opts, args = p.parse_args(args)
if len(args) != 1:
sys.exit(not p.print_help())
kmc_out, = args
fp = open(kmc_out)
for r... | %prog entropy kmc_dump.out
kmc_dump.out contains two columns:
AAAAAAAAAAAGAAGAAAGAAA 34 |
def load_tabs(self):
tab_group = self.get_tabs(self.request, **self.kwargs)
tabs = tab_group.get_tabs()
for tab in [t for t in tabs if issubclass(t.__class__, TableTab)]:
self.table_classes.extend(tab.table_classes)
for table in tab._tables.values():
... | Loads the tab group.
It compiles the table instances for each table attached to
any :class:`horizon.tabs.TableTab` instances on the tab group.
This step is necessary before processing any tab or table actions. |
def abort(*args, **kwargs):
code = kwargs.pop("code", 1)
logger = kwargs.pop("logger", LOG.error if code else LOG.info)
fatal = kwargs.pop("fatal", True)
return_value = fatal
if isinstance(fatal, tuple) and len(fatal) == 2:
fatal, return_value = fatal
if logger and fatal is not No... | Usage:
return abort("...") => will sys.exit() by default
return abort("...", fatal=True) => Will sys.exit()
# Not fatal, but will log/print message:
return abort("...", fatal=False) => Will return False
return abort("...", fatal=(False, None)) => Will return None
return ... |
def sigma_cached(self, psd):
if not hasattr(self, ):
from pycbc.opt import LimitedSizeDict
self._sigmasq = LimitedSizeDict(size_limit=2**5)
key = id(psd)
if not hasattr(psd, ):
psd._sigma_cached_key = {}
if key not in self._sigmasq or id(self) not in psd._sigma_cached_key:... | Cache sigma calculate for use in tandem with the FilterBank class |
def depends (self, d):
self.dependencies_ = unique (self.dependencies_ + d).sort () | Adds additional instances of 'VirtualTarget' that this
one depends on. |
def add_module(self, module, cython=False):
name_module = module.__name__.split()[-1]
short = (
% name_module)
long = (
% module.__name__)
self._short2long[short] = long
for (name_member, member) in vars(module).items():
if se... | Add the given module, its members, and their submembers.
The first examples are based on the site-package |numpy|: which
is passed to method |Substituter.add_module|:
>>> from hydpy.core.autodoctools import Substituter
>>> substituter = Substituter()
>>> import numpy
>>... |
def _hashable_bytes(data):
if isinstance(data, bytes):
return data
elif isinstance(data, str):
return data.encode()
else:
raise TypeError(data) | Coerce strings to hashable bytes. |
def _set_least_batch_id(self, txn_signature):
batch = self._batches_by_txn_id[txn_signature]
least_index = self._index_of_batch(
self._batches_by_id[self._least_batch_id_wo_results].batch)
current_index = self._index_of_batch(batch)
all_prior = False
if c... | Set the first batch id that doesn't have all results.
Args:
txn_signature (str): The txn identifier of the transaction with
results being set. |
def with_reconnect(func):
from pymongo.errors import AutoReconnect
@functools.wraps(func)
def _reconnector(*args, **kwargs):
for _ in range(20):
try:
return func(*args, **kwargs)
except AutoReconnect:
time.sleep(0.250)
raise
... | Handle when AutoReconnect is raised from pymongo. This is the standard error
raised for everything from "host disconnected" to "couldn't connect to host"
and more.
The sleep handles the edge case when the state of a replica set changes, and
the cursor raises AutoReconnect because the master may have ch... |
def stats(self):
status, _, body = self._request(, self.stats_path(),
{: })
if status == 200:
return json.loads(bytes_to_str(body))
else:
return None | Gets performance statistics and server information |
def cysparse_real_type_from_real_cysparse_complex_type(cysparse_type):
r_type = None
if cysparse_type in []:
r_type =
elif cysparse_type in []:
r_type =
elif cysparse_type in []:
r_type =
else:
raise TypeError("Not a recognized complex type")
return r_ty... | Returns the **real** type for the real or imaginary part of a **real** complex type.
For instance:
COMPLEX128_t -> FLOAT64_t
Args:
cysparse: |
def get_mysql_credentials(cfg_file):
try:
parser = ConfigParser.ConfigParser()
cfg_fp = open(cfg_file)
parser.readfp(cfg_fp)
cfg_fp.close()
except ConfigParser.NoOptionError:
cfg_fp.close()
print()
sys.exit(1)
except IOError:
print(, cfg_... | Get the credentials and database name from options in config file. |
def show_grid(images, rows=None, cols=None):
grid = draw_grid(images, rows=rows, cols=cols)
imshow(grid) | Converts the input images to a grid image and shows it in a new window.
dtype support::
minimum of (
:func:`imgaug.imgaug.draw_grid`,
:func:`imgaug.imgaug.imshow`
)
Parameters
----------
images : (N,H,W,3) ndarray or iterable of (H,W,3) array
See :func:... |
def _make_grid_of_axes(self,
bounding_rect=cfg.bounding_rect_default,
num_rows=cfg.num_rows_per_view_default,
num_cols=cfg.num_cols_grid_default,
axis_pad=cfg.axis_pad_default,
commn_an... | Creates a grid of axes bounded within a given rectangle. |
def from_kwargs(cls, **kwargs):
arrays = []
names = []
for p,vals in kwargs.items():
if not isinstance(vals, numpy.ndarray):
if not isinstance(vals, list):
vals = [vals]
vals = numpy.array(vals)
arrays.append(va... | Creates a new instance of self from the given keyword arguments.
Each argument will correspond to a field in the returned array, with
the name of the field given by the keyword, and the value(s) whatever
the keyword was set to. Each keyword may be set to a single value or
a list of value... |
def web(connection, host, port):
from bio2bel.web.application import create_application
app = create_application(connection=connection)
app.run(host=host, port=port) | Run a combine web interface. |
def reflectance_from_tbs(self, sun_zenith, tb_near_ir, tb_thermal, **kwargs):
if hasattr(tb_near_ir, ) or hasattr(tb_thermal, ):
compute = False
else:
compute = True
if hasattr(tb_near_ir, ) or hasattr(tb_thermal, ):
is_masked = True
... | The relfectance calculated is without units and should be between 0 and 1.
Inputs:
sun_zenith: Sun zenith angle for every pixel - in degrees
tb_near_ir: The 3.7 (or 3.9 or equivalent) IR Tb's at every pixel
(Kelvin)
tb_thermal: The 10.8 (or 11 or 12 or equ... |
def password_change(self, wallet, password):
wallet = self._process_value(wallet, )
payload = {"wallet": wallet, "password": password}
resp = self.call(, payload)
return resp[] == | Changes the password for **wallet** to **password**
.. enable_control required
:param wallet: Wallet to change password for
:type wallet: str
:param password: Password to set
:type password: str
:raises: :py:exc:`nano.rpc.RPCException`
>>> rpc.password_change... |
def generate_classified_legend(
analysis,
exposure,
hazard,
use_rounding,
debug_mode):
analysis_row = next(analysis.getFeatures())
thresholds = hazard.keywords.get()
if thresholds:
hazard_unit = hazard.keywords.get()
hazard_unit = defin... | Generate an ordered python structure with the classified symbology.
:param analysis: The analysis layer.
:type analysis: QgsVectorLayer
:param exposure: The exposure layer.
:type exposure: QgsVectorLayer
:param hazard: The hazard layer.
:type hazard: QgsVectorLayer
:param use_rounding: B... |
def _evaluate(self,R,z,phi=0.,t=0.):
l,n = bovy_coords.Rz_to_lambdanu(R,z,ac=self._ac,Delta=self._Delta)
return -1./(nu.sqrt(l) + nu.sqrt(n)) | NAME:
_evaluate
PURPOSE:
evaluate the potential at R,z
INPUT:
R - Galactocentric cylindrical radius
z - vertical height
phi - azimuth
t - time
OUTPUT:
Phi(R,z)
HISTORY:
2015-02-15 - Written - ... |
def plotActivation(self, position=None, time=None, velocity=None):
self.ax1.clear()
y = self.activations["n"] + self.activations["s"] + self.activations["e"] + \
self.activations["w"]
self.ax1.matshow(y.reshape(self.dimensions))
self.ax2.clear()
self.ax2.matshow(self.activationsI.resha... | Plot the activation of the current cell populations. Assumes that
two axes have already been created, ax1 and ax2. If done in a Jupyter
notebook, this plotting will overwrite the old plot.
:param position: The current location of the animal
:param time: The current time in the simulation
:param ve... |
def get_dataset(self, key, info):
datadict = {
1000: [,
,
,
],
500: [,
],
250: []}
platform_name = self.metadata[][][
][][]
info.update({: + platform_name})
... | Read data from file and return the corresponding projectables. |
def _read_by_weight(self, F, att_weights, value):
output = F.batch_dot(att_weights, value)
return output | Read from the value matrix given the attention weights.
Parameters
----------
F : symbol or ndarray
att_weights : Symbol or NDArray
Attention weights.
For single-head attention,
Shape (batch_size, query_length, memory_length).
For mult... |
def for_meters(cls, meter_x, meter_y, zoom):
point = Point.from_meters(meter_x=meter_x, meter_y=meter_y)
pixel_x, pixel_y = point.pixels(zoom=zoom)
return cls.for_pixels(pixel_x=pixel_x, pixel_y=pixel_y, zoom=zoom) | Creates a tile from X Y meters in Spherical Mercator EPSG:900913 |
def multiple_sequence_alignment(seqs_fp, threads=1):
logger = logging.getLogger(__name__)
logger.info( % seqs_fp)
if threads == 0:
threads = -1
if stat(seqs_fp).st_size == 0:
logger.warning( % seqs_fp)
return None
msa_fp = seqs_fp +
params = [, , , , ,
... | Perform multiple sequence alignment on FASTA file using MAFFT.
Parameters
----------
seqs_fp: string
filepath to FASTA file for multiple sequence alignment
threads: integer, optional
number of threads to use. 0 to use all threads
Returns
-------
msa_fp : str
name of... |
def contains(self, string):
vectype = self.weld_type
if isinstance(vectype, WeldVec):
elem_type = vectype.elemType
if isinstance(elem_type, WeldChar):
return SeriesWeld(
grizzly_impl.contains(
self.expr... | Summary
Returns:
TYPE: Description |
def flasher(msg, severity=None):
try:
flash(msg, severity)
except RuntimeError:
if severity == :
logging.error(msg)
else:
logging.info(msg) | Flask's flash if available, logging call if not |
def _fetch(self, key):
save_interp = self.section.main.interpolation
self.section.main.interpolation = False
current_section = self.section
while True:
val = current_section.get(key)
if val is not None and not isinstance(va... | Helper function to fetch values from owning section.
Returns a 2-tuple: the value, and the section where it was found. |
def url(self):
self.render()
return self._apiurl + .join(self._parts()).replace(,) | Returns the rendered URL of the chart |
def set_install_id(filename, install_id):
if get_install_id(filename) is None:
raise InstallNameError(.format(filename))
back_tick([, , install_id, filename]) | Set install id for library named in `filename`
Parameters
----------
filename : str
filename of library
install_id : str
install id for library `filename`
Raises
------
RuntimeError if `filename` has not install id |
def train(self):
self.stamp_start = time.time()
for iteration, batch in tqdm.tqdm(enumerate(self.iter_train),
desc=, total=self.max_iter,
ncols=80):
self.epoch = self.iter_train.epoch
sel... | Train the network using the training dataset.
Parameters
----------
None
Returns
-------
None |
def get_parameters(rq, variables, endpoint, query_metadata, auth=None):
internal_matcher = re.compile("__agg_\d+__")
variable_matcher = re.compile(
"(?P<required>[_]{1,2})(?P<name>[^_]+)_?(?P<type>[a-zA-Z0-9]+)?_?(?P<userdefined>[a-zA-Z0-9]+)?.*$")
parameters = {}
for v in... | ?_name The variable specifies the API mandatory parameter name. The value is incorporated in the query as plain literal.
?__name The parameter name is optional.
?_name_iri The variable is substituted with the parameter value as a IRI (also: number or literal).
?_name_en The parameter value is co... |
def create_blueprint(endpoints):
blueprint = Blueprint(
,
__name__,
url_prefix=,
template_folder=,
static_folder=,
)
@blueprint.errorhandler(PIDDeletedError)
def tombstone_errorhandler(error):
return render_template(
current_app.config[],... | Create Invenio-Records-UI blueprint.
The factory installs one URL route per endpoint defined, and adds an
error handler for rendering tombstones.
:param endpoints: Dictionary of endpoints to be installed. See usage
documentation for further details.
:returns: The initialized blueprint. |
def stop(self):
if not self._providers_registered:
self.queue_consumer.unregister_provider(self)
self._unregistered_from_queue_consumer.send(True) | Stop the RpcConsumer.
The RpcConsumer ordinary unregisters from the QueueConsumer when the
last Rpc subclass unregisters from it. If no providers were registered,
we should unregister from the QueueConsumer as soon as we're asked
to stop. |
def get_object_or_child_by_type(self, *types):
objects = self.get_objects_or_children_by_type(*types)
return objects[0] if any(objects) else None | Get object if child already been read or get child.
Use this method for fast access to objects in case of static configurations.
:param types: requested object types.
:return: all children of the specified types. |
def answer(part, module=):
marks = json.load(open(os.path.join(data_directory, module), ))
return marks[ + str(part+1)] | Returns the answers to the lab classes. |
def import_medusa_data(mat_filename, config_file):
df_emd, df_md = _read_mat_mnu0(mat_filename)
if not isinstance(config_file, np.ndarray):
configs = np.loadtxt(config_file).astype(int)
else:
configs = config_file
print()
quadpole_list = []
if df_emd is not None... | Import measurement data (a .mat file) of the FZJ EIT160 system. This
data format is identified as 'FZJ-EZ-2017'.
Parameters
----------
mat_filename: string
filename to the .mat data file. Note that only MNU0 single-potentials
are supported!
config_file: string
filename for c... |
def escape(identifier, ansi_quotes, should_quote):
if not should_quote(identifier):
return identifier
quote = if ansi_quotes else
identifier = identifier.replace(quote, 2*quote)
return .format(quote, identifier, quote) | Escape identifiers.
ANSI uses single quotes, but many databases use back quotes. |
def is_valid_ipv4 (ip):
if not _ipv4_re.match(ip):
return False
a, b, c, d = [int(i) for i in ip.split(".")]
return a <= 255 and b <= 255 and c <= 255 and d <= 255 | Return True if given ip is a valid IPv4 address. |
def get_api_id(self, lambda_name):
try:
response = self.cf_client.describe_stack_resource(StackName=lambda_name,
LogicalResourceId=)
return response[].get(, None)
except:
try:
... | Given a lambda_name, return the API id. |
def fromdict(dict):
index = dict[]
seed = hb_decode(dict[])
n = dict[]
root = hb_decode(dict[])
hmac = hb_decode(dict[])
timestamp = dict[]
self = State(index, seed, n, root, hmac, timestamp)
return self | Takes a dictionary as an argument and returns a new State object
from the dictionary.
:param dict: the dictionary to convert |
def group_add(self, name, restrict, repos, lces=[], assets=[], queries=[],
policies=[], dashboards=[], credentials=[], description=):
return self.raw_query(, , data={
: [{: i} for i in lces],
: [{: i} for i in assets],
: [{: i} for i in queries],
... | group_add name, restrict, repos |
def verify_connectivity(config):
logger.debug("Verifying Connectivity")
ic = InsightsConnection(config)
try:
branch_info = ic.get_branch_info()
except requests.ConnectionError as e:
logger.debug(e)
logger.debug("Failed to connect to satellite")
return False
excep... | Verify connectivity to satellite server |
def parse_response(self, resp):
p, u = self.getparser()
if hasattr(resp,):
text = resp.text
else:
encoding = requests.utils.get_encoding_from_headers(resp.headers)
if encoding is None:
encoding=
if sys.ver... | Parse the xmlrpc response. |
def validate_path(xj_path):
if not isinstance(xj_path, str):
raise XJPathError()
for path in split(xj_path, ):
if path == :
continue
if path.startswith():
if path == or path == :
continue
try:
int(path[1:])
... | Validates XJ path.
:param str xj_path: XJ Path
:raise: XJPathError if validation fails. |
def next(self):
if not self._zip:
self._zip = zipfile.ZipFile(self._reader(self._blob_key))
self._entries = self._zip.infolist()[self._start_index:self._end_index]
self._entries.reverse()
if not self._entries:
raise StopIteration()
entry = self._entries.pop()
self._st... | Returns the next input from this input reader as (ZipInfo, opener) tuple.
Returns:
The next input from this input reader, in the form of a 2-tuple.
The first element of the tuple is a zipfile.ZipInfo object.
The second element of the tuple is a zero-argument function that, when
called, retu... |
def find_enclosing_bracket_left(self, left_ch, right_ch, start_pos=None):
if self.current_char == left_ch:
return 0
if start_pos is None:
start_pos = 0
else:
start_pos = max(0, start_pos)
stack = 1
for i in range(self.curso... | Find the left bracket enclosing current position. Return the relative
position to the cursor position.
When `start_pos` is given, don't look past the position. |
def cancel(batch_fn, cancel_fn, ops):
canceled_ops = []
error_messages = []
max_batch = 256
total_ops = len(ops)
for first_op in range(0, total_ops, max_batch):
batch_canceled, batch_messages = _cancel_batch(
batch_fn, cancel_fn, ops[first_op:first_op + max_batch])
canceled_ops.... | Cancel operations.
Args:
batch_fn: API-specific batch function.
cancel_fn: API-specific cancel function.
ops: A list of operations to cancel.
Returns:
A list of operations canceled and a list of error messages. |
def get_or_create(name=None, group=None, config=None, extra=0, verbose=0, backend_opts=None):
require(, )
backend_opts = backend_opts or {}
verbose = int(verbose)
extra = int(extra)
if config:
config_fn = common.find_template(config)
config = yaml.load(open(config_fn))
... | Creates a virtual machine instance. |
def SG(self):
rs density at the currently specified conditions.
Examples
--------
>>> Chemical().SG
0.7428160596603596
lsg':
return self.SGg
rho = self.rho
if rho is not None:
return SG(rho)
return None | r'''Specific gravity of the chemical, [dimensionless].
For gas-phase conditions, this is calculated at 15.6 °C (60 °F) and 1
atm for the chemical and the reference fluid, air.
For liquid and solid phase conditions, this is calculated based on a
reference fluid of water at 4°... |
def xy(self):
if self._xy != (None, None):
self._x, self._y = self._xy
if self._x is not None and self._y is not None:
x = self._x
if self._x > 1:
x = self._x / 65555
y = self._y
if self._y > 1:
y = sel... | CIE xy color space coordinates as array [x, y] of real values (0..1). |
def _client_run(self):
self._connection.work()
now = self._counter.get_current_ms()
if self._last_activity_timestamp and not self._was_message_received:
time.sleep(0.05)
if self._timeout > 0:
timespan = now - self._last_activity_times... | MessageReceiver Link is now open - start receiving messages.
Will return True if operation successful and client can remain open for
further work.
:rtype: bool |
def user_return(self, frame, return_value):
if self._wait_for_mainpyfile:
if (self.mainpyfile != self.canonic(frame.f_code.co_filename)
or frame.f_lineno<= 0):
return
self._wait_for_mainpyfile = 0
self._old_Pdb_user_return(frame, return_value) | This function is called when a return trap is set here. |
def get_data_path(cls):
marvin_path = os.environ.get(cls._key)
if not marvin_path:
raise InvalidConfigException()
is_path_created = check_path(marvin_path, create=True)
if not is_path_created:
raise InvalidConfigException()
return marvin_path | Read data path from the following sources in order of priority:
1. Environment variable
If not found raises an exception
:return: str - datapath |
def _parse_request_reply(self):
"waiting for a reply to our request"
(version, reply, _, typ) = struct.unpack(, msg)
if version != 5:
self.reply_error(SocksError(
"Expected version 5, got {}".format(version)))
return
if... | waiting for a reply to our request |
def bold(*content, sep=):
return _md(_join(*content, sep=sep), symbols=MD_SYMBOLS[0]) | Make bold text (Markdown)
:param content:
:param sep:
:return: |
def matplotlib_to_ginga_cmap(cm, name=None):
if name is None:
name = cm.name
arr = cm(np.arange(0, min_cmap_len) / np.float(min_cmap_len - 1))
clst = arr[:, 0:3]
return ColorMap(name, clst) | Convert matplotlib colormap to Ginga's. |
def gather_data(registry):
host = socket.gethostname()
ram_metric = Gauge("memory_usage_bytes", "Memory usage in bytes.",
{: host})
cpu_metric = Gauge("cpu_usage_percent", "CPU usage percent.",
{: host})
registry.register(ram_metric)
... | Gathers the metrics |
def status(self, code=None):
if code is None:
return self.response.status_code
log(_("checking that status code {} is returned...").format(code))
if code != self.response.status_code:
raise Failure(_("expected status code {}, but got {}").format(
... | Check status code in response returned by application.
If ``code`` is not None, assert that ``code`` is returned by application,
else simply return the status code.
:param code: ``code`` to assert that application returns
:type code: int
Example usage::
check50.fla... |
def get_ip_interface_output_interface_ip_address_ipv4(self, **kwargs):
config = ET.Element("config")
get_ip_interface = ET.Element("get_ip_interface")
config = get_ip_interface
output = ET.SubElement(get_ip_interface, "output")
interface = ET.SubElement(output, "interfac... | Auto Generated Code |
def rename(args):
old_name, new_name = args.names
add_tags(resources.ec2.Instance(resolve_instance_id(old_name)), Name=new_name, dry_run=args.dry_run) | Supply two names: Existing instance name or ID, and new name to assign to the instance. |
def _check_fact_ref_eval(cls, cpel_dom):
CHECK_SYSTEM = "check-system"
CHECK_LOCATION = "check-location"
CHECK_ID = "check-id"
checksystemID = cpel_dom.getAttribute(CHECK_SYSTEM)
if (checksystemID == "http://oval.mitre.org/XMLSchema/ovaldefinitions-5"):
... | Returns the result (True, False, Error) of performing the specified
check, unless the check isnt supported, in which case it returns
False. Error is a catch-all for all results other than True and
False.
:param string cpel_dom: XML infoset for the check_fact_ref element.
:retur... |
def deserialize(self, data, fields=None):
if not isinstance(data, (bytes, bytearray)):
return data
return msgpack.unpackb(data, encoding=, object_pairs_hook=decode_to_sorted) | Deserializes msgpack bytes to OrderedDict (in the same sorted order as for serialize)
:param data: the data in bytes
:return: sorted OrderedDict |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.