hexsha
stringlengths
40
40
size
int64
10
805k
ext
stringclasses
6 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
4
176
max_stars_repo_name
stringlengths
7
114
max_stars_repo_head_hexsha
stringlengths
40
40
max_stars_repo_licenses
listlengths
1
10
max_stars_count
int64
1
191k
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
max_issues_repo_path
stringlengths
4
176
max_issues_repo_name
stringlengths
7
114
max_issues_repo_head_hexsha
stringlengths
40
40
max_issues_repo_licenses
listlengths
1
10
max_issues_count
int64
1
48.5k
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
max_forks_repo_path
stringlengths
4
176
max_forks_repo_name
stringlengths
7
114
max_forks_repo_head_hexsha
stringlengths
40
40
max_forks_repo_licenses
listlengths
1
10
max_forks_count
int64
1
105k
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
content
stringlengths
10
805k
avg_line_length
float64
5.53
11k
max_line_length
int64
10
129k
alphanum_fraction
float64
0.13
0.93
content_no_comment
stringlengths
0
449k
is_comment_constant_removed
bool
2 classes
is_sharp_comment_removed
bool
1 class
f70e8f1ee71b250ea1a1671298c70cbb6c8b9589
2,981
py
Python
rovers/fastdownward/experiments/issue717/v2.py
mehrdadzakershahrak/Online-Explanation-Generation
e41ad9b5a390abdaf271562a56105c191e33b74d
[ "MIT" ]
1
2021-09-09T13:03:02.000Z
2021-09-09T13:03:02.000Z
rovers/fastdownward/experiments/issue717/v2.py
mehrdadzakershahrak/Online-Explanation-Generation
e41ad9b5a390abdaf271562a56105c191e33b74d
[ "MIT" ]
null
null
null
rovers/fastdownward/experiments/issue717/v2.py
mehrdadzakershahrak/Online-Explanation-Generation
e41ad9b5a390abdaf271562a56105c191e33b74d
[ "MIT" ]
null
null
null
#! /usr/bin/env python # -*- coding: utf-8 -*- import os from lab.environments import LocalEnvironment, MaiaEnvironment from downward.reports.compare import ComparativeReport from common_setup import IssueConfig, IssueExperiment, is_test_run BENCHMARKS_DIR = os.environ["DOWNWARD_BENCHMARKS"] REVISIONS = ["issue717-v2"] CONFIGS = [ IssueConfig( "lama-first-original", [], driver_options=["--alias", "lama-first"]) ] + [ IssueConfig( "lama-first-new", [], driver_options=["--alias", "lama-first-new"]) ] + [ IssueConfig( "lama-original", [], driver_options=["--alias", "seq-sat-lama-2011"]) ] + [ IssueConfig( "lama-new", [], driver_options=["--alias", "seq-sat-lama-2011-new"]) ] SUITE = [ 'airport', 'assembly', 'barman-sat11-strips', 'barman-sat14-strips', 'blocks', 'cavediving-14-adl', 'childsnack-sat14-strips', 'citycar-sat14-adl', 'depot', 'driverlog', 'elevators-sat08-strips', 'elevators-sat11-strips', 'floortile-sat11-strips', 'floortile-sat14-strips', 'freecell', 'ged-sat14-strips', 'grid', 'gripper', 'hiking-sat14-strips', 'logistics00', 'logistics98', 'maintenance-sat14-adl', 'miconic', 'miconic-fulladl', 'miconic-simpleadl', 'movie', 'mprime', 'mystery', 'nomystery-sat11-strips', 'openstacks', 'openstacks-sat08-adl', 'openstacks-sat08-strips', 'openstacks-sat11-strips', 'openstacks-sat14-strips', 'openstacks-strips', 'optical-telegraphs', 'parcprinter-08-strips', 'parcprinter-sat11-strips', 'parking-sat11-strips', 'parking-sat14-strips', 'pathways', 'pathways-noneg', 'pegsol-08-strips', 'pegsol-sat11-strips', 'philosophers', 'pipesworld-notankage', 'pipesworld-tankage', 'psr-large', 'psr-middle', 'psr-small', 'rovers', 'satellite', 'scanalyzer-08-strips', 'scanalyzer-sat11-strips', 'schedule', 'sokoban-sat08-strips', 'sokoban-sat11-strips', 'storage', 'tetris-sat14-strips', 'thoughtful-sat14-strips', 'tidybot-sat11-strips', 'tpp', 'transport-sat08-strips', 'transport-sat11-strips', 'transport-sat14-strips', 'trucks', 'trucks-strips', 'visitall-sat11-strips', 'visitall-sat14-strips', 'woodworking-sat08-strips', 'woodworking-sat11-strips', 'zenotravel'] ENVIRONMENT = MaiaEnvironment( priority=0, email="cedric.geissmann@unibas.ch") if is_test_run(): SUITE = IssueExperiment.DEFAULT_TEST_SUITE ENVIRONMENT = LocalEnvironment(processes=4) exp = IssueExperiment( revisions=REVISIONS, configs=CONFIGS, environment=ENVIRONMENT, ) exp.add_suite(BENCHMARKS_DIR, SUITE) exp.add_absolute_report_step() algorithm_pairs = [ ('issue717-v2-lama-first-original', 'issue717-v2-lama-first-new', 'Diff lama-first'), ('issue717-v2-lama-original', 'issue717-v2-lama-new', 'Diff lama')] exp.add_report(ComparativeReport( algorithm_pairs, attributes=IssueExperiment.DEFAULT_TABLE_ATTRIBUTES)) exp.add_scatter_plot_step(attributes=["total_time", "memory"]) exp.run_steps()
37.734177
89
0.696746
import os from lab.environments import LocalEnvironment, MaiaEnvironment from downward.reports.compare import ComparativeReport from common_setup import IssueConfig, IssueExperiment, is_test_run BENCHMARKS_DIR = os.environ["DOWNWARD_BENCHMARKS"] REVISIONS = ["issue717-v2"] CONFIGS = [ IssueConfig( "lama-first-original", [], driver_options=["--alias", "lama-first"]) ] + [ IssueConfig( "lama-first-new", [], driver_options=["--alias", "lama-first-new"]) ] + [ IssueConfig( "lama-original", [], driver_options=["--alias", "seq-sat-lama-2011"]) ] + [ IssueConfig( "lama-new", [], driver_options=["--alias", "seq-sat-lama-2011-new"]) ] SUITE = [ 'airport', 'assembly', 'barman-sat11-strips', 'barman-sat14-strips', 'blocks', 'cavediving-14-adl', 'childsnack-sat14-strips', 'citycar-sat14-adl', 'depot', 'driverlog', 'elevators-sat08-strips', 'elevators-sat11-strips', 'floortile-sat11-strips', 'floortile-sat14-strips', 'freecell', 'ged-sat14-strips', 'grid', 'gripper', 'hiking-sat14-strips', 'logistics00', 'logistics98', 'maintenance-sat14-adl', 'miconic', 'miconic-fulladl', 'miconic-simpleadl', 'movie', 'mprime', 'mystery', 'nomystery-sat11-strips', 'openstacks', 'openstacks-sat08-adl', 'openstacks-sat08-strips', 'openstacks-sat11-strips', 'openstacks-sat14-strips', 'openstacks-strips', 'optical-telegraphs', 'parcprinter-08-strips', 'parcprinter-sat11-strips', 'parking-sat11-strips', 'parking-sat14-strips', 'pathways', 'pathways-noneg', 'pegsol-08-strips', 'pegsol-sat11-strips', 'philosophers', 'pipesworld-notankage', 'pipesworld-tankage', 'psr-large', 'psr-middle', 'psr-small', 'rovers', 'satellite', 'scanalyzer-08-strips', 'scanalyzer-sat11-strips', 'schedule', 'sokoban-sat08-strips', 'sokoban-sat11-strips', 'storage', 'tetris-sat14-strips', 'thoughtful-sat14-strips', 'tidybot-sat11-strips', 'tpp', 'transport-sat08-strips', 'transport-sat11-strips', 'transport-sat14-strips', 'trucks', 'trucks-strips', 'visitall-sat11-strips', 'visitall-sat14-strips', 'woodworking-sat08-strips', 'woodworking-sat11-strips', 'zenotravel'] ENVIRONMENT = MaiaEnvironment( priority=0, email="cedric.geissmann@unibas.ch") if is_test_run(): SUITE = IssueExperiment.DEFAULT_TEST_SUITE ENVIRONMENT = LocalEnvironment(processes=4) exp = IssueExperiment( revisions=REVISIONS, configs=CONFIGS, environment=ENVIRONMENT, ) exp.add_suite(BENCHMARKS_DIR, SUITE) exp.add_absolute_report_step() algorithm_pairs = [ ('issue717-v2-lama-first-original', 'issue717-v2-lama-first-new', 'Diff lama-first'), ('issue717-v2-lama-original', 'issue717-v2-lama-new', 'Diff lama')] exp.add_report(ComparativeReport( algorithm_pairs, attributes=IssueExperiment.DEFAULT_TABLE_ATTRIBUTES)) exp.add_scatter_plot_step(attributes=["total_time", "memory"]) exp.run_steps()
true
true
f70e8f84b014e650bd66970655feeddf1ff280c7
13,220
py
Python
pyqmc/dmc.py
gcassella/pyqmc
f7a6e1f656c8eab7ebd72132ee980f77275e3876
[ "MIT" ]
null
null
null
pyqmc/dmc.py
gcassella/pyqmc
f7a6e1f656c8eab7ebd72132ee980f77275e3876
[ "MIT" ]
null
null
null
pyqmc/dmc.py
gcassella/pyqmc
f7a6e1f656c8eab7ebd72132ee980f77275e3876
[ "MIT" ]
null
null
null
import os #import numpy as np import pyqmc.mc as mc import sys import h5py import jax import jax.numpy as jnp import numpy as np from functools import partial def limdrift(g, tau, acyrus=0.25): """ Use Cyrus Umrigar's algorithm to limit the drift near nodes. Args: g: a [nconf,ndim] vector tau: time step acyrus: the maximum magnitude Returns: The vector with the cut off applied and multiplied by tau. """ tot = jnp.linalg.norm(g, axis=1) * acyrus mask = tot > 1e-8 taueff = jnp.ones(tot.shape) * tau taueff = jnp.where( mask, (jnp.sqrt(1 + 2 * tau * tot) - 1) / tot, taueff ) return g * taueff[:, jnp.newaxis] def limdrift_cutoff(g, tau, cutoff=1): """ Limit a vector to have a maximum magnitude of cutoff while maintaining direction Args: g: a [nconf,ndim] vector cutoff: the maximum magnitude Returns: The vector with the cut off applied and multiplied by tau. """ return mc.limdrift(g, cutoff) * tau #@partial(jax.jit, static_argnums=(1,9,10,11)) def dmc_step( key, wf, configs, df, weights, tstep, branchcut_start, branchcut_stop, eref, accumulators, ekey, drift_limiter, ): nconfig, nelec = configs.shape[0:2] #wf.recompute(configs) eloc = accumulators[ekey[0]](configs, wf)[ekey[1]].real acc = jnp.zeros(nelec) for e in range(nelec): # Propose move grad = drift_limiter(jnp.real(wf["gradient"](configs, e, configs[:, e]).T), tstep) key, subkey = jax.random.split(key) gauss = jax.random.normal(subkey, (nconfig, 3))*jnp.sqrt(tstep) newepos = configs[:, e, :] + gauss + grad #newepos = configs.make_irreducible(e, eposnew) # Compute reverse move new_grad = drift_limiter(jnp.real(wf["gradient"](configs, e, newepos).T), tstep) forward = jnp.sum(gauss ** 2, axis=1) backward = jnp.sum((gauss + grad + new_grad) ** 2, axis=1) # forward = np.sum((configs[:, e, :] + grad - eposnew) ** 2, axis=1) # backward = np.sum((eposnew + new_grad - configs[:, e, :]) ** 2, axis=1) t_prob = jnp.exp(1 / (2 * tstep) * (forward - backward)) # Acceptance -- fixed-node: reject if wf changes sign wfratio = wf["testvalue"](configs, e, newepos) ratio = jnp.abs(wfratio) ** 2 * t_prob if not wf["iscomplex"]: ratio *= jnp.sign(wfratio) key, subkey = jax.random.split(key) accept = ratio > jax.random.uniform(subkey, (nconfig,)) # Update wave function proposed = jax.ops.index_update( configs, jax.ops.index[:, e, :], newepos ) configs = jnp.where(accept[:, jnp.newaxis, jnp.newaxis], proposed, configs) #wf.updateinternals(e, newepos, mask=accept) acc = jax.ops.index_update( acc, e, jnp.mean(accept) ) # weights energydat = accumulators[ekey[0]](configs, wf) elocnew = energydat[ekey[1]].real tdamp = limit_timestep( weights, elocnew, eloc, eref, branchcut_start, branchcut_stop ) wmult = jnp.exp(-tstep * 0.5 * tdamp * (eloc + elocnew - 2 * eref)) wmult = jnp.where(wmult > 2.0, 2.0, wmult) weights *= wmult wavg = jnp.mean(weights) avg = {} for k, accumulator in accumulators.items(): dat = accumulator(configs, wf) if k != ekey[0] else energydat for m, res in dat.items(): avg[k + m] = jnp.einsum("...i,i...->...", weights, res) / ( nconfig * wavg ) avg["weight"] = wavg avg["acceptance"] = jnp.mean(acc) df.append(avg) return df, configs, weights def dmc_propagate( key, wf, configs, weights, tstep, branchcut_start, branchcut_stop, eref, nsteps=5, accumulators=None, ekey=("energy", "total"), drift_limiter=limdrift, ): """ Propagate DMC without branching Args: wf: A Wave function-like class. recompute(), gradient(), and updateinternals() are used, as well as anything (such as laplacian() ) used by accumulators configs: Configs object, (nconfig, nelec, 3) - initial coordinates to start calculation. weights: (nconfig,) - initial weights to start calculation tstep: Time step for move proposals. Introduces time step error. nsteps: number of DMC steps to take accumulators: A dictionary of functor objects that take in (coords,wf) and return a dictionary of quantities to be averaged. np.mean(quantity,axis=0) should give the average over configurations. If none, a default energy accumulator will be used. ekey: tuple of strings; energy is needed for DMC weights. Access total energy by accumulators[ekey[0]](configs, wf)[ekey[1] drift_limiter: a function that takes a gradient and a cutoff and returns an adjusted gradient Returns: (df,coords,weights) df: A list of dictionaries nstep long that contains all results from the accumulators. coords: The final coordinates from this calculation. weights: The final weights from this calculation """ assert accumulators is not None, "Need an energy accumulator for DMC" df = [] for _ in range(nsteps): key, subkey = jax.random.split(key) df, configs, weights = dmc_step( subkey, wf, configs, df, weights, tstep, branchcut_start, branchcut_stop, eref, accumulators, ekey, drift_limiter, ) df_ret = {} weight = jnp.asarray([d["weight"] for d in df]) avg_weight = weight / jnp.mean(weight) for k in df[0].keys(): df_ret[k] = jnp.mean(jnp.array([d[k] * w for d, w in zip(df, avg_weight)]), axis=0) df_ret["weight"] = jnp.mean(weight) return df_ret, configs, weights def limit_timestep(weights, elocnew, elocold, eref, start, stop): """ Stabilizes weights by scaling down the effective tstep if the local energy is too far from eref. Args: weights: (nconfigs,) array walker weights elocnew: (nconfigs,) array current local energy of each walker elocold: (nconfigs,) array previous local energy of each walker eref: scalar reference energy that fixes normalization start: scalar number of sigmas to start damping tstep stop: scalar number of sigmas where tstep becomes zero Return: tdamp: scalar Damping factor to multiply timestep; always between 0 and 1. The damping factor is 1 if eref-eloc < branchcut_start*sigma, 0 if eref-eloc > branchcut_stop*sigma, decreases linearly inbetween. """ # JAX does not like this kind of stuff! #if start is None or stop is None: # return 1 #assert ( # stop > start #), "stabilize weights requires stop>start. Invalid stop={0}, start={1}".format( # stop, start #) eloc = jnp.stack([elocnew, elocold]) fbet = jnp.amax(eref - eloc, axis=0) return jnp.clip((1 - (fbet - start)) / (stop - start), 0, 1) def branch(key, configs, weights): """ Perform branching on a set of walkers by stochastic reconfiguration Walkers are resampled with probability proportional to the weights, and the new weights are all set to be equal to the average weight. Args: configs: (nconfig,nelec,3) walker coordinates weights: (nconfig,) walker weights Returns: configs: resampled walker configurations weights: (nconfig,) all weights are equal to average weight """ nconfig = configs.shape[0] wtot = jnp.sum(weights) probability = jnp.cumsum(weights / wtot) key, subkey = jax.random.split(key) base = jax.random.uniform(subkey) newinds = jnp.searchsorted(probability, (base + jnp.arange(nconfig) / nconfig) % 1.0) configs = configs[newinds] weights = jnp.ones((nconfig, ))*wtot/nconfig return configs, weights def dmc_file(hdf_file, data, attr, configs, weights): import pyqmc.hdftools as hdftools npdata = jax.tree_util.tree_map(np.asarray, data) if hdf_file is not None: with h5py.File(hdf_file, "a") as hdf: if "configs" not in hdf.keys(): hdftools.setup_hdf(hdf, npdata, attr) hdf.create_dataset( "configs", configs.shape, chunks=True, maxshape=(None, *configs.shape[1:]), ) if "weights" not in hdf.keys(): hdf.create_dataset("weights", weights.shape) hdftools.append_hdf(hdf, npdata) hdf["configs"].resize(configs.shape) hdf["configs"][...] = configs hdf["weights"][:] = weights def rundmc( key, wf, configs, weights=None, tstep=0.01, nsteps=1000, branchtime=5, stepoffset=0, branchcut_start=3, branchcut_stop=6, drift_limiter=limdrift, verbose=False, accumulators=None, ekey=("energy", "total"), propagate=dmc_propagate, feedback=1.0, hdf_file=None, client=None, npartitions=None, **kwargs, ): """ Run DMC Args: wf: A Wave function-like class. recompute(), gradient(), and updateinternals() are used, as well as anything (such as laplacian() ) used by accumulators configs: (nconfig, nelec, 3) - initial coordinates to start calculation. weights: (nconfig,) - initial weights to start calculation, defaults to uniform. nsteps: number of DMC steps to take tstep: Time step for move proposals. Introduces time step error. branchtime: number of steps to take between branching accumulators: A dictionary of functor objects that take in (coords,wf) and return a dictionary of quantities to be averaged. np.mean(quantity,axis=0) should give the average over configurations. If none, a default energy accumulator will be used. ekey: tuple of strings; energy is needed for DMC weights. Access total energy by accumulators[ekey[0]](configs, wf)[ekey[1] verbose: Print out step information drift_limiter: a function that takes a gradient and a cutoff and returns an adjusted gradient stepoffset: If continuing a run, what to start the step numbering at. Returns: (df,coords,weights) df: A list of dictionaries nstep long that contains all results from the accumulators. coords: The final coordinates from this calculation. weights: The final weights from this calculation """ # Restart from HDF file if hdf_file is not None and os.path.isfile(hdf_file): with h5py.File(hdf_file, "r") as hdf: stepoffset = hdf["step"][-1] + 1 configs.load_hdf(hdf) weights = jnp.array(hdf["weights"]) eref = hdf["eref"][-1] esigma = hdf["esigma"][-1] if verbose: print("Restarted calculation") else: warmup = 2 key, subkey = jax.random.split(key) df, configs = mc.vmc( subkey, wf, configs, accumulators=accumulators, client=client, npartitions=npartitions, verbose=verbose, ) en = df[ekey[0] + ekey[1]][warmup:] eref = jnp.mean(en).real esigma = jnp.sqrt(jnp.var(en) * jnp.mean(df["nconfig"])) if verbose: print("eref start", eref, "esigma", esigma) nconfig = configs.shape[0] if weights is None: weights = jnp.ones(nconfig) npropagate = int(jnp.ceil(nsteps / branchtime)) df = [] for step in range(npropagate): key, subkey = jax.random.split(key) df_, configs, weights = dmc_propagate( subkey, wf, configs, weights, tstep, branchcut_start * esigma, branchcut_stop * esigma, eref=eref, nsteps=branchtime, accumulators=accumulators, ekey=ekey, drift_limiter=drift_limiter, **kwargs, ) df_["eref"] = eref df_["step"] = step + stepoffset df_["esigma"] = esigma df_["tstep"] = tstep df_["weight_std"] = jnp.std(weights) df_["nsteps"] = branchtime dmc_file(hdf_file, df_, {}, configs, weights) # print(df_) df.append(df_) eref = df_[ekey[0] + ekey[1]] - feedback * jnp.log(jnp.mean(weights)) key, subkey = jax.random.split(key) configs, weights = branch(subkey, configs, weights) if verbose: print( "energy", df_[ekey[0] + ekey[1]], "eref", df_["eref"], "sigma(w)", df_["weight_std"], ) df_ret = {} for k in df[0].keys(): df_ret[k] = jnp.asarray([d[k] for d in df]) return df_ret, configs, weights
31.327014
252
0.596747
import os import pyqmc.mc as mc import sys import h5py import jax import jax.numpy as jnp import numpy as np from functools import partial def limdrift(g, tau, acyrus=0.25): tot = jnp.linalg.norm(g, axis=1) * acyrus mask = tot > 1e-8 taueff = jnp.ones(tot.shape) * tau taueff = jnp.where( mask, (jnp.sqrt(1 + 2 * tau * tot) - 1) / tot, taueff ) return g * taueff[:, jnp.newaxis] def limdrift_cutoff(g, tau, cutoff=1): return mc.limdrift(g, cutoff) * tau def dmc_step( key, wf, configs, df, weights, tstep, branchcut_start, branchcut_stop, eref, accumulators, ekey, drift_limiter, ): nconfig, nelec = configs.shape[0:2] eloc = accumulators[ekey[0]](configs, wf)[ekey[1]].real acc = jnp.zeros(nelec) for e in range(nelec): grad = drift_limiter(jnp.real(wf["gradient"](configs, e, configs[:, e]).T), tstep) key, subkey = jax.random.split(key) gauss = jax.random.normal(subkey, (nconfig, 3))*jnp.sqrt(tstep) newepos = configs[:, e, :] + gauss + grad new_grad = drift_limiter(jnp.real(wf["gradient"](configs, e, newepos).T), tstep) forward = jnp.sum(gauss ** 2, axis=1) backward = jnp.sum((gauss + grad + new_grad) ** 2, axis=1) t_prob = jnp.exp(1 / (2 * tstep) * (forward - backward)) wfratio = wf["testvalue"](configs, e, newepos) ratio = jnp.abs(wfratio) ** 2 * t_prob if not wf["iscomplex"]: ratio *= jnp.sign(wfratio) key, subkey = jax.random.split(key) accept = ratio > jax.random.uniform(subkey, (nconfig,)) proposed = jax.ops.index_update( configs, jax.ops.index[:, e, :], newepos ) configs = jnp.where(accept[:, jnp.newaxis, jnp.newaxis], proposed, configs) acc = jax.ops.index_update( acc, e, jnp.mean(accept) ) energydat = accumulators[ekey[0]](configs, wf) elocnew = energydat[ekey[1]].real tdamp = limit_timestep( weights, elocnew, eloc, eref, branchcut_start, branchcut_stop ) wmult = jnp.exp(-tstep * 0.5 * tdamp * (eloc + elocnew - 2 * eref)) wmult = jnp.where(wmult > 2.0, 2.0, wmult) weights *= wmult wavg = jnp.mean(weights) avg = {} for k, accumulator in accumulators.items(): dat = accumulator(configs, wf) if k != ekey[0] else energydat for m, res in dat.items(): avg[k + m] = jnp.einsum("...i,i...->...", weights, res) / ( nconfig * wavg ) avg["weight"] = wavg avg["acceptance"] = jnp.mean(acc) df.append(avg) return df, configs, weights def dmc_propagate( key, wf, configs, weights, tstep, branchcut_start, branchcut_stop, eref, nsteps=5, accumulators=None, ekey=("energy", "total"), drift_limiter=limdrift, ): assert accumulators is not None, "Need an energy accumulator for DMC" df = [] for _ in range(nsteps): key, subkey = jax.random.split(key) df, configs, weights = dmc_step( subkey, wf, configs, df, weights, tstep, branchcut_start, branchcut_stop, eref, accumulators, ekey, drift_limiter, ) df_ret = {} weight = jnp.asarray([d["weight"] for d in df]) avg_weight = weight / jnp.mean(weight) for k in df[0].keys(): df_ret[k] = jnp.mean(jnp.array([d[k] * w for d, w in zip(df, avg_weight)]), axis=0) df_ret["weight"] = jnp.mean(weight) return df_ret, configs, weights def limit_timestep(weights, elocnew, elocold, eref, start, stop): eloc = jnp.stack([elocnew, elocold]) fbet = jnp.amax(eref - eloc, axis=0) return jnp.clip((1 - (fbet - start)) / (stop - start), 0, 1) def branch(key, configs, weights): nconfig = configs.shape[0] wtot = jnp.sum(weights) probability = jnp.cumsum(weights / wtot) key, subkey = jax.random.split(key) base = jax.random.uniform(subkey) newinds = jnp.searchsorted(probability, (base + jnp.arange(nconfig) / nconfig) % 1.0) configs = configs[newinds] weights = jnp.ones((nconfig, ))*wtot/nconfig return configs, weights def dmc_file(hdf_file, data, attr, configs, weights): import pyqmc.hdftools as hdftools npdata = jax.tree_util.tree_map(np.asarray, data) if hdf_file is not None: with h5py.File(hdf_file, "a") as hdf: if "configs" not in hdf.keys(): hdftools.setup_hdf(hdf, npdata, attr) hdf.create_dataset( "configs", configs.shape, chunks=True, maxshape=(None, *configs.shape[1:]), ) if "weights" not in hdf.keys(): hdf.create_dataset("weights", weights.shape) hdftools.append_hdf(hdf, npdata) hdf["configs"].resize(configs.shape) hdf["configs"][...] = configs hdf["weights"][:] = weights def rundmc( key, wf, configs, weights=None, tstep=0.01, nsteps=1000, branchtime=5, stepoffset=0, branchcut_start=3, branchcut_stop=6, drift_limiter=limdrift, verbose=False, accumulators=None, ekey=("energy", "total"), propagate=dmc_propagate, feedback=1.0, hdf_file=None, client=None, npartitions=None, **kwargs, ): if hdf_file is not None and os.path.isfile(hdf_file): with h5py.File(hdf_file, "r") as hdf: stepoffset = hdf["step"][-1] + 1 configs.load_hdf(hdf) weights = jnp.array(hdf["weights"]) eref = hdf["eref"][-1] esigma = hdf["esigma"][-1] if verbose: print("Restarted calculation") else: warmup = 2 key, subkey = jax.random.split(key) df, configs = mc.vmc( subkey, wf, configs, accumulators=accumulators, client=client, npartitions=npartitions, verbose=verbose, ) en = df[ekey[0] + ekey[1]][warmup:] eref = jnp.mean(en).real esigma = jnp.sqrt(jnp.var(en) * jnp.mean(df["nconfig"])) if verbose: print("eref start", eref, "esigma", esigma) nconfig = configs.shape[0] if weights is None: weights = jnp.ones(nconfig) npropagate = int(jnp.ceil(nsteps / branchtime)) df = [] for step in range(npropagate): key, subkey = jax.random.split(key) df_, configs, weights = dmc_propagate( subkey, wf, configs, weights, tstep, branchcut_start * esigma, branchcut_stop * esigma, eref=eref, nsteps=branchtime, accumulators=accumulators, ekey=ekey, drift_limiter=drift_limiter, **kwargs, ) df_["eref"] = eref df_["step"] = step + stepoffset df_["esigma"] = esigma df_["tstep"] = tstep df_["weight_std"] = jnp.std(weights) df_["nsteps"] = branchtime dmc_file(hdf_file, df_, {}, configs, weights) df.append(df_) eref = df_[ekey[0] + ekey[1]] - feedback * jnp.log(jnp.mean(weights)) key, subkey = jax.random.split(key) configs, weights = branch(subkey, configs, weights) if verbose: print( "energy", df_[ekey[0] + ekey[1]], "eref", df_["eref"], "sigma(w)", df_["weight_std"], ) df_ret = {} for k in df[0].keys(): df_ret[k] = jnp.asarray([d[k] for d in df]) return df_ret, configs, weights
true
true
f70e8f8ebf9d810a241e46766340eba377a1d983
2,203
py
Python
kata_solutions/programming_101/_5/seach_string.py
jrj92280/python-katas
74cbe1ca110006c99600ddba6e887f5675c76667
[ "Beerware" ]
2
2019-03-29T02:42:21.000Z
2019-08-06T15:01:19.000Z
kata_solutions/programming_101/_5/seach_string.py
jrj92280/python-katas
74cbe1ca110006c99600ddba6e887f5675c76667
[ "Beerware" ]
null
null
null
kata_solutions/programming_101/_5/seach_string.py
jrj92280/python-katas
74cbe1ca110006c99600ddba6e887f5675c76667
[ "Beerware" ]
1
2019-07-17T15:11:13.000Z
2019-07-17T15:11:13.000Z
# PRIMITIVE DATA TYPES # str - string # bool - boolean # int - integer # float # COMPLEX DATA TYPES # list # dict attendees = ['sara', 'alex', 'justin', 'ryan'] for attendee in attendees: # print(attendee) pass # print(attendees[0]) # key = value employees = { 'sara': 'csa', 'alex': 'it stupport tech', 'justin': 'software ninja', 'ryan': 'numbers nerd', 'robot': str(1) } # print(employees['alex']) # print(employees.get('ilya', 'russian spy')) # for employee_id in employees: # print(employee_id + ' - ' + employees[employee_id]) # for key, value in employees.items(): # print(key + ' - ' + value) # CASTING # str + int = runtime exception # SCOPE # white space in Python defines scope # block of code associated with a control structure # def my_method(): # temp = 1 # print(temp) # CONTROL STRUCTURES # --for loops-- # for {variable_name} in <collection>: # <action> # --logical-- # if <bool>: # pass # elif <bool>: # pass # else <bool>: # pass # -- exception handling -- # try <expression>: # <action> # except [error_type]: # <handle error> # try: # employees['iyla'] # except KeyError: # print('call alex to add access') # print('in exception') # except Exception: # print('here') # else: # print('else') # --assignment-- # = # --comparisons-- # == -> equals # != -> not equals # > -> greater than # >= -> greater than equal # < -> less than # <= -> les than equal """ PRACTICE: print each letter in a given string """ name = 'justin' # for char in name: # print(char) """ PRACTICE: create a function that takes an input, then prints each character of the input """ def print_char(input_name): for char in input_name: print(char) # print_char(name) """ PRACTICE: create a function that takes two inputs, then prints True/False whether or not the first input is contained within the second input """ text_value = 'some input' def search_string(search_value, text_value): return search_value in text_value print(search_string('a', text_value)) # False print(search_string('s', text_value)) # True print(search_string('S', text_value)) # False
16.946154
57
0.63005
attendees = ['sara', 'alex', 'justin', 'ryan'] for attendee in attendees: pass employees = { 'sara': 'csa', 'alex': 'it stupport tech', 'justin': 'software ninja', 'ryan': 'numbers nerd', 'robot': str(1) } name = 'justin' def print_char(input_name): for char in input_name: print(char) text_value = 'some input' def search_string(search_value, text_value): return search_value in text_value print(search_string('a', text_value)) print(search_string('s', text_value)) print(search_string('S', text_value))
true
true
f70e910ed88c79aa27e0e316c3f8283ce1ca89cc
332
py
Python
VTK/vtk_7.1.1_x64_Release/lib/python2.7/site-packages/vtk/vtkIOParallelXML.py
jiaguobing/FastCAE
2348ab87e83fe5c704e4c998cf391229c25ac5d5
[ "BSD-3-Clause" ]
2
2020-02-21T01:04:35.000Z
2020-02-21T03:35:37.000Z
VTK/vtk_7.1.1_x64_Release/lib/python2.7/site-packages/vtk/vtkIOParallelXML.py
Sunqia/FastCAE
cbc023fe07b6e306ceefae8b8bd7c12bc1562acb
[ "BSD-3-Clause" ]
1
2020-03-06T04:49:42.000Z
2020-03-06T04:49:42.000Z
VTK/vtk_7.1.1_x64_Release/lib/python2.7/site-packages/vtk/vtkIOParallelXML.py
Sunqia/FastCAE
cbc023fe07b6e306ceefae8b8bd7c12bc1562acb
[ "BSD-3-Clause" ]
1
2021-11-21T13:03:26.000Z
2021-11-21T13:03:26.000Z
from __future__ import absolute_import try: # use relative import for installed modules from .vtkIOParallelXMLPython import * except ImportError: # during build and testing, the modules will be elsewhere, # e.g. in lib directory or Release/Debug config directories from vtkIOParallelXMLPython import *
33.2
64
0.746988
from __future__ import absolute_import try: from .vtkIOParallelXMLPython import * except ImportError: from vtkIOParallelXMLPython import *
true
true
f70e920f44195f22e9f47b7ea1e5c90c3fce06cf
6,854
py
Python
cement/cli/contrib/yaml/reader.py
tomekr/cement
fece8629c48bcd598fd61d8aa7457a5df4c4f831
[ "BSD-3-Clause" ]
5,421
2018-09-24T08:04:06.000Z
2022-03-31T20:02:37.000Z
cement/cli/contrib/yaml/reader.py
tomekr/cement
fece8629c48bcd598fd61d8aa7457a5df4c4f831
[ "BSD-3-Clause" ]
3,243
2017-02-07T15:30:01.000Z
2022-03-31T16:42:19.000Z
cement/cli/contrib/yaml/reader.py
tomekr/cement
fece8629c48bcd598fd61d8aa7457a5df4c4f831
[ "BSD-3-Clause" ]
429
2015-06-27T19:34:21.000Z
2022-03-23T11:02:51.000Z
# This module contains abstractions for the input stream. You don't have to # looks further, there are no pretty code. # # We define two classes here. # # Mark(source, line, column) # It's just a record and its only use is producing nice error messages. # Parser does not use it for any other purposes. # # Reader(source, data) # Reader determines the encoding of `data` and converts it to unicode. # Reader provides the following methods and attributes: # reader.peek(length=1) - return the next `length` characters # reader.forward(length=1) - move the current position to `length` characters. # reader.index - the number of the current character. # reader.line, stream.column - the line and the column of the current character. __all__ = ['Reader', 'ReaderError'] from .error import YAMLError, Mark import codecs, re class ReaderError(YAMLError): def __init__(self, name, position, character, encoding, reason): self.name = name self.character = character self.position = position self.encoding = encoding self.reason = reason def __str__(self): if isinstance(self.character, bytes): return "'%s' codec can't decode byte #x%02x: %s\n" \ " in \"%s\", position %d" \ % (self.encoding, ord(self.character), self.reason, self.name, self.position) else: return "unacceptable character #x%04x: %s\n" \ " in \"%s\", position %d" \ % (self.character, self.reason, self.name, self.position) class Reader(object): # Reader: # - determines the data encoding and converts it to a unicode string, # - checks if characters are in allowed range, # - adds '\0' to the end. # Reader accepts # - a `bytes` object, # - a `str` object, # - a file-like object with its `read` method returning `str`, # - a file-like object with its `read` method returning `unicode`. # Yeah, it's ugly and slow. def __init__(self, stream): self.name = None self.stream = None self.stream_pointer = 0 self.eof = True self.buffer = '' self.pointer = 0 self.raw_buffer = None self.raw_decode = None self.encoding = None self.index = 0 self.line = 0 self.column = 0 if isinstance(stream, str): self.name = "<unicode string>" self.check_printable(stream) self.buffer = stream+'\0' elif isinstance(stream, bytes): self.name = "<byte string>" self.raw_buffer = stream self.determine_encoding() else: self.stream = stream self.name = getattr(stream, 'name', "<file>") self.eof = False self.raw_buffer = None self.determine_encoding() def peek(self, index=0): try: return self.buffer[self.pointer+index] except IndexError: self.update(index+1) return self.buffer[self.pointer+index] def prefix(self, length=1): if self.pointer+length >= len(self.buffer): self.update(length) return self.buffer[self.pointer:self.pointer+length] def forward(self, length=1): if self.pointer+length+1 >= len(self.buffer): self.update(length+1) while length: ch = self.buffer[self.pointer] self.pointer += 1 self.index += 1 if ch in '\n\x85\u2028\u2029' \ or (ch == '\r' and self.buffer[self.pointer] != '\n'): self.line += 1 self.column = 0 elif ch != '\uFEFF': self.column += 1 length -= 1 def get_mark(self): if self.stream is None: return Mark(self.name, self.index, self.line, self.column, self.buffer, self.pointer) else: return Mark(self.name, self.index, self.line, self.column, None, None) def determine_encoding(self): while not self.eof and (self.raw_buffer is None or len(self.raw_buffer) < 2): self.update_raw() if isinstance(self.raw_buffer, bytes): if self.raw_buffer.startswith(codecs.BOM_UTF16_LE): self.raw_decode = codecs.utf_16_le_decode self.encoding = 'utf-16-le' elif self.raw_buffer.startswith(codecs.BOM_UTF16_BE): self.raw_decode = codecs.utf_16_be_decode self.encoding = 'utf-16-be' else: self.raw_decode = codecs.utf_8_decode self.encoding = 'utf-8' self.update(1) NON_PRINTABLE = re.compile('[^\x09\x0A\x0D\x20-\x7E\x85\xA0-\uD7FF\uE000-\uFFFD]') def check_printable(self, data): match = self.NON_PRINTABLE.search(data) if match: character = match.group() position = self.index+(len(self.buffer)-self.pointer)+match.start() raise ReaderError(self.name, position, ord(character), 'unicode', "special characters are not allowed") def update(self, length): if self.raw_buffer is None: return self.buffer = self.buffer[self.pointer:] self.pointer = 0 while len(self.buffer) < length: if not self.eof: self.update_raw() if self.raw_decode is not None: try: data, converted = self.raw_decode(self.raw_buffer, 'strict', self.eof) except UnicodeDecodeError as exc: character = self.raw_buffer[exc.start] if self.stream is not None: position = self.stream_pointer-len(self.raw_buffer)+exc.start else: position = exc.start raise ReaderError(self.name, position, character, exc.encoding, exc.reason) else: data = self.raw_buffer converted = len(data) self.check_printable(data) self.buffer += data self.raw_buffer = self.raw_buffer[converted:] if self.eof: self.buffer += '\0' self.raw_buffer = None break def update_raw(self, size=4096): data = self.stream.read(size) if self.raw_buffer is None: self.raw_buffer = data else: self.raw_buffer += data self.stream_pointer += len(data) if not data: self.eof = True #try: # import psyco # psyco.bind(Reader) #except ImportError: # pass
35.512953
86
0.556026
# looks further, there are no pretty code. # # We define two classes here. # # Mark(source, line, column) # It's just a record and its only use is producing nice error messages. __all__ = ['Reader', 'ReaderError'] from .error import YAMLError, Mark import codecs, re class ReaderError(YAMLError): def __init__(self, name, position, character, encoding, reason): self.name = name self.character = character self.position = position self.encoding = encoding self.reason = reason def __str__(self): if isinstance(self.character, bytes): return "'%s' codec can't decode byte #x%02x: %s\n" \ " in \"%s\", position %d" \ % (self.encoding, ord(self.character), self.reason, self.name, self.position) else: return "unacceptable character #x%04x: %s\n" \ " in \"%s\", position %d" \ % (self.character, self.reason, self.name, self.position) class Reader(object): # Reader: # - determines the data encoding and converts it to a unicode string, # - checks if characters are in allowed range, # - adds '\0' to the end. # Reader accepts # - a `bytes` object, # - a `str` object, # - a file-like object with its `read` method returning `str`, # - a file-like object with its `read` method returning `unicode`. # Yeah, it's ugly and slow. def __init__(self, stream): self.name = None self.stream = None self.stream_pointer = 0 self.eof = True self.buffer = '' self.pointer = 0 self.raw_buffer = None self.raw_decode = None self.encoding = None self.index = 0 self.line = 0 self.column = 0 if isinstance(stream, str): self.name = "<unicode string>" self.check_printable(stream) self.buffer = stream+'\0' elif isinstance(stream, bytes): self.name = "<byte string>" self.raw_buffer = stream self.determine_encoding() else: self.stream = stream self.name = getattr(stream, 'name', "<file>") self.eof = False self.raw_buffer = None self.determine_encoding() def peek(self, index=0): try: return self.buffer[self.pointer+index] except IndexError: self.update(index+1) return self.buffer[self.pointer+index] def prefix(self, length=1): if self.pointer+length >= len(self.buffer): self.update(length) return self.buffer[self.pointer:self.pointer+length] def forward(self, length=1): if self.pointer+length+1 >= len(self.buffer): self.update(length+1) while length: ch = self.buffer[self.pointer] self.pointer += 1 self.index += 1 if ch in '\n\x85\u2028\u2029' \ or (ch == '\r' and self.buffer[self.pointer] != '\n'): self.line += 1 self.column = 0 elif ch != '\uFEFF': self.column += 1 length -= 1 def get_mark(self): if self.stream is None: return Mark(self.name, self.index, self.line, self.column, self.buffer, self.pointer) else: return Mark(self.name, self.index, self.line, self.column, None, None) def determine_encoding(self): while not self.eof and (self.raw_buffer is None or len(self.raw_buffer) < 2): self.update_raw() if isinstance(self.raw_buffer, bytes): if self.raw_buffer.startswith(codecs.BOM_UTF16_LE): self.raw_decode = codecs.utf_16_le_decode self.encoding = 'utf-16-le' elif self.raw_buffer.startswith(codecs.BOM_UTF16_BE): self.raw_decode = codecs.utf_16_be_decode self.encoding = 'utf-16-be' else: self.raw_decode = codecs.utf_8_decode self.encoding = 'utf-8' self.update(1) NON_PRINTABLE = re.compile('[^\x09\x0A\x0D\x20-\x7E\x85\xA0-\uD7FF\uE000-\uFFFD]') def check_printable(self, data): match = self.NON_PRINTABLE.search(data) if match: character = match.group() position = self.index+(len(self.buffer)-self.pointer)+match.start() raise ReaderError(self.name, position, ord(character), 'unicode', "special characters are not allowed") def update(self, length): if self.raw_buffer is None: return self.buffer = self.buffer[self.pointer:] self.pointer = 0 while len(self.buffer) < length: if not self.eof: self.update_raw() if self.raw_decode is not None: try: data, converted = self.raw_decode(self.raw_buffer, 'strict', self.eof) except UnicodeDecodeError as exc: character = self.raw_buffer[exc.start] if self.stream is not None: position = self.stream_pointer-len(self.raw_buffer)+exc.start else: position = exc.start raise ReaderError(self.name, position, character, exc.encoding, exc.reason) else: data = self.raw_buffer converted = len(data) self.check_printable(data) self.buffer += data self.raw_buffer = self.raw_buffer[converted:] if self.eof: self.buffer += '\0' self.raw_buffer = None break def update_raw(self, size=4096): data = self.stream.read(size) if self.raw_buffer is None: self.raw_buffer = data else: self.raw_buffer += data self.stream_pointer += len(data) if not data: self.eof = True
true
true
f70e9236da0f5a31f9d5918d8f43bff0313a10a6
8,037
py
Python
posthog/api/paths.py
almog27/posthog
705d12f78b73c680b3ebc125cbc5b46c53e24102
[ "MIT" ]
1
2020-07-02T12:25:41.000Z
2020-07-02T12:25:41.000Z
posthog/api/paths.py
almog27/posthog
705d12f78b73c680b3ebc125cbc5b46c53e24102
[ "MIT" ]
null
null
null
posthog/api/paths.py
almog27/posthog
705d12f78b73c680b3ebc125cbc5b46c53e24102
[ "MIT" ]
1
2020-06-24T07:59:41.000Z
2020-06-24T07:59:41.000Z
from rest_framework import viewsets, request from rest_framework.response import Response from rest_framework.decorators import action from posthog.models import Event, Filter from posthog.utils import request_to_date_query, dict_from_cursor_fetchall from django.db.models import OuterRef from django.db import connection from typing import Optional from django.db.models.expressions import Window from django.db.models.functions import Lag from django.db.models import F, Q from django.db import connection import json # At the moment, paths don't support users changing distinct_ids midway through. # See: https://github.com/PostHog/posthog/issues/185 class PathsViewSet(viewsets.ViewSet): def _event_subquery(self, event: str, key: str): return Event.objects.filter(pk=OuterRef(event)).values(key)[:1] def _determine_path_type(self, request): requested_type = request.GET.get('type', None) # Default event: Optional[str] = "$pageview" event_filter = {"event":event} path_type = "properties->> \'$current_url\'" start_comparator = "{} ~".format(path_type) # determine requested type if requested_type: if requested_type == "$screen": event = "$screen" event_filter = {"event":event} path_type = "properties->> \'$screen_name\'" start_comparator = "{} ~".format(path_type) elif requested_type == "$autocapture": event = "$autocapture" event_filter = {"event":event} path_type = "tag_name_source" start_comparator = "group_id =" elif requested_type == "custom_event": event = None event_filter = {} path_type = "event" start_comparator = "event =" return event, path_type, event_filter, start_comparator @action(methods=['GET'], detail=False) def elements(self, request: request.Request): team = request.user.team_set.get() all_events = Event.objects.filter(team=team, event="$autocapture") all_events_SQL, sql_params = all_events.query.sql_with_params() elements_readble = '\ SELECT tag_name_source as name, group_id as id FROM (SELECT \'<\' || e."tag_name" || \'> \' || e."text" as tag_name_source, e."text" as text_source, e.group_id FROM "posthog_element" e\ JOIN ( SELECT group_id, MIN("posthog_element"."order") as minOrder FROM "posthog_element" GROUP BY group_id) e2 ON e.order = e2.minOrder AND e.group_id = e2.group_id) as element\ JOIN (SELECT id, hash, count FROM posthog_elementgroup as g JOIN (SELECT count(*), elements_hash from ({}) as a group by elements_hash) as e on g.hash = e.elements_hash) as outer_group ON element.group_id = outer_group.id where text_source <> \'\' order by count DESC limit 20\ '.format(all_events_SQL) cursor = connection.cursor() cursor.execute(elements_readble, sql_params) rows = dict_from_cursor_fetchall(cursor) return Response(rows) def _apply_start_point(self, start_comparator: str, query_string: str, start_point:str) -> str: marked = '\ SELECT *, CASE WHEN {} \'{}\' THEN timestamp ELSE NULL END as mark from ({}) as sessionified\ '.format(start_comparator, start_point, query_string) marked_plus = '\ SELECT *, MIN(mark) OVER (\ PARTITION BY distinct_id\ , session ORDER BY timestamp\ ) AS max from ({}) as marked order by session\ '.format(marked) sessionified = '\ SELECT * FROM ({}) as something where timestamp >= max \ '.format(marked_plus) return sessionified def _add_elements(self, query_string: str) -> str: element = 'SELECT \'<\'|| e."tag_name" || \'> \' || e."text" as tag_name_source, e."text" as text_source FROM "posthog_element" e JOIN \ ( SELECT group_id, MIN("posthog_element"."order") as minOrder FROM "posthog_element" GROUP BY group_id) e2 ON e.order = e2.minOrder AND e.group_id = e2.group_id where e.group_id = v2.group_id' element_group = 'SELECT g."id" as group_id FROM "posthog_elementgroup" g where v1."elements_hash" = g."hash"' sessions_sql = 'SELECT * FROM ({}) as v1 JOIN LATERAL ({}) as v2 on true JOIN LATERAL ({}) as v3 on true'.format(query_string, element_group, element) return sessions_sql # FIXME: Timestamp is timezone aware timestamp, date range uses naive date. # To avoid unexpected results should convert date range to timestamps with timezone. def list(self, request): team = request.user.team_set.get() resp = [] date_query = request_to_date_query(request.GET, exact=False) event, path_type, event_filter, start_comparator = self._determine_path_type(request) properties = request.GET.get('properties') start_point = request.GET.get('start') sessions = Event.objects.add_person_id(team.pk).filter( team=team, **(event_filter), **date_query )\ .filter(~Q(event__in=['$autocapture', '$pageview', '$identify', '$pageleave']) if event is None else Q())\ .filter(Filter(data={'properties': json.loads(properties)}).properties_to_Q(team_id=team.pk) if properties else Q())\ .annotate(previous_timestamp=Window( expression=Lag('timestamp', default=None), partition_by=F('distinct_id'), order_by=F('timestamp').asc() )) sessions_sql, sessions_sql_params = sessions.query.sql_with_params() if event == "$autocapture": sessions_sql = self._add_elements(query_string=sessions_sql) events_notated = '\ SELECT *, CASE WHEN EXTRACT(\'EPOCH\' FROM (timestamp - previous_timestamp)) >= (60 * 30) OR previous_timestamp IS NULL THEN 1 ELSE 0 END AS new_session\ FROM ({}) AS inner_sessions\ '.format(sessions_sql) sessionified = '\ SELECT events_notated.*, SUM(new_session) OVER (\ ORDER BY distinct_id\ ,timestamp\ ) AS session\ FROM ({}) as events_notated\ '.format(events_notated) if start_point: sessionified = self._apply_start_point(start_comparator=start_comparator, query_string=sessionified, start_point=start_point) final = '\ SELECT {} as path_type, id, sessionified.session\ ,ROW_NUMBER() OVER (\ PARTITION BY distinct_id\ ,session ORDER BY timestamp\ ) AS event_number\ FROM ({}) as sessionified\ '.format(path_type, sessionified) counts = '\ SELECT event_number || \'_\' || path_type as target_event, id as target_id, LAG(event_number || \'_\' || path_type, 1) OVER (\ PARTITION BY session\ ) AS source_event , LAG(id, 1) OVER (\ PARTITION BY session\ ) AS source_id from \ ({}) as final\ where event_number <= 4\ '.format(final) cursor = connection.cursor() cursor.execute('\ SELECT source_event, target_event, MAX(target_id), MAX(source_id), count(*) from ({}) as counts\ where source_event is not null and target_event is not null\ group by source_event, target_event order by count desc limit 20\ '.format(counts), sessions_sql_params) rows = cursor.fetchall() for row in rows: resp.append({ 'source': row[0], 'target': row[1], 'target_id': row[2], 'source_id': row[3], 'value': row[4] }) resp = sorted(resp, key=lambda x: x['value'], reverse=True) return Response(resp)
45.925714
295
0.616026
from rest_framework import viewsets, request from rest_framework.response import Response from rest_framework.decorators import action from posthog.models import Event, Filter from posthog.utils import request_to_date_query, dict_from_cursor_fetchall from django.db.models import OuterRef from django.db import connection from typing import Optional from django.db.models.expressions import Window from django.db.models.functions import Lag from django.db.models import F, Q from django.db import connection import json # See: https://github.com/PostHog/posthog/issues/185 class PathsViewSet(viewsets.ViewSet): def _event_subquery(self, event: str, key: str): return Event.objects.filter(pk=OuterRef(event)).values(key)[:1] def _determine_path_type(self, request): requested_type = request.GET.get('type', None) # Default event: Optional[str] = "$pageview" event_filter = {"event":event} path_type = "properties->> \'$current_url\'" start_comparator = "{} ~".format(path_type) # determine requested type if requested_type: if requested_type == "$screen": event = "$screen" event_filter = {"event":event} path_type = "properties->> \'$screen_name\'" start_comparator = "{} ~".format(path_type) elif requested_type == "$autocapture": event = "$autocapture" event_filter = {"event":event} path_type = "tag_name_source" start_comparator = "group_id =" elif requested_type == "custom_event": event = None event_filter = {} path_type = "event" start_comparator = "event =" return event, path_type, event_filter, start_comparator @action(methods=['GET'], detail=False) def elements(self, request: request.Request): team = request.user.team_set.get() all_events = Event.objects.filter(team=team, event="$autocapture") all_events_SQL, sql_params = all_events.query.sql_with_params() elements_readble = '\ SELECT tag_name_source as name, group_id as id FROM (SELECT \'<\' || e."tag_name" || \'> \' || e."text" as tag_name_source, e."text" as text_source, e.group_id FROM "posthog_element" e\ JOIN ( SELECT group_id, MIN("posthog_element"."order") as minOrder FROM "posthog_element" GROUP BY group_id) e2 ON e.order = e2.minOrder AND e.group_id = e2.group_id) as element\ JOIN (SELECT id, hash, count FROM posthog_elementgroup as g JOIN (SELECT count(*), elements_hash from ({}) as a group by elements_hash) as e on g.hash = e.elements_hash) as outer_group ON element.group_id = outer_group.id where text_source <> \'\' order by count DESC limit 20\ '.format(all_events_SQL) cursor = connection.cursor() cursor.execute(elements_readble, sql_params) rows = dict_from_cursor_fetchall(cursor) return Response(rows) def _apply_start_point(self, start_comparator: str, query_string: str, start_point:str) -> str: marked = '\ SELECT *, CASE WHEN {} \'{}\' THEN timestamp ELSE NULL END as mark from ({}) as sessionified\ '.format(start_comparator, start_point, query_string) marked_plus = '\ SELECT *, MIN(mark) OVER (\ PARTITION BY distinct_id\ , session ORDER BY timestamp\ ) AS max from ({}) as marked order by session\ '.format(marked) sessionified = '\ SELECT * FROM ({}) as something where timestamp >= max \ '.format(marked_plus) return sessionified def _add_elements(self, query_string: str) -> str: element = 'SELECT \'<\'|| e."tag_name" || \'> \' || e."text" as tag_name_source, e."text" as text_source FROM "posthog_element" e JOIN \ ( SELECT group_id, MIN("posthog_element"."order") as minOrder FROM "posthog_element" GROUP BY group_id) e2 ON e.order = e2.minOrder AND e.group_id = e2.group_id where e.group_id = v2.group_id' element_group = 'SELECT g."id" as group_id FROM "posthog_elementgroup" g where v1."elements_hash" = g."hash"' sessions_sql = 'SELECT * FROM ({}) as v1 JOIN LATERAL ({}) as v2 on true JOIN LATERAL ({}) as v3 on true'.format(query_string, element_group, element) return sessions_sql # FIXME: Timestamp is timezone aware timestamp, date range uses naive date. # To avoid unexpected results should convert date range to timestamps with timezone. def list(self, request): team = request.user.team_set.get() resp = [] date_query = request_to_date_query(request.GET, exact=False) event, path_type, event_filter, start_comparator = self._determine_path_type(request) properties = request.GET.get('properties') start_point = request.GET.get('start') sessions = Event.objects.add_person_id(team.pk).filter( team=team, **(event_filter), **date_query )\ .filter(~Q(event__in=['$autocapture', '$pageview', '$identify', '$pageleave']) if event is None else Q())\ .filter(Filter(data={'properties': json.loads(properties)}).properties_to_Q(team_id=team.pk) if properties else Q())\ .annotate(previous_timestamp=Window( expression=Lag('timestamp', default=None), partition_by=F('distinct_id'), order_by=F('timestamp').asc() )) sessions_sql, sessions_sql_params = sessions.query.sql_with_params() if event == "$autocapture": sessions_sql = self._add_elements(query_string=sessions_sql) events_notated = '\ SELECT *, CASE WHEN EXTRACT(\'EPOCH\' FROM (timestamp - previous_timestamp)) >= (60 * 30) OR previous_timestamp IS NULL THEN 1 ELSE 0 END AS new_session\ FROM ({}) AS inner_sessions\ '.format(sessions_sql) sessionified = '\ SELECT events_notated.*, SUM(new_session) OVER (\ ORDER BY distinct_id\ ,timestamp\ ) AS session\ FROM ({}) as events_notated\ '.format(events_notated) if start_point: sessionified = self._apply_start_point(start_comparator=start_comparator, query_string=sessionified, start_point=start_point) final = '\ SELECT {} as path_type, id, sessionified.session\ ,ROW_NUMBER() OVER (\ PARTITION BY distinct_id\ ,session ORDER BY timestamp\ ) AS event_number\ FROM ({}) as sessionified\ '.format(path_type, sessionified) counts = '\ SELECT event_number || \'_\' || path_type as target_event, id as target_id, LAG(event_number || \'_\' || path_type, 1) OVER (\ PARTITION BY session\ ) AS source_event , LAG(id, 1) OVER (\ PARTITION BY session\ ) AS source_id from \ ({}) as final\ where event_number <= 4\ '.format(final) cursor = connection.cursor() cursor.execute('\ SELECT source_event, target_event, MAX(target_id), MAX(source_id), count(*) from ({}) as counts\ where source_event is not null and target_event is not null\ group by source_event, target_event order by count desc limit 20\ '.format(counts), sessions_sql_params) rows = cursor.fetchall() for row in rows: resp.append({ 'source': row[0], 'target': row[1], 'target_id': row[2], 'source_id': row[3], 'value': row[4] }) resp = sorted(resp, key=lambda x: x['value'], reverse=True) return Response(resp)
true
true
f70e9242f485b280107689f59ae366e11677da84
366
py
Python
maldi-learn/maldi_learn/preprocessing/__init__.py
sebastianbalzer/maldi_PIKE
032bfcf1cd9b482ff851d68faaa2cf967aaf62a8
[ "BSD-3-Clause" ]
4
2021-01-05T20:26:32.000Z
2022-02-18T16:31:01.000Z
maldi-learn/maldi_learn/preprocessing/__init__.py
sebastianbalzer/maldi_PIKE
032bfcf1cd9b482ff851d68faaa2cf967aaf62a8
[ "BSD-3-Clause" ]
null
null
null
maldi-learn/maldi_learn/preprocessing/__init__.py
sebastianbalzer/maldi_PIKE
032bfcf1cd9b482ff851d68faaa2cf967aaf62a8
[ "BSD-3-Clause" ]
3
2020-04-23T15:56:50.000Z
2021-07-15T00:43:35.000Z
"""Preprocessing of MALDI-TOF spectra.""" from .generic import SubsetPeaksTransformer from .normalization import TotalIonCurrentNormalizer from .normalization import ScaleNormalizer from .topological import TopologicalPeakFiltering __all__ = [ 'ScaleNormalizer', 'SubsetPeaksTransformer', 'TopologicalPeakFiltering', 'TotalIonCurrentNormalizer' ]
24.4
52
0.800546
from .generic import SubsetPeaksTransformer from .normalization import TotalIonCurrentNormalizer from .normalization import ScaleNormalizer from .topological import TopologicalPeakFiltering __all__ = [ 'ScaleNormalizer', 'SubsetPeaksTransformer', 'TopologicalPeakFiltering', 'TotalIonCurrentNormalizer' ]
true
true
f70e92de59d7f2269db39f2876140d100f76b4fc
1,794
py
Python
aiida_quantumespresso/cli/calculations/q2r.py
lin-cp/aiida-quantumespresso
55f2bc8c137a69be24709a119bc285c700997907
[ "MIT" ]
null
null
null
aiida_quantumespresso/cli/calculations/q2r.py
lin-cp/aiida-quantumespresso
55f2bc8c137a69be24709a119bc285c700997907
[ "MIT" ]
null
null
null
aiida_quantumespresso/cli/calculations/q2r.py
lin-cp/aiida-quantumespresso
55f2bc8c137a69be24709a119bc285c700997907
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- """Command line scripts to launch a `Q2rCalculation` for testing and demonstration purposes.""" from aiida.cmdline.params import options as options_core from aiida.cmdline.params import types from aiida.cmdline.utils import decorators import click from . import cmd_launch from ..utils import launch, options @cmd_launch.command('q2r') @options_core.CODE(required=True, type=types.CodeParamType(entry_point='quantumespresso.q2r')) @options_core.CALCULATION(required=True) @options.MAX_NUM_MACHINES() @options.MAX_WALLCLOCK_SECONDS() @options.WITH_MPI() @options.DAEMON() @decorators.with_dbenv() def launch_calculation(code, calculation, max_num_machines, max_wallclock_seconds, with_mpi, daemon): """Run a Q2rCalculation.""" from aiida.plugins import CalculationFactory from aiida_quantumespresso.utils.resources import get_default_options # Check that the parent calculation node comes from quantumespresso.ph. # I cannot move this check into the option declaration, because CalcJobNode is not subclassed by the specific # calculation plugins (only Process is), and there is no feature yet to filter by the associated process_type. expected_process_type = 'aiida.calculations:quantumespresso.ph' if calculation.process_type != expected_process_type: raise click.BadParameter( f'input calculation node has process_type: {calculation.process_type}; should be {expected_process_type}' ) inputs = { 'code': code, 'parent_folder': calculation.outputs.remote_folder, 'metadata': { 'options': get_default_options(max_num_machines, max_wallclock_seconds, with_mpi), } } launch.launch_process(CalculationFactory('quantumespresso.q2r'), daemon, **inputs)
40.772727
117
0.756968
from aiida.cmdline.params import options as options_core from aiida.cmdline.params import types from aiida.cmdline.utils import decorators import click from . import cmd_launch from ..utils import launch, options @cmd_launch.command('q2r') @options_core.CODE(required=True, type=types.CodeParamType(entry_point='quantumespresso.q2r')) @options_core.CALCULATION(required=True) @options.MAX_NUM_MACHINES() @options.MAX_WALLCLOCK_SECONDS() @options.WITH_MPI() @options.DAEMON() @decorators.with_dbenv() def launch_calculation(code, calculation, max_num_machines, max_wallclock_seconds, with_mpi, daemon): from aiida.plugins import CalculationFactory from aiida_quantumespresso.utils.resources import get_default_options expected_process_type = 'aiida.calculations:quantumespresso.ph' if calculation.process_type != expected_process_type: raise click.BadParameter( f'input calculation node has process_type: {calculation.process_type}; should be {expected_process_type}' ) inputs = { 'code': code, 'parent_folder': calculation.outputs.remote_folder, 'metadata': { 'options': get_default_options(max_num_machines, max_wallclock_seconds, with_mpi), } } launch.launch_process(CalculationFactory('quantumespresso.q2r'), daemon, **inputs)
true
true
f70e93de17fb2ce6b4647d52aec3b4fba78bd341
13,265
py
Python
test_tracetools/test/test_pub_sub.py
Henrycious/ros2_tracing
d201ce4b4720ad71c922d5dfa30183631d6f3597
[ "Apache-2.0" ]
null
null
null
test_tracetools/test/test_pub_sub.py
Henrycious/ros2_tracing
d201ce4b4720ad71c922d5dfa30183631d6f3597
[ "Apache-2.0" ]
null
null
null
test_tracetools/test/test_pub_sub.py
Henrycious/ros2_tracing
d201ce4b4720ad71c922d5dfa30183631d6f3597
[ "Apache-2.0" ]
null
null
null
# Copyright 2021 Christophe Bedard # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import unittest from launch.actions import SetEnvironmentVariable from tracetools_test.case import TraceTestCase from tracetools_trace.tools import tracepoints as tp class TestPubSub(TraceTestCase): def __init__(self, *args) -> None: super().__init__( *args, session_name_prefix='session-test-pub-sub', events_ros=[ tp.rmw_publisher_init, tp.rcl_publisher_init, tp.rmw_publish, tp.rcl_publish, tp.rclcpp_publish, tp.rmw_subscription_init, tp.rcl_subscription_init, tp.rclcpp_subscription_init, tp.rclcpp_subscription_callback_added, tp.callback_start, tp.callback_end, ], package='test_tracetools', nodes=['test_ping', 'test_pong'], # Need rmw_cyclonedds_cpp for the rmw instrumentation additional_actions=SetEnvironmentVariable('RMW_IMPLEMENTATION', 'rmw_cyclonedds_cpp'), ) def test_all(self): # Check events as set self.assertEventsSet(self._events_ros) # Get publisher init events & publisher handles of test topics rmw_pub_init_events = self.get_events_with_name(tp.rmw_publisher_init) rmw_sub_init_events = self.get_events_with_name(tp.rmw_subscription_init) publisher_init_events = self.get_events_with_name(tp.rcl_publisher_init) ping_publisher_init_events = self.get_events_with_field_value( 'topic_name', '/ping', publisher_init_events, ) pong_publisher_init_events = self.get_events_with_field_value( 'topic_name', '/pong', publisher_init_events, ) self.assertNumEventsEqual(ping_publisher_init_events, 1) self.assertNumEventsEqual(pong_publisher_init_events, 1) ping_publisher_init_event = ping_publisher_init_events[0] pong_publisher_init_event = pong_publisher_init_events[0] ping_pub_handle = self.get_field(ping_publisher_init_event, 'publisher_handle') ping_rmw_pub_handle = self.get_field(ping_publisher_init_event, 'rmw_publisher_handle') pong_pub_handle = self.get_field(pong_publisher_init_event, 'publisher_handle') pong_rmw_pub_handle = self.get_field(pong_publisher_init_event, 'rmw_publisher_handle') # Find corresponding rmw_pub_init events ping_rmw_pub_init_events = self.get_events_with_field_value( 'rmw_publisher_handle', ping_rmw_pub_handle, rmw_pub_init_events, ) pong_rmw_pub_init_events = self.get_events_with_field_value( 'rmw_publisher_handle', pong_rmw_pub_handle, rmw_pub_init_events, ) self.assertNumEventsEqual(ping_rmw_pub_init_events, 1) self.assertNumEventsEqual(pong_rmw_pub_init_events, 1) ping_rmw_pub_init_event = ping_rmw_pub_init_events[0] pong_rmw_pub_init_event = pong_rmw_pub_init_events[0] # Check publisher init order (rmw then rcl) self.assertEventOrder([ ping_rmw_pub_init_event, ping_publisher_init_event, ]) self.assertEventOrder([ pong_rmw_pub_init_event, pong_publisher_init_event, ]) # Get corresponding rmw/rcl/rclcpp publish events for ping & pong rcl_publish_events = self.get_events_with_name(tp.rcl_publish) ping_rcl_pub_events = self.get_events_with_field_value( 'publisher_handle', ping_pub_handle, rcl_publish_events, ) pong_rcl_pub_events = self.get_events_with_field_value( 'publisher_handle', pong_pub_handle, rcl_publish_events, ) self.assertNumEventsEqual(ping_rcl_pub_events, 1) self.assertNumEventsEqual(pong_rcl_pub_events, 1) ping_rcl_pub_event = ping_rcl_pub_events[0] pong_rcl_pub_event = pong_rcl_pub_events[0] rclcpp_publish_events = self.get_events_with_name(tp.rclcpp_publish) rmw_publish_events = self.get_events_with_name(tp.rmw_publish) ping_pub_message = self.get_field(ping_rcl_pub_event, 'message') pong_pub_message = self.get_field(pong_rcl_pub_event, 'message') ping_rclcpp_pub_events = self.get_events_with_field_value( 'message', ping_pub_message, rclcpp_publish_events, ) pong_rclcpp_pub_events = self.get_events_with_field_value( 'message', pong_pub_message, rclcpp_publish_events, ) ping_rmw_pub_events = self.get_events_with_field_value( 'message', ping_pub_message, rmw_publish_events, ) pong_rmw_pub_events = self.get_events_with_field_value( 'message', pong_pub_message, rmw_publish_events, ) self.assertNumEventsEqual(ping_rclcpp_pub_events, 1) self.assertNumEventsEqual(pong_rclcpp_pub_events, 1) self.assertNumEventsEqual(ping_rmw_pub_events, 1) self.assertNumEventsEqual(pong_rmw_pub_events, 1) ping_rclcpp_pub_event = ping_rclcpp_pub_events[0] pong_rclcpp_pub_event = pong_rclcpp_pub_events[0] ping_rmw_pub_event = ping_rmw_pub_events[0] pong_rmw_pub_event = pong_rmw_pub_events[0] # Get subscription init events & subscription handles of test topics rcl_subscription_init_events = self.get_events_with_name(tp.rcl_subscription_init) ping_rcl_subscription_init_events = self.get_events_with_field_value( 'topic_name', '/ping', rcl_subscription_init_events, ) pong_rcl_subscription_init_events = self.get_events_with_field_value( 'topic_name', '/pong', rcl_subscription_init_events, ) self.assertNumEventsEqual(ping_rcl_subscription_init_events, 1) self.assertNumEventsEqual(pong_rcl_subscription_init_events, 1) ping_rcl_subscription_init_event = ping_rcl_subscription_init_events[0] pong_rcl_subscription_init_event = pong_rcl_subscription_init_events[0] ping_sub_handle = self.get_field(ping_rcl_subscription_init_event, 'subscription_handle') ping_rmw_sub_handle = self.get_field( ping_rcl_subscription_init_event, 'rmw_subscription_handle') pong_sub_handle = self.get_field(pong_rcl_subscription_init_event, 'subscription_handle') pong_rmw_sub_handle = self.get_field( pong_rcl_subscription_init_event, 'rmw_subscription_handle') # Find corresponding rmw_sub_init events ping_rmw_sub_init_events = self.get_events_with_field_value( 'rmw_subscription_handle', ping_rmw_sub_handle, rmw_sub_init_events, ) pong_rmw_sub_init_events = self.get_events_with_field_value( 'rmw_subscription_handle', pong_rmw_sub_handle, rmw_sub_init_events, ) self.assertNumEventsEqual(ping_rmw_sub_init_events, 1) self.assertNumEventsEqual(pong_rmw_sub_init_events, 1) ping_rmw_sub_init_event = ping_rmw_sub_init_events[0] pong_rmw_sub_init_event = pong_rmw_sub_init_events[0] # Get corresponding subscription objects rclcpp_subscription_init_events = self.get_events_with_name( tp.rclcpp_subscription_init, ) ping_rclcpp_subscription_init_events = self.get_events_with_field_value( 'subscription_handle', ping_sub_handle, rclcpp_subscription_init_events, ) pong_rclcpp_subscription_init_events = self.get_events_with_field_value( 'subscription_handle', pong_sub_handle, rclcpp_subscription_init_events, ) self.assertNumEventsEqual(ping_rclcpp_subscription_init_events, 1) self.assertNumEventsEqual(pong_rclcpp_subscription_init_events, 1) ping_rclcpp_subscription_init_event = ping_rclcpp_subscription_init_events[0] pong_rclcpp_subscription_init_event = pong_rclcpp_subscription_init_events[0] ping_sub_object = self.get_field(ping_rclcpp_subscription_init_event, 'subscription') pong_sub_object = self.get_field(pong_rclcpp_subscription_init_event, 'subscription') # Get corresponding subscription callback objects rclcpp_subscription_callback_events = self.get_events_with_name( tp.rclcpp_subscription_callback_added, ) ping_rclcpp_subscription_callback_events = self.get_events_with_field_value( 'subscription', ping_sub_object, rclcpp_subscription_callback_events, ) pong_rclcpp_subscription_callback_events = self.get_events_with_field_value( 'subscription', pong_sub_object, rclcpp_subscription_callback_events, ) self.assertNumEventsEqual(ping_rclcpp_subscription_callback_events, 1) self.assertNumEventsEqual(pong_rclcpp_subscription_callback_events, 1) ping_rclcpp_subscription_callback_event = ping_rclcpp_subscription_callback_events[0] pong_rclcpp_subscription_callback_event = pong_rclcpp_subscription_callback_events[0] ping_callback_object = self.get_field(ping_rclcpp_subscription_callback_event, 'callback') pong_callback_object = self.get_field(pong_rclcpp_subscription_callback_event, 'callback') # Check subscription init order self.assertEventOrder([ ping_rmw_sub_init_event, ping_rcl_subscription_init_event, ping_rclcpp_subscription_init_event, ping_rclcpp_subscription_callback_event, ]) self.assertEventOrder([ pong_rmw_sub_init_event, pong_rcl_subscription_init_event, pong_rclcpp_subscription_init_event, pong_rclcpp_subscription_callback_event, ]) # Get corresponding callback start/end events callback_start_events = self.get_events_with_name(tp.callback_start) callback_end_events = self.get_events_with_name(tp.callback_end) ping_callback_start_events = self.get_events_with_field_value( 'callback', ping_callback_object, callback_start_events, ) pong_callback_start_events = self.get_events_with_field_value( 'callback', pong_callback_object, callback_start_events, ) ping_callback_end_events = self.get_events_with_field_value( 'callback', ping_callback_object, callback_end_events, ) pong_callback_end_events = self.get_events_with_field_value( 'callback', pong_callback_object, callback_end_events, ) self.assertNumEventsEqual(ping_callback_start_events, 1) self.assertNumEventsEqual(pong_callback_start_events, 1) self.assertNumEventsEqual(ping_callback_end_events, 1) self.assertNumEventsEqual(pong_callback_end_events, 1) ping_callback_start_event = ping_callback_start_events[0] pong_callback_start_event = pong_callback_start_events[0] ping_callback_end_event = ping_callback_end_events[0] pong_callback_end_event = pong_callback_end_events[0] # Check pub/sub order: # * /ping pub rclcpp_publish # * /ping pub rcl_publish # * /ping pub rmw_publish # * /ping sub callback_start # * /pong pub rclcpp_publish # * /pong pub rcl_publish # * /pong pub rmw_publish # ... # * /ping sub callback_end # ... we shouldn't necessarily expect the /pong callback to start # before the /ping callback has ended # * /pong sub callback_start # * /pong sub callback_end self.assertEventOrder([ ping_rclcpp_pub_event, ping_rcl_pub_event, ping_rmw_pub_event, ping_callback_start_event, pong_rclcpp_pub_event, pong_rcl_pub_event, pong_rmw_pub_event, ping_callback_end_event, ]) self.assertEventOrder([ pong_rclcpp_pub_event, pong_rcl_pub_event, pong_rmw_pub_event, pong_callback_start_event, pong_callback_end_event, ]) if __name__ == '__main__': unittest.main()
42.380192
98
0.684357
import unittest from launch.actions import SetEnvironmentVariable from tracetools_test.case import TraceTestCase from tracetools_trace.tools import tracepoints as tp class TestPubSub(TraceTestCase): def __init__(self, *args) -> None: super().__init__( *args, session_name_prefix='session-test-pub-sub', events_ros=[ tp.rmw_publisher_init, tp.rcl_publisher_init, tp.rmw_publish, tp.rcl_publish, tp.rclcpp_publish, tp.rmw_subscription_init, tp.rcl_subscription_init, tp.rclcpp_subscription_init, tp.rclcpp_subscription_callback_added, tp.callback_start, tp.callback_end, ], package='test_tracetools', nodes=['test_ping', 'test_pong'], additional_actions=SetEnvironmentVariable('RMW_IMPLEMENTATION', 'rmw_cyclonedds_cpp'), ) def test_all(self): self.assertEventsSet(self._events_ros) rmw_pub_init_events = self.get_events_with_name(tp.rmw_publisher_init) rmw_sub_init_events = self.get_events_with_name(tp.rmw_subscription_init) publisher_init_events = self.get_events_with_name(tp.rcl_publisher_init) ping_publisher_init_events = self.get_events_with_field_value( 'topic_name', '/ping', publisher_init_events, ) pong_publisher_init_events = self.get_events_with_field_value( 'topic_name', '/pong', publisher_init_events, ) self.assertNumEventsEqual(ping_publisher_init_events, 1) self.assertNumEventsEqual(pong_publisher_init_events, 1) ping_publisher_init_event = ping_publisher_init_events[0] pong_publisher_init_event = pong_publisher_init_events[0] ping_pub_handle = self.get_field(ping_publisher_init_event, 'publisher_handle') ping_rmw_pub_handle = self.get_field(ping_publisher_init_event, 'rmw_publisher_handle') pong_pub_handle = self.get_field(pong_publisher_init_event, 'publisher_handle') pong_rmw_pub_handle = self.get_field(pong_publisher_init_event, 'rmw_publisher_handle') ping_rmw_pub_init_events = self.get_events_with_field_value( 'rmw_publisher_handle', ping_rmw_pub_handle, rmw_pub_init_events, ) pong_rmw_pub_init_events = self.get_events_with_field_value( 'rmw_publisher_handle', pong_rmw_pub_handle, rmw_pub_init_events, ) self.assertNumEventsEqual(ping_rmw_pub_init_events, 1) self.assertNumEventsEqual(pong_rmw_pub_init_events, 1) ping_rmw_pub_init_event = ping_rmw_pub_init_events[0] pong_rmw_pub_init_event = pong_rmw_pub_init_events[0] self.assertEventOrder([ ping_rmw_pub_init_event, ping_publisher_init_event, ]) self.assertEventOrder([ pong_rmw_pub_init_event, pong_publisher_init_event, ]) rcl_publish_events = self.get_events_with_name(tp.rcl_publish) ping_rcl_pub_events = self.get_events_with_field_value( 'publisher_handle', ping_pub_handle, rcl_publish_events, ) pong_rcl_pub_events = self.get_events_with_field_value( 'publisher_handle', pong_pub_handle, rcl_publish_events, ) self.assertNumEventsEqual(ping_rcl_pub_events, 1) self.assertNumEventsEqual(pong_rcl_pub_events, 1) ping_rcl_pub_event = ping_rcl_pub_events[0] pong_rcl_pub_event = pong_rcl_pub_events[0] rclcpp_publish_events = self.get_events_with_name(tp.rclcpp_publish) rmw_publish_events = self.get_events_with_name(tp.rmw_publish) ping_pub_message = self.get_field(ping_rcl_pub_event, 'message') pong_pub_message = self.get_field(pong_rcl_pub_event, 'message') ping_rclcpp_pub_events = self.get_events_with_field_value( 'message', ping_pub_message, rclcpp_publish_events, ) pong_rclcpp_pub_events = self.get_events_with_field_value( 'message', pong_pub_message, rclcpp_publish_events, ) ping_rmw_pub_events = self.get_events_with_field_value( 'message', ping_pub_message, rmw_publish_events, ) pong_rmw_pub_events = self.get_events_with_field_value( 'message', pong_pub_message, rmw_publish_events, ) self.assertNumEventsEqual(ping_rclcpp_pub_events, 1) self.assertNumEventsEqual(pong_rclcpp_pub_events, 1) self.assertNumEventsEqual(ping_rmw_pub_events, 1) self.assertNumEventsEqual(pong_rmw_pub_events, 1) ping_rclcpp_pub_event = ping_rclcpp_pub_events[0] pong_rclcpp_pub_event = pong_rclcpp_pub_events[0] ping_rmw_pub_event = ping_rmw_pub_events[0] pong_rmw_pub_event = pong_rmw_pub_events[0] rcl_subscription_init_events = self.get_events_with_name(tp.rcl_subscription_init) ping_rcl_subscription_init_events = self.get_events_with_field_value( 'topic_name', '/ping', rcl_subscription_init_events, ) pong_rcl_subscription_init_events = self.get_events_with_field_value( 'topic_name', '/pong', rcl_subscription_init_events, ) self.assertNumEventsEqual(ping_rcl_subscription_init_events, 1) self.assertNumEventsEqual(pong_rcl_subscription_init_events, 1) ping_rcl_subscription_init_event = ping_rcl_subscription_init_events[0] pong_rcl_subscription_init_event = pong_rcl_subscription_init_events[0] ping_sub_handle = self.get_field(ping_rcl_subscription_init_event, 'subscription_handle') ping_rmw_sub_handle = self.get_field( ping_rcl_subscription_init_event, 'rmw_subscription_handle') pong_sub_handle = self.get_field(pong_rcl_subscription_init_event, 'subscription_handle') pong_rmw_sub_handle = self.get_field( pong_rcl_subscription_init_event, 'rmw_subscription_handle') ping_rmw_sub_init_events = self.get_events_with_field_value( 'rmw_subscription_handle', ping_rmw_sub_handle, rmw_sub_init_events, ) pong_rmw_sub_init_events = self.get_events_with_field_value( 'rmw_subscription_handle', pong_rmw_sub_handle, rmw_sub_init_events, ) self.assertNumEventsEqual(ping_rmw_sub_init_events, 1) self.assertNumEventsEqual(pong_rmw_sub_init_events, 1) ping_rmw_sub_init_event = ping_rmw_sub_init_events[0] pong_rmw_sub_init_event = pong_rmw_sub_init_events[0] rclcpp_subscription_init_events = self.get_events_with_name( tp.rclcpp_subscription_init, ) ping_rclcpp_subscription_init_events = self.get_events_with_field_value( 'subscription_handle', ping_sub_handle, rclcpp_subscription_init_events, ) pong_rclcpp_subscription_init_events = self.get_events_with_field_value( 'subscription_handle', pong_sub_handle, rclcpp_subscription_init_events, ) self.assertNumEventsEqual(ping_rclcpp_subscription_init_events, 1) self.assertNumEventsEqual(pong_rclcpp_subscription_init_events, 1) ping_rclcpp_subscription_init_event = ping_rclcpp_subscription_init_events[0] pong_rclcpp_subscription_init_event = pong_rclcpp_subscription_init_events[0] ping_sub_object = self.get_field(ping_rclcpp_subscription_init_event, 'subscription') pong_sub_object = self.get_field(pong_rclcpp_subscription_init_event, 'subscription') rclcpp_subscription_callback_events = self.get_events_with_name( tp.rclcpp_subscription_callback_added, ) ping_rclcpp_subscription_callback_events = self.get_events_with_field_value( 'subscription', ping_sub_object, rclcpp_subscription_callback_events, ) pong_rclcpp_subscription_callback_events = self.get_events_with_field_value( 'subscription', pong_sub_object, rclcpp_subscription_callback_events, ) self.assertNumEventsEqual(ping_rclcpp_subscription_callback_events, 1) self.assertNumEventsEqual(pong_rclcpp_subscription_callback_events, 1) ping_rclcpp_subscription_callback_event = ping_rclcpp_subscription_callback_events[0] pong_rclcpp_subscription_callback_event = pong_rclcpp_subscription_callback_events[0] ping_callback_object = self.get_field(ping_rclcpp_subscription_callback_event, 'callback') pong_callback_object = self.get_field(pong_rclcpp_subscription_callback_event, 'callback') self.assertEventOrder([ ping_rmw_sub_init_event, ping_rcl_subscription_init_event, ping_rclcpp_subscription_init_event, ping_rclcpp_subscription_callback_event, ]) self.assertEventOrder([ pong_rmw_sub_init_event, pong_rcl_subscription_init_event, pong_rclcpp_subscription_init_event, pong_rclcpp_subscription_callback_event, ]) callback_start_events = self.get_events_with_name(tp.callback_start) callback_end_events = self.get_events_with_name(tp.callback_end) ping_callback_start_events = self.get_events_with_field_value( 'callback', ping_callback_object, callback_start_events, ) pong_callback_start_events = self.get_events_with_field_value( 'callback', pong_callback_object, callback_start_events, ) ping_callback_end_events = self.get_events_with_field_value( 'callback', ping_callback_object, callback_end_events, ) pong_callback_end_events = self.get_events_with_field_value( 'callback', pong_callback_object, callback_end_events, ) self.assertNumEventsEqual(ping_callback_start_events, 1) self.assertNumEventsEqual(pong_callback_start_events, 1) self.assertNumEventsEqual(ping_callback_end_events, 1) self.assertNumEventsEqual(pong_callback_end_events, 1) ping_callback_start_event = ping_callback_start_events[0] pong_callback_start_event = pong_callback_start_events[0] ping_callback_end_event = ping_callback_end_events[0] pong_callback_end_event = pong_callback_end_events[0] # before the /ping callback has ended # * /pong sub callback_start # * /pong sub callback_end self.assertEventOrder([ ping_rclcpp_pub_event, ping_rcl_pub_event, ping_rmw_pub_event, ping_callback_start_event, pong_rclcpp_pub_event, pong_rcl_pub_event, pong_rmw_pub_event, ping_callback_end_event, ]) self.assertEventOrder([ pong_rclcpp_pub_event, pong_rcl_pub_event, pong_rmw_pub_event, pong_callback_start_event, pong_callback_end_event, ]) if __name__ == '__main__': unittest.main()
true
true
f70e941ad501593c69646c1d51172a53cac84224
3,438
py
Python
IPython/utils/tests/test_jsonutil.py
flexlee/ipython
7528fbd76073c90262b9ac127de57c4c59b23a5c
[ "BSD-3-Clause-Clear" ]
1
2018-09-24T13:45:40.000Z
2018-09-24T13:45:40.000Z
IPython/utils/tests/test_jsonutil.py
andreasjansson/ipython
09b4311726f46945b936c699f7a6489d74d7397f
[ "BSD-3-Clause-Clear" ]
null
null
null
IPython/utils/tests/test_jsonutil.py
andreasjansson/ipython
09b4311726f46945b936c699f7a6489d74d7397f
[ "BSD-3-Clause-Clear" ]
1
2020-05-03T10:25:12.000Z
2020-05-03T10:25:12.000Z
"""Test suite for our JSON utilities. """ #----------------------------------------------------------------------------- # Copyright (C) 2010-2011 The IPython Development Team # # Distributed under the terms of the BSD License. The full license is in # the file COPYING.txt, distributed as part of this software. #----------------------------------------------------------------------------- #----------------------------------------------------------------------------- # Imports #----------------------------------------------------------------------------- # stdlib import json from base64 import decodestring # third party import nose.tools as nt # our own from IPython.testing import decorators as dec from ..jsonutil import json_clean, encode_images from ..py3compat import unicode_to_str, str_to_bytes #----------------------------------------------------------------------------- # Test functions #----------------------------------------------------------------------------- def test(): # list of input/expected output. Use None for the expected output if it # can be the same as the input. pairs = [(1, None), # start with scalars (1.0, None), ('a', None), (True, None), (False, None), (None, None), # complex numbers for now just go to strings, as otherwise they # are unserializable (1j, '1j'), # Containers ([1, 2], None), ((1, 2), [1, 2]), (set([1, 2]), [1, 2]), (dict(x=1), None), ({'x': 1, 'y':[1,2,3], '1':'int'}, None), # More exotic objects ((x for x in range(3)), [0, 1, 2]), (iter([1, 2]), [1, 2]), ] for val, jval in pairs: if jval is None: jval = val out = json_clean(val) # validate our cleanup nt.assert_equal(out, jval) # and ensure that what we return, indeed encodes cleanly json.loads(json.dumps(out)) @dec.parametric def test_encode_images(): # invalid data, but the header and footer are from real files pngdata = b'\x89PNG\r\n\x1a\nblahblahnotactuallyvalidIEND\xaeB`\x82' jpegdata = b'\xff\xd8\xff\xe0\x00\x10JFIFblahblahjpeg(\xa0\x0f\xff\xd9' fmt = { 'image/png' : pngdata, 'image/jpeg' : jpegdata, } encoded = encode_images(fmt) for key, value in fmt.iteritems(): # encoded has unicode, want bytes decoded = decodestring(encoded[key].encode('ascii')) yield nt.assert_equal(decoded, value) encoded2 = encode_images(encoded) yield nt.assert_equal(encoded, encoded2) b64_str = {} for key, encoded in encoded.iteritems(): b64_str[key] = unicode_to_str(encoded) encoded3 = encode_images(b64_str) yield nt.assert_equal(encoded3, b64_str) for key, value in fmt.iteritems(): # encoded3 has str, want bytes decoded = decodestring(str_to_bytes(encoded3[key])) yield nt.assert_equal(decoded, value) def test_lambda(): jc = json_clean(lambda : 1) assert isinstance(jc, str) assert '<lambda>' in jc json.dumps(jc) def test_exception(): bad_dicts = [{1:'number', '1':'string'}, {True:'bool', 'True':'string'}, ] for d in bad_dicts: nt.assert_raises(ValueError, json_clean, d)
32.742857
78
0.506981
import json from base64 import decodestring import nose.tools as nt from IPython.testing import decorators as dec from ..jsonutil import json_clean, encode_images from ..py3compat import unicode_to_str, str_to_bytes def test(): pairs = [(1, None), (1.0, None), ('a', None), (True, None), (False, None), (None, None), (1j, '1j'), ([1, 2], None), ((1, 2), [1, 2]), (set([1, 2]), [1, 2]), (dict(x=1), None), ({'x': 1, 'y':[1,2,3], '1':'int'}, None), ((x for x in range(3)), [0, 1, 2]), (iter([1, 2]), [1, 2]), ] for val, jval in pairs: if jval is None: jval = val out = json_clean(val) nt.assert_equal(out, jval) json.loads(json.dumps(out)) @dec.parametric def test_encode_images(): pngdata = b'\x89PNG\r\n\x1a\nblahblahnotactuallyvalidIEND\xaeB`\x82' jpegdata = b'\xff\xd8\xff\xe0\x00\x10JFIFblahblahjpeg(\xa0\x0f\xff\xd9' fmt = { 'image/png' : pngdata, 'image/jpeg' : jpegdata, } encoded = encode_images(fmt) for key, value in fmt.iteritems(): decoded = decodestring(encoded[key].encode('ascii')) yield nt.assert_equal(decoded, value) encoded2 = encode_images(encoded) yield nt.assert_equal(encoded, encoded2) b64_str = {} for key, encoded in encoded.iteritems(): b64_str[key] = unicode_to_str(encoded) encoded3 = encode_images(b64_str) yield nt.assert_equal(encoded3, b64_str) for key, value in fmt.iteritems(): decoded = decodestring(str_to_bytes(encoded3[key])) yield nt.assert_equal(decoded, value) def test_lambda(): jc = json_clean(lambda : 1) assert isinstance(jc, str) assert '<lambda>' in jc json.dumps(jc) def test_exception(): bad_dicts = [{1:'number', '1':'string'}, {True:'bool', 'True':'string'}, ] for d in bad_dicts: nt.assert_raises(ValueError, json_clean, d)
true
true
f70e941e15a39ccab967128409f18a00fc06093a
2,673
py
Python
SpotifyDownloader/SpotifyWebAPI.py
Hugo4400/Spotify-Downloader
a128a401d39b16df397f56e07802bd6e65e1710e
[ "MIT" ]
9
2021-06-16T08:44:07.000Z
2022-02-11T17:19:05.000Z
SpotifyDownloader/SpotifyWebAPI.py
Hugo4400/Spotify-Downloader
a128a401d39b16df397f56e07802bd6e65e1710e
[ "MIT" ]
3
2021-06-16T08:21:18.000Z
2021-12-12T14:49:15.000Z
SpotifyDownloader/SpotifyWebAPI.py
Hugo4400/Spotify-Downloader
a128a401d39b16df397f56e07802bd6e65e1710e
[ "MIT" ]
2
2021-06-17T22:08:22.000Z
2021-07-08T21:45:47.000Z
import requests from urllib.parse import urlencode from urllib.request import urlopen from urllib.error import HTTPError import re import json from base64 import b64encode def get_playlists(spotify_url): with open('MY_SECRETS.json', 'r') as f: spotify_key = json.load(f)['SPOTIFY_KEY'] playlist_id = spotify_url.split('/')[-1].split('?')[0] r = requests.get(f"https://api.spotify.com/v1/playlists/{playlist_id}", headers={'Authorization': f'Bearer {spotify_key}'}) if r.status_code == 400 or r.status_code == 401: raise TypeError('Invalid Spotify Token') returned_tracks = {} playlist_name = r.json()['name'] r = requests.get(f"https://api.spotify.com/v1/playlists/{playlist_id}/tracks", headers={'Authorization': f'Bearer {spotify_key}'}) data = r.json() tracks = data['items'] while data['next']: r = requests.get(data['next'], headers={'Authorization': f'Bearer {spotify_key}'}) data = r.json() tracks = tracks + data['items'] for track in tracks: song_name = track['track']['name'] artists = [] for artist in track['track']['artists']: artists.append(artist['name']) artist_name = ' '.join(artists) try: query_string = urlencode({'search_query': artist_name + ' ' + song_name}) htm_content = urlopen('http://www.youtube.com/results?' + query_string) search_results = re.findall(r'/watch\?v=(.{11})', htm_content.read().decode()) returned_tracks.update({f'{song_name}': f'http://www.youtube.com/watch?v={search_results[0]}'}) except HTTPError: print(f'Couldn\'t download "{song_name}", continuing') continue return playlist_name, returned_tracks def get_access_token(): with open('MY_SECRETS.json', 'r') as f: load_file = json.load(f) spotify_client_id = load_file['spotify_client_id'] spotify_client_secret = load_file['spotify_client_secret'] headers = { 'Authorization': f'Basic {b64encode(f"{spotify_client_id}:{spotify_client_secret}".encode()).decode()}', } data = { 'grant_type': 'client_credentials' } r = requests.post('https://accounts.spotify.com/api/token', headers=headers, data=data) token = r.json()['access_token'] updated_dict = { "spotify_client_id": f"{spotify_client_id}", "spotify_client_secret": f"{spotify_client_secret}", "SPOTIFY_KEY": token } with open('MY_SECRETS.json', 'w') as f: json.dump(updated_dict, f)
33
135
0.616162
import requests from urllib.parse import urlencode from urllib.request import urlopen from urllib.error import HTTPError import re import json from base64 import b64encode def get_playlists(spotify_url): with open('MY_SECRETS.json', 'r') as f: spotify_key = json.load(f)['SPOTIFY_KEY'] playlist_id = spotify_url.split('/')[-1].split('?')[0] r = requests.get(f"https://api.spotify.com/v1/playlists/{playlist_id}", headers={'Authorization': f'Bearer {spotify_key}'}) if r.status_code == 400 or r.status_code == 401: raise TypeError('Invalid Spotify Token') returned_tracks = {} playlist_name = r.json()['name'] r = requests.get(f"https://api.spotify.com/v1/playlists/{playlist_id}/tracks", headers={'Authorization': f'Bearer {spotify_key}'}) data = r.json() tracks = data['items'] while data['next']: r = requests.get(data['next'], headers={'Authorization': f'Bearer {spotify_key}'}) data = r.json() tracks = tracks + data['items'] for track in tracks: song_name = track['track']['name'] artists = [] for artist in track['track']['artists']: artists.append(artist['name']) artist_name = ' '.join(artists) try: query_string = urlencode({'search_query': artist_name + ' ' + song_name}) htm_content = urlopen('http://www.youtube.com/results?' + query_string) search_results = re.findall(r'/watch\?v=(.{11})', htm_content.read().decode()) returned_tracks.update({f'{song_name}': f'http://www.youtube.com/watch?v={search_results[0]}'}) except HTTPError: print(f'Couldn\'t download "{song_name}", continuing') continue return playlist_name, returned_tracks def get_access_token(): with open('MY_SECRETS.json', 'r') as f: load_file = json.load(f) spotify_client_id = load_file['spotify_client_id'] spotify_client_secret = load_file['spotify_client_secret'] headers = { 'Authorization': f'Basic {b64encode(f"{spotify_client_id}:{spotify_client_secret}".encode()).decode()}', } data = { 'grant_type': 'client_credentials' } r = requests.post('https://accounts.spotify.com/api/token', headers=headers, data=data) token = r.json()['access_token'] updated_dict = { "spotify_client_id": f"{spotify_client_id}", "spotify_client_secret": f"{spotify_client_secret}", "SPOTIFY_KEY": token } with open('MY_SECRETS.json', 'w') as f: json.dump(updated_dict, f)
true
true
f70e94b149473aa9ed65477fe96bca78b862bbc6
5,033
py
Python
src/collective/solr/manager.py
adrianschulz/collective.solr
2d76fe01a02174d383fdc335d38ee52afa8bfa27
[ "ZPL-1.1" ]
null
null
null
src/collective/solr/manager.py
adrianschulz/collective.solr
2d76fe01a02174d383fdc335d38ee52afa8bfa27
[ "ZPL-1.1" ]
1
2020-05-20T06:02:00.000Z
2020-07-14T13:36:58.000Z
src/collective/solr/manager.py
adrianschulz/collective.solr
2d76fe01a02174d383fdc335d38ee52afa8bfa27
[ "ZPL-1.1" ]
null
null
null
# -*- coding: utf-8 -*- from collective.solr.interfaces import ISolrConnectionManager from collective.solr.interfaces import IZCMLSolrConnectionConfig from collective.solr.local import getLocal from collective.solr.local import setLocal from collective.solr.solr import SolrConnection from collective.solr.utils import getConfig from collective.solr.utils import isActive from six.moves.http_client import CannotSendRequest from six.moves.http_client import ResponseNotReady from logging import getLogger from socket import error from zope.component import queryUtility from zope.interface import implementer from plone.registry.interfaces import IRegistry from zope.component import getUtility import six logger = getLogger("collective.solr.manager") marker = object() @implementer(IZCMLSolrConnectionConfig) class ZCMLSolrConnectionConfig(object): """Connection values that can be configured through zcml""" def __init__(self, host, port, base): self.host = "%s:%d" % (host, port) self.base = base @implementer(ISolrConnectionManager) class SolrConnectionManager(object): """ a thread-local connection manager for solr """ lock = False def __init__(self, active=None): if isinstance(active, bool): self.setHost(active=active) def setHost(self, active=False, host="localhost", port=8983, base="/solr/plone"): """ set connection parameters """ config = getConfig() config.active = active config.host = six.text_type(host) config.port = port config.base = six.text_type(base) self.closeConnection(clearSchema=True) def closeConnection(self, clearSchema=False): """ close the current connection, if any """ logger.debug("closing connection") conn = getLocal("connection") if conn is not None: conn.close() setLocal("connection", None) if clearSchema: setLocal("schema", None) def getConnection(self): """ returns an existing connection or opens one """ if not isActive(): return None conn = getLocal("connection") if conn is not None: return conn zcmlconfig = queryUtility(IZCMLSolrConnectionConfig) registry = getUtility(IRegistry) config_host = registry["collective.solr.host"] if zcmlconfig is not None: # use connection parameters defined in zcml... logger.debug("opening connection to %s", zcmlconfig.host) conn = SolrConnection( host=zcmlconfig.host, solrBase=zcmlconfig.base, persistent=True ) setLocal("connection", conn) elif config_host is not None: # otherwise use connection parameters defined in control panel... config_port = registry["collective.solr.port"] config_base = registry["collective.solr.base"] host = "%s:%d" % (config_host, config_port) logger.debug("opening connection to %s", host) conn = SolrConnection(host=host, solrBase=config_base, persistent=True) setLocal("connection", conn) return conn def getSchema(self): """ returns the currently used schema or fetches it """ schema = getLocal("schema") if schema is None: conn = self.getConnection() if conn is not None: logger.debug("getting schema from solr") self.setSearchTimeout() try: schema = conn.get_schema() setLocal("schema", schema) except (error, CannotSendRequest, ResponseNotReady): logger.exception("exception while getting schema") return schema def setTimeout(self, timeout, lock=marker): """ set the timeout on the current (or to be opened) connection to the given value """ update = not self.lock # update if not locked... if lock is not marker: self.lock = bool(lock) update = True # ...or changed logger.debug("%ssetting timeout lock", lock and "" or "re") if update: conn = self.getConnection() if conn is not None: logger.debug("setting timeout to %s", timeout) conn.setTimeout(timeout) def setIndexTimeout(self): """ set the timeout on the current (or to be opened) connection to the value specified for indexing operations """ registry = getUtility(IRegistry) index_timeout = registry["collective.solr.index_timeout"] self.setTimeout(index_timeout or None) def setSearchTimeout(self): """ set the timeout on the current (or to be opened) connection to the value specified for search operations """ registry = getUtility(IRegistry) search_timeout = registry["collective.solr.search_timeout"] self.setTimeout(search_timeout or None)
38.419847
85
0.642956
from collective.solr.interfaces import ISolrConnectionManager from collective.solr.interfaces import IZCMLSolrConnectionConfig from collective.solr.local import getLocal from collective.solr.local import setLocal from collective.solr.solr import SolrConnection from collective.solr.utils import getConfig from collective.solr.utils import isActive from six.moves.http_client import CannotSendRequest from six.moves.http_client import ResponseNotReady from logging import getLogger from socket import error from zope.component import queryUtility from zope.interface import implementer from plone.registry.interfaces import IRegistry from zope.component import getUtility import six logger = getLogger("collective.solr.manager") marker = object() @implementer(IZCMLSolrConnectionConfig) class ZCMLSolrConnectionConfig(object): def __init__(self, host, port, base): self.host = "%s:%d" % (host, port) self.base = base @implementer(ISolrConnectionManager) class SolrConnectionManager(object): lock = False def __init__(self, active=None): if isinstance(active, bool): self.setHost(active=active) def setHost(self, active=False, host="localhost", port=8983, base="/solr/plone"): config = getConfig() config.active = active config.host = six.text_type(host) config.port = port config.base = six.text_type(base) self.closeConnection(clearSchema=True) def closeConnection(self, clearSchema=False): logger.debug("closing connection") conn = getLocal("connection") if conn is not None: conn.close() setLocal("connection", None) if clearSchema: setLocal("schema", None) def getConnection(self): if not isActive(): return None conn = getLocal("connection") if conn is not None: return conn zcmlconfig = queryUtility(IZCMLSolrConnectionConfig) registry = getUtility(IRegistry) config_host = registry["collective.solr.host"] if zcmlconfig is not None: logger.debug("opening connection to %s", zcmlconfig.host) conn = SolrConnection( host=zcmlconfig.host, solrBase=zcmlconfig.base, persistent=True ) setLocal("connection", conn) elif config_host is not None: config_port = registry["collective.solr.port"] config_base = registry["collective.solr.base"] host = "%s:%d" % (config_host, config_port) logger.debug("opening connection to %s", host) conn = SolrConnection(host=host, solrBase=config_base, persistent=True) setLocal("connection", conn) return conn def getSchema(self): schema = getLocal("schema") if schema is None: conn = self.getConnection() if conn is not None: logger.debug("getting schema from solr") self.setSearchTimeout() try: schema = conn.get_schema() setLocal("schema", schema) except (error, CannotSendRequest, ResponseNotReady): logger.exception("exception while getting schema") return schema def setTimeout(self, timeout, lock=marker): update = not self.lock if lock is not marker: self.lock = bool(lock) update = True logger.debug("%ssetting timeout lock", lock and "" or "re") if update: conn = self.getConnection() if conn is not None: logger.debug("setting timeout to %s", timeout) conn.setTimeout(timeout) def setIndexTimeout(self): registry = getUtility(IRegistry) index_timeout = registry["collective.solr.index_timeout"] self.setTimeout(index_timeout or None) def setSearchTimeout(self): registry = getUtility(IRegistry) search_timeout = registry["collective.solr.search_timeout"] self.setTimeout(search_timeout or None)
true
true
f70e95585af52e0b03e70818b18ac47fb4f46f63
26,001
py
Python
evaluation/macrobenchmark/workload/models/classification.py
DelphianCalamity/PrivateKube
14f575e77021ab7baca30f4061140ec83bdc96a7
[ "Apache-2.0" ]
9
2021-06-16T00:22:45.000Z
2021-11-25T07:19:11.000Z
evaluation/macrobenchmark/workload/models/classification.py
DelphianCalamity/PrivateKube
14f575e77021ab7baca30f4061140ec83bdc96a7
[ "Apache-2.0" ]
2
2021-11-14T10:42:43.000Z
2022-03-16T03:43:22.000Z
evaluation/macrobenchmark/workload/models/classification.py
DelphianCalamity/PrivateKube
14f575e77021ab7baca30f4061140ec83bdc96a7
[ "Apache-2.0" ]
3
2021-04-08T08:08:48.000Z
2021-12-24T01:42:20.000Z
import sys, os, shutil import h5py import time import io import random import tempfile from tqdm import tqdm from absl import app, flags, logging from ray.util.multiprocessing import Pool import gcsfs import numpy as np from pathlib import Path from sklearn.linear_model import LogisticRegression from sklearn.ensemble import RandomForestClassifier from sklearn.svm import LinearSVC from sklearn.feature_extraction.text import CountVectorizer from sklearn.feature_extraction.text import TfidfTransformer from sklearn.naive_bayes import MultinomialNB from sklearn.pipeline import Pipeline from sklearn.linear_model import SGDClassifier import torchtext import torch import torch.optim as optim from torch.optim.lr_scheduler import ReduceLROnPlateau import torch.nn as nn from transformers import BertTokenizer, BertModel, BertForSequenceClassification import opacus from privatekube.experiments.datasets import ( EventLevelDataset, split_review_batch, UserTimeLevelDataset, select_blocks_by_timeframe, ) from privatekube.experiments.utils import ( build_flags, flags_to_dict, load_yaml, results_to_dict, save_yaml, save_model, binary_accuracy, multiclass_accuracy, epoch_time, ) from privatekube.privacy.text import build_public_vocab from privatekube.privacy.rdp import ( compute_noise_from_target_epsilon, ALPHAS, compute_rdp_sgm, ) import models DEFAULT_DATA_PATH = Path(__file__).resolve().parent.parent.parent.joinpath("data") # Define default args dataset_args = { "n_blocks": 200, "max_text_len": 140, "vocab_size": 10_000, "n_blocks_test": 200, } input_path_args = { "dataset_dir": "", "dataset_monofile": "", "block_counts": str(DEFAULT_DATA_PATH.joinpath("block_counts.yaml")), "emb_path": str(DEFAULT_DATA_PATH.joinpath(".vector_cache")), } model_args = { "task": "product", "model": "bow", "embedding_dim": 100, "hidden_dim_1": 240, "hidden_dim_2": 195, "hidden_dim": 100, "dropout": 0.25, } training_args = { "device": "cuda", "learning_rate": 0.01, "dp": 0, "dp_eval": 0, "user_level": 0, "epsilon": 5.0, "delta": 1e-5, "n_epochs": 15, "batch_size": 64, "virtual_batch_multiplier": 2, "adaptive_batch_size": 1, "noise": -1.0, "timeframe_days": 0, "learning_rate_scheduler": 1, "dynamic_clipping": 0, "max_grad_norm": 1.0, "per_layer_clipping": 0, "n_workers": 6, "non_dp_batch_size": 256, } output_args = { "log_path": "", "model_path": "", "metrics_path": "", } build_flags(dataset_args, model_args, training_args, input_path_args, output_args) FLAGS = flags.FLAGS np.random.seed(0) def build_split_dataset(): block_dir = tempfile.mkdtemp() test_block_dir = tempfile.mkdtemp() if FLAGS.dataset_dir[0:5] == "gs://": os.system( "gcloud auth activate-service-account --key-file=$GOOGLE_APPLICATION_CREDENTIALS" ) fs = gcsfs.GCSFileSystem( project=os.get_env("GCP_PROJECT"), token="google_default" ) # Get the local Gcloud token logging.info("Listing bucket files.") all_blocks = list( map( lambda blob: os.path.basename(blob["name"]), fs.listdir(FLAGS.dataset_dir), ) ) logging.info(f"Got {len(all_blocks)} blocks.") logging.warning(f"The evaluation set is not fixed.") elif FLAGS.dataset_dir == "": logging.info("Listing the block names.") all_blocks = list(load_yaml(FLAGS.block_counts).keys()) else: all_blocks = os.listdir(FLAGS.dataset_dir) logging.info(f"Selecting {FLAGS.n_blocks_test} test blocks (fixed randomness).") test_blocks = np.random.choice(all_blocks, FLAGS.n_blocks_test, replace=False) for tb in test_blocks: all_blocks.remove(tb) # Use every user to the maximum. def sort_by_user(block_name): if block_name.endswith(".h5"): block_name = block_name[: -len(".h5")] name = block_name.split("-") user_slice = int(name[1]) return user_slice logging.info( f"Selecting as few users as possible.\n Pseudorandom and deterministic (hashed user ids)." ) selected_blocks = sorted(all_blocks, key=sort_by_user)[0 : FLAGS.n_blocks] if FLAGS.dataset_dir[0:5] == "gs://": pool = Pool() bucket_path = FLAGS.dataset_dir def download_datasource(block_name): block_path = os.path.join(bucket_path, block_name) dest = os.path.join(block_dir, block_name) os.system(f"gsutil cp {block_path} {dest}") return logging.warning("Downloading the blocks in parallel.") b = pool.map(download_datasource, selected_blocks) pool.close() pool.join() block_names = None test_block_names = None elif FLAGS.dataset_dir == "": block_dir = None test_block_dir = None block_names = selected_blocks test_block_names = test_blocks else: for b in selected_blocks: os.symlink(os.path.join(FLAGS.dataset_dir, b), os.path.join(block_dir, b)) for b in test_blocks: os.symlink( os.path.join(FLAGS.dataset_dir, b), os.path.join(test_block_dir, b) ) block_names = None test_block_names = None # Store for the logs FLAGS.dataset_dir = block_dir if not FLAGS.dataset_monofile: if FLAGS.model == "bert": from_h5 = DEFAULT_DATA_PATH.joinpath("reviews.h5") else: from_h5 = DEFAULT_DATA_PATH.joinpath("reviews_custom_vocab.h5") else: from_h5 = FLAGS.dataset_monofile if FLAGS.dp and FLAGS.user_level: train_data = UserTimeLevelDataset( blocks_dir=block_dir, timeframe=FLAGS.timeframe_days * 86400, from_h5=from_h5, block_names=block_names, ) else: train_data = EventLevelDataset( blocks_dir=block_dir, from_h5=from_h5, block_names=block_names, ) test_data = EventLevelDataset( blocks_dir=test_block_dir, from_h5=from_h5, block_names=test_block_names, ) test_data, valid_data = test_data.split([0.75, 0.25]) logging.info(f"Test size: {len(test_data)}\n Valid size: {len(valid_data)}") # Values from the preprocessing # (max text len doesn't matter here) text_field = torchtext.data.Field( batch_first=True, use_vocab=True, init_token="<bos>", eos_token="<eos>", pad_token="<pad>", unk_token="<unk>", include_lengths=True, ) build_public_vocab( text_field, max_size=FLAGS.vocab_size - 4, vectors=f"glove.6B.{FLAGS.embedding_dim}d", unk_init=torch.Tensor.normal_, vectors_cache=FLAGS.emb_path, ) return train_data, test_data, valid_data, text_field def compute_optimal_batch_size(real_batch_size, dataset_len): logging.info( f"Computing the optimal batch size. Dataset {dataset_len}, real batch {real_batch_size}" ) # Under approximate optimal_batch_size = int(np.sqrt(dataset_len)) if optimal_batch_size <= real_batch_size: return optimal_batch_size, 0 else: return (real_batch_size, optimal_batch_size // real_batch_size) def count_parameters(model): return sum(p.numel() for p in model.parameters() if p.requires_grad) def build_model(text_field): INPUT_DIM = len(text_field.vocab) word_embeddings = text_field.vocab.vectors PAD_IDX = text_field.vocab.stoi[text_field.pad_token] UNK_IDX = text_field.vocab.stoi[text_field.unk_token] if FLAGS.task == "sentiment": output_dim = 1 elif FLAGS.task == "product": output_dim = 11 if FLAGS.model == "lstm": model = models.LSTMClassifier( batch_size=FLAGS.batch_size, output_size=output_dim, hidden_size=FLAGS.hidden_dim, vocab_size=INPUT_DIM, embedding_length=FLAGS.embedding_dim, weights=word_embeddings, dropout=FLAGS.dropout, dp=FLAGS.dp, ) elif FLAGS.model == "bow": model = models.NBOW( input_dim=word_embeddings.shape[0], emb_dim=FLAGS.embedding_dim, output_dim=output_dim, pad_idx=PAD_IDX, word_embeddings=word_embeddings, ) elif FLAGS.model == "feedforward": model = models.FeedforwardModel( vocab_size=INPUT_DIM, embedding_dim=FLAGS.embedding_dim, pad_idx=PAD_IDX, H_1=FLAGS.hidden_dim_1, H_2=FLAGS.hidden_dim_2, D_out=output_dim, word_embeddings=word_embeddings, ) elif FLAGS.model == "bert": # The dataset has been preprocessed with the bert tokenizer, so the indices should be correct logging.info(f"Pad and unk index {PAD_IDX, UNK_IDX}") model = models.FineTunedBert.build_new(output_dim=output_dim) logging.info( f"Model {FLAGS.model} has {count_parameters(model)} trainable parameters." ) # Bert has its own pretrained embeddings return model pretrained_embeddings = text_field.vocab.vectors model.embedding.weight.data.copy_(pretrained_embeddings) model.embedding.weight.data[UNK_IDX] = torch.zeros(FLAGS.embedding_dim) model.embedding.weight.data[PAD_IDX] = torch.zeros(FLAGS.embedding_dim) logging.info( f"Model {FLAGS.model} has {count_parameters(model)} trainable parameters." ) return model def train(model, iterator, optimizer, criterion, accuracy_fn): epoch_loss = 0 epoch_acc = 0 model.train() optimizer.zero_grad() for i, batch in enumerate(tqdm(iterator)): # batch = batch.to(FLAGS.device) if FLAGS.task == "sentiment": data, label = split_review_batch( batch, label_feature="binary_rating", max_text_len=FLAGS.max_text_len, include_len=True, vocab_size=FLAGS.vocab_size, custom_vocab=(FLAGS.model != "bert"), ) text_lengths, text = data elif FLAGS.task == "product": text, label = split_review_batch( batch, label_feature="category", max_text_len=FLAGS.max_text_len, vocab_size=FLAGS.vocab_size, custom_vocab=(FLAGS.model != "bert"), ) text = text.to(device=FLAGS.device, dtype=torch.long) label = ( label.to(device=FLAGS.device, dtype=torch.long) if FLAGS.task == "product" else label.to(device=FLAGS.device, dtype=torch.float) ) if FLAGS.model == "lstm": hidden = model.init_hidden(batch_size=len(batch)) if isinstance(hidden, tuple): hidden = ( hidden[0].to(FLAGS.device), hidden[1].to(FLAGS.device), ) else: hidden = hidden.to(FLAGS.device) outputs = model(text, hidden) elif FLAGS.model == "bert": PAD_IDX = 0 inputs = { "input_ids": text, "labels": label, "attention_mask": torch.where( text == PAD_IDX, torch.zeros_like(text), torch.ones_like(text) ), } # logging.info(f"Inputs {inputs}") # The model outputs loss, logits outputs = model(**inputs)[1] # logging.info(f"Outputs {outputs}") else: outputs = model(text) # logging.info(f"Outputs {outputs}") if FLAGS.task == "sentiment": outputs = outputs.squeeze(1) loss = criterion(outputs, label) acc = accuracy_fn(outputs.detach(), label) loss.backward() if FLAGS.dp and FLAGS.virtual_batch_multiplier > 1: # NOTE: step is not called at every minibatch, so the RDP accountant need to know this if (i + 1) % FLAGS.virtual_batch_multiplier == 0 or (i + 1) == len( iterator ): # For the (virtual_batch_multiplier)th batch, call a clip-noise-step optimizer.step() optimizer.zero_grad() else: # For the first (virtual_batch_multiplier - 1) batches, just accumulate the gradients optimizer.virtual_step() else: # Regular optimizer step (either non-DP or DP with no virtual step) optimizer.step() optimizer.zero_grad() epoch_loss += loss.item() # epoch_loss += loss.detach().item() epoch_acc += acc.item() return epoch_loss / len(iterator), epoch_acc / len(iterator) def evaluate(model, iterator, criterion, accuracy_fn): epoch_loss = 0 epoch_acc = 0 model.eval() with torch.no_grad(): for batch in iterator: # batch = batch.to(FLAGS.device) if FLAGS.task == "sentiment": data, label = split_review_batch( batch, label_feature="binary_rating", max_text_len=FLAGS.max_text_len, include_len=True, vocab_size=FLAGS.vocab_size, custom_vocab=(FLAGS.model != "bert"), ) text_lengths, text = data elif FLAGS.task == "product": text, label = split_review_batch( batch, label_feature="category", max_text_len=FLAGS.max_text_len, vocab_size=FLAGS.vocab_size, custom_vocab=(FLAGS.model != "bert"), ) text = text.to(device=FLAGS.device, dtype=torch.long) label = ( label.to(device=FLAGS.device, dtype=torch.long) if FLAGS.task == "product" else label.to(device=FLAGS.device, dtype=torch.float) ) if FLAGS.model == "lstm": hidden = model.init_hidden(batch_size=len(batch)) if isinstance(hidden, tuple): hidden = ( hidden[0].to(FLAGS.device), hidden[1].to(FLAGS.device), ) else: hidden = hidden.to(FLAGS.device) outputs = model(text, hidden) elif FLAGS.model == "bert": PAD_IDX = 0 inputs = { "input_ids": text, "labels": label, "attention_mask": torch.where( text == PAD_IDX, torch.zeros_like(text), torch.ones_like(text) ), } outputs = model(**inputs)[1] else: outputs = model(text) if FLAGS.task == "sentiment": outputs = outputs.squeeze(1) # print(f"Training. Outputs: {outputs}, labels: {batch.label}") loss = criterion(outputs, label) acc = accuracy_fn(outputs, label) epoch_loss += loss.item() epoch_acc += acc.item() return epoch_loss / len(iterator), epoch_acc / len(iterator) def train_validate( train_data, valid_data, model, optimizer, criterion, accuracy_fn, scheduler ): validation_accuracy_epochs = [] validation_loss_epochs = [] training_loss_epochs = [] training_accuracy_epochs = [] logging.info(f"n workers: {FLAGS.n_workers}") train_iterator = torch.utils.data.DataLoader( train_data, batch_size=FLAGS.batch_size, shuffle=True, num_workers=FLAGS.n_workers, drop_last=True, ) valid_iterator = torch.utils.data.DataLoader( valid_data, batch_size=FLAGS.batch_size, shuffle=True, num_workers=FLAGS.n_workers, drop_last=False, ) criterion = criterion.to(FLAGS.device) best_valid_loss = float("inf") for epoch in range(FLAGS.n_epochs): start_time = time.time() logging.info(f"Starting epoch {epoch + 1}.") train_loss, train_acc = train( model, train_iterator, optimizer, criterion, accuracy_fn ) valid_loss, valid_acc = evaluate(model, valid_iterator, criterion, accuracy_fn) end_time = time.time() epoch_mins, epoch_secs = epoch_time(start_time, end_time) if valid_loss < best_valid_loss: best_valid_loss = valid_loss torch.save(model.state_dict(), "tut2-model.pt") logging.info(f"Epoch: {epoch+1:02} | Epoch Time: {epoch_mins}m {epoch_secs}s") logging.info( f"\tTrain Loss: {train_loss:.3f} | Train Acc: {train_acc*100:.2f}%" ) scheduler.step(train_loss) logging.info( f"\t Val. Loss: {valid_loss:.3f} | Val. Acc: {valid_acc*100:.2f}%" ) validation_accuracy_epochs.append(valid_acc) validation_loss_epochs.append(valid_loss) training_loss_epochs.append(train_loss) training_accuracy_epochs.append(train_acc) return ( training_loss_epochs, training_accuracy_epochs, validation_loss_epochs, validation_accuracy_epochs, ) def count_parameters(model): return sum(p.numel() for p in model.parameters() if p.requires_grad) def main(argv): start_time = time.time() # Convert flags for the epsilon = -1 shortcut if FLAGS.dp and FLAGS.epsilon < 0 and FLAGS.noise < 0: FLAGS.dp = False # No multiprocessing for large datasets (save RAM) if FLAGS.n_blocks > 50_000: logging.info(f"Large dataset, we use a single thread for the loader.") FLAGS.n_workers = 0 # Build the dataset, either event level or user level train_data, test_data, valid_data, text_field = build_split_dataset() logging.info( f"Number of samples for training: {len(train_data)}, validation: {len(valid_data)} and testing: {len(test_data)}" ) # Adapt the batch size and the virtual step size, unless it has been specified manually if FLAGS.dp and FLAGS.adaptive_batch_size and FLAGS.virtual_batch_multiplier <= 0: FLAGS.batch_size, FLAGS.virtual_batch_multiplier = compute_optimal_batch_size( FLAGS.batch_size, len(train_data) ) logging.info( f"Using real batch {FLAGS.batch_size} with multiplier {FLAGS.virtual_batch_multiplier}" ) if not FLAGS.dp: FLAGS.batch_size = FLAGS.non_dp_batch_size # Prepare the model and optimizer model = build_model(text_field).to(FLAGS.device) logging.info(f"Number of trainable parameters: {count_parameters(model)}") # optimizer = optim.Adam(model.parameters()) optimizer = optim.AdamW(model.parameters(), lr=FLAGS.learning_rate, eps=1e-8) scheduler = ReduceLROnPlateau(optimizer, mode="min", patience=3) # train_it = torch.utils.data.DataLoader( # train_data, # batch_size=2048, # shuffle=False, # num_workers=FLAGS.n_workers, # drop_last=False, # ) # counts = {} # for i in range(11): # counts[i] = 0 # for b in train_it: # for cat in b[:, 3]: # counts[int(cat)] += 1 # s = sum(counts.values()) # for cat, count in counts.items(): # counts[cat] = count / s # logging.info(counts) if FLAGS.task == "sentiment": criterion = nn.BCEWithLogitsLoss().to(FLAGS.device) accuracy_fn = binary_accuracy # automotive: 0.03036145803296712 # books: 0.41258122723567553 # cds: 0.012897189083383703 # clothing: 0.2025265712144095 # games: 0.031613111956201506 # groceries: 0.01949595483554337 # home: 0.119920985593197 # movies: 0.0484712255807162 # pets: 0.03665525816121956 # sports: 0.04961580907019007 # tools: 0.035861209236496445 elif FLAGS.task == "product": # criterion = nn.CrossEntropyLoss( # weight=torch.Tensor( # [0.05, 0.035, 0.03, 0.035, 0.05, 0.02, 0.12, 0.01, 0.03, 0.20, 0.41] # ) # ) criterion = nn.CrossEntropyLoss() accuracy_fn = multiclass_accuracy # Plug Opacus if DP training is activated if FLAGS.dp: if FLAGS.noise >= 0: logging.info(f"User-provided noise: {FLAGS.noise}.") else: logging.info("Computing noise for the given parameters.") FLAGS.noise = compute_noise_from_target_epsilon( target_epsilon=FLAGS.epsilon, target_delta=FLAGS.delta, epochs=FLAGS.n_epochs, batch_size=FLAGS.batch_size * FLAGS.virtual_batch_multiplier if FLAGS.virtual_batch_multiplier > 0 else FLAGS.batch_size, dataset_size=len(train_data), alphas=ALPHAS, ) logging.info(f"Noise computed from RDP budget: {FLAGS.noise}.") # NOTE: when user-level DP is activated, the training dataset __len__ method returns # the number of users, and the DataLoader calls the batch-of-user method that overrides # the regular __getitem__ method # WARNING: fishy non-DP adaptive clipping privacy_engine = opacus.PrivacyEngine( module=model, batch_size=FLAGS.batch_size * FLAGS.virtual_batch_multiplier if FLAGS.virtual_batch_multiplier > 0 else FLAGS.batch_size, sample_size=len(train_data), alphas=ALPHAS, noise_multiplier=FLAGS.noise, max_grad_norm=FLAGS.max_grad_norm, experimental=bool(FLAGS.dynamic_clipping), clipping_method=FLAGS.dynamic_clipping, clip_per_layer=bool(FLAGS.per_layer_clipping), ) privacy_engine.attach(optimizer) # Do the actual training t = time.time() ( training_loss_epochs, training_accuracy_epochs, validation_loss_epochs, validation_accuracy_epochs, ) = train_validate( train_data, valid_data, model, optimizer, criterion, accuracy_fn, scheduler ) training_time = time.time() - t if FLAGS.dp: epsilon_consumed, best_alpha = optimizer.privacy_engine.get_privacy_spent( FLAGS.delta ) epsilon_consumed = float(epsilon_consumed) best_alpha = float(best_alpha) logging.info(f"Best alpha: {best_alpha}") rdp_epsilons_consumed = ( optimizer.privacy_engine.get_renyi_divergence() * optimizer.privacy_engine.steps ).tolist() logging.info(f"RDP budget consumed: {rdp_epsilons_consumed} for orders.") # Identical to planned budget when we don't have early stopping # rdp_epsilon_planned = compute_rdp_sgm( # epochs=FLAGS.n_epochs, # batch_size=FLAGS.batch_size * FLAGS.virtual_batch_multiplier # if FLAGS.virtual_batch_multiplier > 0 # else FLAGS.batch_size, # dataset_size=len(train_data), # noise=FLAGS.noise, # alphas=ALPHAS, # ) # logging.info(f"Planned RDP budget: {rdp_epsilon_planned}") else: epsilon_consumed = None rdp_epsilons_consumed = None best_alpha = None # Evaluate the model (non-DP evaluation here) testing_size = len(test_data) test_iterator = torch.utils.data.DataLoader( test_data, batch_size=FLAGS.batch_size, shuffle=True, num_workers=FLAGS.n_workers, drop_last=False, ) final_loss, final_accuracy = evaluate(model, test_iterator, criterion, accuracy_fn) # Collect the metrics and the logs logs = { "training_time": training_time, "total_time": time.time() - start_time, "test_size": testing_size, "n_trainable_parameters": count_parameters(model), } # Update the logs with the training data if isinstance(train_data, UserTimeLevelDataset): logs["train_size"] = train_data.get_n_events() logs["n_train_users"] = len(train_data) else: logs["train_size"] = len(train_data) logs.update( flags_to_dict(dataset_args, model_args, training_args) ) # Dump the configuration flags metrics = { "accuracy": final_accuracy, "training_loss_epochs": training_loss_epochs, "training_accuracy_epochs": training_accuracy_epochs, "validation_loss_epochs": validation_loss_epochs, "validation_accuracy_epochs": validation_accuracy_epochs, "loss": final_loss, "epsilon": epsilon_consumed, "target_epsilon": FLAGS.epsilon, "alphas": ALPHAS, "rdp_epsilons": rdp_epsilons_consumed, "best_alpha": best_alpha, # "dataset_files": os.listdir(FLAGS.dataset_dir), } # Save or logging.info the outputs # Useless to separate for our experiments if FLAGS.metrics_path != "": save_yaml(FLAGS.metrics_path, metrics) logging.info(f"Saved metrics: {FLAGS.metrics_path}") else: logging.info("Metrics not saved but concatenated to the logs.") logs.update(metrics) if FLAGS.log_path != "": save_yaml(FLAGS.log_path, logs) logging.info(f"Saved logs: {FLAGS.log_path}") if FLAGS.model_path != "": save_model(FLAGS.model_path, model) logging.info(f"Saved model: {FLAGS.model_path}") logging.info(logs) logging.info(metrics) if __name__ == "__main__": app.run(main)
32.379826
121
0.614323
import sys, os, shutil import h5py import time import io import random import tempfile from tqdm import tqdm from absl import app, flags, logging from ray.util.multiprocessing import Pool import gcsfs import numpy as np from pathlib import Path from sklearn.linear_model import LogisticRegression from sklearn.ensemble import RandomForestClassifier from sklearn.svm import LinearSVC from sklearn.feature_extraction.text import CountVectorizer from sklearn.feature_extraction.text import TfidfTransformer from sklearn.naive_bayes import MultinomialNB from sklearn.pipeline import Pipeline from sklearn.linear_model import SGDClassifier import torchtext import torch import torch.optim as optim from torch.optim.lr_scheduler import ReduceLROnPlateau import torch.nn as nn from transformers import BertTokenizer, BertModel, BertForSequenceClassification import opacus from privatekube.experiments.datasets import ( EventLevelDataset, split_review_batch, UserTimeLevelDataset, select_blocks_by_timeframe, ) from privatekube.experiments.utils import ( build_flags, flags_to_dict, load_yaml, results_to_dict, save_yaml, save_model, binary_accuracy, multiclass_accuracy, epoch_time, ) from privatekube.privacy.text import build_public_vocab from privatekube.privacy.rdp import ( compute_noise_from_target_epsilon, ALPHAS, compute_rdp_sgm, ) import models DEFAULT_DATA_PATH = Path(__file__).resolve().parent.parent.parent.joinpath("data") dataset_args = { "n_blocks": 200, "max_text_len": 140, "vocab_size": 10_000, "n_blocks_test": 200, } input_path_args = { "dataset_dir": "", "dataset_monofile": "", "block_counts": str(DEFAULT_DATA_PATH.joinpath("block_counts.yaml")), "emb_path": str(DEFAULT_DATA_PATH.joinpath(".vector_cache")), } model_args = { "task": "product", "model": "bow", "embedding_dim": 100, "hidden_dim_1": 240, "hidden_dim_2": 195, "hidden_dim": 100, "dropout": 0.25, } training_args = { "device": "cuda", "learning_rate": 0.01, "dp": 0, "dp_eval": 0, "user_level": 0, "epsilon": 5.0, "delta": 1e-5, "n_epochs": 15, "batch_size": 64, "virtual_batch_multiplier": 2, "adaptive_batch_size": 1, "noise": -1.0, "timeframe_days": 0, "learning_rate_scheduler": 1, "dynamic_clipping": 0, "max_grad_norm": 1.0, "per_layer_clipping": 0, "n_workers": 6, "non_dp_batch_size": 256, } output_args = { "log_path": "", "model_path": "", "metrics_path": "", } build_flags(dataset_args, model_args, training_args, input_path_args, output_args) FLAGS = flags.FLAGS np.random.seed(0) def build_split_dataset(): block_dir = tempfile.mkdtemp() test_block_dir = tempfile.mkdtemp() if FLAGS.dataset_dir[0:5] == "gs://": os.system( "gcloud auth activate-service-account --key-file=$GOOGLE_APPLICATION_CREDENTIALS" ) fs = gcsfs.GCSFileSystem( project=os.get_env("GCP_PROJECT"), token="google_default" ) logging.info("Listing bucket files.") all_blocks = list( map( lambda blob: os.path.basename(blob["name"]), fs.listdir(FLAGS.dataset_dir), ) ) logging.info(f"Got {len(all_blocks)} blocks.") logging.warning(f"The evaluation set is not fixed.") elif FLAGS.dataset_dir == "": logging.info("Listing the block names.") all_blocks = list(load_yaml(FLAGS.block_counts).keys()) else: all_blocks = os.listdir(FLAGS.dataset_dir) logging.info(f"Selecting {FLAGS.n_blocks_test} test blocks (fixed randomness).") test_blocks = np.random.choice(all_blocks, FLAGS.n_blocks_test, replace=False) for tb in test_blocks: all_blocks.remove(tb) def sort_by_user(block_name): if block_name.endswith(".h5"): block_name = block_name[: -len(".h5")] name = block_name.split("-") user_slice = int(name[1]) return user_slice logging.info( f"Selecting as few users as possible.\n Pseudorandom and deterministic (hashed user ids)." ) selected_blocks = sorted(all_blocks, key=sort_by_user)[0 : FLAGS.n_blocks] if FLAGS.dataset_dir[0:5] == "gs://": pool = Pool() bucket_path = FLAGS.dataset_dir def download_datasource(block_name): block_path = os.path.join(bucket_path, block_name) dest = os.path.join(block_dir, block_name) os.system(f"gsutil cp {block_path} {dest}") return logging.warning("Downloading the blocks in parallel.") b = pool.map(download_datasource, selected_blocks) pool.close() pool.join() block_names = None test_block_names = None elif FLAGS.dataset_dir == "": block_dir = None test_block_dir = None block_names = selected_blocks test_block_names = test_blocks else: for b in selected_blocks: os.symlink(os.path.join(FLAGS.dataset_dir, b), os.path.join(block_dir, b)) for b in test_blocks: os.symlink( os.path.join(FLAGS.dataset_dir, b), os.path.join(test_block_dir, b) ) block_names = None test_block_names = None FLAGS.dataset_dir = block_dir if not FLAGS.dataset_monofile: if FLAGS.model == "bert": from_h5 = DEFAULT_DATA_PATH.joinpath("reviews.h5") else: from_h5 = DEFAULT_DATA_PATH.joinpath("reviews_custom_vocab.h5") else: from_h5 = FLAGS.dataset_monofile if FLAGS.dp and FLAGS.user_level: train_data = UserTimeLevelDataset( blocks_dir=block_dir, timeframe=FLAGS.timeframe_days * 86400, from_h5=from_h5, block_names=block_names, ) else: train_data = EventLevelDataset( blocks_dir=block_dir, from_h5=from_h5, block_names=block_names, ) test_data = EventLevelDataset( blocks_dir=test_block_dir, from_h5=from_h5, block_names=test_block_names, ) test_data, valid_data = test_data.split([0.75, 0.25]) logging.info(f"Test size: {len(test_data)}\n Valid size: {len(valid_data)}") text_field = torchtext.data.Field( batch_first=True, use_vocab=True, init_token="<bos>", eos_token="<eos>", pad_token="<pad>", unk_token="<unk>", include_lengths=True, ) build_public_vocab( text_field, max_size=FLAGS.vocab_size - 4, vectors=f"glove.6B.{FLAGS.embedding_dim}d", unk_init=torch.Tensor.normal_, vectors_cache=FLAGS.emb_path, ) return train_data, test_data, valid_data, text_field def compute_optimal_batch_size(real_batch_size, dataset_len): logging.info( f"Computing the optimal batch size. Dataset {dataset_len}, real batch {real_batch_size}" ) # Under approximate optimal_batch_size = int(np.sqrt(dataset_len)) if optimal_batch_size <= real_batch_size: return optimal_batch_size, 0 else: return (real_batch_size, optimal_batch_size // real_batch_size) def count_parameters(model): return sum(p.numel() for p in model.parameters() if p.requires_grad) def build_model(text_field): INPUT_DIM = len(text_field.vocab) word_embeddings = text_field.vocab.vectors PAD_IDX = text_field.vocab.stoi[text_field.pad_token] UNK_IDX = text_field.vocab.stoi[text_field.unk_token] if FLAGS.task == "sentiment": output_dim = 1 elif FLAGS.task == "product": output_dim = 11 if FLAGS.model == "lstm": model = models.LSTMClassifier( batch_size=FLAGS.batch_size, output_size=output_dim, hidden_size=FLAGS.hidden_dim, vocab_size=INPUT_DIM, embedding_length=FLAGS.embedding_dim, weights=word_embeddings, dropout=FLAGS.dropout, dp=FLAGS.dp, ) elif FLAGS.model == "bow": model = models.NBOW( input_dim=word_embeddings.shape[0], emb_dim=FLAGS.embedding_dim, output_dim=output_dim, pad_idx=PAD_IDX, word_embeddings=word_embeddings, ) elif FLAGS.model == "feedforward": model = models.FeedforwardModel( vocab_size=INPUT_DIM, embedding_dim=FLAGS.embedding_dim, pad_idx=PAD_IDX, H_1=FLAGS.hidden_dim_1, H_2=FLAGS.hidden_dim_2, D_out=output_dim, word_embeddings=word_embeddings, ) elif FLAGS.model == "bert": # The dataset has been preprocessed with the bert tokenizer, so the indices should be correct logging.info(f"Pad and unk index {PAD_IDX, UNK_IDX}") model = models.FineTunedBert.build_new(output_dim=output_dim) logging.info( f"Model {FLAGS.model} has {count_parameters(model)} trainable parameters." ) # Bert has its own pretrained embeddings return model pretrained_embeddings = text_field.vocab.vectors model.embedding.weight.data.copy_(pretrained_embeddings) model.embedding.weight.data[UNK_IDX] = torch.zeros(FLAGS.embedding_dim) model.embedding.weight.data[PAD_IDX] = torch.zeros(FLAGS.embedding_dim) logging.info( f"Model {FLAGS.model} has {count_parameters(model)} trainable parameters." ) return model def train(model, iterator, optimizer, criterion, accuracy_fn): epoch_loss = 0 epoch_acc = 0 model.train() optimizer.zero_grad() for i, batch in enumerate(tqdm(iterator)): # batch = batch.to(FLAGS.device) if FLAGS.task == "sentiment": data, label = split_review_batch( batch, label_feature="binary_rating", max_text_len=FLAGS.max_text_len, include_len=True, vocab_size=FLAGS.vocab_size, custom_vocab=(FLAGS.model != "bert"), ) text_lengths, text = data elif FLAGS.task == "product": text, label = split_review_batch( batch, label_feature="category", max_text_len=FLAGS.max_text_len, vocab_size=FLAGS.vocab_size, custom_vocab=(FLAGS.model != "bert"), ) text = text.to(device=FLAGS.device, dtype=torch.long) label = ( label.to(device=FLAGS.device, dtype=torch.long) if FLAGS.task == "product" else label.to(device=FLAGS.device, dtype=torch.float) ) if FLAGS.model == "lstm": hidden = model.init_hidden(batch_size=len(batch)) if isinstance(hidden, tuple): hidden = ( hidden[0].to(FLAGS.device), hidden[1].to(FLAGS.device), ) else: hidden = hidden.to(FLAGS.device) outputs = model(text, hidden) elif FLAGS.model == "bert": PAD_IDX = 0 inputs = { "input_ids": text, "labels": label, "attention_mask": torch.where( text == PAD_IDX, torch.zeros_like(text), torch.ones_like(text) ), } # logging.info(f"Inputs {inputs}") # The model outputs loss, logits outputs = model(**inputs)[1] # logging.info(f"Outputs {outputs}") else: outputs = model(text) # logging.info(f"Outputs {outputs}") if FLAGS.task == "sentiment": outputs = outputs.squeeze(1) loss = criterion(outputs, label) acc = accuracy_fn(outputs.detach(), label) loss.backward() if FLAGS.dp and FLAGS.virtual_batch_multiplier > 1: # NOTE: step is not called at every minibatch, so the RDP accountant need to know this if (i + 1) % FLAGS.virtual_batch_multiplier == 0 or (i + 1) == len( iterator ): # For the (virtual_batch_multiplier)th batch, call a clip-noise-step optimizer.step() optimizer.zero_grad() else: # For the first (virtual_batch_multiplier - 1) batches, just accumulate the gradients optimizer.virtual_step() else: # Regular optimizer step (either non-DP or DP with no virtual step) optimizer.step() optimizer.zero_grad() epoch_loss += loss.item() # epoch_loss += loss.detach().item() epoch_acc += acc.item() return epoch_loss / len(iterator), epoch_acc / len(iterator) def evaluate(model, iterator, criterion, accuracy_fn): epoch_loss = 0 epoch_acc = 0 model.eval() with torch.no_grad(): for batch in iterator: # batch = batch.to(FLAGS.device) if FLAGS.task == "sentiment": data, label = split_review_batch( batch, label_feature="binary_rating", max_text_len=FLAGS.max_text_len, include_len=True, vocab_size=FLAGS.vocab_size, custom_vocab=(FLAGS.model != "bert"), ) text_lengths, text = data elif FLAGS.task == "product": text, label = split_review_batch( batch, label_feature="category", max_text_len=FLAGS.max_text_len, vocab_size=FLAGS.vocab_size, custom_vocab=(FLAGS.model != "bert"), ) text = text.to(device=FLAGS.device, dtype=torch.long) label = ( label.to(device=FLAGS.device, dtype=torch.long) if FLAGS.task == "product" else label.to(device=FLAGS.device, dtype=torch.float) ) if FLAGS.model == "lstm": hidden = model.init_hidden(batch_size=len(batch)) if isinstance(hidden, tuple): hidden = ( hidden[0].to(FLAGS.device), hidden[1].to(FLAGS.device), ) else: hidden = hidden.to(FLAGS.device) outputs = model(text, hidden) elif FLAGS.model == "bert": PAD_IDX = 0 inputs = { "input_ids": text, "labels": label, "attention_mask": torch.where( text == PAD_IDX, torch.zeros_like(text), torch.ones_like(text) ), } outputs = model(**inputs)[1] else: outputs = model(text) if FLAGS.task == "sentiment": outputs = outputs.squeeze(1) # print(f"Training. Outputs: {outputs}, labels: {batch.label}") loss = criterion(outputs, label) acc = accuracy_fn(outputs, label) epoch_loss += loss.item() epoch_acc += acc.item() return epoch_loss / len(iterator), epoch_acc / len(iterator) def train_validate( train_data, valid_data, model, optimizer, criterion, accuracy_fn, scheduler ): validation_accuracy_epochs = [] validation_loss_epochs = [] training_loss_epochs = [] training_accuracy_epochs = [] logging.info(f"n workers: {FLAGS.n_workers}") train_iterator = torch.utils.data.DataLoader( train_data, batch_size=FLAGS.batch_size, shuffle=True, num_workers=FLAGS.n_workers, drop_last=True, ) valid_iterator = torch.utils.data.DataLoader( valid_data, batch_size=FLAGS.batch_size, shuffle=True, num_workers=FLAGS.n_workers, drop_last=False, ) criterion = criterion.to(FLAGS.device) best_valid_loss = float("inf") for epoch in range(FLAGS.n_epochs): start_time = time.time() logging.info(f"Starting epoch {epoch + 1}.") train_loss, train_acc = train( model, train_iterator, optimizer, criterion, accuracy_fn ) valid_loss, valid_acc = evaluate(model, valid_iterator, criterion, accuracy_fn) end_time = time.time() epoch_mins, epoch_secs = epoch_time(start_time, end_time) if valid_loss < best_valid_loss: best_valid_loss = valid_loss torch.save(model.state_dict(), "tut2-model.pt") logging.info(f"Epoch: {epoch+1:02} | Epoch Time: {epoch_mins}m {epoch_secs}s") logging.info( f"\tTrain Loss: {train_loss:.3f} | Train Acc: {train_acc*100:.2f}%" ) scheduler.step(train_loss) logging.info( f"\t Val. Loss: {valid_loss:.3f} | Val. Acc: {valid_acc*100:.2f}%" ) validation_accuracy_epochs.append(valid_acc) validation_loss_epochs.append(valid_loss) training_loss_epochs.append(train_loss) training_accuracy_epochs.append(train_acc) return ( training_loss_epochs, training_accuracy_epochs, validation_loss_epochs, validation_accuracy_epochs, ) def count_parameters(model): return sum(p.numel() for p in model.parameters() if p.requires_grad) def main(argv): start_time = time.time() # Convert flags for the epsilon = -1 shortcut if FLAGS.dp and FLAGS.epsilon < 0 and FLAGS.noise < 0: FLAGS.dp = False # No multiprocessing for large datasets (save RAM) if FLAGS.n_blocks > 50_000: logging.info(f"Large dataset, we use a single thread for the loader.") FLAGS.n_workers = 0 # Build the dataset, either event level or user level train_data, test_data, valid_data, text_field = build_split_dataset() logging.info( f"Number of samples for training: {len(train_data)}, validation: {len(valid_data)} and testing: {len(test_data)}" ) # Adapt the batch size and the virtual step size, unless it has been specified manually if FLAGS.dp and FLAGS.adaptive_batch_size and FLAGS.virtual_batch_multiplier <= 0: FLAGS.batch_size, FLAGS.virtual_batch_multiplier = compute_optimal_batch_size( FLAGS.batch_size, len(train_data) ) logging.info( f"Using real batch {FLAGS.batch_size} with multiplier {FLAGS.virtual_batch_multiplier}" ) if not FLAGS.dp: FLAGS.batch_size = FLAGS.non_dp_batch_size # Prepare the model and optimizer model = build_model(text_field).to(FLAGS.device) logging.info(f"Number of trainable parameters: {count_parameters(model)}") # optimizer = optim.Adam(model.parameters()) optimizer = optim.AdamW(model.parameters(), lr=FLAGS.learning_rate, eps=1e-8) scheduler = ReduceLROnPlateau(optimizer, mode="min", patience=3) # train_it = torch.utils.data.DataLoader( # train_data, # batch_size=2048, # shuffle=False, # num_workers=FLAGS.n_workers, # drop_last=False, # ) # counts = {} # for i in range(11): # counts[i] = 0 # for b in train_it: # for cat in b[:, 3]: # counts[int(cat)] += 1 # s = sum(counts.values()) # for cat, count in counts.items(): # counts[cat] = count / s # logging.info(counts) if FLAGS.task == "sentiment": criterion = nn.BCEWithLogitsLoss().to(FLAGS.device) accuracy_fn = binary_accuracy # automotive: 0.03036145803296712 # books: 0.41258122723567553 # cds: 0.012897189083383703 # clothing: 0.2025265712144095 # games: 0.031613111956201506 # groceries: 0.01949595483554337 # home: 0.119920985593197 # movies: 0.0484712255807162 # pets: 0.03665525816121956 # sports: 0.04961580907019007 # tools: 0.035861209236496445 elif FLAGS.task == "product": # criterion = nn.CrossEntropyLoss( # weight=torch.Tensor( # [0.05, 0.035, 0.03, 0.035, 0.05, 0.02, 0.12, 0.01, 0.03, 0.20, 0.41] # ) # ) criterion = nn.CrossEntropyLoss() accuracy_fn = multiclass_accuracy # Plug Opacus if DP training is activated if FLAGS.dp: if FLAGS.noise >= 0: logging.info(f"User-provided noise: {FLAGS.noise}.") else: logging.info("Computing noise for the given parameters.") FLAGS.noise = compute_noise_from_target_epsilon( target_epsilon=FLAGS.epsilon, target_delta=FLAGS.delta, epochs=FLAGS.n_epochs, batch_size=FLAGS.batch_size * FLAGS.virtual_batch_multiplier if FLAGS.virtual_batch_multiplier > 0 else FLAGS.batch_size, dataset_size=len(train_data), alphas=ALPHAS, ) logging.info(f"Noise computed from RDP budget: {FLAGS.noise}.") # NOTE: when user-level DP is activated, the training dataset __len__ method returns # the number of users, and the DataLoader calls the batch-of-user method that overrides # the regular __getitem__ method # WARNING: fishy non-DP adaptive clipping privacy_engine = opacus.PrivacyEngine( module=model, batch_size=FLAGS.batch_size * FLAGS.virtual_batch_multiplier if FLAGS.virtual_batch_multiplier > 0 else FLAGS.batch_size, sample_size=len(train_data), alphas=ALPHAS, noise_multiplier=FLAGS.noise, max_grad_norm=FLAGS.max_grad_norm, experimental=bool(FLAGS.dynamic_clipping), clipping_method=FLAGS.dynamic_clipping, clip_per_layer=bool(FLAGS.per_layer_clipping), ) privacy_engine.attach(optimizer) # Do the actual training t = time.time() ( training_loss_epochs, training_accuracy_epochs, validation_loss_epochs, validation_accuracy_epochs, ) = train_validate( train_data, valid_data, model, optimizer, criterion, accuracy_fn, scheduler ) training_time = time.time() - t if FLAGS.dp: epsilon_consumed, best_alpha = optimizer.privacy_engine.get_privacy_spent( FLAGS.delta ) epsilon_consumed = float(epsilon_consumed) best_alpha = float(best_alpha) logging.info(f"Best alpha: {best_alpha}") rdp_epsilons_consumed = ( optimizer.privacy_engine.get_renyi_divergence() * optimizer.privacy_engine.steps ).tolist() logging.info(f"RDP budget consumed: {rdp_epsilons_consumed} for orders.") # Identical to planned budget when we don't have early stopping else: epsilon_consumed = None rdp_epsilons_consumed = None best_alpha = None testing_size = len(test_data) test_iterator = torch.utils.data.DataLoader( test_data, batch_size=FLAGS.batch_size, shuffle=True, num_workers=FLAGS.n_workers, drop_last=False, ) final_loss, final_accuracy = evaluate(model, test_iterator, criterion, accuracy_fn) logs = { "training_time": training_time, "total_time": time.time() - start_time, "test_size": testing_size, "n_trainable_parameters": count_parameters(model), } if isinstance(train_data, UserTimeLevelDataset): logs["train_size"] = train_data.get_n_events() logs["n_train_users"] = len(train_data) else: logs["train_size"] = len(train_data) logs.update( flags_to_dict(dataset_args, model_args, training_args) ) metrics = { "accuracy": final_accuracy, "training_loss_epochs": training_loss_epochs, "training_accuracy_epochs": training_accuracy_epochs, "validation_loss_epochs": validation_loss_epochs, "validation_accuracy_epochs": validation_accuracy_epochs, "loss": final_loss, "epsilon": epsilon_consumed, "target_epsilon": FLAGS.epsilon, "alphas": ALPHAS, "rdp_epsilons": rdp_epsilons_consumed, "best_alpha": best_alpha, } if FLAGS.metrics_path != "": save_yaml(FLAGS.metrics_path, metrics) logging.info(f"Saved metrics: {FLAGS.metrics_path}") else: logging.info("Metrics not saved but concatenated to the logs.") logs.update(metrics) if FLAGS.log_path != "": save_yaml(FLAGS.log_path, logs) logging.info(f"Saved logs: {FLAGS.log_path}") if FLAGS.model_path != "": save_model(FLAGS.model_path, model) logging.info(f"Saved model: {FLAGS.model_path}") logging.info(logs) logging.info(metrics) if __name__ == "__main__": app.run(main)
true
true
f70e95956c365ef8a6f6d6a43deb9ae598d92d2e
8,683
py
Python
xonsh/xoreutils/uptime.py
halloleo/xonsh
3a34d9f2e347d5b4518d42a25a8b4080001ec748
[ "BSD-2-Clause-FreeBSD" ]
1
2020-10-30T15:13:54.000Z
2020-10-30T15:13:54.000Z
xonsh/xoreutils/uptime.py
halloleo/xonsh
3a34d9f2e347d5b4518d42a25a8b4080001ec748
[ "BSD-2-Clause-FreeBSD" ]
null
null
null
xonsh/xoreutils/uptime.py
halloleo/xonsh
3a34d9f2e347d5b4518d42a25a8b4080001ec748
[ "BSD-2-Clause-FreeBSD" ]
null
null
null
""" Provides a cross-platform way to figure out the system uptime. Should work on damned near any operating system you can realistically expect to be asked to write Python code for. If this module is invoked as a stand-alone script, it will print the current uptime in a human-readable format, or display an error message if it can't, to standard output. This file was forked from the uptime project: https://github.com/Cairnarvon/uptime Copyright (c) 2012, Koen Crolla, All rights reserved. """ import os import sys import time import ctypes import struct import typing as tp import xonsh.platform as xp import xonsh.lazyimps as xlimps import xonsh.lazyasd as xl _BOOTTIME: tp.Optional[float] = None def _uptime_osx(): """Returns the uptime on mac / darwin.""" global _BOOTTIME bt = xlimps.macutils.sysctlbyname(b"kern.boottime", return_str=False) if len(bt) == 4: bt = struct.unpack_from("@hh", bt) elif len(bt) == 8: bt = struct.unpack_from("@ii", bt) elif len(bt) == 16: bt = struct.unpack_from("@qq", bt) else: raise ValueError("length of boot time not understood: " + repr(bt)) bt = bt[0] + bt[1] * 1e-6 if bt == 0.0: return None _BOOTTIME = bt return time.time() - bt def _uptime_linux(): """Returns uptime in seconds or None, on Linux.""" # With procfs try: with open("/proc/uptime") as f: up = float(f.readline().split()[0]) return up except (OSError, ValueError): pass buf = ctypes.create_string_buffer(128) # 64 suffices on 32-bit, whatever. if xp.LIBC.sysinfo(buf) < 0: return None up = struct.unpack_from("@l", buf.raw)[0] if up < 0: up = None return up def _boottime_linux(): """A way to figure out the boot time directly on Linux.""" global _BOOTTIME try: with open("/proc/stat") as f: for line in f: if line.startswith("btime"): _BOOTTIME = float(line.split()[1]) return _BOOTTIME except (OSError, IndexError): return None def _uptime_amiga(): """Returns uptime in seconds or None, on AmigaOS.""" global _BOOTTIME try: _BOOTTIME = os.stat("RAM:").st_ctime return time.time() - _BOOTTIME except (NameError, OSError): return None def _uptime_beos(): """Returns uptime in seconds on None, on BeOS/Haiku.""" if not hasattr(xp.LIBC, "system_time"): return None xp.LIBC.system_time.restype = ctypes.c_int64 return xp.LIBC.system_time() / 1000000.0 def _uptime_bsd(): """Returns uptime in seconds or None, on BSD (including OS X).""" global _BOOTTIME if not hasattr(xp.LIBC, "sysctlbyname"): # Not BSD. return None # Determine how much space we need for the response. sz = ctypes.c_uint(0) xp.LIBC.sysctlbyname("kern.boottime", None, ctypes.byref(sz), None, 0) if sz.value != struct.calcsize("@LL"): # Unexpected, let's give up. return None # For real now. buf = ctypes.create_string_buffer(sz.value) xp.LIBC.sysctlbyname("kern.boottime", buf, ctypes.byref(sz), None, 0) sec, usec = struct.unpack_from("@LL", buf.raw) # OS X disagrees what that second value is. if usec > 1000000: usec = 0.0 _BOOTTIME = sec + usec / 1000000.0 up = time.time() - _BOOTTIME if up < 0: up = None return up def _uptime_minix(): """Returns uptime in seconds or None, on MINIX.""" try: with open("/proc/uptime") as f: up = float(f.read()) return up except (OSError, ValueError): return None def _uptime_plan9(): """Returns uptime in seconds or None, on Plan 9.""" # Apparently Plan 9 only has Python 2.2, which I'm not prepared to # support. Maybe some Linuxes implement /dev/time, though, someone was # talking about it somewhere. try: # The time file holds one 32-bit number representing the sec- # onds since start of epoch and three 64-bit numbers, repre- # senting nanoseconds since start of epoch, clock ticks, and # clock frequency. # -- cons(3) with open("/dev/time") as f: s, ns, ct, cf = f.read().split() return float(ct) / float(cf) except (OSError, ValueError): return None def _uptime_solaris(): """Returns uptime in seconds or None, on Solaris.""" global _BOOTTIME try: kstat = ctypes.CDLL("libkstat.so") except (AttributeError, OSError): return None # kstat doesn't have uptime, but it does have boot time. # Unfortunately, getting at it isn't perfectly straightforward. # First, let's pretend to be kstat.h # Constant KSTAT_STRLEN = 31 # According to every kstat.h I could find. # Data structures class anon_union(ctypes.Union): # The ``value'' union in kstat_named_t actually has a bunch more # members, but we're only using it for boot_time, so we only need # the padding and the one we're actually using. _fields_ = [("c", ctypes.c_char * 16), ("time", ctypes.c_int)] class kstat_named_t(ctypes.Structure): _fields_ = [ ("name", ctypes.c_char * KSTAT_STRLEN), ("data_type", ctypes.c_char), ("value", anon_union), ] # Function signatures kstat.kstat_open.restype = ctypes.c_void_p kstat.kstat_lookup.restype = ctypes.c_void_p kstat.kstat_lookup.argtypes = [ ctypes.c_void_p, ctypes.c_char_p, ctypes.c_int, ctypes.c_char_p, ] kstat.kstat_read.restype = ctypes.c_int kstat.kstat_read.argtypes = [ctypes.c_void_p, ctypes.c_void_p, ctypes.c_void_p] kstat.kstat_data_lookup.restype = ctypes.POINTER(kstat_named_t) kstat.kstat_data_lookup.argtypes = [ctypes.c_void_p, ctypes.c_char_p] # Now, let's do something useful. # Initialise kstat control structure. kc = kstat.kstat_open() if not kc: return None # We're looking for unix:0:system_misc:boot_time. ksp = kstat.kstat_lookup(kc, "unix", 0, "system_misc") if ksp and kstat.kstat_read(kc, ksp, None) != -1: data = kstat.kstat_data_lookup(ksp, "boot_time") if data: _BOOTTIME = data.contents.value.time # Clean-up. kstat.kstat_close(kc) if _BOOTTIME is not None: return time.time() - _BOOTTIME return None def _uptime_syllable(): """Returns uptime in seconds or None, on Syllable.""" global _BOOTTIME try: _BOOTTIME = os.stat("/dev/pty/mst/pty0").st_mtime return time.time() - _BOOTTIME except (NameError, OSError): return None def _uptime_windows(): """ Returns uptime in seconds or None, on Windows. Warning: may return incorrect answers after 49.7 days on versions older than Vista. """ if hasattr(xp.LIBC, "GetTickCount64"): # Vista/Server 2008 or later. xp.LIBC.GetTickCount64.restype = ctypes.c_uint64 return xp.LIBC.GetTickCount64() / 1000.0 if hasattr(xp.LIBC, "GetTickCount"): # WinCE and Win2k or later; gives wrong answers after 49.7 days. xp.LIBC.GetTickCount.restype = ctypes.c_uint32 return xp.LIBC.GetTickCount() / 1000.0 return None @xl.lazyobject def _UPTIME_FUNCS(): return { "amiga": _uptime_amiga, "aros12": _uptime_amiga, "beos5": _uptime_beos, "cygwin": _uptime_linux, "darwin": _uptime_osx, "haiku1": _uptime_beos, "linux": _uptime_linux, "linux-armv71": _uptime_linux, "linux2": _uptime_linux, "minix3": _uptime_minix, "sunos5": _uptime_solaris, "syllable": _uptime_syllable, "win32": _uptime_windows, "wince": _uptime_windows, } def uptime(): """Returns uptime in seconds if even remotely possible, or None if not.""" if _BOOTTIME is not None: return time.time() - _BOOTTIME up = _UPTIME_FUNCS.get(sys.platform, _uptime_bsd)() if up is None: up = ( _uptime_bsd() or _uptime_plan9() or _uptime_linux() or _uptime_windows() or _uptime_solaris() or _uptime_beos() or _uptime_amiga() or _uptime_syllable() or _uptime_osx() ) return up def boottime(): """Returns boot time if remotely possible, or None if not.""" global _BOOTTIME if _BOOTTIME is None: up = uptime() if up is None: return None _BOOTTIME = time.time() - up return _BOOTTIME
30.466667
83
0.625014
import os import sys import time import ctypes import struct import typing as tp import xonsh.platform as xp import xonsh.lazyimps as xlimps import xonsh.lazyasd as xl _BOOTTIME: tp.Optional[float] = None def _uptime_osx(): global _BOOTTIME bt = xlimps.macutils.sysctlbyname(b"kern.boottime", return_str=False) if len(bt) == 4: bt = struct.unpack_from("@hh", bt) elif len(bt) == 8: bt = struct.unpack_from("@ii", bt) elif len(bt) == 16: bt = struct.unpack_from("@qq", bt) else: raise ValueError("length of boot time not understood: " + repr(bt)) bt = bt[0] + bt[1] * 1e-6 if bt == 0.0: return None _BOOTTIME = bt return time.time() - bt def _uptime_linux(): try: with open("/proc/uptime") as f: up = float(f.readline().split()[0]) return up except (OSError, ValueError): pass buf = ctypes.create_string_buffer(128) if xp.LIBC.sysinfo(buf) < 0: return None up = struct.unpack_from("@l", buf.raw)[0] if up < 0: up = None return up def _boottime_linux(): global _BOOTTIME try: with open("/proc/stat") as f: for line in f: if line.startswith("btime"): _BOOTTIME = float(line.split()[1]) return _BOOTTIME except (OSError, IndexError): return None def _uptime_amiga(): global _BOOTTIME try: _BOOTTIME = os.stat("RAM:").st_ctime return time.time() - _BOOTTIME except (NameError, OSError): return None def _uptime_beos(): if not hasattr(xp.LIBC, "system_time"): return None xp.LIBC.system_time.restype = ctypes.c_int64 return xp.LIBC.system_time() / 1000000.0 def _uptime_bsd(): global _BOOTTIME if not hasattr(xp.LIBC, "sysctlbyname"): return None sz = ctypes.c_uint(0) xp.LIBC.sysctlbyname("kern.boottime", None, ctypes.byref(sz), None, 0) if sz.value != struct.calcsize("@LL"): return None # For real now. buf = ctypes.create_string_buffer(sz.value) xp.LIBC.sysctlbyname("kern.boottime", buf, ctypes.byref(sz), None, 0) sec, usec = struct.unpack_from("@LL", buf.raw) # OS X disagrees what that second value is. if usec > 1000000: usec = 0.0 _BOOTTIME = sec + usec / 1000000.0 up = time.time() - _BOOTTIME if up < 0: up = None return up def _uptime_minix(): try: with open("/proc/uptime") as f: up = float(f.read()) return up except (OSError, ValueError): return None def _uptime_plan9(): # Apparently Plan 9 only has Python 2.2, which I'm not prepared to try: with open("/dev/time") as f: s, ns, ct, cf = f.read().split() return float(ct) / float(cf) except (OSError, ValueError): return None def _uptime_solaris(): global _BOOTTIME try: kstat = ctypes.CDLL("libkstat.so") except (AttributeError, OSError): return None # Unfortunately, getting at it isn't perfectly straightforward. # Constant KSTAT_STRLEN = 31 # According to every kstat.h I could find. # Data structures class anon_union(ctypes.Union): # The ``value'' union in kstat_named_t actually has a bunch more # members, but we're only using it for boot_time, so we only need _fields_ = [("c", ctypes.c_char * 16), ("time", ctypes.c_int)] class kstat_named_t(ctypes.Structure): _fields_ = [ ("name", ctypes.c_char * KSTAT_STRLEN), ("data_type", ctypes.c_char), ("value", anon_union), ] # Function signatures kstat.kstat_open.restype = ctypes.c_void_p kstat.kstat_lookup.restype = ctypes.c_void_p kstat.kstat_lookup.argtypes = [ ctypes.c_void_p, ctypes.c_char_p, ctypes.c_int, ctypes.c_char_p, ] kstat.kstat_read.restype = ctypes.c_int kstat.kstat_read.argtypes = [ctypes.c_void_p, ctypes.c_void_p, ctypes.c_void_p] kstat.kstat_data_lookup.restype = ctypes.POINTER(kstat_named_t) kstat.kstat_data_lookup.argtypes = [ctypes.c_void_p, ctypes.c_char_p] # Now, let's do something useful. kc = kstat.kstat_open() if not kc: return None ksp = kstat.kstat_lookup(kc, "unix", 0, "system_misc") if ksp and kstat.kstat_read(kc, ksp, None) != -1: data = kstat.kstat_data_lookup(ksp, "boot_time") if data: _BOOTTIME = data.contents.value.time # Clean-up. kstat.kstat_close(kc) if _BOOTTIME is not None: return time.time() - _BOOTTIME return None def _uptime_syllable(): global _BOOTTIME try: _BOOTTIME = os.stat("/dev/pty/mst/pty0").st_mtime return time.time() - _BOOTTIME except (NameError, OSError): return None def _uptime_windows(): if hasattr(xp.LIBC, "GetTickCount64"): # Vista/Server 2008 or later. xp.LIBC.GetTickCount64.restype = ctypes.c_uint64 return xp.LIBC.GetTickCount64() / 1000.0 if hasattr(xp.LIBC, "GetTickCount"): # WinCE and Win2k or later; gives wrong answers after 49.7 days. xp.LIBC.GetTickCount.restype = ctypes.c_uint32 return xp.LIBC.GetTickCount() / 1000.0 return None @xl.lazyobject def _UPTIME_FUNCS(): return { "amiga": _uptime_amiga, "aros12": _uptime_amiga, "beos5": _uptime_beos, "cygwin": _uptime_linux, "darwin": _uptime_osx, "haiku1": _uptime_beos, "linux": _uptime_linux, "linux-armv71": _uptime_linux, "linux2": _uptime_linux, "minix3": _uptime_minix, "sunos5": _uptime_solaris, "syllable": _uptime_syllable, "win32": _uptime_windows, "wince": _uptime_windows, } def uptime(): if _BOOTTIME is not None: return time.time() - _BOOTTIME up = _UPTIME_FUNCS.get(sys.platform, _uptime_bsd)() if up is None: up = ( _uptime_bsd() or _uptime_plan9() or _uptime_linux() or _uptime_windows() or _uptime_solaris() or _uptime_beos() or _uptime_amiga() or _uptime_syllable() or _uptime_osx() ) return up def boottime(): global _BOOTTIME if _BOOTTIME is None: up = uptime() if up is None: return None _BOOTTIME = time.time() - up return _BOOTTIME
true
true
f70e96217e805d497d912559b61045f8ea83f841
6,295
py
Python
xarray/core/nanops.py
LEGOS-CTOH/xarray
d543d09aaa7fdfc4f5f92edcd4e3c0af1207c95b
[ "Apache-2.0" ]
1
2021-02-12T02:15:33.000Z
2021-02-12T02:15:33.000Z
xarray/core/nanops.py
LEGOS-CTOH/xarray
d543d09aaa7fdfc4f5f92edcd4e3c0af1207c95b
[ "Apache-2.0" ]
null
null
null
xarray/core/nanops.py
LEGOS-CTOH/xarray
d543d09aaa7fdfc4f5f92edcd4e3c0af1207c95b
[ "Apache-2.0" ]
null
null
null
import numpy as np from . import dtypes, nputils, utils from .duck_array_ops import _dask_or_eager_func, count, fillna, isnull, where_method from .pycompat import dask_array_type try: import dask.array as dask_array except ImportError: dask_array = None def _replace_nan(a, val): """ replace nan in a by val, and returns the replaced array and the nan position """ mask = isnull(a) return where_method(val, mask, a), mask def _maybe_null_out(result, axis, mask, min_count=1): """ xarray version of pandas.core.nanops._maybe_null_out """ if hasattr(axis, "__len__"): # if tuple or list raise ValueError( "min_count is not available for reduction with more than one dimensions." ) if axis is not None and getattr(result, "ndim", False): null_mask = (mask.shape[axis] - mask.sum(axis) - min_count) < 0 if null_mask.any(): dtype, fill_value = dtypes.maybe_promote(result.dtype) result = result.astype(dtype) result[null_mask] = fill_value elif getattr(result, "dtype", None) not in dtypes.NAT_TYPES: null_mask = mask.size - mask.sum() if null_mask < min_count: result = np.nan return result def _nan_argminmax_object(func, fill_value, value, axis=None, **kwargs): """ In house nanargmin, nanargmax for object arrays. Always return integer type """ valid_count = count(value, axis=axis) value = fillna(value, fill_value) data = _dask_or_eager_func(func)(value, axis=axis, **kwargs) # TODO This will evaluate dask arrays and might be costly. if (valid_count == 0).any(): raise ValueError("All-NaN slice encountered") return data def _nan_minmax_object(func, fill_value, value, axis=None, **kwargs): """ In house nanmin and nanmax for object array """ valid_count = count(value, axis=axis) filled_value = fillna(value, fill_value) data = getattr(np, func)(filled_value, axis=axis, **kwargs) if not hasattr(data, "dtype"): # scalar case data = fill_value if valid_count == 0 else data # we've computed a single min, max value of type object. # don't let np.array turn a tuple back into an array return utils.to_0d_object_array(data) return where_method(data, valid_count != 0) def nanmin(a, axis=None, out=None): if a.dtype.kind == "O": return _nan_minmax_object("min", dtypes.get_pos_infinity(a.dtype), a, axis) module = dask_array if isinstance(a, dask_array_type) else nputils return module.nanmin(a, axis=axis) def nanmax(a, axis=None, out=None): if a.dtype.kind == "O": return _nan_minmax_object("max", dtypes.get_neg_infinity(a.dtype), a, axis) module = dask_array if isinstance(a, dask_array_type) else nputils return module.nanmax(a, axis=axis) def nanargmin(a, axis=None): if a.dtype.kind == "O": fill_value = dtypes.get_pos_infinity(a.dtype) return _nan_argminmax_object("argmin", fill_value, a, axis=axis) module = dask_array if isinstance(a, dask_array_type) else nputils return module.nanargmin(a, axis=axis) def nanargmax(a, axis=None): if a.dtype.kind == "O": fill_value = dtypes.get_neg_infinity(a.dtype) return _nan_argminmax_object("argmax", fill_value, a, axis=axis) module = dask_array if isinstance(a, dask_array_type) else nputils return module.nanargmax(a, axis=axis) def nansum(a, axis=None, dtype=None, out=None, min_count=None): a, mask = _replace_nan(a, 0) result = _dask_or_eager_func("sum")(a, axis=axis, dtype=dtype) if min_count is not None: return _maybe_null_out(result, axis, mask, min_count) else: return result def _nanmean_ddof_object(ddof, value, axis=None, dtype=None, **kwargs): """ In house nanmean. ddof argument will be used in _nanvar method """ from .duck_array_ops import count, fillna, _dask_or_eager_func, where_method valid_count = count(value, axis=axis) value = fillna(value, 0) # As dtype inference is impossible for object dtype, we assume float # https://github.com/dask/dask/issues/3162 if dtype is None and value.dtype.kind == "O": dtype = value.dtype if value.dtype.kind in ["cf"] else float data = _dask_or_eager_func("sum")(value, axis=axis, dtype=dtype, **kwargs) data = data / (valid_count - ddof) return where_method(data, valid_count != 0) def nanmean(a, axis=None, dtype=None, out=None): if a.dtype.kind == "O": return _nanmean_ddof_object(0, a, axis=axis, dtype=dtype) if isinstance(a, dask_array_type): return dask_array.nanmean(a, axis=axis, dtype=dtype) return np.nanmean(a, axis=axis, dtype=dtype) def nanmedian(a, axis=None, out=None): return _dask_or_eager_func("nanmedian", eager_module=nputils)(a, axis=axis) def _nanvar_object(value, axis=None, ddof=0, keepdims=False, **kwargs): value_mean = _nanmean_ddof_object( ddof=0, value=value, axis=axis, keepdims=True, **kwargs ) squared = (value.astype(value_mean.dtype) - value_mean) ** 2 return _nanmean_ddof_object(ddof, squared, axis=axis, keepdims=keepdims, **kwargs) def nanvar(a, axis=None, dtype=None, out=None, ddof=0): if a.dtype.kind == "O": return _nanvar_object(a, axis=axis, dtype=dtype, ddof=ddof) return _dask_or_eager_func("nanvar", eager_module=nputils)( a, axis=axis, dtype=dtype, ddof=ddof ) def nanstd(a, axis=None, dtype=None, out=None, ddof=0): return _dask_or_eager_func("nanstd", eager_module=nputils)( a, axis=axis, dtype=dtype, ddof=ddof ) def nanprod(a, axis=None, dtype=None, out=None, min_count=None): a, mask = _replace_nan(a, 1) result = _dask_or_eager_func("nanprod")(a, axis=axis, dtype=dtype, out=out) if min_count is not None: return _maybe_null_out(result, axis, mask, min_count) else: return result def nancumsum(a, axis=None, dtype=None, out=None): return _dask_or_eager_func("nancumsum", eager_module=nputils)( a, axis=axis, dtype=dtype ) def nancumprod(a, axis=None, dtype=None, out=None): return _dask_or_eager_func("nancumprod", eager_module=nputils)( a, axis=axis, dtype=dtype )
33.306878
86
0.678793
import numpy as np from . import dtypes, nputils, utils from .duck_array_ops import _dask_or_eager_func, count, fillna, isnull, where_method from .pycompat import dask_array_type try: import dask.array as dask_array except ImportError: dask_array = None def _replace_nan(a, val): mask = isnull(a) return where_method(val, mask, a), mask def _maybe_null_out(result, axis, mask, min_count=1): if hasattr(axis, "__len__"): raise ValueError( "min_count is not available for reduction with more than one dimensions." ) if axis is not None and getattr(result, "ndim", False): null_mask = (mask.shape[axis] - mask.sum(axis) - min_count) < 0 if null_mask.any(): dtype, fill_value = dtypes.maybe_promote(result.dtype) result = result.astype(dtype) result[null_mask] = fill_value elif getattr(result, "dtype", None) not in dtypes.NAT_TYPES: null_mask = mask.size - mask.sum() if null_mask < min_count: result = np.nan return result def _nan_argminmax_object(func, fill_value, value, axis=None, **kwargs): valid_count = count(value, axis=axis) value = fillna(value, fill_value) data = _dask_or_eager_func(func)(value, axis=axis, **kwargs) if (valid_count == 0).any(): raise ValueError("All-NaN slice encountered") return data def _nan_minmax_object(func, fill_value, value, axis=None, **kwargs): valid_count = count(value, axis=axis) filled_value = fillna(value, fill_value) data = getattr(np, func)(filled_value, axis=axis, **kwargs) if not hasattr(data, "dtype"): data = fill_value if valid_count == 0 else data # don't let np.array turn a tuple back into an array return utils.to_0d_object_array(data) return where_method(data, valid_count != 0) def nanmin(a, axis=None, out=None): if a.dtype.kind == "O": return _nan_minmax_object("min", dtypes.get_pos_infinity(a.dtype), a, axis) module = dask_array if isinstance(a, dask_array_type) else nputils return module.nanmin(a, axis=axis) def nanmax(a, axis=None, out=None): if a.dtype.kind == "O": return _nan_minmax_object("max", dtypes.get_neg_infinity(a.dtype), a, axis) module = dask_array if isinstance(a, dask_array_type) else nputils return module.nanmax(a, axis=axis) def nanargmin(a, axis=None): if a.dtype.kind == "O": fill_value = dtypes.get_pos_infinity(a.dtype) return _nan_argminmax_object("argmin", fill_value, a, axis=axis) module = dask_array if isinstance(a, dask_array_type) else nputils return module.nanargmin(a, axis=axis) def nanargmax(a, axis=None): if a.dtype.kind == "O": fill_value = dtypes.get_neg_infinity(a.dtype) return _nan_argminmax_object("argmax", fill_value, a, axis=axis) module = dask_array if isinstance(a, dask_array_type) else nputils return module.nanargmax(a, axis=axis) def nansum(a, axis=None, dtype=None, out=None, min_count=None): a, mask = _replace_nan(a, 0) result = _dask_or_eager_func("sum")(a, axis=axis, dtype=dtype) if min_count is not None: return _maybe_null_out(result, axis, mask, min_count) else: return result def _nanmean_ddof_object(ddof, value, axis=None, dtype=None, **kwargs): from .duck_array_ops import count, fillna, _dask_or_eager_func, where_method valid_count = count(value, axis=axis) value = fillna(value, 0) if dtype is None and value.dtype.kind == "O": dtype = value.dtype if value.dtype.kind in ["cf"] else float data = _dask_or_eager_func("sum")(value, axis=axis, dtype=dtype, **kwargs) data = data / (valid_count - ddof) return where_method(data, valid_count != 0) def nanmean(a, axis=None, dtype=None, out=None): if a.dtype.kind == "O": return _nanmean_ddof_object(0, a, axis=axis, dtype=dtype) if isinstance(a, dask_array_type): return dask_array.nanmean(a, axis=axis, dtype=dtype) return np.nanmean(a, axis=axis, dtype=dtype) def nanmedian(a, axis=None, out=None): return _dask_or_eager_func("nanmedian", eager_module=nputils)(a, axis=axis) def _nanvar_object(value, axis=None, ddof=0, keepdims=False, **kwargs): value_mean = _nanmean_ddof_object( ddof=0, value=value, axis=axis, keepdims=True, **kwargs ) squared = (value.astype(value_mean.dtype) - value_mean) ** 2 return _nanmean_ddof_object(ddof, squared, axis=axis, keepdims=keepdims, **kwargs) def nanvar(a, axis=None, dtype=None, out=None, ddof=0): if a.dtype.kind == "O": return _nanvar_object(a, axis=axis, dtype=dtype, ddof=ddof) return _dask_or_eager_func("nanvar", eager_module=nputils)( a, axis=axis, dtype=dtype, ddof=ddof ) def nanstd(a, axis=None, dtype=None, out=None, ddof=0): return _dask_or_eager_func("nanstd", eager_module=nputils)( a, axis=axis, dtype=dtype, ddof=ddof ) def nanprod(a, axis=None, dtype=None, out=None, min_count=None): a, mask = _replace_nan(a, 1) result = _dask_or_eager_func("nanprod")(a, axis=axis, dtype=dtype, out=out) if min_count is not None: return _maybe_null_out(result, axis, mask, min_count) else: return result def nancumsum(a, axis=None, dtype=None, out=None): return _dask_or_eager_func("nancumsum", eager_module=nputils)( a, axis=axis, dtype=dtype ) def nancumprod(a, axis=None, dtype=None, out=None): return _dask_or_eager_func("nancumprod", eager_module=nputils)( a, axis=axis, dtype=dtype )
true
true
f70e963b0297a39ae5a52c25d46cbe5293ab6756
911
py
Python
mover.py
ritiek/shapes-classifier
1dab7ce174bae2cd95d603ada5bfc9a59911c8e6
[ "MIT" ]
5
2018-09-17T18:27:21.000Z
2020-06-01T14:12:33.000Z
mover.py
ritiek/shapes-classifier
1dab7ce174bae2cd95d603ada5bfc9a59911c8e6
[ "MIT" ]
null
null
null
mover.py
ritiek/shapes-classifier
1dab7ce174bae2cd95d603ada5bfc9a59911c8e6
[ "MIT" ]
1
2020-08-11T02:43:19.000Z
2020-08-11T02:43:19.000Z
import random import os def move_all(data_type, shape): dirpath = os.path.join(data_type, shape) os.makedirs(dirpath, exist_ok=True) for filename in os.listdir(shape): if filename.endswith('.png'): os.rename(os.path.join(shape, filename), os.path.join(data_type, shape, filename)) def move_data(data_type, shape, count): dirpath = os.path.join(data_type, shape) os.makedirs(dirpath, exist_ok=True) for x in random.sample(range(1, 3700), count): filename = '{}.png'.format(x) os.rename(os.path.join(shape, filename), os.path.join(data_type, shape, filename)) move_data('train', 'circle', 3000) move_data('train', 'square', 3000) move_data('train', 'star', 3000) move_data('train', 'triangle', 3000) move_all('test', 'circle') move_all('test', 'square') move_all('test', 'star') move_all('test', 'triangle')
30.366667
63
0.646542
import random import os def move_all(data_type, shape): dirpath = os.path.join(data_type, shape) os.makedirs(dirpath, exist_ok=True) for filename in os.listdir(shape): if filename.endswith('.png'): os.rename(os.path.join(shape, filename), os.path.join(data_type, shape, filename)) def move_data(data_type, shape, count): dirpath = os.path.join(data_type, shape) os.makedirs(dirpath, exist_ok=True) for x in random.sample(range(1, 3700), count): filename = '{}.png'.format(x) os.rename(os.path.join(shape, filename), os.path.join(data_type, shape, filename)) move_data('train', 'circle', 3000) move_data('train', 'square', 3000) move_data('train', 'star', 3000) move_data('train', 'triangle', 3000) move_all('test', 'circle') move_all('test', 'square') move_all('test', 'star') move_all('test', 'triangle')
true
true
f70e96c8cec0e377e59eeeec0b0401e2ea4cd129
15,679
py
Python
main_verbq_working.py
thilinicooray/mac-network-pytorch
0e4bf3f7f301570b652490f697758361c866f3c1
[ "MIT" ]
null
null
null
main_verbq_working.py
thilinicooray/mac-network-pytorch
0e4bf3f7f301570b652490f697758361c866f3c1
[ "MIT" ]
null
null
null
main_verbq_working.py
thilinicooray/mac-network-pytorch
0e4bf3f7f301570b652490f697758361c866f3c1
[ "MIT" ]
null
null
null
import torch #from imsitu_encoder_verbq import imsitu_encoder from imsitu_encoder_roleqverbq_embdhz import imsitu_encoder from imsitu_loader import imsitu_loader_roleq_updated from imsitu_scorer_log import imsitu_scorer import json import model_verbq_working import os import utils import time import random #from torchviz import make_dot #from graphviz import Digraph def train(model, train_loader, dev_loader, traindev_loader, optimizer, scheduler, max_epoch, model_dir, encoder, gpu_mode, clip_norm, lr_max, model_name, args,eval_frequency=4): model.train() train_loss = 0 total_steps = 0 print_freq = 400 dev_score_list = [] time_all = time.time() if model.gpu_mode >= 0 : ngpus = 2 device_array = [i for i in range(0,ngpus)] pmodel = torch.nn.DataParallel(model, device_ids=device_array) else: pmodel = model #pmodel = model '''if scheduler.get_lr()[0] < lr_max: scheduler.step()''' top1 = imsitu_scorer(encoder, 1, 3) top5 = imsitu_scorer(encoder, 5, 3) '''print('init param data check :') for f in model.parameters(): if f.requires_grad: print(f.data.size())''' for epoch in range(max_epoch): #print('current sample : ', i, img.size(), verb.size(), roles.size(), labels.size()) #sizes batch_size*3*height*width, batch*504*1, batch*6*190*1, batch*3*6*lebale_count*1 mx = len(train_loader) for i, (id, img, verb, labels) in enumerate(train_loader): #print("epoch{}-{}/{} batches\r".format(epoch,i+1,mx)) , t0 = time.time() t1 = time.time() total_steps += 1 if gpu_mode >= 0: img = torch.autograd.Variable(img.cuda()) verb = torch.autograd.Variable(verb.cuda()) labels = torch.autograd.Variable(labels.cuda()) else: img = torch.autograd.Variable(img) verb = torch.autograd.Variable(verb) labels = torch.autograd.Variable(labels) '''print('all inputs') print(img) print('=========================================================================') print(verb) print('=========================================================================') print(roles) print('=========================================================================') print(labels)''' verb_predict, loss = pmodel(img, verb, labels) #verb_predict, rol1pred, role_predict = pmodel.forward_eval5(img) #print ("forward time = {}".format(time.time() - t1)) t1 = time.time() '''g = make_dot(verb_predict, model.state_dict()) g.view()''' #loss = model.calculate_loss(verb_predict, verb) #loss = model.calculate_eval_loss_new(verb_predict, verb, rol1pred, labels, args) #loss = loss_ * random.random() #try random loss #print ("loss time = {}".format(time.time() - t1)) t1 = time.time() #print('current loss = ', loss) if gpu_mode >= 0 : #loss.backward(torch.ones([2,1]).to(torch.device('cuda'))) loss.mean().backward() else: loss.backward() #loss.backward() #print ("backward time = {}".format(time.time() - t1)) torch.nn.utils.clip_grad_norm_(model.parameters(), clip_norm) '''for param in filter(lambda p: p.requires_grad,model.parameters()): print(param.grad.data.sum())''' #start debugger #import pdb; pdb.set_trace() optimizer.step() '''print('grad check after:') for f in model.conv.parameters(): print('data is') print(f.data [0][0]) #print('grad is') #print(f.grad[0][0].item()) break''' optimizer.zero_grad() train_loss += float(loss.mean()) #top1.add_point_eval5(verb_predict, verb, role_predict, labels) #top5.add_point_eval5(verb_predict, verb, role_predict, labels) top1.add_point_verb_only_eval(id, verb_predict, verb) top5.add_point_verb_only_eval(id, verb_predict, verb) if total_steps % print_freq == 0: top1_a = top1.get_average_results() top5_a = top5.get_average_results() print ("{},{},{}, {} , {}, loss = {:.2f}, avg loss = {:.2f}" .format(total_steps-1,epoch,i, utils.format_dict(top1_a, "{:.2f}", "1-"), utils.format_dict(top5_a,"{:.2f}","5-"), loss.mean().item(), train_loss / ((total_steps-1)%eval_frequency) )) if total_steps % eval_frequency == 0: top1, top5, val_loss = eval(model, dev_loader, encoder, gpu_mode) model.train() top1_avg = top1.get_average_results() top5_avg = top5.get_average_results() avg_score = top1_avg["verb"] + top1_avg["value"] + top1_avg["value-all"] + top5_avg["verb"] + \ top5_avg["value"] + top5_avg["value-all"] avg_score /= 8 print ('Dev {} average :{:.2f} {} {}'.format(total_steps-1, avg_score*100, utils.format_dict(top1_avg,'{:.2f}', '1-'), utils.format_dict(top5_avg, '{:.2f}', '5-'))) #print('Dev loss :', val_loss) dev_score_list.append(avg_score) max_score = max(dev_score_list) if max_score == dev_score_list[-1]: torch.save(model.state_dict(), model_dir + "/{}_verbq_iter0_change.model".format( model_name)) print ('New best model saved! {0}'.format(max_score)) #eval on the trainset '''top1, top5, val_loss = eval(model, traindev_loader, encoder, gpu_mode) model.train() top1_avg = top1.get_average_results() top5_avg = top5.get_average_results() avg_score = top1_avg["verb"] + top1_avg["value"] + top1_avg["value-all"] + top5_avg["verb"] + \ top5_avg["value"] + top5_avg["value-all"] + top5_avg["value*"] + top5_avg["value-all*"] avg_score /= 8 print ('TRAINDEV {} average :{:.2f} {} {}'.format(total_steps-1, avg_score*100, utils.format_dict(top1_avg,'{:.2f}', '1-'), utils.format_dict(top5_avg, '{:.2f}', '5-')))''' print('current train loss', train_loss) train_loss = 0 top1 = imsitu_scorer(encoder, 1, 3) top5 = imsitu_scorer(encoder, 5, 3) del verb_predict, loss, img, verb, labels #break print('Epoch ', epoch, ' completed!') scheduler.step() #break def eval(model, dev_loader, encoder, gpu_mode, write_to_file = False): model.eval() val_loss = 0 print ('evaluating model...') top1 = imsitu_scorer(encoder, 1, 3, write_to_file) top5 = imsitu_scorer(encoder, 5, 3) with torch.no_grad(): mx = len(dev_loader) for i, (img_id, img, verb, labels) in enumerate(dev_loader): #print("{}/{} batches\r".format(i+1,mx)) , '''im_data = torch.squeeze(im_data,0) im_info = torch.squeeze(im_info,0) gt_boxes = torch.squeeze(gt_boxes,0) num_boxes = torch.squeeze(num_boxes,0) verb = torch.squeeze(verb,0) roles = torch.squeeze(roles,0) labels = torch.squeeze(labels,0)''' if gpu_mode >= 0: img = torch.autograd.Variable(img.cuda()) verb = torch.autograd.Variable(verb.cuda()) labels = torch.autograd.Variable(labels.cuda()) else: img = torch.autograd.Variable(img) verb = torch.autograd.Variable(verb) labels = torch.autograd.Variable(labels) verb_predict, _= model(img, verb, labels) '''loss = model.calculate_eval_loss(verb_predict, verb, role_predict, labels) val_loss += loss.item()''' top1.add_point_verb_only_eval(img_id, verb_predict, verb) top5.add_point_verb_only_eval(img_id, verb_predict, verb) del img, verb, labels break #return top1, top5, val_loss/mx return top1, top5, 0 def main(): import argparse parser = argparse.ArgumentParser(description="imsitu VSRL. Training, evaluation and prediction.") parser.add_argument("--gpuid", default=-1, help="put GPU id > -1 in GPU mode", type=int) #parser.add_argument("--command", choices = ["train", "eval", "resume", 'predict'], required = True) parser.add_argument('--resume_training', action='store_true', help='Resume training from the model [resume_model]') parser.add_argument('--resume_model', type=str, default='', help='The model we resume') parser.add_argument('--verb_module', type=str, default='', help='pretrained verb module') parser.add_argument('--role_module', type=str, default='', help='pretrained role module') parser.add_argument('--train_role', action='store_true', help='cnn fix, verb fix, role train from the scratch') parser.add_argument('--finetune_verb', action='store_true', help='cnn fix, verb finetune, role train from the scratch') parser.add_argument('--finetune_cnn', action='store_true', help='cnn finetune, verb finetune, role train from the scratch') parser.add_argument('--output_dir', type=str, default='./trained_models', help='Location to output the model') parser.add_argument('--evaluate', action='store_true', help='Only use the testing mode') parser.add_argument('--test', action='store_true', help='Only use the testing mode') parser.add_argument('--dataset_folder', type=str, default='./imSitu', help='Location of annotations') parser.add_argument('--imgset_dir', type=str, default='./resized_256', help='Location of original images') parser.add_argument('--frcnn_feat_dir', type=str, help='Location of output from detectron') #todo: train role module separately with gt verbs args = parser.parse_args() batch_size = 640 #lr = 5e-6 lr = 0.0001 lr_max = 5e-4 lr_gamma = 0.1 lr_step = 15 clip_norm = 0.5 weight_decay = 1e-4 n_epoch = 500 n_worker = 3 #dataset_folder = 'imSitu' #imgset_folder = 'resized_256' dataset_folder = args.dataset_folder imgset_folder = args.imgset_dir print('model spec :, top down att with role q ') train_set = json.load(open(dataset_folder + "/updated_train_new.json")) imsitu_roleq = json.load(open("imsitu_data/imsitu_questions_prev.json")) verb_templates = json.load(open("imsitu_data/verb_questions_template_new.json")) encoder = imsitu_encoder(train_set, imsitu_roleq, verb_templates) model = model_verbq_working.BaseModel(encoder, args.gpuid) # To group up the features #cnn_features, role_features = utils.group_features_noun(model) cnn_features, role_features = utils.group_features_noun(model) train_set = imsitu_loader_roleq_updated(imgset_folder, train_set, encoder, model.train_preprocess()) train_loader = torch.utils.data.DataLoader(train_set, batch_size=4, shuffle=True, num_workers=n_worker) dev_set = json.load(open(dataset_folder +"/dev.json")) dev_set = imsitu_loader_roleq_updated(imgset_folder, dev_set, encoder, model.dev_preprocess()) dev_loader = torch.utils.data.DataLoader(dev_set, batch_size=4, shuffle=True, num_workers=n_worker) test_set = json.load(open(dataset_folder +"/test.json")) test_set = imsitu_loader_roleq_updated(imgset_folder, test_set, encoder, model.dev_preprocess()) test_loader = torch.utils.data.DataLoader(test_set, batch_size=64, shuffle=True, num_workers=n_worker) traindev_set = json.load(open(dataset_folder +"/dev.json")) traindev_set = imsitu_loader_roleq_updated(imgset_folder, traindev_set, encoder, model.dev_preprocess()) traindev_loader = torch.utils.data.DataLoader(traindev_set, batch_size=8, shuffle=True, num_workers=n_worker) #utils.load_net(args.verb_module, [model.verb_module]) #utils.load_net(args.role_module, [model.role_module]) model_name = 'train_full' if not os.path.exists(args.output_dir): os.mkdir(args.output_dir) torch.manual_seed(1234) if args.gpuid >= 0: #print('GPU enabled') model.cuda() torch.cuda.manual_seed(1234) torch.backends.cudnn.deterministic = True optimizer = torch.optim.Adamax([ {'params': cnn_features, 'lr': 5e-5}, {'params': role_features} ], lr=1e-3) #optimizer = torch.optim.Adam(model.parameters(), lr=lr, weight_decay=weight_decay) #scheduler = torch.optim.lr_scheduler.StepLR(optimizer, step_size=lr_step, gamma=lr_gamma) #gradient clipping, grad check scheduler = torch.optim.lr_scheduler.ExponentialLR(optimizer, gamma=0.9) if args.evaluate: top1, top5, val_loss = eval(model, dev_loader, encoder, args.gpuid, write_to_file = True) top1_avg = top1.get_average_results() top5_avg = top5.get_average_results() avg_score = top1_avg["verb"] + top1_avg["value"] + top1_avg["value-all"] + top5_avg["verb"] + \ top5_avg["value"] + top5_avg["value-all"] + top5_avg["value*"] + top5_avg["value-all*"] avg_score /= 8 print ('Dev average :{:.2f} {} {}'.format( avg_score*100, utils.format_dict(top1_avg,'{:.2f}', '1-'), utils.format_dict(top5_avg, '{:.2f}', '5-'))) #write results to csv file role_dict = top1.role_dict fail_val_all = top1.value_all_dict pass_val_dict = top1.vall_all_correct with open('role_pred_data.json', 'w') as fp: json.dump(role_dict, fp, indent=4) with open('fail_val_all.json', 'w') as fp: json.dump(fail_val_all, fp, indent=4) with open('pass_val_all.json', 'w') as fp: json.dump(pass_val_dict, fp, indent=4) print('Writing predictions to file completed !') elif args.test: top1, top5, val_loss = eval(model, test_loader, encoder, args.gpuid, write_to_file = True) top1_avg = top1.get_average_results() top5_avg = top5.get_average_results() avg_score = top1_avg["verb"] + top1_avg["value"] + top1_avg["value-all"] + top5_avg["verb"] + \ top5_avg["value"] + top5_avg["value-all"] + top5_avg["value*"] + top5_avg["value-all*"] avg_score /= 8 print ('Test average :{:.2f} {} {}'.format( avg_score*100, utils.format_dict(top1_avg,'{:.2f}', '1-'), utils.format_dict(top5_avg, '{:.2f}', '5-'))) else: print('Model training started!') train(model, train_loader, dev_loader, traindev_loader, optimizer, scheduler, n_epoch, args.output_dir, encoder, args.gpuid, clip_norm, lr_max, model_name, args) if __name__ == "__main__": main()
39.794416
177
0.586389
import torch from imsitu_encoder_roleqverbq_embdhz import imsitu_encoder from imsitu_loader import imsitu_loader_roleq_updated from imsitu_scorer_log import imsitu_scorer import json import model_verbq_working import os import utils import time import random def train(model, train_loader, dev_loader, traindev_loader, optimizer, scheduler, max_epoch, model_dir, encoder, gpu_mode, clip_norm, lr_max, model_name, args,eval_frequency=4): model.train() train_loss = 0 total_steps = 0 print_freq = 400 dev_score_list = [] time_all = time.time() if model.gpu_mode >= 0 : ngpus = 2 device_array = [i for i in range(0,ngpus)] pmodel = torch.nn.DataParallel(model, device_ids=device_array) else: pmodel = model top1 = imsitu_scorer(encoder, 1, 3) top5 = imsitu_scorer(encoder, 5, 3) for epoch in range(max_epoch): mx = len(train_loader) for i, (id, img, verb, labels) in enumerate(train_loader): t0 = time.time() t1 = time.time() total_steps += 1 if gpu_mode >= 0: img = torch.autograd.Variable(img.cuda()) verb = torch.autograd.Variable(verb.cuda()) labels = torch.autograd.Variable(labels.cuda()) else: img = torch.autograd.Variable(img) verb = torch.autograd.Variable(verb) labels = torch.autograd.Variable(labels) verb_predict, loss = pmodel(img, verb, labels) t1 = time.time() t1 = time.time() if gpu_mode >= 0 : loss.mean().backward() else: loss.backward() torch.nn.utils.clip_grad_norm_(model.parameters(), clip_norm) optimizer.step() optimizer.zero_grad() train_loss += float(loss.mean()) top1.add_point_verb_only_eval(id, verb_predict, verb) top5.add_point_verb_only_eval(id, verb_predict, verb) if total_steps % print_freq == 0: top1_a = top1.get_average_results() top5_a = top5.get_average_results() print ("{},{},{}, {} , {}, loss = {:.2f}, avg loss = {:.2f}" .format(total_steps-1,epoch,i, utils.format_dict(top1_a, "{:.2f}", "1-"), utils.format_dict(top5_a,"{:.2f}","5-"), loss.mean().item(), train_loss / ((total_steps-1)%eval_frequency) )) if total_steps % eval_frequency == 0: top1, top5, val_loss = eval(model, dev_loader, encoder, gpu_mode) model.train() top1_avg = top1.get_average_results() top5_avg = top5.get_average_results() avg_score = top1_avg["verb"] + top1_avg["value"] + top1_avg["value-all"] + top5_avg["verb"] + \ top5_avg["value"] + top5_avg["value-all"] avg_score /= 8 print ('Dev {} average :{:.2f} {} {}'.format(total_steps-1, avg_score*100, utils.format_dict(top1_avg,'{:.2f}', '1-'), utils.format_dict(top5_avg, '{:.2f}', '5-'))) dev_score_list.append(avg_score) max_score = max(dev_score_list) if max_score == dev_score_list[-1]: torch.save(model.state_dict(), model_dir + "/{}_verbq_iter0_change.model".format( model_name)) print ('New best model saved! {0}'.format(max_score)) print('current train loss', train_loss) train_loss = 0 top1 = imsitu_scorer(encoder, 1, 3) top5 = imsitu_scorer(encoder, 5, 3) del verb_predict, loss, img, verb, labels print('Epoch ', epoch, ' completed!') scheduler.step() def eval(model, dev_loader, encoder, gpu_mode, write_to_file = False): model.eval() val_loss = 0 print ('evaluating model...') top1 = imsitu_scorer(encoder, 1, 3, write_to_file) top5 = imsitu_scorer(encoder, 5, 3) with torch.no_grad(): mx = len(dev_loader) for i, (img_id, img, verb, labels) in enumerate(dev_loader): if gpu_mode >= 0: img = torch.autograd.Variable(img.cuda()) verb = torch.autograd.Variable(verb.cuda()) labels = torch.autograd.Variable(labels.cuda()) else: img = torch.autograd.Variable(img) verb = torch.autograd.Variable(verb) labels = torch.autograd.Variable(labels) verb_predict, _= model(img, verb, labels) top1.add_point_verb_only_eval(img_id, verb_predict, verb) top5.add_point_verb_only_eval(img_id, verb_predict, verb) del img, verb, labels break return top1, top5, 0 def main(): import argparse parser = argparse.ArgumentParser(description="imsitu VSRL. Training, evaluation and prediction.") parser.add_argument("--gpuid", default=-1, help="put GPU id > -1 in GPU mode", type=int) parser.add_argument('--resume_training', action='store_true', help='Resume training from the model [resume_model]') parser.add_argument('--resume_model', type=str, default='', help='The model we resume') parser.add_argument('--verb_module', type=str, default='', help='pretrained verb module') parser.add_argument('--role_module', type=str, default='', help='pretrained role module') parser.add_argument('--train_role', action='store_true', help='cnn fix, verb fix, role train from the scratch') parser.add_argument('--finetune_verb', action='store_true', help='cnn fix, verb finetune, role train from the scratch') parser.add_argument('--finetune_cnn', action='store_true', help='cnn finetune, verb finetune, role train from the scratch') parser.add_argument('--output_dir', type=str, default='./trained_models', help='Location to output the model') parser.add_argument('--evaluate', action='store_true', help='Only use the testing mode') parser.add_argument('--test', action='store_true', help='Only use the testing mode') parser.add_argument('--dataset_folder', type=str, default='./imSitu', help='Location of annotations') parser.add_argument('--imgset_dir', type=str, default='./resized_256', help='Location of original images') parser.add_argument('--frcnn_feat_dir', type=str, help='Location of output from detectron') args = parser.parse_args() batch_size = 640 lr = 0.0001 lr_max = 5e-4 lr_gamma = 0.1 lr_step = 15 clip_norm = 0.5 weight_decay = 1e-4 n_epoch = 500 n_worker = 3 dataset_folder = args.dataset_folder imgset_folder = args.imgset_dir print('model spec :, top down att with role q ') train_set = json.load(open(dataset_folder + "/updated_train_new.json")) imsitu_roleq = json.load(open("imsitu_data/imsitu_questions_prev.json")) verb_templates = json.load(open("imsitu_data/verb_questions_template_new.json")) encoder = imsitu_encoder(train_set, imsitu_roleq, verb_templates) model = model_verbq_working.BaseModel(encoder, args.gpuid) cnn_features, role_features = utils.group_features_noun(model) train_set = imsitu_loader_roleq_updated(imgset_folder, train_set, encoder, model.train_preprocess()) train_loader = torch.utils.data.DataLoader(train_set, batch_size=4, shuffle=True, num_workers=n_worker) dev_set = json.load(open(dataset_folder +"/dev.json")) dev_set = imsitu_loader_roleq_updated(imgset_folder, dev_set, encoder, model.dev_preprocess()) dev_loader = torch.utils.data.DataLoader(dev_set, batch_size=4, shuffle=True, num_workers=n_worker) test_set = json.load(open(dataset_folder +"/test.json")) test_set = imsitu_loader_roleq_updated(imgset_folder, test_set, encoder, model.dev_preprocess()) test_loader = torch.utils.data.DataLoader(test_set, batch_size=64, shuffle=True, num_workers=n_worker) traindev_set = json.load(open(dataset_folder +"/dev.json")) traindev_set = imsitu_loader_roleq_updated(imgset_folder, traindev_set, encoder, model.dev_preprocess()) traindev_loader = torch.utils.data.DataLoader(traindev_set, batch_size=8, shuffle=True, num_workers=n_worker) model_name = 'train_full' if not os.path.exists(args.output_dir): os.mkdir(args.output_dir) torch.manual_seed(1234) if args.gpuid >= 0: model.cuda() torch.cuda.manual_seed(1234) torch.backends.cudnn.deterministic = True optimizer = torch.optim.Adamax([ {'params': cnn_features, 'lr': 5e-5}, {'params': role_features} ], lr=1e-3) scheduler = torch.optim.lr_scheduler.ExponentialLR(optimizer, gamma=0.9) if args.evaluate: top1, top5, val_loss = eval(model, dev_loader, encoder, args.gpuid, write_to_file = True) top1_avg = top1.get_average_results() top5_avg = top5.get_average_results() avg_score = top1_avg["verb"] + top1_avg["value"] + top1_avg["value-all"] + top5_avg["verb"] + \ top5_avg["value"] + top5_avg["value-all"] + top5_avg["value*"] + top5_avg["value-all*"] avg_score /= 8 print ('Dev average :{:.2f} {} {}'.format( avg_score*100, utils.format_dict(top1_avg,'{:.2f}', '1-'), utils.format_dict(top5_avg, '{:.2f}', '5-'))) role_dict = top1.role_dict fail_val_all = top1.value_all_dict pass_val_dict = top1.vall_all_correct with open('role_pred_data.json', 'w') as fp: json.dump(role_dict, fp, indent=4) with open('fail_val_all.json', 'w') as fp: json.dump(fail_val_all, fp, indent=4) with open('pass_val_all.json', 'w') as fp: json.dump(pass_val_dict, fp, indent=4) print('Writing predictions to file completed !') elif args.test: top1, top5, val_loss = eval(model, test_loader, encoder, args.gpuid, write_to_file = True) top1_avg = top1.get_average_results() top5_avg = top5.get_average_results() avg_score = top1_avg["verb"] + top1_avg["value"] + top1_avg["value-all"] + top5_avg["verb"] + \ top5_avg["value"] + top5_avg["value-all"] + top5_avg["value*"] + top5_avg["value-all*"] avg_score /= 8 print ('Test average :{:.2f} {} {}'.format( avg_score*100, utils.format_dict(top1_avg,'{:.2f}', '1-'), utils.format_dict(top5_avg, '{:.2f}', '5-'))) else: print('Model training started!') train(model, train_loader, dev_loader, traindev_loader, optimizer, scheduler, n_epoch, args.output_dir, encoder, args.gpuid, clip_norm, lr_max, model_name, args) if __name__ == "__main__": main()
true
true
f70e97783eb7a110103eaab5ec9a5c9b120dce42
3,177
py
Python
tests/test_app.py
Divjyot/SentimentAnalysis
f2ad46a3ac18a2c048ffc6b838fa68a2bb8febed
[ "MIT" ]
null
null
null
tests/test_app.py
Divjyot/SentimentAnalysis
f2ad46a3ac18a2c048ffc6b838fa68a2bb8febed
[ "MIT" ]
null
null
null
tests/test_app.py
Divjyot/SentimentAnalysis
f2ad46a3ac18a2c048ffc6b838fa68a2bb8febed
[ "MIT" ]
null
null
null
import pytest from app import create_app @pytest.fixture def request_header_secret(): return "dev" @pytest.fixture def request_body_positive(): return {"query": "I am having a great day!"} @pytest.fixture def request_body_negative(): return {"query": "I am feeling sad today"} @pytest.fixture def http_error_METHOD_NOT_ALLOWED(): return 405 @pytest.fixture def http_error_BAD_REQUEST(): return 400 @pytest.fixture def http_OK(): return 200 @pytest.fixture def flask_client(): app = create_app() with app.test_client() as client: yield client ## TESTS ######### # Index/ Health Check Test def test_health_check(flask_client): res = flask_client.get("/") assert b"up & running" in res.data ## OK REQUESTS Tests #################### def test_predict_positive(flask_client, http_OK, request_body_positive, request_header_secret): res = flask_client.post("/predict", json=request_body_positive, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_OK assert b"POSITIVE" in res.data def test_predict_negative(flask_client, http_OK, request_body_negative, request_header_secret): res = flask_client.post("/predict", json=request_body_negative, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_OK assert b"NEGATIVE" in res.data ## BAD REQUESTS Tests #################### def test_GET_instead_POST(flask_client, http_error_METHOD_NOT_ALLOWED, request_header_secret): res = flask_client.get("/predict", json={"query": ""}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_METHOD_NOT_ALLOWED ## Body def test_None_body(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json=None, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST def test_empty_body(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST ## Query def test_none_query(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={"query": None}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST def test_empty_query(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={"query": ""}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST def test_non_string_numerical(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={"query": 456123}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST def test_non_string_object(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={"query": ["I am happy"]}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST
28.881818
118
0.745672
import pytest from app import create_app @pytest.fixture def request_header_secret(): return "dev" @pytest.fixture def request_body_positive(): return {"query": "I am having a great day!"} @pytest.fixture def request_body_negative(): return {"query": "I am feeling sad today"} @pytest.fixture def http_error_METHOD_NOT_ALLOWED(): return 405 @pytest.fixture def http_error_BAD_REQUEST(): return 400 @pytest.fixture def http_OK(): return 200 @pytest.fixture def flask_client(): app = create_app() with app.test_client() as client: yield client res = flask_client.get("/") assert b"up & running" in res.data ret}) assert res.status_code == http_OK assert b"POSITIVE" in res.data def test_predict_negative(flask_client, http_OK, request_body_negative, request_header_secret): res = flask_client.post("/predict", json=request_body_negative, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_OK assert b"NEGATIVE" in res.data ssert res.status_code == http_error_METHOD_NOT_ALLOWED test_None_body(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json=None, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST def test_empty_body(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST test_none_query(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={"query": None}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST def test_empty_query(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={"query": ""}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST def test_non_string_numerical(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={"query": 456123}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST def test_non_string_object(flask_client, http_error_BAD_REQUEST, request_header_secret): res = flask_client.post("/predict", json={"query": ["I am happy"]}, headers={"Secret-Key": request_header_secret}) assert res.status_code == http_error_BAD_REQUEST
true
true
f70e97c43309e5287f3d18538ed116d9df87ad0f
6,451
py
Python
virt/ansible-latest/lib/python2.7/site-packages/ansible/modules/storage/purestorage/purefb_s3user.py
lakhlaifi/RedHat-Ansible
27c5077cced9d416081fcd5d69ea44bca0317fa4
[ "Apache-2.0" ]
1
2020-03-29T18:41:01.000Z
2020-03-29T18:41:01.000Z
ansible/ansible/modules/storage/purestorage/purefb_s3user.py
SergeyCherepanov/ansible
875711cd2fd6b783c812241c2ed7a954bf6f670f
[ "MIT" ]
7
2020-09-07T17:27:56.000Z
2022-03-02T06:25:46.000Z
ansible/ansible/modules/storage/purestorage/purefb_s3user.py
SergeyCherepanov/ansible
875711cd2fd6b783c812241c2ed7a954bf6f670f
[ "MIT" ]
1
2020-10-30T12:48:24.000Z
2020-10-30T12:48:24.000Z
#!/usr/bin/python # -*- coding: utf-8 -*- # (c) 2018, Simon Dodsley (simon@purestorage.com) # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import absolute_import, division, print_function __metaclass__ = type ANSIBLE_METADATA = {'metadata_version': '1.1', 'status': ['preview'], 'supported_by': 'community'} DOCUMENTATION = r''' --- module: purefb_s3user version_added: '2.8' short_description: Create or delete FlashBlade Object Store account users description: - Create or delete object store account users on a Pure Stoage FlashBlade. author: - Pure Storage Ansible Team (@sdodsley) <pure-ansible-team@purestorage.com> options: state: description: - Create or delete object store account user default: present choices: [ absent, present ] type: str name: description: - The name of object store user type: str account: description: - The name of object store account associated with user type: str access_key: description: - Create secret access key. - Key can be exposed using the I(debug) module type: bool default: true extends_documentation_fragment: - purestorage.fb ''' EXAMPLES = r''' - name: Crrate object store user (with access ID and key) foo in account bar purefb_s3user: name: foo account: bar fb_url: 10.10.10.2 api_token: e31060a7-21fc-e277-6240-25983c6c4592 debug: var: ansible_facts.fb_s3user - name: Delete object store user foo in account bar purefb_s3user: name: foo account: bar state: absent fb_url: 10.10.10.2 api_token: e31060a7-21fc-e277-6240-25983c6c4592 ''' RETURN = r''' ''' HAS_PURITY_FB = True try: from purity_fb import ObjectStoreAccessKey except ImportError: HAS_PURITY_FB = False from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.pure import get_blade, purefb_argument_spec MIN_REQUIRED_API_VERSION = '1.3' def get_s3acc(module, blade): """Return Object Store Account or None""" s3acc = None accts = blade.object_store_accounts.list_object_store_accounts() for acct in range(0, len(accts.items)): if accts.items[acct].name == module.params['account']: s3acc = accts.items[acct] return s3acc def get_s3user(module, blade): """Return Object Store Account or None""" full_user = module.params['account'] + "/" + module.params['name'] s3user = None s3users = blade.object_store_users.list_object_store_users() for user in range(0, len(s3users.items)): if s3users.items[user].name == full_user: s3user = s3users.items[user] return s3user def update_s3user(module, blade): """Update Object Store User""" changed = False s3user_facts = {} user = module.params['account'] + "/" + module.params['name'] if module.params['access_key']: try: result = blade.object_store_access_keys.create_object_store_access_keys( object_store_access_key=ObjectStoreAccessKey(user={'name': user})) s3user_facts['fb_s3user'] = {'user': user, 'access_key': result.items[0].secret_access_key, 'access_id': result.items[0].name} except Exception: delete_s3user(module, blade) module.fail_json(msg='Object Store User {0}: Creation failed'.format(user)) changed = True module.exit_json(changed=changed, ansible_facts=s3user_facts) def create_s3user(module, blade): """Create Object Store Account""" s3user_facts = {} changed = False user = module.params['account'] + "/" + module.params['name'] try: blade.object_store_users.create_object_store_users(names=[user]) if module.params['access_key']: try: result = blade.object_store_access_keys.create_object_store_access_keys( object_store_access_key=ObjectStoreAccessKey(user={'name': user})) s3user_facts['fb_s3user'] = {'user': user, 'access_key': result.items[0].secret_access_key, 'access_id': result.items[0].name} except Exception: delete_s3user(module, blade) module.fail_json(msg='Object Store User {0}: Creation failed'.format(user)) changed = True except Exception: module.fail_json(msg='Object Store User {0}: Creation failed'.format(user)) module.exit_json(changed=changed, ansible_facts=s3user_facts) def delete_s3user(module, blade): """Delete Object Store Account""" changed = False user = module.params['account'] + "/" + module.params['name'] try: blade.object_store_users.delete_object_store_users(names=[user]) changed = True except Exception: module.fail_json(msg='Object Store Account {0}: Deletion failed'.format(module.params['name'])) module.exit_json(changed=changed) def main(): argument_spec = purefb_argument_spec() argument_spec.update(dict( name=dict(required=True, type='str'), account=dict(required=True, type='str'), access_key=dict(default='true', type='bool'), state=dict(default='present', choices=['present', 'absent']), )) module = AnsibleModule(argument_spec, supports_check_mode=False) if not HAS_PURITY_FB: module.fail_json(msg='purity_fb sdk is required for this module') state = module.params['state'] blade = get_blade(module) versions = blade.api_version.list_versions().versions if MIN_REQUIRED_API_VERSION not in versions: module.fail_json(msg='FlashBlade REST version not supported. Minimum version required: {0}'.format(MIN_REQUIRED_API_VERSION)) s3acc = get_s3acc(module, blade) if not s3acc: module.fail_json(msg='Object Store Account {0} does not exist'.format(module.params['account'])) s3user = get_s3user(module, blade) if state == 'absent' and s3user: delete_s3user(module, blade) elif state == 'present' and s3user: update_s3user(module, blade) elif not s3user and state == 'present': create_s3user(module, blade) else: module.exit_json(changed=False) if __name__ == '__main__': main()
32.094527
133
0.656952
from __future__ import absolute_import, division, print_function __metaclass__ = type ANSIBLE_METADATA = {'metadata_version': '1.1', 'status': ['preview'], 'supported_by': 'community'} DOCUMENTATION = r''' --- module: purefb_s3user version_added: '2.8' short_description: Create or delete FlashBlade Object Store account users description: - Create or delete object store account users on a Pure Stoage FlashBlade. author: - Pure Storage Ansible Team (@sdodsley) <pure-ansible-team@purestorage.com> options: state: description: - Create or delete object store account user default: present choices: [ absent, present ] type: str name: description: - The name of object store user type: str account: description: - The name of object store account associated with user type: str access_key: description: - Create secret access key. - Key can be exposed using the I(debug) module type: bool default: true extends_documentation_fragment: - purestorage.fb ''' EXAMPLES = r''' - name: Crrate object store user (with access ID and key) foo in account bar purefb_s3user: name: foo account: bar fb_url: 10.10.10.2 api_token: e31060a7-21fc-e277-6240-25983c6c4592 debug: var: ansible_facts.fb_s3user - name: Delete object store user foo in account bar purefb_s3user: name: foo account: bar state: absent fb_url: 10.10.10.2 api_token: e31060a7-21fc-e277-6240-25983c6c4592 ''' RETURN = r''' ''' HAS_PURITY_FB = True try: from purity_fb import ObjectStoreAccessKey except ImportError: HAS_PURITY_FB = False from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.pure import get_blade, purefb_argument_spec MIN_REQUIRED_API_VERSION = '1.3' def get_s3acc(module, blade): s3acc = None accts = blade.object_store_accounts.list_object_store_accounts() for acct in range(0, len(accts.items)): if accts.items[acct].name == module.params['account']: s3acc = accts.items[acct] return s3acc def get_s3user(module, blade): full_user = module.params['account'] + "/" + module.params['name'] s3user = None s3users = blade.object_store_users.list_object_store_users() for user in range(0, len(s3users.items)): if s3users.items[user].name == full_user: s3user = s3users.items[user] return s3user def update_s3user(module, blade): changed = False s3user_facts = {} user = module.params['account'] + "/" + module.params['name'] if module.params['access_key']: try: result = blade.object_store_access_keys.create_object_store_access_keys( object_store_access_key=ObjectStoreAccessKey(user={'name': user})) s3user_facts['fb_s3user'] = {'user': user, 'access_key': result.items[0].secret_access_key, 'access_id': result.items[0].name} except Exception: delete_s3user(module, blade) module.fail_json(msg='Object Store User {0}: Creation failed'.format(user)) changed = True module.exit_json(changed=changed, ansible_facts=s3user_facts) def create_s3user(module, blade): s3user_facts = {} changed = False user = module.params['account'] + "/" + module.params['name'] try: blade.object_store_users.create_object_store_users(names=[user]) if module.params['access_key']: try: result = blade.object_store_access_keys.create_object_store_access_keys( object_store_access_key=ObjectStoreAccessKey(user={'name': user})) s3user_facts['fb_s3user'] = {'user': user, 'access_key': result.items[0].secret_access_key, 'access_id': result.items[0].name} except Exception: delete_s3user(module, blade) module.fail_json(msg='Object Store User {0}: Creation failed'.format(user)) changed = True except Exception: module.fail_json(msg='Object Store User {0}: Creation failed'.format(user)) module.exit_json(changed=changed, ansible_facts=s3user_facts) def delete_s3user(module, blade): changed = False user = module.params['account'] + "/" + module.params['name'] try: blade.object_store_users.delete_object_store_users(names=[user]) changed = True except Exception: module.fail_json(msg='Object Store Account {0}: Deletion failed'.format(module.params['name'])) module.exit_json(changed=changed) def main(): argument_spec = purefb_argument_spec() argument_spec.update(dict( name=dict(required=True, type='str'), account=dict(required=True, type='str'), access_key=dict(default='true', type='bool'), state=dict(default='present', choices=['present', 'absent']), )) module = AnsibleModule(argument_spec, supports_check_mode=False) if not HAS_PURITY_FB: module.fail_json(msg='purity_fb sdk is required for this module') state = module.params['state'] blade = get_blade(module) versions = blade.api_version.list_versions().versions if MIN_REQUIRED_API_VERSION not in versions: module.fail_json(msg='FlashBlade REST version not supported. Minimum version required: {0}'.format(MIN_REQUIRED_API_VERSION)) s3acc = get_s3acc(module, blade) if not s3acc: module.fail_json(msg='Object Store Account {0} does not exist'.format(module.params['account'])) s3user = get_s3user(module, blade) if state == 'absent' and s3user: delete_s3user(module, blade) elif state == 'present' and s3user: update_s3user(module, blade) elif not s3user and state == 'present': create_s3user(module, blade) else: module.exit_json(changed=False) if __name__ == '__main__': main()
true
true
f70e988f4dafef4a16d2ad7c3426c886babdf411
2,642
py
Python
internal/twirptest/service_pb2_twirp.py
DennisSSDev/twirp
c27eaf7ba72b359deaed686bc1e2a486214059ca
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
internal/twirptest/service_pb2_twirp.py
DennisSSDev/twirp
c27eaf7ba72b359deaed686bc1e2a486214059ca
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
internal/twirptest/service_pb2_twirp.py
DennisSSDev/twirp
c27eaf7ba72b359deaed686bc1e2a486214059ca
[ "ECL-2.0", "Apache-2.0" ]
null
null
null
# Code generated by protoc-gen-twirp_python v7.1.0, DO NOT EDIT. # source: service.proto try: import httplib from urllib2 import Request, HTTPError, urlopen except ImportError: import http.client as httplib from urllib.request import Request, urlopen from urllib.error import HTTPError import json from google.protobuf import symbol_database as _symbol_database import sys _sym_db = _symbol_database.Default() class TwirpException(httplib.HTTPException): def __init__(self, code, message, meta): self.code = code self.message = message self.meta = meta super(TwirpException, self).__init__(message) @classmethod def from_http_err(cls, err): try: jsonerr = json.load(err) code = jsonerr["code"] msg = jsonerr["msg"] meta = jsonerr.get("meta") if meta is None: meta = {} except: code = "internal" msg = "Error from intermediary with HTTP status code {} {}".format( err.code, httplib.responses[err.code], ) meta = {} return cls(code, msg, meta) class HaberdasherClient(object): """ A Haberdasher makes hats for clients. """ def __init__(self, server_address): """Creates a new client for the Haberdasher service. Args: server_address: The address of the server to send requests to, in the full protocol://host:port form. """ if sys.version_info[0] > 2: self.__target = server_address else: self.__target = server_address.encode('ascii') self.__service_name = "twirp.internal.twirptest.Haberdasher" def __make_request(self, body, full_method): req = Request( url=self.__target + "/twirp" + full_method, data=body, headers={"Content-Type": "application/protobuf"}, ) try: resp = urlopen(req) except HTTPError as err: raise TwirpException.from_http_err(err) return resp.read() def make_hat(self, size): """ MakeHat produces a hat of mysterious, randomly-selected color! """ serialize = _sym_db.GetSymbol("twirp.internal.twirptest.Size").SerializeToString deserialize = _sym_db.GetSymbol("twirp.internal.twirptest.Hat").FromString full_method = "/{}/{}".format(self.__service_name, "MakeHat") body = serialize(size) resp_str = self.__make_request(body=body, full_method=full_method) return deserialize(resp_str)
31.082353
88
0.613929
try: import httplib from urllib2 import Request, HTTPError, urlopen except ImportError: import http.client as httplib from urllib.request import Request, urlopen from urllib.error import HTTPError import json from google.protobuf import symbol_database as _symbol_database import sys _sym_db = _symbol_database.Default() class TwirpException(httplib.HTTPException): def __init__(self, code, message, meta): self.code = code self.message = message self.meta = meta super(TwirpException, self).__init__(message) @classmethod def from_http_err(cls, err): try: jsonerr = json.load(err) code = jsonerr["code"] msg = jsonerr["msg"] meta = jsonerr.get("meta") if meta is None: meta = {} except: code = "internal" msg = "Error from intermediary with HTTP status code {} {}".format( err.code, httplib.responses[err.code], ) meta = {} return cls(code, msg, meta) class HaberdasherClient(object): def __init__(self, server_address): if sys.version_info[0] > 2: self.__target = server_address else: self.__target = server_address.encode('ascii') self.__service_name = "twirp.internal.twirptest.Haberdasher" def __make_request(self, body, full_method): req = Request( url=self.__target + "/twirp" + full_method, data=body, headers={"Content-Type": "application/protobuf"}, ) try: resp = urlopen(req) except HTTPError as err: raise TwirpException.from_http_err(err) return resp.read() def make_hat(self, size): serialize = _sym_db.GetSymbol("twirp.internal.twirptest.Size").SerializeToString deserialize = _sym_db.GetSymbol("twirp.internal.twirptest.Hat").FromString full_method = "/{}/{}".format(self.__service_name, "MakeHat") body = serialize(size) resp_str = self.__make_request(body=body, full_method=full_method) return deserialize(resp_str)
true
true
f70e98a26c7f36c76b2da78699df18fa1459660b
8,717
py
Python
dropbox/packing-your-dropbox/packurbox.py
ramsay/ramsay-snippets
d64e30a7d7b64eedab46150babb2bb675b3da8ff
[ "MIT" ]
null
null
null
dropbox/packing-your-dropbox/packurbox.py
ramsay/ramsay-snippets
d64e30a7d7b64eedab46150babb2bb675b3da8ff
[ "MIT" ]
null
null
null
dropbox/packing-your-dropbox/packurbox.py
ramsay/ramsay-snippets
d64e30a7d7b64eedab46150babb2bb675b3da8ff
[ "MIT" ]
null
null
null
#!/usr/bin/python """ Robert Ramsay <robert.alan.ramsay@gmail.com> Packing your Dropbox When you're working with petabytes of data, you have to store files wherever they can fit. All of us here at Dropbox are always searching for more ways to efficiently pack data into smaller and more manageable chunks. The fun begins when you bend the rules a little bit and visualize it in two dimensions. You'll be given a list of rectangular "files" that you'll need to pack into as small a "Dropbox" as possible. The dimensions of each file will be specified by a tuple (width, height), both of which will be integers. The output of your function should be the area of the smallest rectangular Dropbox that can enclose all of them without any overlap. Files can be rotated 90(deg) if it helps. Bonus points if you can draw pictures of the winning configurations along the way. While drawing pictures, any files sharing dimensions should be considered identical/interchangeable. Input Your program must read a small integer N (1 <= N <= 100) from stdin representing the maximum number of files to consider, followed by the width and height of each file, one per line. Output Output should be simply be the area of the smallest containing Dropbox. If you want to print pretty pictures, send that to stderr. Only the output on stdout will be judged. Sample Input 3 8 8 4 3 3 4 Sample Output 88 """ #from __future__ import print_function import sys class DropBox: w = 0 h = 0 x = 0 y = 0 def __init__(self,vector=None, w=0, h=0): if vector: self.w, self.h = vector else: self.w = w self.h = h def rotate(self): t = self.w self.w = self.h self.h = t def align(self): if self.w > self.h: self.rotate() return self.h #free space = (lowest left x, lowest left y, width, height) def fit(size, free, box): x, y, w, h = free box.x = x box.y = y if h < box.h and w < box.w: # Our box will not fit inside the current freespace. size = (size[0]+box.w-w, size[1]+box.h-h) x += box.w w = 0 h = box.h elif w < box.w: size = (size[0] + box.w - w, size[1]) w = box.w y += box.h h -= box.h elif h < box.h: x += box.w w -= box.w else: box.rotate() if w < box.w: size = (size[0] + box.w - w, size[1]) w = box.w y += box.h h -= box.h else: x += box.w w -= box.w free = (x, y, w, h) return size, free def pretty(boxes,w,h): '''Pretty print the list of boxes''' print >> sys.stderr, str(w) + 'x' + str(h) + ':' graph = [[' ' for l in range(h+1)] for m in range(w+1)] for box in boxes: try: # Vertices graph[box.x][box.y] = '+' graph[box.x+box.w][box.y] = '+' graph[box.x][box.y+box.h] = '+' graph[box.x+box.w][box.y+box.h] = '+' # Edges for x in range(box.x+1, box.x+box.w): graph[x][box.y] = '|' graph[x][box.y+box.h] = '|' for y in range(box.y+1, box.y+box.h): graph[box.x][y] = '-' graph[box.x+box.w][y] = '-' except Exception as e: print >> sys.stderr, "Box (", box.x, box.y, box.w, box.h, ") is outside bounds (", w, h,")" raise e print >> sys.stderr, '\n'.join([''.join(row) for row in graph]) def pack(boxes): #Align all the boxes and sort them by height lagest to smallest boxes.sort(key=lambda box: box.align(), reverse=True) size = (0, 0) #free = (left, lower, width, height) free = (0, 0, 0, 0) for box in boxes: size, free = fit(size, free, box) pretty(boxes, size[0], size[1]) return size[0]*size[1] class DropNode: left = None # Left Edge is the parent. vertex = None # We can store at most one Box right = None # Right Edge is the child. direction = [1,0] # direction is the identity ray def __init__(self,vertex=None, left=None, right=None): self.vertex = vertex self.left = left if self.left: self.left.right = self self.right = right if self.right: w = self.right.width() h = self.right.height() if self.vertex.w > self.vertex.h: # An increase in width costs less than an increase in height # if width is already greater. self.direction = [0,1] if w < h: self.right.rotate() else: self.direction = [0,1] if h < w: self.right.rotate() self.right.left = self def rotate(self): self.direction.reverse() if self.vertex: self.vertex.rotate() if self.right: self.right.rotate() def width(self): w = 0 if self.vertex is not None: w = self.vertex.w if self.right is not None: if self.direction[0]: w += self.right.width() return w def height(self): h = 0 if self.vertex is not None: h = self.vertex.h if self.right is not None: if self.direction[1]: h += self.right.height() return h def packtree(node, boxes): '''This is a recursive pack algorithm, similar to a binary search tree.''' if node is None: node = DropNode() if not boxes: # Stack empty. while node.left: node = node.left return node # Return root if node is None: #RootNode print >> sys.stderr, "root node", boxes[-1] return packtree(DropNode(boxes.pop(0)), boxes) if node.vertex is None: # Not sure if I agree with this. print >> sys.stderr, "curious" node.vertex = boxes.pop() return packtree(node, boxes) # Make comparisons simpler left = (max(boxes[0].w, boxes[0].h), min(boxes[0].w, boxes[0].h)) w = node.width() h = node.height() right = (max(w, h), min(w, h)) print >> sys.stderr, "left", left, "right", right, if left[0] > right[0]: print >> sys.stderr, "insert left" if node.left: return packtree(node.left, boxes) else: return packtree(DropNode(boxes.pop(0),None,node), boxes) #if left[0] < right[1]: # print >> sys.stderr, "insert right" # if node.right: # return packtree(node.right, boxes) # else: # return packtree(DropNode(boxes.pop(0),node),boxes) print >> sys.stderr, "insert middle" return packtree(DropNode(boxes.pop(0), node.left, node), boxes) def prettytree(tree): '''Pretty print the list of boxes''' w = tree.width() h = tree.height() print >> sys.stderr, str(w) + 'x' + str(h) + ':' graph = [[' ' for l in range(h+1)] for m in range(w+1)] vx = 0 vy = 0 i = 0 node = tree while node.right: i += 1 print >> sys.stderr, '.', if node.vertex is None: print >> sys.stderr, "Empty Vertex" node = node.right continue try: vw = tree.vertex.w vh = tree.vertex.h # Vertices graph[vx][vy] = '+' graph[vx+vw][vy] = '+' graph[vx][vy+vh] = '+' graph[vx+vw][vy+vh] = '+' # Edges for x in range(vx+1, vx+vw): graph[x][vy] = '|' graph[x][vy+vh] = '|' for y in range(vy+1, vy+vh): graph[vx][y] = '-' graph[vx+vw][y] = '-' vx += tree.direction[0]*vw vy += tree.direction[1]*vh except Exception as e: raise e node = node.right print >> sys.stderr print >> sys.stderr, '\n'.join([''.join(row) for row in graph]) if __name__ == '__main__': import sys inp = input() #Number of boxes try: boxcount = int(inp) if boxcount < 1 or boxcount > 100: raise except: sys.exit("Box count must be between 1 and 100 (inclusive)") boxes = [] for i in range(boxcount): inp = raw_input('') #Box: width height box = DropBox() try: w, h = inp.split(" ") box.w = int(w) box.h = int(h) except: sys.exit("Box definition should be integers seperated "\ "by whitespace") boxes.append(box) print(pack(boxes)) sys.exit()
31.356115
574
0.533326
import sys class DropBox: w = 0 h = 0 x = 0 y = 0 def __init__(self,vector=None, w=0, h=0): if vector: self.w, self.h = vector else: self.w = w self.h = h def rotate(self): t = self.w self.w = self.h self.h = t def align(self): if self.w > self.h: self.rotate() return self.h def fit(size, free, box): x, y, w, h = free box.x = x box.y = y if h < box.h and w < box.w: size = (size[0]+box.w-w, size[1]+box.h-h) x += box.w w = 0 h = box.h elif w < box.w: size = (size[0] + box.w - w, size[1]) w = box.w y += box.h h -= box.h elif h < box.h: x += box.w w -= box.w else: box.rotate() if w < box.w: size = (size[0] + box.w - w, size[1]) w = box.w y += box.h h -= box.h else: x += box.w w -= box.w free = (x, y, w, h) return size, free def pretty(boxes,w,h): print >> sys.stderr, str(w) + 'x' + str(h) + ':' graph = [[' ' for l in range(h+1)] for m in range(w+1)] for box in boxes: try: graph[box.x][box.y] = '+' graph[box.x+box.w][box.y] = '+' graph[box.x][box.y+box.h] = '+' graph[box.x+box.w][box.y+box.h] = '+' for x in range(box.x+1, box.x+box.w): graph[x][box.y] = '|' graph[x][box.y+box.h] = '|' for y in range(box.y+1, box.y+box.h): graph[box.x][y] = '-' graph[box.x+box.w][y] = '-' except Exception as e: print >> sys.stderr, "Box (", box.x, box.y, box.w, box.h, ") is outside bounds (", w, h,")" raise e print >> sys.stderr, '\n'.join([''.join(row) for row in graph]) def pack(boxes): boxes.sort(key=lambda box: box.align(), reverse=True) size = (0, 0) free = (0, 0, 0, 0) for box in boxes: size, free = fit(size, free, box) pretty(boxes, size[0], size[1]) return size[0]*size[1] class DropNode: left = None vertex = None right = None direction = [1,0] def __init__(self,vertex=None, left=None, right=None): self.vertex = vertex self.left = left if self.left: self.left.right = self self.right = right if self.right: w = self.right.width() h = self.right.height() if self.vertex.w > self.vertex.h: self.direction = [0,1] if w < h: self.right.rotate() else: self.direction = [0,1] if h < w: self.right.rotate() self.right.left = self def rotate(self): self.direction.reverse() if self.vertex: self.vertex.rotate() if self.right: self.right.rotate() def width(self): w = 0 if self.vertex is not None: w = self.vertex.w if self.right is not None: if self.direction[0]: w += self.right.width() return w def height(self): h = 0 if self.vertex is not None: h = self.vertex.h if self.right is not None: if self.direction[1]: h += self.right.height() return h def packtree(node, boxes): if node is None: node = DropNode() if not boxes: while node.left: node = node.left return node if node is None: print >> sys.stderr, "root node", boxes[-1] return packtree(DropNode(boxes.pop(0)), boxes) if node.vertex is None: print >> sys.stderr, "curious" node.vertex = boxes.pop() return packtree(node, boxes) left = (max(boxes[0].w, boxes[0].h), min(boxes[0].w, boxes[0].h)) w = node.width() h = node.height() right = (max(w, h), min(w, h)) print >> sys.stderr, "left", left, "right", right, if left[0] > right[0]: print >> sys.stderr, "insert left" if node.left: return packtree(node.left, boxes) else: return packtree(DropNode(boxes.pop(0),None,node), boxes) print >> sys.stderr, "insert middle" return packtree(DropNode(boxes.pop(0), node.left, node), boxes) def prettytree(tree): w = tree.width() h = tree.height() print >> sys.stderr, str(w) + 'x' + str(h) + ':' graph = [[' ' for l in range(h+1)] for m in range(w+1)] vx = 0 vy = 0 i = 0 node = tree while node.right: i += 1 print >> sys.stderr, '.', if node.vertex is None: print >> sys.stderr, "Empty Vertex" node = node.right continue try: vw = tree.vertex.w vh = tree.vertex.h graph[vx][vy] = '+' graph[vx+vw][vy] = '+' graph[vx][vy+vh] = '+' graph[vx+vw][vy+vh] = '+' for x in range(vx+1, vx+vw): graph[x][vy] = '|' graph[x][vy+vh] = '|' for y in range(vy+1, vy+vh): graph[vx][y] = '-' graph[vx+vw][y] = '-' vx += tree.direction[0]*vw vy += tree.direction[1]*vh except Exception as e: raise e node = node.right print >> sys.stderr print >> sys.stderr, '\n'.join([''.join(row) for row in graph]) if __name__ == '__main__': import sys inp = input() try: boxcount = int(inp) if boxcount < 1 or boxcount > 100: raise except: sys.exit("Box count must be between 1 and 100 (inclusive)") boxes = [] for i in range(boxcount): inp = raw_input('') box = DropBox() try: w, h = inp.split(" ") box.w = int(w) box.h = int(h) except: sys.exit("Box definition should be integers seperated "\ "by whitespace") boxes.append(box) print(pack(boxes)) sys.exit()
true
true
f70e997410788ed72a9b2883d54249e487064f32
2,675
py
Python
libralli/circcuitpython/adafruit-circuitpython-bundle-7.x-mpy-20211225/examples/progressbar_magtag_simpletest.py
Yarik9008/SoftAcademic
118c9dc4620ca444c1557edd141a838820577202
[ "MIT" ]
5
2020-04-14T20:50:30.000Z
2021-10-19T18:49:46.000Z
libralli/circcuitpython/adafruit-circuitpython-bundle-7.x-mpy-20211225/examples/progressbar_magtag_simpletest.py
Yarik9008/SoftAcademic
118c9dc4620ca444c1557edd141a838820577202
[ "MIT" ]
25
2020-07-17T03:05:56.000Z
2022-01-06T16:00:51.000Z
libralli/circcuitpython/adafruit-circuitpython-bundle-7.x-mpy-20211225/examples/progressbar_magtag_simpletest.py
Yarik9008/SoftAcademic
118c9dc4620ca444c1557edd141a838820577202
[ "MIT" ]
10
2020-01-07T20:14:56.000Z
2021-11-14T20:34:43.000Z
# SPDX-FileCopyrightText: 2021 ladyada for Adafruit Industries # SPDX-License-Identifier: MIT """ Basic progressbar example script adapted for use on MagTag. """ import time import board import displayio import digitalio from adafruit_progressbar.progressbar import HorizontalProgressBar # use built in display (PyPortal, PyGamer, PyBadge, CLUE, etc.) # see guide for setting up external displays (TFT / OLED breakouts, RGB matrices, etc.) # https://learn.adafruit.com/circuitpython-display-support-using-displayio/display-and-display-bus display = board.DISPLAY time.sleep(display.time_to_refresh) # B/up button will be used to increase the progress up_btn = digitalio.DigitalInOut(board.BUTTON_B) up_btn.direction = digitalio.Direction.INPUT up_btn.pull = digitalio.Pull.UP # C/down button will be used to increase the progress down_btn = digitalio.DigitalInOut(board.BUTTON_C) down_btn.direction = digitalio.Direction.INPUT down_btn.pull = digitalio.Pull.UP # Make the display context splash = displayio.Group() display.show(splash) # set progress bar width and height relative to board's display BAR_WIDTH = display.width - 40 BAR_HEIGHT = 30 x = display.width // 2 - BAR_WIDTH // 2 y = display.height // 3 # Create a new progress_bar object at (x, y) progress_bar = HorizontalProgressBar( (x, y), (BAR_WIDTH, BAR_HEIGHT), bar_color=0xFFFFFF, outline_color=0xAAAAAA, fill_color=0x777777, ) # Append progress_bar to the splash group splash.append(progress_bar) # Get a random starting value within our min/max range current_progress = time.monotonic() % 101 print(current_progress) progress_bar.value = current_progress # refresh the display display.refresh() value_incrementor = 3 prev_up = up_btn.value prev_down = down_btn.value while True: cur_up = up_btn.value cur_down = down_btn.value do_refresh = False # if up_btn was just pressed down if not cur_up and prev_up: current_progress += value_incrementor # Wrap if we get over the maximum value if current_progress > progress_bar.maximum: current_progress = progress_bar.minimum do_refresh = True if not cur_down and prev_down: current_progress -= value_incrementor # Wrap if we get below the minimum value if current_progress < progress_bar.minimum: current_progress = progress_bar.maximum do_refresh = True if do_refresh: print(current_progress) progress_bar.value = current_progress time.sleep(display.time_to_refresh) display.refresh() time.sleep(display.time_to_refresh) prev_up = cur_up prev_down = cur_down
27.864583
98
0.741682
import time import board import displayio import digitalio from adafruit_progressbar.progressbar import HorizontalProgressBar display = board.DISPLAY time.sleep(display.time_to_refresh) up_btn = digitalio.DigitalInOut(board.BUTTON_B) up_btn.direction = digitalio.Direction.INPUT up_btn.pull = digitalio.Pull.UP down_btn = digitalio.DigitalInOut(board.BUTTON_C) down_btn.direction = digitalio.Direction.INPUT down_btn.pull = digitalio.Pull.UP splash = displayio.Group() display.show(splash) BAR_WIDTH = display.width - 40 BAR_HEIGHT = 30 x = display.width // 2 - BAR_WIDTH // 2 y = display.height // 3 # Create a new progress_bar object at (x, y) progress_bar = HorizontalProgressBar( (x, y), (BAR_WIDTH, BAR_HEIGHT), bar_color=0xFFFFFF, outline_color=0xAAAAAA, fill_color=0x777777, ) # Append progress_bar to the splash group splash.append(progress_bar) # Get a random starting value within our min/max range current_progress = time.monotonic() % 101 print(current_progress) progress_bar.value = current_progress # refresh the display display.refresh() value_incrementor = 3 prev_up = up_btn.value prev_down = down_btn.value while True: cur_up = up_btn.value cur_down = down_btn.value do_refresh = False # if up_btn was just pressed down if not cur_up and prev_up: current_progress += value_incrementor # Wrap if we get over the maximum value if current_progress > progress_bar.maximum: current_progress = progress_bar.minimum do_refresh = True if not cur_down and prev_down: current_progress -= value_incrementor # Wrap if we get below the minimum value if current_progress < progress_bar.minimum: current_progress = progress_bar.maximum do_refresh = True if do_refresh: print(current_progress) progress_bar.value = current_progress time.sleep(display.time_to_refresh) display.refresh() time.sleep(display.time_to_refresh) prev_up = cur_up prev_down = cur_down
true
true
f70e99a6d83ec20f3d35576e5d254633a5aa413c
19,416
py
Python
ServidorPython/python32_web/Lib/site-packages/nbconvert/nbconvertapp.py
mak213k/Servidor_automatizado_python
4403ef8027a2f814220baacc95856cf5fbf01d21
[ "MIT" ]
4
2019-07-26T11:32:22.000Z
2019-09-11T05:34:59.000Z
ServidorPython/python32_web/Lib/site-packages/nbconvert/nbconvertapp.py
mak213k/Servidor_automatizado_python
4403ef8027a2f814220baacc95856cf5fbf01d21
[ "MIT" ]
1
2021-09-02T17:44:02.000Z
2021-09-02T17:44:02.000Z
ServidorPython/python32_web/Lib/site-packages/nbconvert/nbconvertapp.py
mak213k/Servidor_automatizado_python
4403ef8027a2f814220baacc95856cf5fbf01d21
[ "MIT" ]
2
2019-08-28T14:57:54.000Z
2019-11-26T16:18:30.000Z
#!/usr/bin/env python """NbConvert is a utility for conversion of .ipynb files. Command-line interface for the NbConvert conversion utility. """ # Copyright (c) IPython Development Team. # Distributed under the terms of the Modified BSD License. from __future__ import print_function import logging import sys import os import glob from jupyter_core.application import JupyterApp, base_aliases, base_flags from traitlets.config import catch_config_error, Configurable from traitlets import ( Unicode, List, Instance, DottedObjectName, Type, Bool, default, observe, ) from traitlets.utils.importstring import import_item from .exporters.base import get_export_names, get_exporter from nbconvert import exporters, preprocessors, writers, postprocessors, __version__ from .utils.base import NbConvertBase from .utils.exceptions import ConversionException from .utils.io import unicode_stdin_stream #----------------------------------------------------------------------------- #Classes and functions #----------------------------------------------------------------------------- class DottedOrNone(DottedObjectName): """A string holding a valid dotted object name in Python, such as A.b3._c Also allows for None type. """ default_value = u'' def validate(self, obj, value): if value is not None and len(value) > 0: return super(DottedOrNone, self).validate(obj, value) else: return value nbconvert_aliases = {} nbconvert_aliases.update(base_aliases) nbconvert_aliases.update({ 'to' : 'NbConvertApp.export_format', 'template' : 'TemplateExporter.template_file', 'writer' : 'NbConvertApp.writer_class', 'post': 'NbConvertApp.postprocessor_class', 'output': 'NbConvertApp.output_base', 'output-dir': 'FilesWriter.build_directory', 'reveal-prefix': 'SlidesExporter.reveal_url_prefix', 'nbformat': 'NotebookExporter.nbformat_version', }) nbconvert_flags = {} nbconvert_flags.update(base_flags) nbconvert_flags.update({ 'execute' : ( {'ExecutePreprocessor' : {'enabled' : True}}, "Execute the notebook prior to export." ), 'allow-errors' : ( {'ExecutePreprocessor' : {'allow_errors' : True}}, ("Continue notebook execution even if one of the cells throws " "an error and include the error message in the cell output " "(the default behaviour is to abort conversion). This flag " "is only relevant if '--execute' was specified, too.") ), 'stdin' : ( {'NbConvertApp' : { 'from_stdin' : True, } }, "read a single notebook file from stdin. Write the resulting notebook with default basename 'notebook.*'" ), 'stdout' : ( {'NbConvertApp' : {'writer_class' : "StdoutWriter"}}, "Write notebook output to stdout instead of files." ), 'inplace' : ( { 'NbConvertApp' : { 'use_output_suffix' : False, 'export_format' : 'notebook', }, 'FilesWriter' : {'build_directory': ''}, }, """Run nbconvert in place, overwriting the existing notebook (only relevant when converting to notebook format)""" ), 'clear-output' : ( { 'NbConvertApp' : { 'use_output_suffix' : False, 'export_format' : 'notebook', }, 'FilesWriter' : {'build_directory': ''}, 'ClearOutputPreprocessor' : {'enabled' : True}, }, """Clear output of current file and save in place, overwriting the existing notebook. """ ), 'no-prompt' : ( {'TemplateExporter' : { 'exclude_input_prompt' : True, 'exclude_output_prompt' : True, } }, "Exclude input and output prompts from converted document." ), 'no-input' : ( {'TemplateExporter' : { 'exclude_output_prompt' : True, 'exclude_input': True, } }, """Exclude input cells and output prompts from converted document. This mode is ideal for generating code-free reports.""" ), }) class NbConvertApp(JupyterApp): """Application used to convert from notebook file type (``*.ipynb``)""" version = __version__ name = 'jupyter-nbconvert' aliases = nbconvert_aliases flags = nbconvert_flags @default('log_level') def _log_level_default(self): return logging.INFO classes = List() @default('classes') def _classes_default(self): classes = [NbConvertBase] for pkg in (exporters, preprocessors, writers, postprocessors): for name in dir(pkg): cls = getattr(pkg, name) if isinstance(cls, type) and issubclass(cls, Configurable): classes.append(cls) return classes description = Unicode( u"""This application is used to convert notebook files (*.ipynb) to various other formats. WARNING: THE COMMANDLINE INTERFACE MAY CHANGE IN FUTURE RELEASES.""") output_base = Unicode('', help='''overwrite base name use for output files. can only be used when converting one notebook at a time. ''').tag(config=True) use_output_suffix = Bool( True, help="""Whether to apply a suffix prior to the extension (only relevant when converting to notebook format). The suffix is determined by the exporter, and is usually '.nbconvert'.""" ).tag(config=True) output_files_dir = Unicode('{notebook_name}_files', help='''Directory to copy extra files (figures) to. '{notebook_name}' in the string will be converted to notebook basename''' ).tag(config=True) examples = Unicode(u""" The simplest way to use nbconvert is > jupyter nbconvert mynotebook.ipynb which will convert mynotebook.ipynb to the default format (probably HTML). You can specify the export format with `--to`. Options include {formats}. > jupyter nbconvert --to latex mynotebook.ipynb Both HTML and LaTeX support multiple output templates. LaTeX includes 'base', 'article' and 'report'. HTML includes 'basic' and 'full'. You can specify the flavor of the format used. > jupyter nbconvert --to html --template basic mynotebook.ipynb You can also pipe the output to stdout, rather than a file > jupyter nbconvert mynotebook.ipynb --stdout PDF is generated via latex > jupyter nbconvert mynotebook.ipynb --to pdf You can get (and serve) a Reveal.js-powered slideshow > jupyter nbconvert myslides.ipynb --to slides --post serve Multiple notebooks can be given at the command line in a couple of different ways: > jupyter nbconvert notebook*.ipynb > jupyter nbconvert notebook1.ipynb notebook2.ipynb or you can specify the notebooks list in a config file, containing:: c.NbConvertApp.notebooks = ["my_notebook.ipynb"] > jupyter nbconvert --config mycfg.py """.format(formats=get_export_names())) # Writer specific variables writer = Instance('nbconvert.writers.base.WriterBase', help="""Instance of the writer class used to write the results of the conversion.""", allow_none=True) writer_class = DottedObjectName('FilesWriter', help="""Writer class used to write the results of the conversion""").tag(config=True) writer_aliases = {'fileswriter': 'nbconvert.writers.files.FilesWriter', 'debugwriter': 'nbconvert.writers.debug.DebugWriter', 'stdoutwriter': 'nbconvert.writers.stdout.StdoutWriter'} writer_factory = Type(allow_none=True) @observe('writer_class') def _writer_class_changed(self, change): new = change['new'] if new.lower() in self.writer_aliases: new = self.writer_aliases[new.lower()] self.writer_factory = import_item(new) # Post-processor specific variables postprocessor = Instance('nbconvert.postprocessors.base.PostProcessorBase', help="""Instance of the PostProcessor class used to write the results of the conversion.""", allow_none=True) postprocessor_class = DottedOrNone( help="""PostProcessor class used to write the results of the conversion""" ).tag(config=True) postprocessor_aliases = {'serve': 'nbconvert.postprocessors.serve.ServePostProcessor'} postprocessor_factory = Type(None, allow_none=True) @observe('postprocessor_class') def _postprocessor_class_changed(self, change): new = change['new'] if new.lower() in self.postprocessor_aliases: new = self.postprocessor_aliases[new.lower()] if new: self.postprocessor_factory = import_item(new) ipywidgets_base_url = Unicode("https://unpkg.com/", help="URL base for ipywidgets package").tag(config=True) export_format = Unicode( 'html', allow_none=False, help="""The export format to be used, either one of the built-in formats {formats} or a dotted object name that represents the import path for an `Exporter` class""".format(formats=get_export_names()) ).tag(config=True) notebooks = List([], help="""List of notebooks to convert. Wildcards are supported. Filenames passed positionally will be added to the list. """ ).tag(config=True) from_stdin = Bool(False, help="read a single notebook from stdin.").tag(config=True) @catch_config_error def initialize(self, argv=None): """Initialize application, notebooks, writer, and postprocessor""" self.init_syspath() super(NbConvertApp, self).initialize(argv) self.init_notebooks() self.init_writer() self.init_postprocessor() def init_syspath(self): """Add the cwd to the sys.path ($PYTHONPATH)""" sys.path.insert(0, os.getcwd()) def init_notebooks(self): """Construct the list of notebooks. If notebooks are passed on the command-line, they override (rather than add) notebooks specified in config files. Glob each notebook to replace notebook patterns with filenames. """ # Specifying notebooks on the command-line overrides (rather than # adds) the notebook list if self.extra_args: patterns = self.extra_args else: patterns = self.notebooks # Use glob to replace all the notebook patterns with filenames. filenames = [] for pattern in patterns: # Use glob to find matching filenames. Allow the user to convert # notebooks without having to type the extension. globbed_files = glob.glob(pattern) globbed_files.extend(glob.glob(pattern + '.ipynb')) if not globbed_files: self.log.warning("pattern %r matched no files", pattern) for filename in globbed_files: if not filename in filenames: filenames.append(filename) self.notebooks = filenames def init_writer(self): """Initialize the writer (which is stateless)""" self._writer_class_changed({ 'new': self.writer_class }) self.writer = self.writer_factory(parent=self) if hasattr(self.writer, 'build_directory') and self.writer.build_directory != '': self.use_output_suffix = False def init_postprocessor(self): """Initialize the postprocessor (which is stateless)""" self._postprocessor_class_changed({'new': self.postprocessor_class}) if self.postprocessor_factory: self.postprocessor = self.postprocessor_factory(parent=self) def start(self): """Run start after initialization process has completed""" super(NbConvertApp, self).start() self.convert_notebooks() def init_single_notebook_resources(self, notebook_filename): """Step 1: Initialize resources This initializes the resources dictionary for a single notebook. Returns ------- dict resources dictionary for a single notebook that MUST include the following keys: - config_dir: the location of the Jupyter config directory - unique_key: the notebook name - output_files_dir: a directory where output files (not including the notebook itself) should be saved """ basename = os.path.basename(notebook_filename) notebook_name = basename[:basename.rfind('.')] if self.output_base: # strip duplicate extension from output_base, to avoid Basename.ext.ext if getattr(self.exporter, 'file_extension', False): base, ext = os.path.splitext(self.output_base) if ext == self.exporter.file_extension: self.output_base = base notebook_name = self.output_base self.log.debug("Notebook name is '%s'", notebook_name) # first initialize the resources we want to use resources = {} resources['config_dir'] = self.config_dir resources['unique_key'] = notebook_name output_files_dir = (self.output_files_dir .format(notebook_name=notebook_name)) resources['output_files_dir'] = output_files_dir resources['ipywidgets_base_url'] = self.ipywidgets_base_url return resources def export_single_notebook(self, notebook_filename, resources, input_buffer=None): """Step 2: Export the notebook Exports the notebook to a particular format according to the specified exporter. This function returns the output and (possibly modified) resources from the exporter. Parameters ---------- notebook_filename : str name of notebook file. resources : dict input_buffer : readable file-like object returning unicode. if not None, notebook_filename is ignored Returns ------- output dict resources (possibly modified) """ try: if input_buffer is not None: output, resources = self.exporter.from_file(input_buffer, resources=resources) else: output, resources = self.exporter.from_filename(notebook_filename, resources=resources) except ConversionException: self.log.error("Error while converting '%s'", notebook_filename, exc_info=True) self.exit(1) return output, resources def write_single_notebook(self, output, resources): """Step 3: Write the notebook to file This writes output from the exporter to file using the specified writer. It returns the results from the writer. Parameters ---------- output : resources : dict resources for a single notebook including name, config directory and directory to save output Returns ------- file results from the specified writer output of exporter """ if 'unique_key' not in resources: raise KeyError("unique_key MUST be specified in the resources, but it is not") notebook_name = resources['unique_key'] if self.use_output_suffix and not self.output_base: notebook_name += resources.get('output_suffix', '') write_results = self.writer.write( output, resources, notebook_name=notebook_name) return write_results def postprocess_single_notebook(self, write_results): """Step 4: Post-process the written file Only used if a postprocessor has been specified. After the converted notebook is written to a file in Step 3, this post-processes the notebook. """ # Post-process if post processor has been defined. if hasattr(self, 'postprocessor') and self.postprocessor: self.postprocessor(write_results) def convert_single_notebook(self, notebook_filename, input_buffer=None): """Convert a single notebook. Performs the following steps: 1. Initialize notebook resources 2. Export the notebook to a particular format 3. Write the exported notebook to file 4. (Maybe) postprocess the written file Parameters ---------- notebook_filename : str input_buffer : If input_buffer is not None, conversion is done and the buffer is used as source into a file basenamed by the notebook_filename argument. """ if input_buffer is None: self.log.info("Converting notebook %s to %s", notebook_filename, self.export_format) else: self.log.info("Converting notebook into %s", self.export_format) resources = self.init_single_notebook_resources(notebook_filename) output, resources = self.export_single_notebook(notebook_filename, resources, input_buffer=input_buffer) write_results = self.write_single_notebook(output, resources) self.postprocess_single_notebook(write_results) def convert_notebooks(self): """Convert the notebooks in the self.notebook traitlet """ # check that the output base isn't specified if there is more than # one notebook to convert if self.output_base != '' and len(self.notebooks) > 1: self.log.error( """ UsageError: --output flag or `NbConvertApp.output_base` config option cannot be used when converting multiple notebooks. """ ) self.exit(1) # initialize the exporter cls = get_exporter(self.export_format) self.exporter = cls(config=self.config) # no notebooks to convert! if len(self.notebooks) == 0 and not self.from_stdin: self.print_help() sys.exit(-1) # convert each notebook if not self.from_stdin: for notebook_filename in self.notebooks: self.convert_single_notebook(notebook_filename) else: input_buffer = unicode_stdin_stream() # default name when conversion from stdin self.convert_single_notebook("notebook.ipynb", input_buffer=input_buffer) #----------------------------------------------------------------------------- # Main entry point #----------------------------------------------------------------------------- main = launch_new_instance = NbConvertApp.launch_instance
37.410405
113
0.613103
from __future__ import print_function import logging import sys import os import glob from jupyter_core.application import JupyterApp, base_aliases, base_flags from traitlets.config import catch_config_error, Configurable from traitlets import ( Unicode, List, Instance, DottedObjectName, Type, Bool, default, observe, ) from traitlets.utils.importstring import import_item from .exporters.base import get_export_names, get_exporter from nbconvert import exporters, preprocessors, writers, postprocessors, __version__ from .utils.base import NbConvertBase from .utils.exceptions import ConversionException from .utils.io import unicode_stdin_stream class DottedOrNone(DottedObjectName): default_value = u'' def validate(self, obj, value): if value is not None and len(value) > 0: return super(DottedOrNone, self).validate(obj, value) else: return value nbconvert_aliases = {} nbconvert_aliases.update(base_aliases) nbconvert_aliases.update({ 'to' : 'NbConvertApp.export_format', 'template' : 'TemplateExporter.template_file', 'writer' : 'NbConvertApp.writer_class', 'post': 'NbConvertApp.postprocessor_class', 'output': 'NbConvertApp.output_base', 'output-dir': 'FilesWriter.build_directory', 'reveal-prefix': 'SlidesExporter.reveal_url_prefix', 'nbformat': 'NotebookExporter.nbformat_version', }) nbconvert_flags = {} nbconvert_flags.update(base_flags) nbconvert_flags.update({ 'execute' : ( {'ExecutePreprocessor' : {'enabled' : True}}, "Execute the notebook prior to export." ), 'allow-errors' : ( {'ExecutePreprocessor' : {'allow_errors' : True}}, ("Continue notebook execution even if one of the cells throws " "an error and include the error message in the cell output " "(the default behaviour is to abort conversion). This flag " "is only relevant if '--execute' was specified, too.") ), 'stdin' : ( {'NbConvertApp' : { 'from_stdin' : True, } }, "read a single notebook file from stdin. Write the resulting notebook with default basename 'notebook.*'" ), 'stdout' : ( {'NbConvertApp' : {'writer_class' : "StdoutWriter"}}, "Write notebook output to stdout instead of files." ), 'inplace' : ( { 'NbConvertApp' : { 'use_output_suffix' : False, 'export_format' : 'notebook', }, 'FilesWriter' : {'build_directory': ''}, }, """Run nbconvert in place, overwriting the existing notebook (only relevant when converting to notebook format)""" ), 'clear-output' : ( { 'NbConvertApp' : { 'use_output_suffix' : False, 'export_format' : 'notebook', }, 'FilesWriter' : {'build_directory': ''}, 'ClearOutputPreprocessor' : {'enabled' : True}, }, """Clear output of current file and save in place, overwriting the existing notebook. """ ), 'no-prompt' : ( {'TemplateExporter' : { 'exclude_input_prompt' : True, 'exclude_output_prompt' : True, } }, "Exclude input and output prompts from converted document." ), 'no-input' : ( {'TemplateExporter' : { 'exclude_output_prompt' : True, 'exclude_input': True, } }, """Exclude input cells and output prompts from converted document. This mode is ideal for generating code-free reports.""" ), }) class NbConvertApp(JupyterApp): version = __version__ name = 'jupyter-nbconvert' aliases = nbconvert_aliases flags = nbconvert_flags @default('log_level') def _log_level_default(self): return logging.INFO classes = List() @default('classes') def _classes_default(self): classes = [NbConvertBase] for pkg in (exporters, preprocessors, writers, postprocessors): for name in dir(pkg): cls = getattr(pkg, name) if isinstance(cls, type) and issubclass(cls, Configurable): classes.append(cls) return classes description = Unicode( u"""This application is used to convert notebook files (*.ipynb) to various other formats. WARNING: THE COMMANDLINE INTERFACE MAY CHANGE IN FUTURE RELEASES.""") output_base = Unicode('', help='''overwrite base name use for output files. can only be used when converting one notebook at a time. ''').tag(config=True) use_output_suffix = Bool( True, help="""Whether to apply a suffix prior to the extension (only relevant when converting to notebook format). The suffix is determined by the exporter, and is usually '.nbconvert'.""" ).tag(config=True) output_files_dir = Unicode('{notebook_name}_files', help='''Directory to copy extra files (figures) to. '{notebook_name}' in the string will be converted to notebook basename''' ).tag(config=True) examples = Unicode(u""" The simplest way to use nbconvert is > jupyter nbconvert mynotebook.ipynb which will convert mynotebook.ipynb to the default format (probably HTML). You can specify the export format with `--to`. Options include {formats}. > jupyter nbconvert --to latex mynotebook.ipynb Both HTML and LaTeX support multiple output templates. LaTeX includes 'base', 'article' and 'report'. HTML includes 'basic' and 'full'. You can specify the flavor of the format used. > jupyter nbconvert --to html --template basic mynotebook.ipynb You can also pipe the output to stdout, rather than a file > jupyter nbconvert mynotebook.ipynb --stdout PDF is generated via latex > jupyter nbconvert mynotebook.ipynb --to pdf You can get (and serve) a Reveal.js-powered slideshow > jupyter nbconvert myslides.ipynb --to slides --post serve Multiple notebooks can be given at the command line in a couple of different ways: > jupyter nbconvert notebook*.ipynb > jupyter nbconvert notebook1.ipynb notebook2.ipynb or you can specify the notebooks list in a config file, containing:: c.NbConvertApp.notebooks = ["my_notebook.ipynb"] > jupyter nbconvert --config mycfg.py """.format(formats=get_export_names())) writer = Instance('nbconvert.writers.base.WriterBase', help="""Instance of the writer class used to write the results of the conversion.""", allow_none=True) writer_class = DottedObjectName('FilesWriter', help="""Writer class used to write the results of the conversion""").tag(config=True) writer_aliases = {'fileswriter': 'nbconvert.writers.files.FilesWriter', 'debugwriter': 'nbconvert.writers.debug.DebugWriter', 'stdoutwriter': 'nbconvert.writers.stdout.StdoutWriter'} writer_factory = Type(allow_none=True) @observe('writer_class') def _writer_class_changed(self, change): new = change['new'] if new.lower() in self.writer_aliases: new = self.writer_aliases[new.lower()] self.writer_factory = import_item(new) postprocessor = Instance('nbconvert.postprocessors.base.PostProcessorBase', help="""Instance of the PostProcessor class used to write the results of the conversion.""", allow_none=True) postprocessor_class = DottedOrNone( help="""PostProcessor class used to write the results of the conversion""" ).tag(config=True) postprocessor_aliases = {'serve': 'nbconvert.postprocessors.serve.ServePostProcessor'} postprocessor_factory = Type(None, allow_none=True) @observe('postprocessor_class') def _postprocessor_class_changed(self, change): new = change['new'] if new.lower() in self.postprocessor_aliases: new = self.postprocessor_aliases[new.lower()] if new: self.postprocessor_factory = import_item(new) ipywidgets_base_url = Unicode("https://unpkg.com/", help="URL base for ipywidgets package").tag(config=True) export_format = Unicode( 'html', allow_none=False, help="""The export format to be used, either one of the built-in formats {formats} or a dotted object name that represents the import path for an `Exporter` class""".format(formats=get_export_names()) ).tag(config=True) notebooks = List([], help="""List of notebooks to convert. Wildcards are supported. Filenames passed positionally will be added to the list. """ ).tag(config=True) from_stdin = Bool(False, help="read a single notebook from stdin.").tag(config=True) @catch_config_error def initialize(self, argv=None): self.init_syspath() super(NbConvertApp, self).initialize(argv) self.init_notebooks() self.init_writer() self.init_postprocessor() def init_syspath(self): sys.path.insert(0, os.getcwd()) def init_notebooks(self): if self.extra_args: patterns = self.extra_args else: patterns = self.notebooks filenames = [] for pattern in patterns: globbed_files = glob.glob(pattern) globbed_files.extend(glob.glob(pattern + '.ipynb')) if not globbed_files: self.log.warning("pattern %r matched no files", pattern) for filename in globbed_files: if not filename in filenames: filenames.append(filename) self.notebooks = filenames def init_writer(self): self._writer_class_changed({ 'new': self.writer_class }) self.writer = self.writer_factory(parent=self) if hasattr(self.writer, 'build_directory') and self.writer.build_directory != '': self.use_output_suffix = False def init_postprocessor(self): self._postprocessor_class_changed({'new': self.postprocessor_class}) if self.postprocessor_factory: self.postprocessor = self.postprocessor_factory(parent=self) def start(self): super(NbConvertApp, self).start() self.convert_notebooks() def init_single_notebook_resources(self, notebook_filename): basename = os.path.basename(notebook_filename) notebook_name = basename[:basename.rfind('.')] if self.output_base: if getattr(self.exporter, 'file_extension', False): base, ext = os.path.splitext(self.output_base) if ext == self.exporter.file_extension: self.output_base = base notebook_name = self.output_base self.log.debug("Notebook name is '%s'", notebook_name) resources = {} resources['config_dir'] = self.config_dir resources['unique_key'] = notebook_name output_files_dir = (self.output_files_dir .format(notebook_name=notebook_name)) resources['output_files_dir'] = output_files_dir resources['ipywidgets_base_url'] = self.ipywidgets_base_url return resources def export_single_notebook(self, notebook_filename, resources, input_buffer=None): try: if input_buffer is not None: output, resources = self.exporter.from_file(input_buffer, resources=resources) else: output, resources = self.exporter.from_filename(notebook_filename, resources=resources) except ConversionException: self.log.error("Error while converting '%s'", notebook_filename, exc_info=True) self.exit(1) return output, resources def write_single_notebook(self, output, resources): if 'unique_key' not in resources: raise KeyError("unique_key MUST be specified in the resources, but it is not") notebook_name = resources['unique_key'] if self.use_output_suffix and not self.output_base: notebook_name += resources.get('output_suffix', '') write_results = self.writer.write( output, resources, notebook_name=notebook_name) return write_results def postprocess_single_notebook(self, write_results): if hasattr(self, 'postprocessor') and self.postprocessor: self.postprocessor(write_results) def convert_single_notebook(self, notebook_filename, input_buffer=None): if input_buffer is None: self.log.info("Converting notebook %s to %s", notebook_filename, self.export_format) else: self.log.info("Converting notebook into %s", self.export_format) resources = self.init_single_notebook_resources(notebook_filename) output, resources = self.export_single_notebook(notebook_filename, resources, input_buffer=input_buffer) write_results = self.write_single_notebook(output, resources) self.postprocess_single_notebook(write_results) def convert_notebooks(self): # one notebook to convert if self.output_base != '' and len(self.notebooks) > 1: self.log.error( """ UsageError: --output flag or `NbConvertApp.output_base` config option cannot be used when converting multiple notebooks. """ ) self.exit(1) # initialize the exporter cls = get_exporter(self.export_format) self.exporter = cls(config=self.config) # no notebooks to convert! if len(self.notebooks) == 0 and not self.from_stdin: self.print_help() sys.exit(-1) # convert each notebook if not self.from_stdin: for notebook_filename in self.notebooks: self.convert_single_notebook(notebook_filename) else: input_buffer = unicode_stdin_stream() # default name when conversion from stdin self.convert_single_notebook("notebook.ipynb", input_buffer=input_buffer) #----------------------------------------------------------------------------- # Main entry point #----------------------------------------------------------------------------- main = launch_new_instance = NbConvertApp.launch_instance
true
true
f70e9ab4127f82a7bc61abb5b1f4d8af7ef4fcd5
122
py
Python
contributed_traders/util.py
andrewsonin/abides_dev
e8a9c8450bbbe98597f31767362c86eb193597a0
[ "BSD-3-Clause" ]
null
null
null
contributed_traders/util.py
andrewsonin/abides_dev
e8a9c8450bbbe98597f31767362c86eb193597a0
[ "BSD-3-Clause" ]
null
null
null
contributed_traders/util.py
andrewsonin/abides_dev
e8a9c8450bbbe98597f31767362c86eb193597a0
[ "BSD-3-Clause" ]
null
null
null
#!/usr/bin/env python3 from pathlib import Path def get_file(fname): return Path(__file__).resolve().parent / fname
17.428571
50
0.729508
from pathlib import Path def get_file(fname): return Path(__file__).resolve().parent / fname
true
true
f70e9b3086a20a853e14284409d0e7e61f4feaf3
326
py
Python
utils/bytes_utils.py
linyuan0213/qb-bot
be8a172d7d72a7e3878f89ba341c3b5f36e3d76c
[ "MIT" ]
3
2021-08-28T18:14:54.000Z
2022-01-19T18:22:59.000Z
utils/bytes_utils.py
linyuan0213/qb-bot
be8a172d7d72a7e3878f89ba341c3b5f36e3d76c
[ "MIT" ]
null
null
null
utils/bytes_utils.py
linyuan0213/qb-bot
be8a172d7d72a7e3878f89ba341c3b5f36e3d76c
[ "MIT" ]
1
2021-07-16T08:37:03.000Z
2021-07-16T08:37:03.000Z
def bytes_to_human(n): symbols = ('KB', 'MB', 'GB', 'TB', 'PB', 'EB') prefix = {} for i, s in enumerate(symbols): prefix[s] = 1 << (i + 1) * 10 for s in reversed(symbols): if n >= prefix[s]: value = float(n) / prefix[s] return '%.1f%s' % (value, s) return '%sB' % n
29.636364
50
0.46319
def bytes_to_human(n): symbols = ('KB', 'MB', 'GB', 'TB', 'PB', 'EB') prefix = {} for i, s in enumerate(symbols): prefix[s] = 1 << (i + 1) * 10 for s in reversed(symbols): if n >= prefix[s]: value = float(n) / prefix[s] return '%.1f%s' % (value, s) return '%sB' % n
true
true
f70e9b6e0d6c5f3a42f4bb7ff0f34b00d662bc27
1,338
py
Python
spartan/examples/ssvd/qr.py
GabrielWen/spartan
ce3bf7f2bb551d7f996a1884acef819b620cc854
[ "Apache-2.0" ]
156
2015-01-10T21:54:25.000Z
2021-10-17T14:13:57.000Z
spartan/examples/ssvd/qr.py
GabrielWen/spartan
ce3bf7f2bb551d7f996a1884acef819b620cc854
[ "Apache-2.0" ]
8
2015-01-05T16:34:18.000Z
2015-12-11T08:12:28.000Z
spartan/examples/ssvd/qr.py
GabrielWen/spartan
ce3bf7f2bb551d7f996a1884acef819b620cc854
[ "Apache-2.0" ]
24
2015-01-10T21:55:48.000Z
2021-04-14T08:09:34.000Z
import spartan from spartan import expr, core import numpy as np from sys import stderr def qr(Y): ''' Compute the thin qr factorization of a matrix. Factor the matrix Y as QR, where Q is orthonormal and R is upper-triangular. Parameters ---------- Y: Spartan array of shape (M, K). Notes ---------- Y'Y must fit in memory. Y is a Spartan array of shape (M, K). Since this QR decomposition is mainly used in Stochastic SVD, K will be the rank of the matrix of shape (M, N) and the assumption is that the rank K should be far less than M or N. Returns ------- Q : Spartan array of shape (M, K). R : Numpy array of shape (K, K). ''' # Since the K should be far less than M. So the matrix multiplication # should be the bottleneck instead of local cholesky decomposition and # finding inverse of R. So we just parallelize the matrix mulitplication. # If K is really large, we may consider using our Spartan cholesky # decomposition, but for now, we use numpy version, it works fine. # YTY = Y'Y. YTY has shape of (K, K). YTY = expr.dot(expr.transpose(Y), Y).optimized().glom() # Do cholesky decomposition and get R. R = np.linalg.cholesky(YTY).T # Find the inverse of R inv_R = np.linalg.inv(R) # Q = Y * inv(R) Q = expr.dot(Y, inv_R).optimized().evaluate() return Q, R
29.086957
75
0.675635
import spartan from spartan import expr, core import numpy as np from sys import stderr def qr(Y): YTY = expr.dot(expr.transpose(Y), Y).optimized().glom() # Do cholesky decomposition and get R. R = np.linalg.cholesky(YTY).T # Find the inverse of R inv_R = np.linalg.inv(R) # Q = Y * inv(R) Q = expr.dot(Y, inv_R).optimized().evaluate() return Q, R
true
true
f70e9bc29651fbb0c65ebc831bb2635b14150dfd
3,633
py
Python
servers/migrations/0012_auto__add_field_server_samba_base_folder.py
PolyLAN/azimut-gestion
0b076570d7ceeef8cfd9c8e65fd0495aa16d3c8a
[ "MIT" ]
null
null
null
servers/migrations/0012_auto__add_field_server_samba_base_folder.py
PolyLAN/azimut-gestion
0b076570d7ceeef8cfd9c8e65fd0495aa16d3c8a
[ "MIT" ]
null
null
null
servers/migrations/0012_auto__add_field_server_samba_base_folder.py
PolyLAN/azimut-gestion
0b076570d7ceeef8cfd9c8e65fd0495aa16d3c8a
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Adding field 'Server.samba_base_folder' db.add_column(u'servers_server', 'samba_base_folder', self.gf('django.db.models.fields.CharField')(default='', max_length=255, null=True, blank=True), keep_default=False) def backwards(self, orm): # Deleting field 'Server.samba_base_folder' db.delete_column(u'servers_server', 'samba_base_folder') models = { u'servers.server': { 'Meta': {'object_name': 'Server'}, 'external_hostname_for_vms_creation': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'external_interface': ('django.db.models.fields.CharField', [], {'max_length': '15', 'null': 'True', 'blank': 'True'}), 'external_ip': ('django.db.models.fields.IPAddressField', [], {'max_length': '15', 'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'internal_ip': ('django.db.models.fields.IPAddressField', [], {'max_length': '15', 'null': 'True', 'blank': 'True'}), 'is_proxmox': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'is_vm': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'keymanger_name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '255'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'ngnix_server': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'ngnixed_server_set'", 'null': 'True', 'to': u"orm['servers.Server']"}), 'proxmox_node_name': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '255', 'null': 'True', 'blank': 'True'}), 'samba_base_folder': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '255', 'null': 'True', 'blank': 'True'}), 'samba_management': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'ssh_connection_string_from_backup': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'ssh_connection_string_from_gestion': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'vm_host': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'server_set'", 'null': 'True', 'to': u"orm['servers.Server']"}) }, u'servers.serveruser': { 'Meta': {'object_name': 'ServerUser'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'server': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['servers.Server']"}) }, u'servers.sshkey': { 'Meta': {'object_name': 'SshKey'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'key': ('django.db.models.fields.TextField', [], {}), 'server': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['servers.Server']"}), 'user': ('django.db.models.fields.CharField', [], {'max_length': '255'}) } } complete_apps = ['servers']
63.736842
186
0.57886
import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): db.add_column(u'servers_server', 'samba_base_folder', self.gf('django.db.models.fields.CharField')(default='', max_length=255, null=True, blank=True), keep_default=False) def backwards(self, orm): db.delete_column(u'servers_server', 'samba_base_folder') models = { u'servers.server': { 'Meta': {'object_name': 'Server'}, 'external_hostname_for_vms_creation': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'external_interface': ('django.db.models.fields.CharField', [], {'max_length': '15', 'null': 'True', 'blank': 'True'}), 'external_ip': ('django.db.models.fields.IPAddressField', [], {'max_length': '15', 'null': 'True', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'internal_ip': ('django.db.models.fields.IPAddressField', [], {'max_length': '15', 'null': 'True', 'blank': 'True'}), 'is_proxmox': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'is_vm': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'keymanger_name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '255'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'ngnix_server': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'ngnixed_server_set'", 'null': 'True', 'to': u"orm['servers.Server']"}), 'proxmox_node_name': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '255', 'null': 'True', 'blank': 'True'}), 'samba_base_folder': ('django.db.models.fields.CharField', [], {'default': "''", 'max_length': '255', 'null': 'True', 'blank': 'True'}), 'samba_management': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'ssh_connection_string_from_backup': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'ssh_connection_string_from_gestion': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}), 'vm_host': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'server_set'", 'null': 'True', 'to': u"orm['servers.Server']"}) }, u'servers.serveruser': { 'Meta': {'object_name': 'ServerUser'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}), 'server': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['servers.Server']"}) }, u'servers.sshkey': { 'Meta': {'object_name': 'SshKey'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'key': ('django.db.models.fields.TextField', [], {}), 'server': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['servers.Server']"}), 'user': ('django.db.models.fields.CharField', [], {'max_length': '255'}) } } complete_apps = ['servers']
true
true
f70e9c09777356b17130746cb212b29360ca2fcc
3,204
py
Python
storm/__init__.py
silky/storm
91fe53a5edb52d4784fbba134a614aae964ded81
[ "MIT" ]
1
2019-06-27T11:34:13.000Z
2019-06-27T11:34:13.000Z
storm/__init__.py
silky/storm
91fe53a5edb52d4784fbba134a614aae964ded81
[ "MIT" ]
null
null
null
storm/__init__.py
silky/storm
91fe53a5edb52d4784fbba134a614aae964ded81
[ "MIT" ]
null
null
null
# -*- coding: utf8 -*- from ssh_config import ConfigParser from exceptions import StormValueError from operator import itemgetter import getpass __version__ = '0.5.2' class Storm(object): def __init__(self, ssh_config_file=None): self.ssh_config = ConfigParser(ssh_config_file) self.ssh_config.load() def add_entry(self, name, host, user, port, id_file, custom_options=[]): if self.is_host_in(name): raise StormValueError('{0} is already in your sshconfig. use storm edit command to modify.'.format(name)) options = self.get_options(host, user, port, id_file, custom_options) self.ssh_config.add_host(name, options) self.ssh_config.write_to_ssh_config() return True def edit_entry(self, name, host, user, port, id_file, custom_options=[]): if not self.is_host_in(name): raise StormValueError('{0} doesn\'t exists in your sshconfig. use storm add command to add.'.format(name)) options = self.get_options(host, user, port, id_file, custom_options) self.ssh_config.update_host(name, options) self.ssh_config.write_to_ssh_config() return True def delete_entry(self, name): self.ssh_config.delete_host(name) self.ssh_config.write_to_ssh_config() return True def list_entries(self, order=False, only_servers=False): config_data = self.ssh_config.config_data # required for the web api. if only_servers: for index, value in enumerate(config_data): print value if value.get('type') and value.get("type") != 'entry': del config_data[index] if order: config_data = sorted(config_data, key=itemgetter("host")) return config_data def delete_all_entries(self): self.ssh_config.delete_all_hosts() return True def search_host(self, search_string): results = self.ssh_config.search_host(search_string) formatted_results = [] for host_entry in results: formatted_results.append(" {0} -> {1}@{2}:{3}\n".format( host_entry.get("host"), host_entry.get("options").get("user", getpass.getuser()), host_entry.get("options").get("hostname"), host_entry.get("options").get("port", 22), )) return formatted_results def get_options(self, host, user, port, id_file, custom_options): options = { 'hostname': host, 'user': user, 'port': port, } if id_file: options.update({ 'identityfile': id_file, }) if len(custom_options) > 0: for custom_option in custom_options: if '=' in custom_option: key, value = custom_option.split("=")[0:2] options.update({ key: value, }) return options def is_host_in(self, host): for host_ in self.ssh_config.config_data: if host_.get("host") == host: return True return False
29.943925
118
0.594569
from ssh_config import ConfigParser from exceptions import StormValueError from operator import itemgetter import getpass __version__ = '0.5.2' class Storm(object): def __init__(self, ssh_config_file=None): self.ssh_config = ConfigParser(ssh_config_file) self.ssh_config.load() def add_entry(self, name, host, user, port, id_file, custom_options=[]): if self.is_host_in(name): raise StormValueError('{0} is already in your sshconfig. use storm edit command to modify.'.format(name)) options = self.get_options(host, user, port, id_file, custom_options) self.ssh_config.add_host(name, options) self.ssh_config.write_to_ssh_config() return True def edit_entry(self, name, host, user, port, id_file, custom_options=[]): if not self.is_host_in(name): raise StormValueError('{0} doesn\'t exists in your sshconfig. use storm add command to add.'.format(name)) options = self.get_options(host, user, port, id_file, custom_options) self.ssh_config.update_host(name, options) self.ssh_config.write_to_ssh_config() return True def delete_entry(self, name): self.ssh_config.delete_host(name) self.ssh_config.write_to_ssh_config() return True def list_entries(self, order=False, only_servers=False): config_data = self.ssh_config.config_data # required for the web api. if only_servers: for index, value in enumerate(config_data): print value if value.get('type') and value.get("type") != 'entry': del config_data[index] if order: config_data = sorted(config_data, key=itemgetter("host")) return config_data def delete_all_entries(self): self.ssh_config.delete_all_hosts() return True def search_host(self, search_string): results = self.ssh_config.search_host(search_string) formatted_results = [] for host_entry in results: formatted_results.append(" {0} -> {1}@{2}:{3}\n".format( host_entry.get("host"), host_entry.get("options").get("user", getpass.getuser()), host_entry.get("options").get("hostname"), host_entry.get("options").get("port", 22), )) return formatted_results def get_options(self, host, user, port, id_file, custom_options): options = { 'hostname': host, 'user': user, 'port': port, } if id_file: options.update({ 'identityfile': id_file, }) if len(custom_options) > 0: for custom_option in custom_options: if '=' in custom_option: key, value = custom_option.split("=")[0:2] options.update({ key: value, }) return options def is_host_in(self, host): for host_ in self.ssh_config.config_data: if host_.get("host") == host: return True return False
false
true
f70e9c0b8bac3d670748990df3ff0e02f0a7f8ad
15,619
py
Python
zerver/lib/cache.py
dehnert/zulip
f5935e81c7cf2f11ff4ccfcd31d2a1061b8d7ff5
[ "Apache-2.0" ]
null
null
null
zerver/lib/cache.py
dehnert/zulip
f5935e81c7cf2f11ff4ccfcd31d2a1061b8d7ff5
[ "Apache-2.0" ]
null
null
null
zerver/lib/cache.py
dehnert/zulip
f5935e81c7cf2f11ff4ccfcd31d2a1061b8d7ff5
[ "Apache-2.0" ]
null
null
null
from __future__ import absolute_import from __future__ import print_function from functools import wraps from django.core.cache import cache as djcache from django.core.cache import caches from django.conf import settings from django.db.models import Q from django.core.cache.backends.base import BaseCache from typing import Any, Callable, Iterable, Optional, Union, TypeVar from zerver.lib.utils import statsd, statsd_key, make_safe_digest import subprocess import time import base64 import random import sys import os import os.path import hashlib import six from six import text_type if False: from zerver.models import UserProfile, Realm, Message # These modules have to be imported for type annotations but # they cannot be imported at runtime due to cyclic dependency. FuncT = TypeVar('FuncT', bound=Callable[..., Any]) remote_cache_time_start = 0.0 remote_cache_total_time = 0.0 remote_cache_total_requests = 0 def get_remote_cache_time(): # type: () -> float return remote_cache_total_time def get_remote_cache_requests(): # type: () -> int return remote_cache_total_requests def remote_cache_stats_start(): # type: () -> None global remote_cache_time_start remote_cache_time_start = time.time() def remote_cache_stats_finish(): # type: () -> None global remote_cache_total_time global remote_cache_total_requests global remote_cache_time_start remote_cache_total_requests += 1 remote_cache_total_time += (time.time() - remote_cache_time_start) def get_or_create_key_prefix(): # type: () -> text_type if settings.TEST_SUITE: # This sets the prefix mostly for the benefit of the JS tests. # The Python tests overwrite KEY_PREFIX on each test. return u'test_suite:%s:' % (text_type(os.getpid()),) # directory `var` should exist in production subprocess.check_call(["mkdir", "-p", os.path.join(settings.DEPLOY_ROOT, "var")]) filename = os.path.join(settings.DEPLOY_ROOT, "var", "remote_cache_prefix") try: fd = os.open(filename, os.O_CREAT | os.O_EXCL | os.O_RDWR, 0o444) random_hash = hashlib.sha256(text_type(random.getrandbits(256)).encode('utf-8')).digest() prefix = base64.b16encode(random_hash)[:32].decode('utf-8').lower() + ':' # This does close the underlying file with os.fdopen(fd, 'w') as f: f.write(prefix + "\n") except OSError: # The file already exists tries = 1 while tries < 10: with open(filename, 'r') as f: prefix = f.readline()[:-1] if len(prefix) == 33: break tries += 1 prefix = '' time.sleep(0.5) if not prefix: print("Could not read remote cache key prefix file") sys.exit(1) return prefix KEY_PREFIX = get_or_create_key_prefix() # type: text_type def bounce_key_prefix_for_testing(test_name): # type: (text_type) -> None global KEY_PREFIX KEY_PREFIX = test_name + u':' + text_type(os.getpid()) + u':' def get_cache_backend(cache_name): # type: (Optional[str]) -> BaseCache if cache_name is None: return djcache return caches[cache_name] def cache_with_key(keyfunc, cache_name=None, timeout=None, with_statsd_key=None): # type: (Any, Optional[str], Optional[int], Optional[str]) -> Any # This function can't be typed perfectly because returning a generic function # isn't supported in mypy - https://github.com/python/mypy/issues/1551. """Decorator which applies Django caching to a function. Decorator argument is a function which computes a cache key from the original function's arguments. You are responsible for avoiding collisions with other uses of this decorator or other uses of caching.""" def decorator(func): # type: (Callable[..., Any]) -> (Callable[..., Any]) @wraps(func) def func_with_caching(*args, **kwargs): # type: (*Any, **Any) -> Callable[..., Any] key = keyfunc(*args, **kwargs) val = cache_get(key, cache_name=cache_name) extra = "" if cache_name == 'database': extra = ".dbcache" if with_statsd_key is not None: metric_key = with_statsd_key else: metric_key = statsd_key(key) status = "hit" if val is not None else "miss" statsd.incr("cache%s.%s.%s" % (extra, metric_key, status)) # Values are singleton tuples so that we can distinguish # a result of None from a missing key. if val is not None: return val[0] val = func(*args, **kwargs) cache_set(key, val, cache_name=cache_name, timeout=timeout) return val return func_with_caching return decorator def cache_set(key, val, cache_name=None, timeout=None): # type: (text_type, Any, Optional[str], Optional[int]) -> None remote_cache_stats_start() cache_backend = get_cache_backend(cache_name) cache_backend.set(KEY_PREFIX + key, (val,), timeout=timeout) remote_cache_stats_finish() def cache_get(key, cache_name=None): # type: (text_type, Optional[str]) -> Any remote_cache_stats_start() cache_backend = get_cache_backend(cache_name) ret = cache_backend.get(KEY_PREFIX + key) remote_cache_stats_finish() return ret def cache_get_many(keys, cache_name=None): # type: (List[text_type], Optional[str]) -> Dict[text_type, Any] keys = [KEY_PREFIX + key for key in keys] remote_cache_stats_start() ret = get_cache_backend(cache_name).get_many(keys) remote_cache_stats_finish() return dict([(key[len(KEY_PREFIX):], value) for key, value in ret.items()]) def cache_set_many(items, cache_name=None, timeout=None): # type: (Dict[text_type, Any], Optional[str], Optional[int]) -> None new_items = {} for key in items: new_items[KEY_PREFIX + key] = items[key] items = new_items remote_cache_stats_start() get_cache_backend(cache_name).set_many(items, timeout=timeout) remote_cache_stats_finish() def cache_delete(key, cache_name=None): # type: (text_type, Optional[str]) -> None remote_cache_stats_start() get_cache_backend(cache_name).delete(KEY_PREFIX + key) remote_cache_stats_finish() def cache_delete_many(items, cache_name=None): # type: (Iterable[text_type], Optional[str]) -> None remote_cache_stats_start() get_cache_backend(cache_name).delete_many( KEY_PREFIX + item for item in items) remote_cache_stats_finish() # Required Arguments are as follows: # * object_ids: The list of object ids to look up # * cache_key_function: object_id => cache key # * query_function: [object_ids] => [objects from database] # Optional keyword arguments: # * setter: Function to call before storing items to cache (e.g. compression) # * extractor: Function to call on items returned from cache # (e.g. decompression). Should be the inverse of the setter # function. # * id_fetcher: Function mapping an object from database => object_id # (in case we're using a key more complex than obj.id) # * cache_transformer: Function mapping an object from database => # value for cache (in case the values that we're caching are some # function of the objects, not the objects themselves) ObjKT = TypeVar('ObjKT', int, text_type) ItemT = Any # https://github.com/python/mypy/issues/1721 CompressedItemT = Any # https://github.com/python/mypy/issues/1721 def generic_bulk_cached_fetch(cache_key_function, # type: Callable[[ObjKT], text_type] query_function, # type: Callable[[List[ObjKT]], Iterable[Any]] object_ids, # type: Iterable[ObjKT] extractor=lambda obj: obj, # type: Callable[[CompressedItemT], ItemT] setter=lambda obj: obj, # type: Callable[[ItemT], CompressedItemT] id_fetcher=lambda obj: obj.id, # type: Callable[[Any], ObjKT] cache_transformer=lambda obj: obj # type: Callable[[Any], ItemT] ): # type: (...) -> Dict[ObjKT, Any] cache_keys = {} # type: Dict[ObjKT, text_type] for object_id in object_ids: cache_keys[object_id] = cache_key_function(object_id) cached_objects = cache_get_many([cache_keys[object_id] for object_id in object_ids]) for (key, val) in cached_objects.items(): cached_objects[key] = extractor(cached_objects[key][0]) needed_ids = [object_id for object_id in object_ids if cache_keys[object_id] not in cached_objects] db_objects = query_function(needed_ids) items_for_remote_cache = {} # type: Dict[text_type, Any] for obj in db_objects: key = cache_keys[id_fetcher(obj)] item = cache_transformer(obj) items_for_remote_cache[key] = (setter(item),) cached_objects[key] = item if len(items_for_remote_cache) > 0: cache_set_many(items_for_remote_cache) return dict((object_id, cached_objects[cache_keys[object_id]]) for object_id in object_ids if cache_keys[object_id] in cached_objects) def cache(func): # type: (FuncT) -> FuncT """Decorator which applies Django caching to a function. Uses a key based on the function's name, filename, and the repr() of its arguments.""" func_uniqifier = '%s-%s' % (func.__code__.co_filename, func.__name__) # type: ignore # https://github.com/python/mypy/issues/1923 @wraps(func) def keyfunc(*args, **kwargs): # type: (*Any, **Any) -> str # Django complains about spaces because memcached rejects them key = func_uniqifier + repr((args, kwargs)) return key.replace('-', '--').replace(' ', '-s') return cache_with_key(keyfunc)(func) def display_recipient_cache_key(recipient_id): # type: (int) -> text_type return u"display_recipient_dict:%d" % (recipient_id,) def user_profile_by_email_cache_key(email): # type: (text_type) -> text_type # See the comment in zerver/lib/avatar_hash.py:gravatar_hash for why we # are proactively encoding email addresses even though they will # with high likelihood be ASCII-only for the foreseeable future. return u'user_profile_by_email:%s' % (make_safe_digest(email.strip()),) def user_profile_by_id_cache_key(user_profile_id): # type: (int) -> text_type return u"user_profile_by_id:%s" % (user_profile_id,) # TODO: Refactor these cache helpers into another file that can import # models.py so that python3-style type annotations can also work. def cache_save_user_profile(user_profile): # type: (UserProfile) -> None cache_set(user_profile_by_id_cache_key(user_profile.id), user_profile, timeout=3600*24*7) active_user_dict_fields = ['id', 'full_name', 'short_name', 'email', 'is_realm_admin', 'is_bot'] # type: List[str] def active_user_dicts_in_realm_cache_key(realm): # type: (Realm) -> text_type return u"active_user_dicts_in_realm:%s" % (realm.id,) active_bot_dict_fields = ['id', 'full_name', 'short_name', 'email', 'default_sending_stream__name', 'default_events_register_stream__name', 'default_all_public_streams', 'api_key', 'bot_owner__email', 'avatar_source'] # type: List[str] def active_bot_dicts_in_realm_cache_key(realm): # type: (Realm) -> text_type return u"active_bot_dicts_in_realm:%s" % (realm.id,) def get_stream_cache_key(stream_name, realm): # type: (text_type, Union[Realm, int]) -> text_type from zerver.models import Realm if isinstance(realm, Realm): realm_id = realm.id else: realm_id = realm return u"stream_by_realm_and_name:%s:%s" % ( realm_id, make_safe_digest(stream_name.strip().lower())) def delete_user_profile_caches(user_profiles): # type: (Iterable[UserProfile]) -> None keys = [] for user_profile in user_profiles: keys.append(user_profile_by_email_cache_key(user_profile.email)) keys.append(user_profile_by_id_cache_key(user_profile.id)) cache_delete_many(keys) # Called by models.py to flush the user_profile cache whenever we save # a user_profile object def flush_user_profile(sender, **kwargs): # type: (Any, **Any) -> None user_profile = kwargs['instance'] delete_user_profile_caches([user_profile]) # Invalidate our active_users_in_realm info dict if any user has changed # the fields in the dict or become (in)active if kwargs.get('update_fields') is None or \ len(set(active_user_dict_fields + ['is_active']) & set(kwargs['update_fields'])) > 0: cache_delete(active_user_dicts_in_realm_cache_key(user_profile.realm)) # Invalidate our active_bots_in_realm info dict if any bot has # changed the fields in the dict or become (in)active if user_profile.is_bot and (kwargs['update_fields'] is None or (set(active_bot_dict_fields + ['is_active']) & set(kwargs['update_fields']))): cache_delete(active_bot_dicts_in_realm_cache_key(user_profile.realm)) # Invalidate realm-wide alert words cache if any user in the realm has changed # alert words if kwargs.get('update_fields') is None or "alert_words" in kwargs['update_fields']: cache_delete(realm_alert_words_cache_key(user_profile.realm)) # Called by models.py to flush various caches whenever we save # a Realm object. The main tricky thing here is that Realm info is # generally cached indirectly through user_profile objects. def flush_realm(sender, **kwargs): # type: (Any, **Any) -> None realm = kwargs['instance'] users = realm.get_active_users() delete_user_profile_caches(users) if realm.deactivated: cache_delete(active_user_dicts_in_realm_cache_key(realm)) cache_delete(active_bot_dicts_in_realm_cache_key(realm)) cache_delete(realm_alert_words_cache_key(realm)) def realm_alert_words_cache_key(realm): # type: (Realm) -> text_type return u"realm_alert_words:%s" % (realm.domain,) # Called by models.py to flush the stream cache whenever we save a stream # object. def flush_stream(sender, **kwargs): # type: (Any, **Any) -> None from zerver.models import UserProfile stream = kwargs['instance'] items_for_remote_cache = {} items_for_remote_cache[get_stream_cache_key(stream.name, stream.realm)] = (stream,) cache_set_many(items_for_remote_cache) if kwargs.get('update_fields') is None or 'name' in kwargs['update_fields'] and \ UserProfile.objects.filter( Q(default_sending_stream=stream) | Q(default_events_register_stream=stream) ).exists(): cache_delete(active_bot_dicts_in_realm_cache_key(stream.realm)) # TODO: Rename to_dict_cache_key_id and to_dict_cache_key def to_dict_cache_key_id(message_id, apply_markdown): # type: (int, bool) -> text_type return u'message_dict:%d:%d' % (message_id, apply_markdown) def to_dict_cache_key(message, apply_markdown): # type: (Message, bool) -> text_type return to_dict_cache_key_id(message.id, apply_markdown) def flush_message(sender, **kwargs): # type: (Any, **Any) -> None message = kwargs['instance'] cache_delete(to_dict_cache_key(message, False)) cache_delete(to_dict_cache_key(message, True))
39.541772
133
0.681414
from __future__ import absolute_import from __future__ import print_function from functools import wraps from django.core.cache import cache as djcache from django.core.cache import caches from django.conf import settings from django.db.models import Q from django.core.cache.backends.base import BaseCache from typing import Any, Callable, Iterable, Optional, Union, TypeVar from zerver.lib.utils import statsd, statsd_key, make_safe_digest import subprocess import time import base64 import random import sys import os import os.path import hashlib import six from six import text_type if False: from zerver.models import UserProfile, Realm, Message FuncT = TypeVar('FuncT', bound=Callable[..., Any]) remote_cache_time_start = 0.0 remote_cache_total_time = 0.0 remote_cache_total_requests = 0 def get_remote_cache_time(): return remote_cache_total_time def get_remote_cache_requests(): return remote_cache_total_requests def remote_cache_stats_start(): global remote_cache_time_start remote_cache_time_start = time.time() def remote_cache_stats_finish(): global remote_cache_total_time global remote_cache_total_requests global remote_cache_time_start remote_cache_total_requests += 1 remote_cache_total_time += (time.time() - remote_cache_time_start) def get_or_create_key_prefix(): if settings.TEST_SUITE: return u'test_suite:%s:' % (text_type(os.getpid()),) subprocess.check_call(["mkdir", "-p", os.path.join(settings.DEPLOY_ROOT, "var")]) filename = os.path.join(settings.DEPLOY_ROOT, "var", "remote_cache_prefix") try: fd = os.open(filename, os.O_CREAT | os.O_EXCL | os.O_RDWR, 0o444) random_hash = hashlib.sha256(text_type(random.getrandbits(256)).encode('utf-8')).digest() prefix = base64.b16encode(random_hash)[:32].decode('utf-8').lower() + ':' with os.fdopen(fd, 'w') as f: f.write(prefix + "\n") except OSError: tries = 1 while tries < 10: with open(filename, 'r') as f: prefix = f.readline()[:-1] if len(prefix) == 33: break tries += 1 prefix = '' time.sleep(0.5) if not prefix: print("Could not read remote cache key prefix file") sys.exit(1) return prefix KEY_PREFIX = get_or_create_key_prefix() def bounce_key_prefix_for_testing(test_name): global KEY_PREFIX KEY_PREFIX = test_name + u':' + text_type(os.getpid()) + u':' def get_cache_backend(cache_name): if cache_name is None: return djcache return caches[cache_name] def cache_with_key(keyfunc, cache_name=None, timeout=None, with_statsd_key=None): # isn't supported in mypy - https://github.com/python/mypy/issues/1551. def decorator(func): @wraps(func) def func_with_caching(*args, **kwargs): key = keyfunc(*args, **kwargs) val = cache_get(key, cache_name=cache_name) extra = "" if cache_name == 'database': extra = ".dbcache" if with_statsd_key is not None: metric_key = with_statsd_key else: metric_key = statsd_key(key) status = "hit" if val is not None else "miss" statsd.incr("cache%s.%s.%s" % (extra, metric_key, status)) if val is not None: return val[0] val = func(*args, **kwargs) cache_set(key, val, cache_name=cache_name, timeout=timeout) return val return func_with_caching return decorator def cache_set(key, val, cache_name=None, timeout=None): remote_cache_stats_start() cache_backend = get_cache_backend(cache_name) cache_backend.set(KEY_PREFIX + key, (val,), timeout=timeout) remote_cache_stats_finish() def cache_get(key, cache_name=None): remote_cache_stats_start() cache_backend = get_cache_backend(cache_name) ret = cache_backend.get(KEY_PREFIX + key) remote_cache_stats_finish() return ret def cache_get_many(keys, cache_name=None): keys = [KEY_PREFIX + key for key in keys] remote_cache_stats_start() ret = get_cache_backend(cache_name).get_many(keys) remote_cache_stats_finish() return dict([(key[len(KEY_PREFIX):], value) for key, value in ret.items()]) def cache_set_many(items, cache_name=None, timeout=None): new_items = {} for key in items: new_items[KEY_PREFIX + key] = items[key] items = new_items remote_cache_stats_start() get_cache_backend(cache_name).set_many(items, timeout=timeout) remote_cache_stats_finish() def cache_delete(key, cache_name=None): remote_cache_stats_start() get_cache_backend(cache_name).delete(KEY_PREFIX + key) remote_cache_stats_finish() def cache_delete_many(items, cache_name=None): remote_cache_stats_start() get_cache_backend(cache_name).delete_many( KEY_PREFIX + item for item in items) remote_cache_stats_finish() # * cache_transformer: Function mapping an object from database => # value for cache (in case the values that we're caching are some ObjKT = TypeVar('ObjKT', int, text_type) ItemT = Any CompressedItemT = Any def generic_bulk_cached_fetch(cache_key_function, query_function, object_ids, extractor=lambda obj: obj, setter=lambda obj: obj, id_fetcher=lambda obj: obj.id, cache_transformer=lambda obj: obj ): cache_keys = {} for object_id in object_ids: cache_keys[object_id] = cache_key_function(object_id) cached_objects = cache_get_many([cache_keys[object_id] for object_id in object_ids]) for (key, val) in cached_objects.items(): cached_objects[key] = extractor(cached_objects[key][0]) needed_ids = [object_id for object_id in object_ids if cache_keys[object_id] not in cached_objects] db_objects = query_function(needed_ids) items_for_remote_cache = {} for obj in db_objects: key = cache_keys[id_fetcher(obj)] item = cache_transformer(obj) items_for_remote_cache[key] = (setter(item),) cached_objects[key] = item if len(items_for_remote_cache) > 0: cache_set_many(items_for_remote_cache) return dict((object_id, cached_objects[cache_keys[object_id]]) for object_id in object_ids if cache_keys[object_id] in cached_objects) def cache(func): func_uniqifier = '%s-%s' % (func.__code__.co_filename, func.__name__) kwargs): key = func_uniqifier + repr((args, kwargs)) return key.replace('-', '--').replace(' ', '-s') return cache_with_key(keyfunc)(func) def display_recipient_cache_key(recipient_id): return u"display_recipient_dict:%d" % (recipient_id,) def user_profile_by_email_cache_key(email): return u'user_profile_by_email:%s' % (make_safe_digest(email.strip()),) def user_profile_by_id_cache_key(user_profile_id): return u"user_profile_by_id:%s" % (user_profile_id,) def cache_save_user_profile(user_profile): cache_set(user_profile_by_id_cache_key(user_profile.id), user_profile, timeout=3600*24*7) active_user_dict_fields = ['id', 'full_name', 'short_name', 'email', 'is_realm_admin', 'is_bot'] def active_user_dicts_in_realm_cache_key(realm): return u"active_user_dicts_in_realm:%s" % (realm.id,) active_bot_dict_fields = ['id', 'full_name', 'short_name', 'email', 'default_sending_stream__name', 'default_events_register_stream__name', 'default_all_public_streams', 'api_key', 'bot_owner__email', 'avatar_source'] def active_bot_dicts_in_realm_cache_key(realm): return u"active_bot_dicts_in_realm:%s" % (realm.id,) def get_stream_cache_key(stream_name, realm): from zerver.models import Realm if isinstance(realm, Realm): realm_id = realm.id else: realm_id = realm return u"stream_by_realm_and_name:%s:%s" % ( realm_id, make_safe_digest(stream_name.strip().lower())) def delete_user_profile_caches(user_profiles): keys = [] for user_profile in user_profiles: keys.append(user_profile_by_email_cache_key(user_profile.email)) keys.append(user_profile_by_id_cache_key(user_profile.id)) cache_delete_many(keys) def flush_user_profile(sender, **kwargs): user_profile = kwargs['instance'] delete_user_profile_caches([user_profile]) if kwargs.get('update_fields') is None or \ len(set(active_user_dict_fields + ['is_active']) & set(kwargs['update_fields'])) > 0: cache_delete(active_user_dicts_in_realm_cache_key(user_profile.realm)) if user_profile.is_bot and (kwargs['update_fields'] is None or (set(active_bot_dict_fields + ['is_active']) & set(kwargs['update_fields']))): cache_delete(active_bot_dicts_in_realm_cache_key(user_profile.realm)) if kwargs.get('update_fields') is None or "alert_words" in kwargs['update_fields']: cache_delete(realm_alert_words_cache_key(user_profile.realm)) def flush_realm(sender, **kwargs): realm = kwargs['instance'] users = realm.get_active_users() delete_user_profile_caches(users) if realm.deactivated: cache_delete(active_user_dicts_in_realm_cache_key(realm)) cache_delete(active_bot_dicts_in_realm_cache_key(realm)) cache_delete(realm_alert_words_cache_key(realm)) def realm_alert_words_cache_key(realm): return u"realm_alert_words:%s" % (realm.domain,) def flush_stream(sender, **kwargs): from zerver.models import UserProfile stream = kwargs['instance'] items_for_remote_cache = {} items_for_remote_cache[get_stream_cache_key(stream.name, stream.realm)] = (stream,) cache_set_many(items_for_remote_cache) if kwargs.get('update_fields') is None or 'name' in kwargs['update_fields'] and \ UserProfile.objects.filter( Q(default_sending_stream=stream) | Q(default_events_register_stream=stream) ).exists(): cache_delete(active_bot_dicts_in_realm_cache_key(stream.realm)) def to_dict_cache_key_id(message_id, apply_markdown): return u'message_dict:%d:%d' % (message_id, apply_markdown) def to_dict_cache_key(message, apply_markdown): return to_dict_cache_key_id(message.id, apply_markdown) def flush_message(sender, **kwargs): message = kwargs['instance'] cache_delete(to_dict_cache_key(message, False)) cache_delete(to_dict_cache_key(message, True))
true
true
f70e9c0f85a4b70d03afc9fad7137e074aea7e36
1,674
py
Python
benchmarks_sphere/paper_jrn_parco_rexi_nonlinear/scalability_space_galewsky_cheyenne_intel/postprocessing.py
valentinaschueller/sweet
27e99c7a110c99deeadee70688c186d82b39ac90
[ "MIT" ]
6
2017-11-20T08:12:46.000Z
2021-03-11T15:32:36.000Z
benchmarks_sphere/paper_jrn_parco_rexi_nonlinear/scalability_space_galewsky_cheyenne_intel/postprocessing.py
valentinaschueller/sweet
27e99c7a110c99deeadee70688c186d82b39ac90
[ "MIT" ]
4
2018-02-02T21:46:33.000Z
2022-01-11T11:10:27.000Z
benchmarks_sphere/paper_jrn_parco_rexi_nonlinear/scalability_space_galewsky_cheyenne_intel/postprocessing.py
valentinaschueller/sweet
27e99c7a110c99deeadee70688c186d82b39ac90
[ "MIT" ]
12
2016-03-01T18:33:34.000Z
2022-02-08T22:20:31.000Z
#! /usr/bin/env python3 from SWEET import * from mule.postprocessing.JobsData import * from mule.postprocessing.JobsDataConsolidate import * from mule.plotting.Plotting import * sys.path.append('../') import pretty_plotting as pp sys.path.pop() # # Load data # j = JobsData('job_bench_*', verbosity=0) # # Create groups # groups = ['runtime.timestepping_method'] c = JobsDataConsolidate(j) job_groups = c.create_groups(groups) print("Groups:") for key, g in job_groups.items(): print(key) tagname_x = 'parallelization.num_threads_per_rank' tagname_y = 'output.simulation_benchmark_timings.main_timestepping' # # Make ready for plotting # d = JobsData_GroupsPlottingScattered( job_groups, tagname_x, tagname_y ) data_plotting = d.get_data_float() # Make pretty for key, data in data_plotting.items(): data['label'] = pp.get_pretty_name(key) # # Plot! # p = Plotting_ScatteredData() p.plot( data_plotting = data_plotting, xlabel = "Number of threads", ylabel = "Wallclock time (seconds)", title = "Wallclock time", outfile = "output_threads_vs_wallclock_time.pdf" ) # # Scalability # for key, values in data_plotting.items(): label = key x_values = values['x_values'] y_values = values['y_values'] # Basis for scalability (number of cores) basis_scalability = 1.0 # Get index of x value for scalability i = x_values.index(basis_scalability) if i == None: raise Exception("Scalability basis not found") # Convert to scalability values['y_values'] = [y_values[i]/y for y in y_values] p.plot( data_plotting, xlabel="Number of threads", ylabel="Scalability", title = "Scalability", outfile="output_threads_vs_scalability.pdf" )
18.6
67
0.737754
from SWEET import * from mule.postprocessing.JobsData import * from mule.postprocessing.JobsDataConsolidate import * from mule.plotting.Plotting import * sys.path.append('../') import pretty_plotting as pp sys.path.pop() j = JobsData('job_bench_*', verbosity=0) groups = ['runtime.timestepping_method'] c = JobsDataConsolidate(j) job_groups = c.create_groups(groups) print("Groups:") for key, g in job_groups.items(): print(key) tagname_x = 'parallelization.num_threads_per_rank' tagname_y = 'output.simulation_benchmark_timings.main_timestepping' d = JobsData_GroupsPlottingScattered( job_groups, tagname_x, tagname_y ) data_plotting = d.get_data_float() for key, data in data_plotting.items(): data['label'] = pp.get_pretty_name(key) p = Plotting_ScatteredData() p.plot( data_plotting = data_plotting, xlabel = "Number of threads", ylabel = "Wallclock time (seconds)", title = "Wallclock time", outfile = "output_threads_vs_wallclock_time.pdf" ) for key, values in data_plotting.items(): label = key x_values = values['x_values'] y_values = values['y_values'] basis_scalability = 1.0 i = x_values.index(basis_scalability) if i == None: raise Exception("Scalability basis not found") values['y_values'] = [y_values[i]/y for y in y_values] p.plot( data_plotting, xlabel="Number of threads", ylabel="Scalability", title = "Scalability", outfile="output_threads_vs_scalability.pdf" )
true
true
f70e9d0ca009d2d068924b1345df34f152420b7c
1,346
py
Python
setup.py
FaizChishtie/MrTopo
e6a674738d8b0a0c56edde2be0ae272ea9e62f1a
[ "MIT" ]
1
2021-01-26T11:01:32.000Z
2021-01-26T11:01:32.000Z
setup.py
FaizChishtie/MrTopo
e6a674738d8b0a0c56edde2be0ae272ea9e62f1a
[ "MIT" ]
null
null
null
setup.py
FaizChishtie/MrTopo
e6a674738d8b0a0c56edde2be0ae272ea9e62f1a
[ "MIT" ]
1
2020-12-23T22:01:09.000Z
2020-12-23T22:01:09.000Z
import setuptools import re with open("README.md", "r") as fh: long_description = fh.read() version = re.search( '^__version__\s*=\s*"(.*)"', open('mrtopo/__main__.py').read(), re.M ).group(1) setuptools.setup( name='mrtopo', version=version, packages=setuptools.find_packages(), url='https://github.com/FaizChishtie/mrtopo', license='MIT', author='faizchishtie', author_email='faizchishtie@gmail.com', description='Mutate Mininet topology files with MrTopo', python_requires='>=3.0', entry_points={'console_scripts': ['mrtopo = mrtopo.cli:cli']}, long_description=long_description, long_description_content_type='text/markdown', install_requires=[ 'mininet', 'click' ], keywords='topology network startup', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Utilities', 'Typing :: Typed', ], )
29.911111
66
0.61367
import setuptools import re with open("README.md", "r") as fh: long_description = fh.read() version = re.search( '^__version__\s*=\s*"(.*)"', open('mrtopo/__main__.py').read(), re.M ).group(1) setuptools.setup( name='mrtopo', version=version, packages=setuptools.find_packages(), url='https://github.com/FaizChishtie/mrtopo', license='MIT', author='faizchishtie', author_email='faizchishtie@gmail.com', description='Mutate Mininet topology files with MrTopo', python_requires='>=3.0', entry_points={'console_scripts': ['mrtopo = mrtopo.cli:cli']}, long_description=long_description, long_description_content_type='text/markdown', install_requires=[ 'mininet', 'click' ], keywords='topology network startup', classifiers=[ 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3.6', 'Programming Language :: Python :: 3.7', 'Programming Language :: Python :: 3.8', 'Programming Language :: Python :: 3.9', 'Topic :: Utilities', 'Typing :: Typed', ], )
true
true
f70e9e8f3f4e25b72d078669cef22e5f17566484
1,495
py
Python
setup.py
ucamhal/jsonlogging
7606abb00c3cdc9536c8947d5ab1e210716e8d47
[ "BSD-2-Clause" ]
1
2017-03-19T12:06:58.000Z
2017-03-19T12:06:58.000Z
setup.py
ucamhal/jsonlogging
7606abb00c3cdc9536c8947d5ab1e210716e8d47
[ "BSD-2-Clause" ]
null
null
null
setup.py
ucamhal/jsonlogging
7606abb00c3cdc9536c8947d5ab1e210716e8d47
[ "BSD-2-Clause" ]
null
null
null
from setuptools import setup def get_version(filename): """ Parse the value of the __version__ var from a Python source file without running/importing the file. """ import re version_pattern = r"^ *__version__ *= *['\"](\d+\.\d+\.\d+)['\"] *$" match = re.search(version_pattern, open(filename).read(), re.MULTILINE) assert match, ("No version found in file: {!r} matching pattern: {!r}" .format(filename, version_pattern)) return match.group(1) setup( name="jsonlogging", description="jsonlogging provides structured log output from the " "logging module in JSON format", author="Hal Blackburn", author_email="hwtb2@cam.ac.uk", url="https://github.com/ucamhal/ravenpy", version=get_version("jsonlogging/__init__.py"), packages=["jsonlogging"], license="BSD", classifiers=[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python", "Topic :: Software Development", "Topic :: System :: Logging" ], long_description=open("README.md").read(), test_suite="jsonlogging.tests.test_all", tests_require="mock >= 1.0.0, < 2.0.0" )
33.977273
75
0.623411
from setuptools import setup def get_version(filename): import re version_pattern = r"^ *__version__ *= *['\"](\d+\.\d+\.\d+)['\"] *$" match = re.search(version_pattern, open(filename).read(), re.MULTILINE) assert match, ("No version found in file: {!r} matching pattern: {!r}" .format(filename, version_pattern)) return match.group(1) setup( name="jsonlogging", description="jsonlogging provides structured log output from the " "logging module in JSON format", author="Hal Blackburn", author_email="hwtb2@cam.ac.uk", url="https://github.com/ucamhal/ravenpy", version=get_version("jsonlogging/__init__.py"), packages=["jsonlogging"], license="BSD", classifiers=[ "Development Status :: 5 - Production/Stable", "Intended Audience :: Developers", "Intended Audience :: System Administrators", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", "Programming Language :: Python :: 2.6", "Programming Language :: Python :: 2.7", "Programming Language :: Python", "Topic :: Software Development", "Topic :: System :: Logging" ], long_description=open("README.md").read(), test_suite="jsonlogging.tests.test_all", tests_require="mock >= 1.0.0, < 2.0.0" )
true
true
f70e9eba1ab08ba41a6d14813985705f6afe0f1e
94,216
py
Python
sympy/plotting/plot.py
Michal-Gagala/sympy
3cc756c2af73b5506102abaeefd1b654e286e2c8
[ "MIT" ]
null
null
null
sympy/plotting/plot.py
Michal-Gagala/sympy
3cc756c2af73b5506102abaeefd1b654e286e2c8
[ "MIT" ]
null
null
null
sympy/plotting/plot.py
Michal-Gagala/sympy
3cc756c2af73b5506102abaeefd1b654e286e2c8
[ "MIT" ]
null
null
null
"""Plotting module for SymPy. A plot is represented by the ``Plot`` class that contains a reference to the backend and a list of the data series to be plotted. The data series are instances of classes meant to simplify getting points and meshes from SymPy expressions. ``plot_backends`` is a dictionary with all the backends. This module gives only the essential. For all the fancy stuff use directly the backend. You can get the backend wrapper for every plot from the ``_backend`` attribute. Moreover the data series classes have various useful methods like ``get_points``, ``get_meshes``, etc, that may be useful if you wish to use another plotting library. Especially if you need publication ready graphs and this module is not enough for you - just get the ``_backend`` attribute and add whatever you want directly to it. In the case of matplotlib (the common way to graph data in python) just copy ``_backend.fig`` which is the figure and ``_backend.ax`` which is the axis and work on them as you would on any other matplotlib object. Simplicity of code takes much greater importance than performance. Do not use it if you care at all about performance. A new backend instance is initialized every time you call ``show()`` and the old one is left to the garbage collector. """ from collections.abc import Callable from sympy.core.basic import Basic from sympy.core.containers import Tuple from sympy.core.expr import Expr from sympy.core.function import arity, Function from sympy.core.symbol import (Dummy, Symbol) from sympy.core.sympify import sympify from sympy.external import import_module from sympy.printing.latex import latex from sympy.utilities.exceptions import sympy_deprecation_warning from sympy.utilities.iterables import is_sequence from .experimental_lambdify import (vectorized_lambdify, lambdify) # N.B. # When changing the minimum module version for matplotlib, please change # the same in the `SymPyDocTestFinder`` in `sympy/testing/runtests.py` # Backend specific imports - textplot from sympy.plotting.textplot import textplot # Global variable # Set to False when running tests / doctests so that the plots don't show. _show = True def unset_show(): """ Disable show(). For use in the tests. """ global _show _show = False def _str_or_latex(label): if isinstance(label, Basic): return latex(label, mode='inline') return str(label) ############################################################################## # The public interface ############################################################################## class Plot: """The central class of the plotting module. Explanation =========== For interactive work the function ``plot`` is better suited. This class permits the plotting of SymPy expressions using numerous backends (matplotlib, textplot, the old pyglet module for sympy, Google charts api, etc). The figure can contain an arbitrary number of plots of SymPy expressions, lists of coordinates of points, etc. Plot has a private attribute _series that contains all data series to be plotted (expressions for lines or surfaces, lists of points, etc (all subclasses of BaseSeries)). Those data series are instances of classes not imported by ``from sympy import *``. The customization of the figure is on two levels. Global options that concern the figure as a whole (eg title, xlabel, scale, etc) and per-data series options (eg name) and aesthetics (eg. color, point shape, line type, etc.). The difference between options and aesthetics is that an aesthetic can be a function of the coordinates (or parameters in a parametric plot). The supported values for an aesthetic are: - None (the backend uses default values) - a constant - a function of one variable (the first coordinate or parameter) - a function of two variables (the first and second coordinate or parameters) - a function of three variables (only in nonparametric 3D plots) Their implementation depends on the backend so they may not work in some backends. If the plot is parametric and the arity of the aesthetic function permits it the aesthetic is calculated over parameters and not over coordinates. If the arity does not permit calculation over parameters the calculation is done over coordinates. Only cartesian coordinates are supported for the moment, but you can use the parametric plots to plot in polar, spherical and cylindrical coordinates. The arguments for the constructor Plot must be subclasses of BaseSeries. Any global option can be specified as a keyword argument. The global options for a figure are: - title : str - xlabel : str or Symbol - ylabel : str or Symbol - zlabel : str or Symbol - legend : bool - xscale : {'linear', 'log'} - yscale : {'linear', 'log'} - axis : bool - axis_center : tuple of two floats or {'center', 'auto'} - xlim : tuple of two floats - ylim : tuple of two floats - aspect_ratio : tuple of two floats or {'auto'} - autoscale : bool - margin : float in [0, 1] - backend : {'default', 'matplotlib', 'text'} or a subclass of BaseBackend - size : optional tuple of two floats, (width, height); default: None The per data series options and aesthetics are: There are none in the base series. See below for options for subclasses. Some data series support additional aesthetics or options: ListSeries, LineOver1DRangeSeries, Parametric2DLineSeries, Parametric3DLineSeries support the following: Aesthetics: - line_color : string, or float, or function, optional Specifies the color for the plot, which depends on the backend being used. For example, if ``MatplotlibBackend`` is being used, then Matplotlib string colors are acceptable ("red", "r", "cyan", "c", ...). Alternatively, we can use a float number `0 < color < 1` wrapped in a string (for example, `line_color="0.5"`) to specify grayscale colors. Alternatively, We can specify a function returning a single float value: this will be used to apply a color-loop (for example, `line_color=lambda x: math.cos(x)`). Note that by setting line_color, it would be applied simultaneously to all the series. options: - label : str - steps : bool - integers_only : bool SurfaceOver2DRangeSeries, ParametricSurfaceSeries support the following: aesthetics: - surface_color : function which returns a float. """ def __init__(self, *args, title=None, xlabel=None, ylabel=None, zlabel=None, aspect_ratio='auto', xlim=None, ylim=None, axis_center='auto', axis=True, xscale='linear', yscale='linear', legend=False, autoscale=True, margin=0, annotations=None, markers=None, rectangles=None, fill=None, backend='default', size=None, **kwargs): super().__init__() # Options for the graph as a whole. # The possible values for each option are described in the docstring of # Plot. They are based purely on convention, no checking is done. self.title = title self.xlabel = xlabel self.ylabel = ylabel self.zlabel = zlabel self.aspect_ratio = aspect_ratio self.axis_center = axis_center self.axis = axis self.xscale = xscale self.yscale = yscale self.legend = legend self.autoscale = autoscale self.margin = margin self.annotations = annotations self.markers = markers self.rectangles = rectangles self.fill = fill # Contains the data objects to be plotted. The backend should be smart # enough to iterate over this list. self._series = [] self._series.extend(args) # The backend type. On every show() a new backend instance is created # in self._backend which is tightly coupled to the Plot instance # (thanks to the parent attribute of the backend). if isinstance(backend, str): self.backend = plot_backends[backend] elif (type(backend) == type) and issubclass(backend, BaseBackend): self.backend = backend else: raise TypeError( "backend must be either a string or a subclass of BaseBackend") is_real = \ lambda lim: all(getattr(i, 'is_real', True) for i in lim) is_finite = \ lambda lim: all(getattr(i, 'is_finite', True) for i in lim) # reduce code repetition def check_and_set(t_name, t): if t: if not is_real(t): raise ValueError( "All numbers from {}={} must be real".format(t_name, t)) if not is_finite(t): raise ValueError( "All numbers from {}={} must be finite".format(t_name, t)) setattr(self, t_name, (float(t[0]), float(t[1]))) self.xlim = None check_and_set("xlim", xlim) self.ylim = None check_and_set("ylim", ylim) self.size = None check_and_set("size", size) def show(self): # TODO move this to the backend (also for save) if hasattr(self, '_backend'): self._backend.close() self._backend = self.backend(self) self._backend.show() def save(self, path): if hasattr(self, '_backend'): self._backend.close() self._backend = self.backend(self) self._backend.save(path) def __str__(self): series_strs = [('[%d]: ' % i) + str(s) for i, s in enumerate(self._series)] return 'Plot object containing:\n' + '\n'.join(series_strs) def __getitem__(self, index): return self._series[index] def __setitem__(self, index, *args): if len(args) == 1 and isinstance(args[0], BaseSeries): self._series[index] = args def __delitem__(self, index): del self._series[index] def append(self, arg): """Adds an element from a plot's series to an existing plot. Examples ======== Consider two ``Plot`` objects, ``p1`` and ``p2``. To add the second plot's first series object to the first, use the ``append`` method, like so: .. plot:: :format: doctest :include-source: True >>> from sympy import symbols >>> from sympy.plotting import plot >>> x = symbols('x') >>> p1 = plot(x*x, show=False) >>> p2 = plot(x, show=False) >>> p1.append(p2[0]) >>> p1 Plot object containing: [0]: cartesian line: x**2 for x over (-10.0, 10.0) [1]: cartesian line: x for x over (-10.0, 10.0) >>> p1.show() See Also ======== extend """ if isinstance(arg, BaseSeries): self._series.append(arg) else: raise TypeError('Must specify element of plot to append.') def extend(self, arg): """Adds all series from another plot. Examples ======== Consider two ``Plot`` objects, ``p1`` and ``p2``. To add the second plot to the first, use the ``extend`` method, like so: .. plot:: :format: doctest :include-source: True >>> from sympy import symbols >>> from sympy.plotting import plot >>> x = symbols('x') >>> p1 = plot(x**2, show=False) >>> p2 = plot(x, -x, show=False) >>> p1.extend(p2) >>> p1 Plot object containing: [0]: cartesian line: x**2 for x over (-10.0, 10.0) [1]: cartesian line: x for x over (-10.0, 10.0) [2]: cartesian line: -x for x over (-10.0, 10.0) >>> p1.show() """ if isinstance(arg, Plot): self._series.extend(arg._series) elif is_sequence(arg): self._series.extend(arg) else: raise TypeError('Expecting Plot or sequence of BaseSeries') class PlotGrid: """This class helps to plot subplots from already created SymPy plots in a single figure. Examples ======== .. plot:: :context: close-figs :format: doctest :include-source: True >>> from sympy import symbols >>> from sympy.plotting import plot, plot3d, PlotGrid >>> x, y = symbols('x, y') >>> p1 = plot(x, x**2, x**3, (x, -5, 5)) >>> p2 = plot((x**2, (x, -6, 6)), (x, (x, -5, 5))) >>> p3 = plot(x**3, (x, -5, 5)) >>> p4 = plot3d(x*y, (x, -5, 5), (y, -5, 5)) Plotting vertically in a single line: .. plot:: :context: close-figs :format: doctest :include-source: True >>> PlotGrid(2, 1, p1, p2) PlotGrid object containing: Plot[0]:Plot object containing: [0]: cartesian line: x for x over (-5.0, 5.0) [1]: cartesian line: x**2 for x over (-5.0, 5.0) [2]: cartesian line: x**3 for x over (-5.0, 5.0) Plot[1]:Plot object containing: [0]: cartesian line: x**2 for x over (-6.0, 6.0) [1]: cartesian line: x for x over (-5.0, 5.0) Plotting horizontally in a single line: .. plot:: :context: close-figs :format: doctest :include-source: True >>> PlotGrid(1, 3, p2, p3, p4) PlotGrid object containing: Plot[0]:Plot object containing: [0]: cartesian line: x**2 for x over (-6.0, 6.0) [1]: cartesian line: x for x over (-5.0, 5.0) Plot[1]:Plot object containing: [0]: cartesian line: x**3 for x over (-5.0, 5.0) Plot[2]:Plot object containing: [0]: cartesian surface: x*y for x over (-5.0, 5.0) and y over (-5.0, 5.0) Plotting in a grid form: .. plot:: :context: close-figs :format: doctest :include-source: True >>> PlotGrid(2, 2, p1, p2, p3, p4) PlotGrid object containing: Plot[0]:Plot object containing: [0]: cartesian line: x for x over (-5.0, 5.0) [1]: cartesian line: x**2 for x over (-5.0, 5.0) [2]: cartesian line: x**3 for x over (-5.0, 5.0) Plot[1]:Plot object containing: [0]: cartesian line: x**2 for x over (-6.0, 6.0) [1]: cartesian line: x for x over (-5.0, 5.0) Plot[2]:Plot object containing: [0]: cartesian line: x**3 for x over (-5.0, 5.0) Plot[3]:Plot object containing: [0]: cartesian surface: x*y for x over (-5.0, 5.0) and y over (-5.0, 5.0) """ def __init__(self, nrows, ncolumns, *args, show=True, size=None, **kwargs): """ Parameters ========== nrows : The number of rows that should be in the grid of the required subplot. ncolumns : The number of columns that should be in the grid of the required subplot. nrows and ncolumns together define the required grid. Arguments ========= A list of predefined plot objects entered in a row-wise sequence i.e. plot objects which are to be in the top row of the required grid are written first, then the second row objects and so on Keyword arguments ================= show : Boolean The default value is set to ``True``. Set show to ``False`` and the function will not display the subplot. The returned instance of the ``PlotGrid`` class can then be used to save or display the plot by calling the ``save()`` and ``show()`` methods respectively. size : (float, float), optional A tuple in the form (width, height) in inches to specify the size of the overall figure. The default value is set to ``None``, meaning the size will be set by the default backend. """ self.nrows = nrows self.ncolumns = ncolumns self._series = [] self.args = args for arg in args: self._series.append(arg._series) self.backend = DefaultBackend self.size = size if show: self.show() def show(self): if hasattr(self, '_backend'): self._backend.close() self._backend = self.backend(self) self._backend.show() def save(self, path): if hasattr(self, '_backend'): self._backend.close() self._backend = self.backend(self) self._backend.save(path) def __str__(self): plot_strs = [('Plot[%d]:' % i) + str(plot) for i, plot in enumerate(self.args)] return 'PlotGrid object containing:\n' + '\n'.join(plot_strs) ############################################################################## # Data Series ############################################################################## #TODO more general way to calculate aesthetics (see get_color_array) ### The base class for all series class BaseSeries: """Base class for the data objects containing stuff to be plotted. Explanation =========== The backend should check if it supports the data series that it's given. (eg TextBackend supports only LineOver1DRange). It's the backend responsibility to know how to use the class of data series that it's given. Some data series classes are grouped (using a class attribute like is_2Dline) according to the api they present (based only on convention). The backend is not obliged to use that api (eg. The LineOver1DRange belongs to the is_2Dline group and presents the get_points method, but the TextBackend does not use the get_points method). """ # Some flags follow. The rationale for using flags instead of checking base # classes is that setting multiple flags is simpler than multiple # inheritance. is_2Dline = False # Some of the backends expect: # - get_points returning 1D np.arrays list_x, list_y # - get_color_array returning 1D np.array (done in Line2DBaseSeries) # with the colors calculated at the points from get_points is_3Dline = False # Some of the backends expect: # - get_points returning 1D np.arrays list_x, list_y, list_y # - get_color_array returning 1D np.array (done in Line2DBaseSeries) # with the colors calculated at the points from get_points is_3Dsurface = False # Some of the backends expect: # - get_meshes returning mesh_x, mesh_y, mesh_z (2D np.arrays) # - get_points an alias for get_meshes is_contour = False # Some of the backends expect: # - get_meshes returning mesh_x, mesh_y, mesh_z (2D np.arrays) # - get_points an alias for get_meshes is_implicit = False # Some of the backends expect: # - get_meshes returning mesh_x (1D array), mesh_y(1D array, # mesh_z (2D np.arrays) # - get_points an alias for get_meshes # Different from is_contour as the colormap in backend will be # different is_parametric = False # The calculation of aesthetics expects: # - get_parameter_points returning one or two np.arrays (1D or 2D) # used for calculation aesthetics def __init__(self): super().__init__() @property def is_3D(self): flags3D = [ self.is_3Dline, self.is_3Dsurface ] return any(flags3D) @property def is_line(self): flagslines = [ self.is_2Dline, self.is_3Dline ] return any(flagslines) ### 2D lines class Line2DBaseSeries(BaseSeries): """A base class for 2D lines. - adding the label, steps and only_integers options - making is_2Dline true - defining get_segments and get_color_array """ is_2Dline = True _dim = 2 def __init__(self): super().__init__() self.label = None self.steps = False self.only_integers = False self.line_color = None def get_data(self): """ Return lists of coordinates for plotting the line. Returns ======= x: list List of x-coordinates y: list List of y-coordinates y: list List of z-coordinates in case of Parametric3DLineSeries """ np = import_module('numpy') points = self.get_points() if self.steps is True: if len(points) == 2: x = np.array((points[0], points[0])).T.flatten()[1:] y = np.array((points[1], points[1])).T.flatten()[:-1] points = (x, y) else: x = np.repeat(points[0], 3)[2:] y = np.repeat(points[1], 3)[:-2] z = np.repeat(points[2], 3)[1:-1] points = (x, y, z) return points def get_segments(self): sympy_deprecation_warning( """ The Line2DBaseSeries.get_segments() method is deprecated. Instead, use the MatplotlibBackend.get_segments() method, or use The get_points() or get_data() methods. """, deprecated_since_version="1.9", active_deprecations_target="deprecated-get-segments") np = import_module('numpy') points = type(self).get_data(self) points = np.ma.array(points).T.reshape(-1, 1, self._dim) return np.ma.concatenate([points[:-1], points[1:]], axis=1) def get_color_array(self): np = import_module('numpy') c = self.line_color if hasattr(c, '__call__'): f = np.vectorize(c) nargs = arity(c) if nargs == 1 and self.is_parametric: x = self.get_parameter_points() return f(centers_of_segments(x)) else: variables = list(map(centers_of_segments, self.get_points())) if nargs == 1: return f(variables[0]) elif nargs == 2: return f(*variables[:2]) else: # only if the line is 3D (otherwise raises an error) return f(*variables) else: return c*np.ones(self.nb_of_points) class List2DSeries(Line2DBaseSeries): """Representation for a line consisting of list of points.""" def __init__(self, list_x, list_y): np = import_module('numpy') super().__init__() self.list_x = np.array(list_x) self.list_y = np.array(list_y) self.label = 'list' def __str__(self): return 'list plot' def get_points(self): return (self.list_x, self.list_y) class LineOver1DRangeSeries(Line2DBaseSeries): """Representation for a line consisting of a SymPy expression over a range.""" def __init__(self, expr, var_start_end, **kwargs): super().__init__() self.expr = sympify(expr) self.label = kwargs.get('label', None) or self.expr self.var = sympify(var_start_end[0]) self.start = float(var_start_end[1]) self.end = float(var_start_end[2]) self.nb_of_points = kwargs.get('nb_of_points', 300) self.adaptive = kwargs.get('adaptive', True) self.depth = kwargs.get('depth', 12) self.line_color = kwargs.get('line_color', None) self.xscale = kwargs.get('xscale', 'linear') def __str__(self): return 'cartesian line: %s for %s over %s' % ( str(self.expr), str(self.var), str((self.start, self.end))) def get_points(self): """ Return lists of coordinates for plotting. Depending on the `adaptive` option, this function will either use an adaptive algorithm or it will uniformly sample the expression over the provided range. Returns ======= x: list List of x-coordinates y: list List of y-coordinates Explanation =========== The adaptive sampling is done by recursively checking if three points are almost collinear. If they are not collinear, then more points are added between those points. References ========== .. [1] Adaptive polygonal approximation of parametric curves, Luiz Henrique de Figueiredo. """ if self.only_integers or not self.adaptive: return self._uniform_sampling() else: f = lambdify([self.var], self.expr) x_coords = [] y_coords = [] np = import_module('numpy') def sample(p, q, depth): """ Samples recursively if three points are almost collinear. For depth < 6, points are added irrespective of whether they satisfy the collinearity condition or not. The maximum depth allowed is 12. """ # Randomly sample to avoid aliasing. random = 0.45 + np.random.rand() * 0.1 if self.xscale == 'log': xnew = 10**(np.log10(p[0]) + random * (np.log10(q[0]) - np.log10(p[0]))) else: xnew = p[0] + random * (q[0] - p[0]) ynew = f(xnew) new_point = np.array([xnew, ynew]) # Maximum depth if depth > self.depth: x_coords.append(q[0]) y_coords.append(q[1]) # Sample irrespective of whether the line is flat till the # depth of 6. We are not using linspace to avoid aliasing. elif depth < 6: sample(p, new_point, depth + 1) sample(new_point, q, depth + 1) # Sample ten points if complex values are encountered # at both ends. If there is a real value in between, then # sample those points further. elif p[1] is None and q[1] is None: if self.xscale == 'log': xarray = np.logspace(p[0], q[0], 10) else: xarray = np.linspace(p[0], q[0], 10) yarray = list(map(f, xarray)) if not all(y is None for y in yarray): for i in range(len(yarray) - 1): if not (yarray[i] is None and yarray[i + 1] is None): sample([xarray[i], yarray[i]], [xarray[i + 1], yarray[i + 1]], depth + 1) # Sample further if one of the end points in None (i.e. a # complex value) or the three points are not almost collinear. elif (p[1] is None or q[1] is None or new_point[1] is None or not flat(p, new_point, q)): sample(p, new_point, depth + 1) sample(new_point, q, depth + 1) else: x_coords.append(q[0]) y_coords.append(q[1]) f_start = f(self.start) f_end = f(self.end) x_coords.append(self.start) y_coords.append(f_start) sample(np.array([self.start, f_start]), np.array([self.end, f_end]), 0) return (x_coords, y_coords) def _uniform_sampling(self): np = import_module('numpy') if self.only_integers is True: if self.xscale == 'log': list_x = np.logspace(int(self.start), int(self.end), num=int(self.end) - int(self.start) + 1) else: list_x = np.linspace(int(self.start), int(self.end), num=int(self.end) - int(self.start) + 1) else: if self.xscale == 'log': list_x = np.logspace(self.start, self.end, num=self.nb_of_points) else: list_x = np.linspace(self.start, self.end, num=self.nb_of_points) f = vectorized_lambdify([self.var], self.expr) list_y = f(list_x) return (list_x, list_y) class Parametric2DLineSeries(Line2DBaseSeries): """Representation for a line consisting of two parametric SymPy expressions over a range.""" is_parametric = True def __init__(self, expr_x, expr_y, var_start_end, **kwargs): super().__init__() self.expr_x = sympify(expr_x) self.expr_y = sympify(expr_y) self.label = kwargs.get('label', None) or \ Tuple(self.expr_x, self.expr_y) self.var = sympify(var_start_end[0]) self.start = float(var_start_end[1]) self.end = float(var_start_end[2]) self.nb_of_points = kwargs.get('nb_of_points', 300) self.adaptive = kwargs.get('adaptive', True) self.depth = kwargs.get('depth', 12) self.line_color = kwargs.get('line_color', None) def __str__(self): return 'parametric cartesian line: (%s, %s) for %s over %s' % ( str(self.expr_x), str(self.expr_y), str(self.var), str((self.start, self.end))) def get_parameter_points(self): np = import_module('numpy') return np.linspace(self.start, self.end, num=self.nb_of_points) def _uniform_sampling(self): param = self.get_parameter_points() fx = vectorized_lambdify([self.var], self.expr_x) fy = vectorized_lambdify([self.var], self.expr_y) list_x = fx(param) list_y = fy(param) return (list_x, list_y) def get_points(self): """ Return lists of coordinates for plotting. Depending on the `adaptive` option, this function will either use an adaptive algorithm or it will uniformly sample the expression over the provided range. Returns ======= x: list List of x-coordinates y: list List of y-coordinates Explanation =========== The adaptive sampling is done by recursively checking if three points are almost collinear. If they are not collinear, then more points are added between those points. References ========== .. [1] Adaptive polygonal approximation of parametric curves, Luiz Henrique de Figueiredo. """ if not self.adaptive: return self._uniform_sampling() f_x = lambdify([self.var], self.expr_x) f_y = lambdify([self.var], self.expr_y) x_coords = [] y_coords = [] def sample(param_p, param_q, p, q, depth): """ Samples recursively if three points are almost collinear. For depth < 6, points are added irrespective of whether they satisfy the collinearity condition or not. The maximum depth allowed is 12. """ # Randomly sample to avoid aliasing. np = import_module('numpy') random = 0.45 + np.random.rand() * 0.1 param_new = param_p + random * (param_q - param_p) xnew = f_x(param_new) ynew = f_y(param_new) new_point = np.array([xnew, ynew]) # Maximum depth if depth > self.depth: x_coords.append(q[0]) y_coords.append(q[1]) # Sample irrespective of whether the line is flat till the # depth of 6. We are not using linspace to avoid aliasing. elif depth < 6: sample(param_p, param_new, p, new_point, depth + 1) sample(param_new, param_q, new_point, q, depth + 1) # Sample ten points if complex values are encountered # at both ends. If there is a real value in between, then # sample those points further. elif ((p[0] is None and q[1] is None) or (p[1] is None and q[1] is None)): param_array = np.linspace(param_p, param_q, 10) x_array = list(map(f_x, param_array)) y_array = list(map(f_y, param_array)) if not all(x is None and y is None for x, y in zip(x_array, y_array)): for i in range(len(y_array) - 1): if ((x_array[i] is not None and y_array[i] is not None) or (x_array[i + 1] is not None and y_array[i + 1] is not None)): point_a = [x_array[i], y_array[i]] point_b = [x_array[i + 1], y_array[i + 1]] sample(param_array[i], param_array[i], point_a, point_b, depth + 1) # Sample further if one of the end points in None (i.e. a complex # value) or the three points are not almost collinear. elif (p[0] is None or p[1] is None or q[1] is None or q[0] is None or not flat(p, new_point, q)): sample(param_p, param_new, p, new_point, depth + 1) sample(param_new, param_q, new_point, q, depth + 1) else: x_coords.append(q[0]) y_coords.append(q[1]) f_start_x = f_x(self.start) f_start_y = f_y(self.start) start = [f_start_x, f_start_y] f_end_x = f_x(self.end) f_end_y = f_y(self.end) end = [f_end_x, f_end_y] x_coords.append(f_start_x) y_coords.append(f_start_y) sample(self.start, self.end, start, end, 0) return x_coords, y_coords ### 3D lines class Line3DBaseSeries(Line2DBaseSeries): """A base class for 3D lines. Most of the stuff is derived from Line2DBaseSeries.""" is_2Dline = False is_3Dline = True _dim = 3 def __init__(self): super().__init__() class Parametric3DLineSeries(Line3DBaseSeries): """Representation for a 3D line consisting of three parametric SymPy expressions and a range.""" is_parametric = True def __init__(self, expr_x, expr_y, expr_z, var_start_end, **kwargs): super().__init__() self.expr_x = sympify(expr_x) self.expr_y = sympify(expr_y) self.expr_z = sympify(expr_z) self.label = kwargs.get('label', None) or \ Tuple(self.expr_x, self.expr_y) self.var = sympify(var_start_end[0]) self.start = float(var_start_end[1]) self.end = float(var_start_end[2]) self.nb_of_points = kwargs.get('nb_of_points', 300) self.line_color = kwargs.get('line_color', None) self._xlim = None self._ylim = None self._zlim = None def __str__(self): return '3D parametric cartesian line: (%s, %s, %s) for %s over %s' % ( str(self.expr_x), str(self.expr_y), str(self.expr_z), str(self.var), str((self.start, self.end))) def get_parameter_points(self): np = import_module('numpy') return np.linspace(self.start, self.end, num=self.nb_of_points) def get_points(self): np = import_module('numpy') param = self.get_parameter_points() fx = vectorized_lambdify([self.var], self.expr_x) fy = vectorized_lambdify([self.var], self.expr_y) fz = vectorized_lambdify([self.var], self.expr_z) list_x = fx(param) list_y = fy(param) list_z = fz(param) list_x = np.array(list_x, dtype=np.float64) list_y = np.array(list_y, dtype=np.float64) list_z = np.array(list_z, dtype=np.float64) list_x = np.ma.masked_invalid(list_x) list_y = np.ma.masked_invalid(list_y) list_z = np.ma.masked_invalid(list_z) self._xlim = (np.amin(list_x), np.amax(list_x)) self._ylim = (np.amin(list_y), np.amax(list_y)) self._zlim = (np.amin(list_z), np.amax(list_z)) return list_x, list_y, list_z ### Surfaces class SurfaceBaseSeries(BaseSeries): """A base class for 3D surfaces.""" is_3Dsurface = True def __init__(self): super().__init__() self.surface_color = None def get_color_array(self): np = import_module('numpy') c = self.surface_color if isinstance(c, Callable): f = np.vectorize(c) nargs = arity(c) if self.is_parametric: variables = list(map(centers_of_faces, self.get_parameter_meshes())) if nargs == 1: return f(variables[0]) elif nargs == 2: return f(*variables) variables = list(map(centers_of_faces, self.get_meshes())) if nargs == 1: return f(variables[0]) elif nargs == 2: return f(*variables[:2]) else: return f(*variables) else: if isinstance(self, SurfaceOver2DRangeSeries): return c*np.ones(min(self.nb_of_points_x, self.nb_of_points_y)) else: return c*np.ones(min(self.nb_of_points_u, self.nb_of_points_v)) class SurfaceOver2DRangeSeries(SurfaceBaseSeries): """Representation for a 3D surface consisting of a SymPy expression and 2D range.""" def __init__(self, expr, var_start_end_x, var_start_end_y, **kwargs): super().__init__() self.expr = sympify(expr) self.var_x = sympify(var_start_end_x[0]) self.start_x = float(var_start_end_x[1]) self.end_x = float(var_start_end_x[2]) self.var_y = sympify(var_start_end_y[0]) self.start_y = float(var_start_end_y[1]) self.end_y = float(var_start_end_y[2]) self.nb_of_points_x = kwargs.get('nb_of_points_x', 50) self.nb_of_points_y = kwargs.get('nb_of_points_y', 50) self.surface_color = kwargs.get('surface_color', None) self._xlim = (self.start_x, self.end_x) self._ylim = (self.start_y, self.end_y) def __str__(self): return ('cartesian surface: %s for' ' %s over %s and %s over %s') % ( str(self.expr), str(self.var_x), str((self.start_x, self.end_x)), str(self.var_y), str((self.start_y, self.end_y))) def get_meshes(self): np = import_module('numpy') mesh_x, mesh_y = np.meshgrid(np.linspace(self.start_x, self.end_x, num=self.nb_of_points_x), np.linspace(self.start_y, self.end_y, num=self.nb_of_points_y)) f = vectorized_lambdify((self.var_x, self.var_y), self.expr) mesh_z = f(mesh_x, mesh_y) mesh_z = np.array(mesh_z, dtype=np.float64) mesh_z = np.ma.masked_invalid(mesh_z) self._zlim = (np.amin(mesh_z), np.amax(mesh_z)) return mesh_x, mesh_y, mesh_z class ParametricSurfaceSeries(SurfaceBaseSeries): """Representation for a 3D surface consisting of three parametric SymPy expressions and a range.""" is_parametric = True def __init__( self, expr_x, expr_y, expr_z, var_start_end_u, var_start_end_v, **kwargs): super().__init__() self.expr_x = sympify(expr_x) self.expr_y = sympify(expr_y) self.expr_z = sympify(expr_z) self.var_u = sympify(var_start_end_u[0]) self.start_u = float(var_start_end_u[1]) self.end_u = float(var_start_end_u[2]) self.var_v = sympify(var_start_end_v[0]) self.start_v = float(var_start_end_v[1]) self.end_v = float(var_start_end_v[2]) self.nb_of_points_u = kwargs.get('nb_of_points_u', 50) self.nb_of_points_v = kwargs.get('nb_of_points_v', 50) self.surface_color = kwargs.get('surface_color', None) def __str__(self): return ('parametric cartesian surface: (%s, %s, %s) for' ' %s over %s and %s over %s') % ( str(self.expr_x), str(self.expr_y), str(self.expr_z), str(self.var_u), str((self.start_u, self.end_u)), str(self.var_v), str((self.start_v, self.end_v))) def get_parameter_meshes(self): np = import_module('numpy') return np.meshgrid(np.linspace(self.start_u, self.end_u, num=self.nb_of_points_u), np.linspace(self.start_v, self.end_v, num=self.nb_of_points_v)) def get_meshes(self): np = import_module('numpy') mesh_u, mesh_v = self.get_parameter_meshes() fx = vectorized_lambdify((self.var_u, self.var_v), self.expr_x) fy = vectorized_lambdify((self.var_u, self.var_v), self.expr_y) fz = vectorized_lambdify((self.var_u, self.var_v), self.expr_z) mesh_x = fx(mesh_u, mesh_v) mesh_y = fy(mesh_u, mesh_v) mesh_z = fz(mesh_u, mesh_v) mesh_x = np.array(mesh_x, dtype=np.float64) mesh_y = np.array(mesh_y, dtype=np.float64) mesh_z = np.array(mesh_z, dtype=np.float64) mesh_x = np.ma.masked_invalid(mesh_x) mesh_y = np.ma.masked_invalid(mesh_y) mesh_z = np.ma.masked_invalid(mesh_z) self._xlim = (np.amin(mesh_x), np.amax(mesh_x)) self._ylim = (np.amin(mesh_y), np.amax(mesh_y)) self._zlim = (np.amin(mesh_z), np.amax(mesh_z)) return mesh_x, mesh_y, mesh_z ### Contours class ContourSeries(BaseSeries): """Representation for a contour plot.""" # The code is mostly repetition of SurfaceOver2DRange. # Presently used in contour_plot function is_contour = True def __init__(self, expr, var_start_end_x, var_start_end_y): super().__init__() self.nb_of_points_x = 50 self.nb_of_points_y = 50 self.expr = sympify(expr) self.var_x = sympify(var_start_end_x[0]) self.start_x = float(var_start_end_x[1]) self.end_x = float(var_start_end_x[2]) self.var_y = sympify(var_start_end_y[0]) self.start_y = float(var_start_end_y[1]) self.end_y = float(var_start_end_y[2]) self.get_points = self.get_meshes self._xlim = (self.start_x, self.end_x) self._ylim = (self.start_y, self.end_y) def __str__(self): return ('contour: %s for ' '%s over %s and %s over %s') % ( str(self.expr), str(self.var_x), str((self.start_x, self.end_x)), str(self.var_y), str((self.start_y, self.end_y))) def get_meshes(self): np = import_module('numpy') mesh_x, mesh_y = np.meshgrid(np.linspace(self.start_x, self.end_x, num=self.nb_of_points_x), np.linspace(self.start_y, self.end_y, num=self.nb_of_points_y)) f = vectorized_lambdify((self.var_x, self.var_y), self.expr) return (mesh_x, mesh_y, f(mesh_x, mesh_y)) ############################################################################## # Backends ############################################################################## class BaseBackend: """Base class for all backends. A backend represents the plotting library, which implements the necessary functionalities in order to use SymPy plotting functions. How the plotting module works: 1. Whenever a plotting function is called, the provided expressions are processed and a list of instances of the `BaseSeries` class is created, containing the necessary information to plot the expressions (eg the expression, ranges, series name, ...). Eventually, these objects will generate the numerical data to be plotted. 2. A Plot object is instantiated, which stores the list of series and the main attributes of the plot (eg axis labels, title, ...). 3. When the "show" command is executed, a new backend is instantiated, which loops through each series object to generate and plot the numerical data. The backend is also going to set the axis labels, title, ..., according to the values stored in the Plot instance. The backend should check if it supports the data series that it's given (eg TextBackend supports only LineOver1DRange). It's the backend responsibility to know how to use the class of data series that it's given. Note that the current implementation of the `*Series` classes is "matplotlib-centric": the numerical data returned by the `get_points` and `get_meshes` methods is meant to be used directly by Matplotlib. Therefore, the new backend will have to pre-process the numerical data to make it compatible with the chosen plotting library. Keep in mind that future SymPy versions may improve the `*Series` classes in order to return numerical data "non-matplotlib-centric", hence if you code a new backend you have the responsibility to check if its working on each SymPy release. Please, explore the `MatplotlibBackend` source code to understand how a backend should be coded. Methods ======= In order to be used by SymPy plotting functions, a backend must implement the following methods: * `show(self)`: used to loop over the data series, generate the numerical data, plot it and set the axis labels, title, ... * save(self, path): used to save the current plot to the specified file path. * close(self): used to close the current plot backend (note: some plotting library does not support this functionality. In that case, just raise a warning). See also ======== MatplotlibBackend """ def __init__(self, parent): super().__init__() self.parent = parent def show(self): raise NotImplementedError def save(self, path): raise NotImplementedError def close(self): raise NotImplementedError # Don't have to check for the success of importing matplotlib in each case; # we will only be using this backend if we can successfully import matploblib class MatplotlibBackend(BaseBackend): """ This class implements the functionalities to use Matplotlib with SymPy plotting functions. """ def __init__(self, parent): super().__init__(parent) self.matplotlib = import_module('matplotlib', import_kwargs={'fromlist': ['pyplot', 'cm', 'collections']}, min_module_version='1.1.0', catch=(RuntimeError,)) self.plt = self.matplotlib.pyplot self.cm = self.matplotlib.cm self.LineCollection = self.matplotlib.collections.LineCollection aspect = getattr(self.parent, 'aspect_ratio', 'auto') if aspect != 'auto': aspect = float(aspect[1]) / aspect[0] if isinstance(self.parent, Plot): nrows, ncolumns = 1, 1 series_list = [self.parent._series] elif isinstance(self.parent, PlotGrid): nrows, ncolumns = self.parent.nrows, self.parent.ncolumns series_list = self.parent._series self.ax = [] self.fig = self.plt.figure(figsize=parent.size) for i, series in enumerate(series_list): are_3D = [s.is_3D for s in series] if any(are_3D) and not all(are_3D): raise ValueError('The matplotlib backend cannot mix 2D and 3D.') elif all(are_3D): # mpl_toolkits.mplot3d is necessary for # projection='3d' mpl_toolkits = import_module('mpl_toolkits', # noqa import_kwargs={'fromlist': ['mplot3d']}) self.ax.append(self.fig.add_subplot(nrows, ncolumns, i + 1, projection='3d', aspect=aspect)) elif not any(are_3D): self.ax.append(self.fig.add_subplot(nrows, ncolumns, i + 1, aspect=aspect)) self.ax[i].spines['left'].set_position('zero') self.ax[i].spines['right'].set_color('none') self.ax[i].spines['bottom'].set_position('zero') self.ax[i].spines['top'].set_color('none') self.ax[i].xaxis.set_ticks_position('bottom') self.ax[i].yaxis.set_ticks_position('left') @staticmethod def get_segments(x, y, z=None): """ Convert two list of coordinates to a list of segments to be used with Matplotlib's LineCollection. Parameters ========== x: list List of x-coordinates y: list List of y-coordinates z: list List of z-coordinates for a 3D line. """ np = import_module('numpy') if z is not None: dim = 3 points = (x, y, z) else: dim = 2 points = (x, y) points = np.ma.array(points).T.reshape(-1, 1, dim) return np.ma.concatenate([points[:-1], points[1:]], axis=1) def _process_series(self, series, ax, parent): np = import_module('numpy') mpl_toolkits = import_module( 'mpl_toolkits', import_kwargs={'fromlist': ['mplot3d']}) # XXX Workaround for matplotlib issue # https://github.com/matplotlib/matplotlib/issues/17130 xlims, ylims, zlims = [], [], [] for s in series: # Create the collections if s.is_2Dline: x, y = s.get_data() if (isinstance(s.line_color, (int, float)) or callable(s.line_color)): segments = self.get_segments(x, y) collection = self.LineCollection(segments) collection.set_array(s.get_color_array()) ax.add_collection(collection) else: lbl = _str_or_latex(s.label) line, = ax.plot(x, y, label=lbl, color=s.line_color) elif s.is_contour: ax.contour(*s.get_meshes()) elif s.is_3Dline: x, y, z = s.get_data() if (isinstance(s.line_color, (int, float)) or callable(s.line_color)): art3d = mpl_toolkits.mplot3d.art3d segments = self.get_segments(x, y, z) collection = art3d.Line3DCollection(segments) collection.set_array(s.get_color_array()) ax.add_collection(collection) else: lbl = _str_or_latex(s.label) ax.plot(x, y, z, label=lbl, color=s.line_color) xlims.append(s._xlim) ylims.append(s._ylim) zlims.append(s._zlim) elif s.is_3Dsurface: x, y, z = s.get_meshes() collection = ax.plot_surface(x, y, z, cmap=getattr(self.cm, 'viridis', self.cm.jet), rstride=1, cstride=1, linewidth=0.1) if isinstance(s.surface_color, (float, int, Callable)): color_array = s.get_color_array() color_array = color_array.reshape(color_array.size) collection.set_array(color_array) else: collection.set_color(s.surface_color) xlims.append(s._xlim) ylims.append(s._ylim) zlims.append(s._zlim) elif s.is_implicit: points = s.get_raster() if len(points) == 2: # interval math plotting x, y = _matplotlib_list(points[0]) ax.fill(x, y, facecolor=s.line_color, edgecolor='None') else: # use contourf or contour depending on whether it is # an inequality or equality. # XXX: ``contour`` plots multiple lines. Should be fixed. ListedColormap = self.matplotlib.colors.ListedColormap colormap = ListedColormap(["white", s.line_color]) xarray, yarray, zarray, plot_type = points if plot_type == 'contour': ax.contour(xarray, yarray, zarray, cmap=colormap, label=_str_or_latex(s.label)) else: ax.contourf(xarray, yarray, zarray, cmap=colormap, label=_str_or_latex(s.label)) else: raise NotImplementedError( '{} is not supported in the SymPy plotting module ' 'with matplotlib backend. Please report this issue.' .format(ax)) Axes3D = mpl_toolkits.mplot3d.Axes3D if not isinstance(ax, Axes3D): ax.autoscale_view( scalex=ax.get_autoscalex_on(), scaley=ax.get_autoscaley_on()) else: # XXX Workaround for matplotlib issue # https://github.com/matplotlib/matplotlib/issues/17130 if xlims: xlims = np.array(xlims) xlim = (np.amin(xlims[:, 0]), np.amax(xlims[:, 1])) ax.set_xlim(xlim) else: ax.set_xlim([0, 1]) if ylims: ylims = np.array(ylims) ylim = (np.amin(ylims[:, 0]), np.amax(ylims[:, 1])) ax.set_ylim(ylim) else: ax.set_ylim([0, 1]) if zlims: zlims = np.array(zlims) zlim = (np.amin(zlims[:, 0]), np.amax(zlims[:, 1])) ax.set_zlim(zlim) else: ax.set_zlim([0, 1]) # Set global options. # TODO The 3D stuff # XXX The order of those is important. if parent.xscale and not isinstance(ax, Axes3D): ax.set_xscale(parent.xscale) if parent.yscale and not isinstance(ax, Axes3D): ax.set_yscale(parent.yscale) if not isinstance(ax, Axes3D) or self.matplotlib.__version__ >= '1.2.0': # XXX in the distant future remove this check ax.set_autoscale_on(parent.autoscale) if parent.axis_center: val = parent.axis_center if isinstance(ax, Axes3D): pass elif val == 'center': ax.spines['left'].set_position('center') ax.spines['bottom'].set_position('center') elif val == 'auto': xl, xh = ax.get_xlim() yl, yh = ax.get_ylim() pos_left = ('data', 0) if xl*xh <= 0 else 'center' pos_bottom = ('data', 0) if yl*yh <= 0 else 'center' ax.spines['left'].set_position(pos_left) ax.spines['bottom'].set_position(pos_bottom) else: ax.spines['left'].set_position(('data', val[0])) ax.spines['bottom'].set_position(('data', val[1])) if not parent.axis: ax.set_axis_off() if parent.legend: if ax.legend(): ax.legend_.set_visible(parent.legend) if parent.margin: ax.set_xmargin(parent.margin) ax.set_ymargin(parent.margin) if parent.title: ax.set_title(parent.title) if parent.xlabel: xlbl = _str_or_latex(parent.xlabel) ax.set_xlabel(xlbl, position=(1, 0)) if parent.ylabel: ylbl = _str_or_latex(parent.ylabel) ax.set_ylabel(ylbl, position=(0, 1)) if isinstance(ax, Axes3D) and parent.zlabel: zlbl = _str_or_latex(parent.zlabel) ax.set_zlabel(zlbl, position=(0, 1)) if parent.annotations: for a in parent.annotations: ax.annotate(**a) if parent.markers: for marker in parent.markers: # make a copy of the marker dictionary # so that it doesn't get altered m = marker.copy() args = m.pop('args') ax.plot(*args, **m) if parent.rectangles: for r in parent.rectangles: rect = self.matplotlib.patches.Rectangle(**r) ax.add_patch(rect) if parent.fill: ax.fill_between(**parent.fill) # xlim and ylim shoulld always be set at last so that plot limits # doesn't get altered during the process. if parent.xlim: ax.set_xlim(parent.xlim) if parent.ylim: ax.set_ylim(parent.ylim) def process_series(self): """ Iterates over every ``Plot`` object and further calls _process_series() """ parent = self.parent if isinstance(parent, Plot): series_list = [parent._series] else: series_list = parent._series for i, (series, ax) in enumerate(zip(series_list, self.ax)): if isinstance(self.parent, PlotGrid): parent = self.parent.args[i] self._process_series(series, ax, parent) def show(self): self.process_series() #TODO after fixing https://github.com/ipython/ipython/issues/1255 # you can uncomment the next line and remove the pyplot.show() call #self.fig.show() if _show: self.fig.tight_layout() self.plt.show() else: self.close() def save(self, path): self.process_series() self.fig.savefig(path) def close(self): self.plt.close(self.fig) class TextBackend(BaseBackend): def __init__(self, parent): super().__init__(parent) def show(self): if not _show: return if len(self.parent._series) != 1: raise ValueError( 'The TextBackend supports only one graph per Plot.') elif not isinstance(self.parent._series[0], LineOver1DRangeSeries): raise ValueError( 'The TextBackend supports only expressions over a 1D range') else: ser = self.parent._series[0] textplot(ser.expr, ser.start, ser.end) def close(self): pass class DefaultBackend(BaseBackend): def __new__(cls, parent): matplotlib = import_module('matplotlib', min_module_version='1.1.0', catch=(RuntimeError,)) if matplotlib: return MatplotlibBackend(parent) else: return TextBackend(parent) plot_backends = { 'matplotlib': MatplotlibBackend, 'text': TextBackend, 'default': DefaultBackend } ############################################################################## # Finding the centers of line segments or mesh faces ############################################################################## def centers_of_segments(array): np = import_module('numpy') return np.mean(np.vstack((array[:-1], array[1:])), 0) def centers_of_faces(array): np = import_module('numpy') return np.mean(np.dstack((array[:-1, :-1], array[1:, :-1], array[:-1, 1:], array[:-1, :-1], )), 2) def flat(x, y, z, eps=1e-3): """Checks whether three points are almost collinear""" np = import_module('numpy') # Workaround plotting piecewise (#8577): # workaround for `lambdify` in `.experimental_lambdify` fails # to return numerical values in some cases. Lower-level fix # in `lambdify` is possible. vector_a = (x - y).astype(np.float64) vector_b = (z - y).astype(np.float64) dot_product = np.dot(vector_a, vector_b) vector_a_norm = np.linalg.norm(vector_a) vector_b_norm = np.linalg.norm(vector_b) cos_theta = dot_product / (vector_a_norm * vector_b_norm) return abs(cos_theta + 1) < eps def _matplotlib_list(interval_list): """ Returns lists for matplotlib ``fill`` command from a list of bounding rectangular intervals """ xlist = [] ylist = [] if len(interval_list): for intervals in interval_list: intervalx = intervals[0] intervaly = intervals[1] xlist.extend([intervalx.start, intervalx.start, intervalx.end, intervalx.end, None]) ylist.extend([intervaly.start, intervaly.end, intervaly.end, intervaly.start, None]) else: #XXX Ugly hack. Matplotlib does not accept empty lists for ``fill`` xlist.extend((None, None, None, None)) ylist.extend((None, None, None, None)) return xlist, ylist ####New API for plotting module #### # TODO: Add color arrays for plots. # TODO: Add more plotting options for 3d plots. # TODO: Adaptive sampling for 3D plots. def plot(*args, show=True, **kwargs): """Plots a function of a single variable as a curve. Parameters ========== args : The first argument is the expression representing the function of single variable to be plotted. The last argument is a 3-tuple denoting the range of the free variable. e.g. ``(x, 0, 5)`` Typical usage examples are in the followings: - Plotting a single expression with a single range. ``plot(expr, range, **kwargs)`` - Plotting a single expression with the default range (-10, 10). ``plot(expr, **kwargs)`` - Plotting multiple expressions with a single range. ``plot(expr1, expr2, ..., range, **kwargs)`` - Plotting multiple expressions with multiple ranges. ``plot((expr1, range1), (expr2, range2), ..., **kwargs)`` It is best practice to specify range explicitly because default range may change in the future if a more advanced default range detection algorithm is implemented. show : bool, optional The default value is set to ``True``. Set show to ``False`` and the function will not display the plot. The returned instance of the ``Plot`` class can then be used to save or display the plot by calling the ``save()`` and ``show()`` methods respectively. line_color : string, or float, or function, optional Specifies the color for the plot. See ``Plot`` to see how to set color for the plots. Note that by setting ``line_color``, it would be applied simultaneously to all the series. title : str, optional Title of the plot. It is set to the latex representation of the expression, if the plot has only one expression. label : str, optional The label of the expression in the plot. It will be used when called with ``legend``. Default is the name of the expression. e.g. ``sin(x)`` xlabel : str or expression, optional Label for the x-axis. ylabel : str or expression, optional Label for the y-axis. xscale : 'linear' or 'log', optional Sets the scaling of the x-axis. yscale : 'linear' or 'log', optional Sets the scaling of the y-axis. axis_center : (float, float), optional Tuple of two floats denoting the coordinates of the center or {'center', 'auto'} xlim : (float, float), optional Denotes the x-axis limits, ``(min, max)```. ylim : (float, float), optional Denotes the y-axis limits, ``(min, max)```. annotations : list, optional A list of dictionaries specifying the type of annotation required. The keys in the dictionary should be equivalent to the arguments of the matplotlib's annotate() function. markers : list, optional A list of dictionaries specifying the type the markers required. The keys in the dictionary should be equivalent to the arguments of the matplotlib's plot() function along with the marker related keyworded arguments. rectangles : list, optional A list of dictionaries specifying the dimensions of the rectangles to be plotted. The keys in the dictionary should be equivalent to the arguments of the matplotlib's patches.Rectangle class. fill : dict, optional A dictionary specifying the type of color filling required in the plot. The keys in the dictionary should be equivalent to the arguments of the matplotlib's fill_between() function. adaptive : bool, optional The default value is set to ``True``. Set adaptive to ``False`` and specify ``nb_of_points`` if uniform sampling is required. The plotting uses an adaptive algorithm which samples recursively to accurately plot. The adaptive algorithm uses a random point near the midpoint of two points that has to be further sampled. Hence the same plots can appear slightly different. depth : int, optional Recursion depth of the adaptive algorithm. A depth of value ``n`` samples a maximum of `2^{n}` points. If the ``adaptive`` flag is set to ``False``, this will be ignored. nb_of_points : int, optional Used when the ``adaptive`` is set to ``False``. The function is uniformly sampled at ``nb_of_points`` number of points. If the ``adaptive`` flag is set to ``True``, this will be ignored. size : (float, float), optional A tuple in the form (width, height) in inches to specify the size of the overall figure. The default value is set to ``None``, meaning the size will be set by the default backend. Examples ======== .. plot:: :context: close-figs :format: doctest :include-source: True >>> from sympy import symbols >>> from sympy.plotting import plot >>> x = symbols('x') Single Plot .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot(x**2, (x, -5, 5)) Plot object containing: [0]: cartesian line: x**2 for x over (-5.0, 5.0) Multiple plots with single range. .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot(x, x**2, x**3, (x, -5, 5)) Plot object containing: [0]: cartesian line: x for x over (-5.0, 5.0) [1]: cartesian line: x**2 for x over (-5.0, 5.0) [2]: cartesian line: x**3 for x over (-5.0, 5.0) Multiple plots with different ranges. .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot((x**2, (x, -6, 6)), (x, (x, -5, 5))) Plot object containing: [0]: cartesian line: x**2 for x over (-6.0, 6.0) [1]: cartesian line: x for x over (-5.0, 5.0) No adaptive sampling. .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot(x**2, adaptive=False, nb_of_points=400) Plot object containing: [0]: cartesian line: x**2 for x over (-10.0, 10.0) See Also ======== Plot, LineOver1DRangeSeries """ args = list(map(sympify, args)) free = set() for a in args: if isinstance(a, Expr): free |= a.free_symbols if len(free) > 1: raise ValueError( 'The same variable should be used in all ' 'univariate expressions being plotted.') x = free.pop() if free else Symbol('x') kwargs.setdefault('xlabel', x) kwargs.setdefault('ylabel', Function('f')(x)) series = [] plot_expr = check_arguments(args, 1, 1) series = [LineOver1DRangeSeries(*arg, **kwargs) for arg in plot_expr] plots = Plot(*series, **kwargs) if show: plots.show() return plots def plot_parametric(*args, show=True, **kwargs): """ Plots a 2D parametric curve. Parameters ========== args Common specifications are: - Plotting a single parametric curve with a range ``plot_parametric((expr_x, expr_y), range)`` - Plotting multiple parametric curves with the same range ``plot_parametric((expr_x, expr_y), ..., range)`` - Plotting multiple parametric curves with different ranges ``plot_parametric((expr_x, expr_y, range), ...)`` ``expr_x`` is the expression representing $x$ component of the parametric function. ``expr_y`` is the expression representing $y$ component of the parametric function. ``range`` is a 3-tuple denoting the parameter symbol, start and stop. For example, ``(u, 0, 5)``. If the range is not specified, then a default range of (-10, 10) is used. However, if the arguments are specified as ``(expr_x, expr_y, range), ...``, you must specify the ranges for each expressions manually. Default range may change in the future if a more advanced algorithm is implemented. adaptive : bool, optional Specifies whether to use the adaptive sampling or not. The default value is set to ``True``. Set adaptive to ``False`` and specify ``nb_of_points`` if uniform sampling is required. depth : int, optional The recursion depth of the adaptive algorithm. A depth of value $n$ samples a maximum of $2^n$ points. nb_of_points : int, optional Used when the ``adaptive`` flag is set to ``False``. Specifies the number of the points used for the uniform sampling. line_color : string, or float, or function, optional Specifies the color for the plot. See ``Plot`` to see how to set color for the plots. Note that by setting ``line_color``, it would be applied simultaneously to all the series. label : str, optional The label of the expression in the plot. It will be used when called with ``legend``. Default is the name of the expression. e.g. ``sin(x)`` xlabel : str, optional Label for the x-axis. ylabel : str, optional Label for the y-axis. xscale : 'linear' or 'log', optional Sets the scaling of the x-axis. yscale : 'linear' or 'log', optional Sets the scaling of the y-axis. axis_center : (float, float), optional Tuple of two floats denoting the coordinates of the center or {'center', 'auto'} xlim : (float, float), optional Denotes the x-axis limits, ``(min, max)```. ylim : (float, float), optional Denotes the y-axis limits, ``(min, max)```. size : (float, float), optional A tuple in the form (width, height) in inches to specify the size of the overall figure. The default value is set to ``None``, meaning the size will be set by the default backend. Examples ======== .. plot:: :context: reset :format: doctest :include-source: True >>> from sympy import plot_parametric, symbols, cos, sin >>> u = symbols('u') A parametric plot with a single expression: .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot_parametric((cos(u), sin(u)), (u, -5, 5)) Plot object containing: [0]: parametric cartesian line: (cos(u), sin(u)) for u over (-5.0, 5.0) A parametric plot with multiple expressions with the same range: .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot_parametric((cos(u), sin(u)), (u, cos(u)), (u, -10, 10)) Plot object containing: [0]: parametric cartesian line: (cos(u), sin(u)) for u over (-10.0, 10.0) [1]: parametric cartesian line: (u, cos(u)) for u over (-10.0, 10.0) A parametric plot with multiple expressions with different ranges for each curve: .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot_parametric((cos(u), sin(u), (u, -5, 5)), ... (cos(u), u, (u, -5, 5))) Plot object containing: [0]: parametric cartesian line: (cos(u), sin(u)) for u over (-5.0, 5.0) [1]: parametric cartesian line: (cos(u), u) for u over (-5.0, 5.0) Notes ===== The plotting uses an adaptive algorithm which samples recursively to accurately plot the curve. The adaptive algorithm uses a random point near the midpoint of two points that has to be further sampled. Hence, repeating the same plot command can give slightly different results because of the random sampling. If there are multiple plots, then the same optional arguments are applied to all the plots drawn in the same canvas. If you want to set these options separately, you can index the returned ``Plot`` object and set it. For example, when you specify ``line_color`` once, it would be applied simultaneously to both series. .. plot:: :context: close-figs :format: doctest :include-source: True >>> from sympy import pi >>> expr1 = (u, cos(2*pi*u)/2 + 1/2) >>> expr2 = (u, sin(2*pi*u)/2 + 1/2) >>> p = plot_parametric(expr1, expr2, (u, 0, 1), line_color='blue') If you want to specify the line color for the specific series, you should index each item and apply the property manually. .. plot:: :context: close-figs :format: doctest :include-source: True >>> p[0].line_color = 'red' >>> p.show() See Also ======== Plot, Parametric2DLineSeries """ args = list(map(sympify, args)) series = [] plot_expr = check_arguments(args, 2, 1) series = [Parametric2DLineSeries(*arg, **kwargs) for arg in plot_expr] plots = Plot(*series, **kwargs) if show: plots.show() return plots def plot3d_parametric_line(*args, show=True, **kwargs): """ Plots a 3D parametric line plot. Usage ===== Single plot: ``plot3d_parametric_line(expr_x, expr_y, expr_z, range, **kwargs)`` If the range is not specified, then a default range of (-10, 10) is used. Multiple plots. ``plot3d_parametric_line((expr_x, expr_y, expr_z, range), ..., **kwargs)`` Ranges have to be specified for every expression. Default range may change in the future if a more advanced default range detection algorithm is implemented. Arguments ========= ``expr_x`` : Expression representing the function along x. ``expr_y`` : Expression representing the function along y. ``expr_z`` : Expression representing the function along z. ``range``: ``(u, 0, 5)``, A 3-tuple denoting the range of the parameter variable. Keyword Arguments ================= Arguments for ``Parametric3DLineSeries`` class. ``nb_of_points``: The range is uniformly sampled at ``nb_of_points`` number of points. Aesthetics: ``line_color``: string, or float, or function, optional Specifies the color for the plot. See ``Plot`` to see how to set color for the plots. Note that by setting ``line_color``, it would be applied simultaneously to all the series. ``label``: str The label to the plot. It will be used when called with ``legend=True`` to denote the function with the given label in the plot. If there are multiple plots, then the same series arguments are applied to all the plots. If you want to set these options separately, you can index the returned ``Plot`` object and set it. Arguments for ``Plot`` class. ``title`` : str. Title of the plot. ``size`` : (float, float), optional A tuple in the form (width, height) in inches to specify the size of the overall figure. The default value is set to ``None``, meaning the size will be set by the default backend. Examples ======== .. plot:: :context: reset :format: doctest :include-source: True >>> from sympy import symbols, cos, sin >>> from sympy.plotting import plot3d_parametric_line >>> u = symbols('u') Single plot. .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot3d_parametric_line(cos(u), sin(u), u, (u, -5, 5)) Plot object containing: [0]: 3D parametric cartesian line: (cos(u), sin(u), u) for u over (-5.0, 5.0) Multiple plots. .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot3d_parametric_line((cos(u), sin(u), u, (u, -5, 5)), ... (sin(u), u**2, u, (u, -5, 5))) Plot object containing: [0]: 3D parametric cartesian line: (cos(u), sin(u), u) for u over (-5.0, 5.0) [1]: 3D parametric cartesian line: (sin(u), u**2, u) for u over (-5.0, 5.0) See Also ======== Plot, Parametric3DLineSeries """ args = list(map(sympify, args)) series = [] plot_expr = check_arguments(args, 3, 1) series = [Parametric3DLineSeries(*arg, **kwargs) for arg in plot_expr] kwargs.setdefault("xlabel", "x") kwargs.setdefault("ylabel", "y") kwargs.setdefault("zlabel", "z") plots = Plot(*series, **kwargs) if show: plots.show() return plots def plot3d(*args, show=True, **kwargs): """ Plots a 3D surface plot. Usage ===== Single plot ``plot3d(expr, range_x, range_y, **kwargs)`` If the ranges are not specified, then a default range of (-10, 10) is used. Multiple plot with the same range. ``plot3d(expr1, expr2, range_x, range_y, **kwargs)`` If the ranges are not specified, then a default range of (-10, 10) is used. Multiple plots with different ranges. ``plot3d((expr1, range_x, range_y), (expr2, range_x, range_y), ..., **kwargs)`` Ranges have to be specified for every expression. Default range may change in the future if a more advanced default range detection algorithm is implemented. Arguments ========= ``expr`` : Expression representing the function along x. ``range_x``: (x, 0, 5), A 3-tuple denoting the range of the x variable. ``range_y``: (y, 0, 5), A 3-tuple denoting the range of the y variable. Keyword Arguments ================= Arguments for ``SurfaceOver2DRangeSeries`` class: ``nb_of_points_x``: int. The x range is sampled uniformly at ``nb_of_points_x`` of points. ``nb_of_points_y``: int. The y range is sampled uniformly at ``nb_of_points_y`` of points. Aesthetics: ``surface_color``: Function which returns a float. Specifies the color for the surface of the plot. See ``sympy.plotting.Plot`` for more details. If there are multiple plots, then the same series arguments are applied to all the plots. If you want to set these options separately, you can index the returned ``Plot`` object and set it. Arguments for ``Plot`` class: ``title`` : str. Title of the plot. ``size`` : (float, float), optional A tuple in the form (width, height) in inches to specify the size of the overall figure. The default value is set to ``None``, meaning the size will be set by the default backend. Examples ======== .. plot:: :context: reset :format: doctest :include-source: True >>> from sympy import symbols >>> from sympy.plotting import plot3d >>> x, y = symbols('x y') Single plot .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot3d(x*y, (x, -5, 5), (y, -5, 5)) Plot object containing: [0]: cartesian surface: x*y for x over (-5.0, 5.0) and y over (-5.0, 5.0) Multiple plots with same range .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot3d(x*y, -x*y, (x, -5, 5), (y, -5, 5)) Plot object containing: [0]: cartesian surface: x*y for x over (-5.0, 5.0) and y over (-5.0, 5.0) [1]: cartesian surface: -x*y for x over (-5.0, 5.0) and y over (-5.0, 5.0) Multiple plots with different ranges. .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot3d((x**2 + y**2, (x, -5, 5), (y, -5, 5)), ... (x*y, (x, -3, 3), (y, -3, 3))) Plot object containing: [0]: cartesian surface: x**2 + y**2 for x over (-5.0, 5.0) and y over (-5.0, 5.0) [1]: cartesian surface: x*y for x over (-3.0, 3.0) and y over (-3.0, 3.0) See Also ======== Plot, SurfaceOver2DRangeSeries """ args = list(map(sympify, args)) series = [] plot_expr = check_arguments(args, 1, 2) series = [SurfaceOver2DRangeSeries(*arg, **kwargs) for arg in plot_expr] kwargs.setdefault("xlabel", series[0].var_x) kwargs.setdefault("ylabel", series[0].var_y) kwargs.setdefault("zlabel", Function('f')(series[0].var_x, series[0].var_y)) plots = Plot(*series, **kwargs) if show: plots.show() return plots def plot3d_parametric_surface(*args, show=True, **kwargs): """ Plots a 3D parametric surface plot. Explanation =========== Single plot. ``plot3d_parametric_surface(expr_x, expr_y, expr_z, range_u, range_v, **kwargs)`` If the ranges is not specified, then a default range of (-10, 10) is used. Multiple plots. ``plot3d_parametric_surface((expr_x, expr_y, expr_z, range_u, range_v), ..., **kwargs)`` Ranges have to be specified for every expression. Default range may change in the future if a more advanced default range detection algorithm is implemented. Arguments ========= ``expr_x``: Expression representing the function along ``x``. ``expr_y``: Expression representing the function along ``y``. ``expr_z``: Expression representing the function along ``z``. ``range_u``: ``(u, 0, 5)``, A 3-tuple denoting the range of the ``u`` variable. ``range_v``: ``(v, 0, 5)``, A 3-tuple denoting the range of the v variable. Keyword Arguments ================= Arguments for ``ParametricSurfaceSeries`` class: ``nb_of_points_u``: int. The ``u`` range is sampled uniformly at ``nb_of_points_v`` of points ``nb_of_points_y``: int. The ``v`` range is sampled uniformly at ``nb_of_points_y`` of points Aesthetics: ``surface_color``: Function which returns a float. Specifies the color for the surface of the plot. See ``sympy.plotting.Plot`` for more details. If there are multiple plots, then the same series arguments are applied for all the plots. If you want to set these options separately, you can index the returned ``Plot`` object and set it. Arguments for ``Plot`` class: ``title`` : str. Title of the plot. ``size`` : (float, float), optional A tuple in the form (width, height) in inches to specify the size of the overall figure. The default value is set to ``None``, meaning the size will be set by the default backend. Examples ======== .. plot:: :context: reset :format: doctest :include-source: True >>> from sympy import symbols, cos, sin >>> from sympy.plotting import plot3d_parametric_surface >>> u, v = symbols('u v') Single plot. .. plot:: :context: close-figs :format: doctest :include-source: True >>> plot3d_parametric_surface(cos(u + v), sin(u - v), u - v, ... (u, -5, 5), (v, -5, 5)) Plot object containing: [0]: parametric cartesian surface: (cos(u + v), sin(u - v), u - v) for u over (-5.0, 5.0) and v over (-5.0, 5.0) See Also ======== Plot, ParametricSurfaceSeries """ args = list(map(sympify, args)) series = [] plot_expr = check_arguments(args, 3, 2) series = [ParametricSurfaceSeries(*arg, **kwargs) for arg in plot_expr] kwargs.setdefault("xlabel", "x") kwargs.setdefault("ylabel", "y") kwargs.setdefault("zlabel", "z") plots = Plot(*series, **kwargs) if show: plots.show() return plots def plot_contour(*args, show=True, **kwargs): """ Draws contour plot of a function Usage ===== Single plot ``plot_contour(expr, range_x, range_y, **kwargs)`` If the ranges are not specified, then a default range of (-10, 10) is used. Multiple plot with the same range. ``plot_contour(expr1, expr2, range_x, range_y, **kwargs)`` If the ranges are not specified, then a default range of (-10, 10) is used. Multiple plots with different ranges. ``plot_contour((expr1, range_x, range_y), (expr2, range_x, range_y), ..., **kwargs)`` Ranges have to be specified for every expression. Default range may change in the future if a more advanced default range detection algorithm is implemented. Arguments ========= ``expr`` : Expression representing the function along x. ``range_x``: (x, 0, 5), A 3-tuple denoting the range of the x variable. ``range_y``: (y, 0, 5), A 3-tuple denoting the range of the y variable. Keyword Arguments ================= Arguments for ``ContourSeries`` class: ``nb_of_points_x``: int. The x range is sampled uniformly at ``nb_of_points_x`` of points. ``nb_of_points_y``: int. The y range is sampled uniformly at ``nb_of_points_y`` of points. Aesthetics: ``surface_color``: Function which returns a float. Specifies the color for the surface of the plot. See ``sympy.plotting.Plot`` for more details. If there are multiple plots, then the same series arguments are applied to all the plots. If you want to set these options separately, you can index the returned ``Plot`` object and set it. Arguments for ``Plot`` class: ``title`` : str. Title of the plot. ``size`` : (float, float), optional A tuple in the form (width, height) in inches to specify the size of the overall figure. The default value is set to ``None``, meaning the size will be set by the default backend. See Also ======== Plot, ContourSeries """ args = list(map(sympify, args)) plot_expr = check_arguments(args, 1, 2) series = [ContourSeries(*arg) for arg in plot_expr] plot_contours = Plot(*series, **kwargs) if len(plot_expr[0].free_symbols) > 2: raise ValueError('Contour Plot cannot Plot for more than two variables.') if show: plot_contours.show() return plot_contours def check_arguments(args, expr_len, nb_of_free_symbols): """ Checks the arguments and converts into tuples of the form (exprs, ranges). Examples ======== .. plot:: :context: reset :format: doctest :include-source: True >>> from sympy import cos, sin, symbols >>> from sympy.plotting.plot import check_arguments >>> x = symbols('x') >>> check_arguments([cos(x), sin(x)], 2, 1) [(cos(x), sin(x), (x, -10, 10))] >>> check_arguments([x, x**2], 1, 1) [(x, (x, -10, 10)), (x**2, (x, -10, 10))] """ if not args: return [] if expr_len > 1 and isinstance(args[0], Expr): # Multiple expressions same range. # The arguments are tuples when the expression length is # greater than 1. if len(args) < expr_len: raise ValueError("len(args) should not be less than expr_len") for i in range(len(args)): if isinstance(args[i], Tuple): break else: i = len(args) + 1 exprs = Tuple(*args[:i]) free_symbols = list(set().union(*[e.free_symbols for e in exprs])) if len(args) == expr_len + nb_of_free_symbols: #Ranges given plots = [exprs + Tuple(*args[expr_len:])] else: default_range = Tuple(-10, 10) ranges = [] for symbol in free_symbols: ranges.append(Tuple(symbol) + default_range) for i in range(len(free_symbols) - nb_of_free_symbols): ranges.append(Tuple(Dummy()) + default_range) plots = [exprs + Tuple(*ranges)] return plots if isinstance(args[0], Expr) or (isinstance(args[0], Tuple) and len(args[0]) == expr_len and expr_len != 3): # Cannot handle expressions with number of expression = 3. It is # not possible to differentiate between expressions and ranges. #Series of plots with same range for i in range(len(args)): if isinstance(args[i], Tuple) and len(args[i]) != expr_len: break if not isinstance(args[i], Tuple): args[i] = Tuple(args[i]) else: i = len(args) + 1 exprs = args[:i] assert all(isinstance(e, Expr) for expr in exprs for e in expr) free_symbols = list(set().union(*[e.free_symbols for expr in exprs for e in expr])) if len(free_symbols) > nb_of_free_symbols: raise ValueError("The number of free_symbols in the expression " "is greater than %d" % nb_of_free_symbols) if len(args) == i + nb_of_free_symbols and isinstance(args[i], Tuple): ranges = Tuple(*[range_expr for range_expr in args[ i:i + nb_of_free_symbols]]) plots = [expr + ranges for expr in exprs] return plots else: # Use default ranges. default_range = Tuple(-10, 10) ranges = [] for symbol in free_symbols: ranges.append(Tuple(symbol) + default_range) for i in range(nb_of_free_symbols - len(free_symbols)): ranges.append(Tuple(Dummy()) + default_range) ranges = Tuple(*ranges) plots = [expr + ranges for expr in exprs] return plots elif isinstance(args[0], Tuple) and len(args[0]) == expr_len + nb_of_free_symbols: # Multiple plots with different ranges. for arg in args: for i in range(expr_len): if not isinstance(arg[i], Expr): raise ValueError("Expected an expression, given %s" % str(arg[i])) for i in range(nb_of_free_symbols): if not len(arg[i + expr_len]) == 3: raise ValueError("The ranges should be a tuple of " "length 3, got %s" % str(arg[i + expr_len])) return args
35.905488
128
0.565721
from collections.abc import Callable from sympy.core.basic import Basic from sympy.core.containers import Tuple from sympy.core.expr import Expr from sympy.core.function import arity, Function from sympy.core.symbol import (Dummy, Symbol) from sympy.core.sympify import sympify from sympy.external import import_module from sympy.printing.latex import latex from sympy.utilities.exceptions import sympy_deprecation_warning from sympy.utilities.iterables import is_sequence from .experimental_lambdify import (vectorized_lambdify, lambdify) from sympy.plotting.textplot import textplot _show = True def unset_show(): global _show _show = False def _str_or_latex(label): if isinstance(label, Basic): return latex(label, mode='inline') return str(label) ############################################################################## # The public interface ############################################################################## class Plot: def __init__(self, *args, title=None, xlabel=None, ylabel=None, zlabel=None, aspect_ratio='auto', xlim=None, ylim=None, axis_center='auto', axis=True, xscale='linear', yscale='linear', legend=False, autoscale=True, margin=0, annotations=None, markers=None, rectangles=None, fill=None, backend='default', size=None, **kwargs): super().__init__() # Options for the graph as a whole. # The possible values for each option are described in the docstring of # Plot. They are based purely on convention, no checking is done. self.title = title self.xlabel = xlabel self.ylabel = ylabel self.zlabel = zlabel self.aspect_ratio = aspect_ratio self.axis_center = axis_center self.axis = axis self.xscale = xscale self.yscale = yscale self.legend = legend self.autoscale = autoscale self.margin = margin self.annotations = annotations self.markers = markers self.rectangles = rectangles self.fill = fill # Contains the data objects to be plotted. The backend should be smart # enough to iterate over this list. self._series = [] self._series.extend(args) # The backend type. On every show() a new backend instance is created # in self._backend which is tightly coupled to the Plot instance # (thanks to the parent attribute of the backend). if isinstance(backend, str): self.backend = plot_backends[backend] elif (type(backend) == type) and issubclass(backend, BaseBackend): self.backend = backend else: raise TypeError( "backend must be either a string or a subclass of BaseBackend") is_real = \ lambda lim: all(getattr(i, 'is_real', True) for i in lim) is_finite = \ lambda lim: all(getattr(i, 'is_finite', True) for i in lim) # reduce code repetition def check_and_set(t_name, t): if t: if not is_real(t): raise ValueError( "All numbers from {}={} must be real".format(t_name, t)) if not is_finite(t): raise ValueError( "All numbers from {}={} must be finite".format(t_name, t)) setattr(self, t_name, (float(t[0]), float(t[1]))) self.xlim = None check_and_set("xlim", xlim) self.ylim = None check_and_set("ylim", ylim) self.size = None check_and_set("size", size) def show(self): # TODO move this to the backend (also for save) if hasattr(self, '_backend'): self._backend.close() self._backend = self.backend(self) self._backend.show() def save(self, path): if hasattr(self, '_backend'): self._backend.close() self._backend = self.backend(self) self._backend.save(path) def __str__(self): series_strs = [('[%d]: ' % i) + str(s) for i, s in enumerate(self._series)] return 'Plot object containing:\n' + '\n'.join(series_strs) def __getitem__(self, index): return self._series[index] def __setitem__(self, index, *args): if len(args) == 1 and isinstance(args[0], BaseSeries): self._series[index] = args def __delitem__(self, index): del self._series[index] def append(self, arg): if isinstance(arg, BaseSeries): self._series.append(arg) else: raise TypeError('Must specify element of plot to append.') def extend(self, arg): if isinstance(arg, Plot): self._series.extend(arg._series) elif is_sequence(arg): self._series.extend(arg) else: raise TypeError('Expecting Plot or sequence of BaseSeries') class PlotGrid: def __init__(self, nrows, ncolumns, *args, show=True, size=None, **kwargs): self.nrows = nrows self.ncolumns = ncolumns self._series = [] self.args = args for arg in args: self._series.append(arg._series) self.backend = DefaultBackend self.size = size if show: self.show() def show(self): if hasattr(self, '_backend'): self._backend.close() self._backend = self.backend(self) self._backend.show() def save(self, path): if hasattr(self, '_backend'): self._backend.close() self._backend = self.backend(self) self._backend.save(path) def __str__(self): plot_strs = [('Plot[%d]:' % i) + str(plot) for i, plot in enumerate(self.args)] return 'PlotGrid object containing:\n' + '\n'.join(plot_strs) ############################################################################## # Data Series ############################################################################## #TODO more general way to calculate aesthetics (see get_color_array) ### The base class for all series class BaseSeries: # Some flags follow. The rationale for using flags instead of checking base # classes is that setting multiple flags is simpler than multiple # inheritance. is_2Dline = False # Some of the backends expect: # - get_points returning 1D np.arrays list_x, list_y # - get_color_array returning 1D np.array (done in Line2DBaseSeries) # with the colors calculated at the points from get_points is_3Dline = False # Some of the backends expect: # - get_points returning 1D np.arrays list_x, list_y, list_y # - get_color_array returning 1D np.array (done in Line2DBaseSeries) # with the colors calculated at the points from get_points is_3Dsurface = False # Some of the backends expect: # - get_meshes returning mesh_x, mesh_y, mesh_z (2D np.arrays) # - get_points an alias for get_meshes is_contour = False # Some of the backends expect: # - get_meshes returning mesh_x, mesh_y, mesh_z (2D np.arrays) # - get_points an alias for get_meshes is_implicit = False # Some of the backends expect: # - get_meshes returning mesh_x (1D array), mesh_y(1D array, # mesh_z (2D np.arrays) # - get_points an alias for get_meshes # Different from is_contour as the colormap in backend will be # different is_parametric = False # The calculation of aesthetics expects: # - get_parameter_points returning one or two np.arrays (1D or 2D) # used for calculation aesthetics def __init__(self): super().__init__() @property def is_3D(self): flags3D = [ self.is_3Dline, self.is_3Dsurface ] return any(flags3D) @property def is_line(self): flagslines = [ self.is_2Dline, self.is_3Dline ] return any(flagslines) ### 2D lines class Line2DBaseSeries(BaseSeries): is_2Dline = True _dim = 2 def __init__(self): super().__init__() self.label = None self.steps = False self.only_integers = False self.line_color = None def get_data(self): np = import_module('numpy') points = self.get_points() if self.steps is True: if len(points) == 2: x = np.array((points[0], points[0])).T.flatten()[1:] y = np.array((points[1], points[1])).T.flatten()[:-1] points = (x, y) else: x = np.repeat(points[0], 3)[2:] y = np.repeat(points[1], 3)[:-2] z = np.repeat(points[2], 3)[1:-1] points = (x, y, z) return points def get_segments(self): sympy_deprecation_warning( """ The Line2DBaseSeries.get_segments() method is deprecated. Instead, use the MatplotlibBackend.get_segments() method, or use The get_points() or get_data() methods. """, deprecated_since_version="1.9", active_deprecations_target="deprecated-get-segments") np = import_module('numpy') points = type(self).get_data(self) points = np.ma.array(points).T.reshape(-1, 1, self._dim) return np.ma.concatenate([points[:-1], points[1:]], axis=1) def get_color_array(self): np = import_module('numpy') c = self.line_color if hasattr(c, '__call__'): f = np.vectorize(c) nargs = arity(c) if nargs == 1 and self.is_parametric: x = self.get_parameter_points() return f(centers_of_segments(x)) else: variables = list(map(centers_of_segments, self.get_points())) if nargs == 1: return f(variables[0]) elif nargs == 2: return f(*variables[:2]) else: # only if the line is 3D (otherwise raises an error) return f(*variables) else: return c*np.ones(self.nb_of_points) class List2DSeries(Line2DBaseSeries): def __init__(self, list_x, list_y): np = import_module('numpy') super().__init__() self.list_x = np.array(list_x) self.list_y = np.array(list_y) self.label = 'list' def __str__(self): return 'list plot' def get_points(self): return (self.list_x, self.list_y) class LineOver1DRangeSeries(Line2DBaseSeries): def __init__(self, expr, var_start_end, **kwargs): super().__init__() self.expr = sympify(expr) self.label = kwargs.get('label', None) or self.expr self.var = sympify(var_start_end[0]) self.start = float(var_start_end[1]) self.end = float(var_start_end[2]) self.nb_of_points = kwargs.get('nb_of_points', 300) self.adaptive = kwargs.get('adaptive', True) self.depth = kwargs.get('depth', 12) self.line_color = kwargs.get('line_color', None) self.xscale = kwargs.get('xscale', 'linear') def __str__(self): return 'cartesian line: %s for %s over %s' % ( str(self.expr), str(self.var), str((self.start, self.end))) def get_points(self): if self.only_integers or not self.adaptive: return self._uniform_sampling() else: f = lambdify([self.var], self.expr) x_coords = [] y_coords = [] np = import_module('numpy') def sample(p, q, depth): """ Samples recursively if three points are almost collinear. For depth < 6, points are added irrespective of whether they satisfy the collinearity condition or not. The maximum depth allowed is 12. """ # Randomly sample to avoid aliasing. random = 0.45 + np.random.rand() * 0.1 if self.xscale == 'log': xnew = 10**(np.log10(p[0]) + random * (np.log10(q[0]) - np.log10(p[0]))) else: xnew = p[0] + random * (q[0] - p[0]) ynew = f(xnew) new_point = np.array([xnew, ynew]) # Maximum depth if depth > self.depth: x_coords.append(q[0]) y_coords.append(q[1]) # Sample irrespective of whether the line is flat till the # depth of 6. We are not using linspace to avoid aliasing. elif depth < 6: sample(p, new_point, depth + 1) sample(new_point, q, depth + 1) # Sample ten points if complex values are encountered # at both ends. If there is a real value in between, then # sample those points further. elif p[1] is None and q[1] is None: if self.xscale == 'log': xarray = np.logspace(p[0], q[0], 10) else: xarray = np.linspace(p[0], q[0], 10) yarray = list(map(f, xarray)) if not all(y is None for y in yarray): for i in range(len(yarray) - 1): if not (yarray[i] is None and yarray[i + 1] is None): sample([xarray[i], yarray[i]], [xarray[i + 1], yarray[i + 1]], depth + 1) # Sample further if one of the end points in None (i.e. a # complex value) or the three points are not almost collinear. elif (p[1] is None or q[1] is None or new_point[1] is None or not flat(p, new_point, q)): sample(p, new_point, depth + 1) sample(new_point, q, depth + 1) else: x_coords.append(q[0]) y_coords.append(q[1]) f_start = f(self.start) f_end = f(self.end) x_coords.append(self.start) y_coords.append(f_start) sample(np.array([self.start, f_start]), np.array([self.end, f_end]), 0) return (x_coords, y_coords) def _uniform_sampling(self): np = import_module('numpy') if self.only_integers is True: if self.xscale == 'log': list_x = np.logspace(int(self.start), int(self.end), num=int(self.end) - int(self.start) + 1) else: list_x = np.linspace(int(self.start), int(self.end), num=int(self.end) - int(self.start) + 1) else: if self.xscale == 'log': list_x = np.logspace(self.start, self.end, num=self.nb_of_points) else: list_x = np.linspace(self.start, self.end, num=self.nb_of_points) f = vectorized_lambdify([self.var], self.expr) list_y = f(list_x) return (list_x, list_y) class Parametric2DLineSeries(Line2DBaseSeries): is_parametric = True def __init__(self, expr_x, expr_y, var_start_end, **kwargs): super().__init__() self.expr_x = sympify(expr_x) self.expr_y = sympify(expr_y) self.label = kwargs.get('label', None) or \ Tuple(self.expr_x, self.expr_y) self.var = sympify(var_start_end[0]) self.start = float(var_start_end[1]) self.end = float(var_start_end[2]) self.nb_of_points = kwargs.get('nb_of_points', 300) self.adaptive = kwargs.get('adaptive', True) self.depth = kwargs.get('depth', 12) self.line_color = kwargs.get('line_color', None) def __str__(self): return 'parametric cartesian line: (%s, %s) for %s over %s' % ( str(self.expr_x), str(self.expr_y), str(self.var), str((self.start, self.end))) def get_parameter_points(self): np = import_module('numpy') return np.linspace(self.start, self.end, num=self.nb_of_points) def _uniform_sampling(self): param = self.get_parameter_points() fx = vectorized_lambdify([self.var], self.expr_x) fy = vectorized_lambdify([self.var], self.expr_y) list_x = fx(param) list_y = fy(param) return (list_x, list_y) def get_points(self): if not self.adaptive: return self._uniform_sampling() f_x = lambdify([self.var], self.expr_x) f_y = lambdify([self.var], self.expr_y) x_coords = [] y_coords = [] def sample(param_p, param_q, p, q, depth): # Randomly sample to avoid aliasing. np = import_module('numpy') random = 0.45 + np.random.rand() * 0.1 param_new = param_p + random * (param_q - param_p) xnew = f_x(param_new) ynew = f_y(param_new) new_point = np.array([xnew, ynew]) # Maximum depth if depth > self.depth: x_coords.append(q[0]) y_coords.append(q[1]) # Sample irrespective of whether the line is flat till the # depth of 6. We are not using linspace to avoid aliasing. elif depth < 6: sample(param_p, param_new, p, new_point, depth + 1) sample(param_new, param_q, new_point, q, depth + 1) # Sample ten points if complex values are encountered # at both ends. If there is a real value in between, then # sample those points further. elif ((p[0] is None and q[1] is None) or (p[1] is None and q[1] is None)): param_array = np.linspace(param_p, param_q, 10) x_array = list(map(f_x, param_array)) y_array = list(map(f_y, param_array)) if not all(x is None and y is None for x, y in zip(x_array, y_array)): for i in range(len(y_array) - 1): if ((x_array[i] is not None and y_array[i] is not None) or (x_array[i + 1] is not None and y_array[i + 1] is not None)): point_a = [x_array[i], y_array[i]] point_b = [x_array[i + 1], y_array[i + 1]] sample(param_array[i], param_array[i], point_a, point_b, depth + 1) # Sample further if one of the end points in None (i.e. a complex # value) or the three points are not almost collinear. elif (p[0] is None or p[1] is None or q[1] is None or q[0] is None or not flat(p, new_point, q)): sample(param_p, param_new, p, new_point, depth + 1) sample(param_new, param_q, new_point, q, depth + 1) else: x_coords.append(q[0]) y_coords.append(q[1]) f_start_x = f_x(self.start) f_start_y = f_y(self.start) start = [f_start_x, f_start_y] f_end_x = f_x(self.end) f_end_y = f_y(self.end) end = [f_end_x, f_end_y] x_coords.append(f_start_x) y_coords.append(f_start_y) sample(self.start, self.end, start, end, 0) return x_coords, y_coords ### 3D lines class Line3DBaseSeries(Line2DBaseSeries): is_2Dline = False is_3Dline = True _dim = 3 def __init__(self): super().__init__() class Parametric3DLineSeries(Line3DBaseSeries): is_parametric = True def __init__(self, expr_x, expr_y, expr_z, var_start_end, **kwargs): super().__init__() self.expr_x = sympify(expr_x) self.expr_y = sympify(expr_y) self.expr_z = sympify(expr_z) self.label = kwargs.get('label', None) or \ Tuple(self.expr_x, self.expr_y) self.var = sympify(var_start_end[0]) self.start = float(var_start_end[1]) self.end = float(var_start_end[2]) self.nb_of_points = kwargs.get('nb_of_points', 300) self.line_color = kwargs.get('line_color', None) self._xlim = None self._ylim = None self._zlim = None def __str__(self): return '3D parametric cartesian line: (%s, %s, %s) for %s over %s' % ( str(self.expr_x), str(self.expr_y), str(self.expr_z), str(self.var), str((self.start, self.end))) def get_parameter_points(self): np = import_module('numpy') return np.linspace(self.start, self.end, num=self.nb_of_points) def get_points(self): np = import_module('numpy') param = self.get_parameter_points() fx = vectorized_lambdify([self.var], self.expr_x) fy = vectorized_lambdify([self.var], self.expr_y) fz = vectorized_lambdify([self.var], self.expr_z) list_x = fx(param) list_y = fy(param) list_z = fz(param) list_x = np.array(list_x, dtype=np.float64) list_y = np.array(list_y, dtype=np.float64) list_z = np.array(list_z, dtype=np.float64) list_x = np.ma.masked_invalid(list_x) list_y = np.ma.masked_invalid(list_y) list_z = np.ma.masked_invalid(list_z) self._xlim = (np.amin(list_x), np.amax(list_x)) self._ylim = (np.amin(list_y), np.amax(list_y)) self._zlim = (np.amin(list_z), np.amax(list_z)) return list_x, list_y, list_z ### Surfaces class SurfaceBaseSeries(BaseSeries): is_3Dsurface = True def __init__(self): super().__init__() self.surface_color = None def get_color_array(self): np = import_module('numpy') c = self.surface_color if isinstance(c, Callable): f = np.vectorize(c) nargs = arity(c) if self.is_parametric: variables = list(map(centers_of_faces, self.get_parameter_meshes())) if nargs == 1: return f(variables[0]) elif nargs == 2: return f(*variables) variables = list(map(centers_of_faces, self.get_meshes())) if nargs == 1: return f(variables[0]) elif nargs == 2: return f(*variables[:2]) else: return f(*variables) else: if isinstance(self, SurfaceOver2DRangeSeries): return c*np.ones(min(self.nb_of_points_x, self.nb_of_points_y)) else: return c*np.ones(min(self.nb_of_points_u, self.nb_of_points_v)) class SurfaceOver2DRangeSeries(SurfaceBaseSeries): def __init__(self, expr, var_start_end_x, var_start_end_y, **kwargs): super().__init__() self.expr = sympify(expr) self.var_x = sympify(var_start_end_x[0]) self.start_x = float(var_start_end_x[1]) self.end_x = float(var_start_end_x[2]) self.var_y = sympify(var_start_end_y[0]) self.start_y = float(var_start_end_y[1]) self.end_y = float(var_start_end_y[2]) self.nb_of_points_x = kwargs.get('nb_of_points_x', 50) self.nb_of_points_y = kwargs.get('nb_of_points_y', 50) self.surface_color = kwargs.get('surface_color', None) self._xlim = (self.start_x, self.end_x) self._ylim = (self.start_y, self.end_y) def __str__(self): return ('cartesian surface: %s for' ' %s over %s and %s over %s') % ( str(self.expr), str(self.var_x), str((self.start_x, self.end_x)), str(self.var_y), str((self.start_y, self.end_y))) def get_meshes(self): np = import_module('numpy') mesh_x, mesh_y = np.meshgrid(np.linspace(self.start_x, self.end_x, num=self.nb_of_points_x), np.linspace(self.start_y, self.end_y, num=self.nb_of_points_y)) f = vectorized_lambdify((self.var_x, self.var_y), self.expr) mesh_z = f(mesh_x, mesh_y) mesh_z = np.array(mesh_z, dtype=np.float64) mesh_z = np.ma.masked_invalid(mesh_z) self._zlim = (np.amin(mesh_z), np.amax(mesh_z)) return mesh_x, mesh_y, mesh_z class ParametricSurfaceSeries(SurfaceBaseSeries): is_parametric = True def __init__( self, expr_x, expr_y, expr_z, var_start_end_u, var_start_end_v, **kwargs): super().__init__() self.expr_x = sympify(expr_x) self.expr_y = sympify(expr_y) self.expr_z = sympify(expr_z) self.var_u = sympify(var_start_end_u[0]) self.start_u = float(var_start_end_u[1]) self.end_u = float(var_start_end_u[2]) self.var_v = sympify(var_start_end_v[0]) self.start_v = float(var_start_end_v[1]) self.end_v = float(var_start_end_v[2]) self.nb_of_points_u = kwargs.get('nb_of_points_u', 50) self.nb_of_points_v = kwargs.get('nb_of_points_v', 50) self.surface_color = kwargs.get('surface_color', None) def __str__(self): return ('parametric cartesian surface: (%s, %s, %s) for' ' %s over %s and %s over %s') % ( str(self.expr_x), str(self.expr_y), str(self.expr_z), str(self.var_u), str((self.start_u, self.end_u)), str(self.var_v), str((self.start_v, self.end_v))) def get_parameter_meshes(self): np = import_module('numpy') return np.meshgrid(np.linspace(self.start_u, self.end_u, num=self.nb_of_points_u), np.linspace(self.start_v, self.end_v, num=self.nb_of_points_v)) def get_meshes(self): np = import_module('numpy') mesh_u, mesh_v = self.get_parameter_meshes() fx = vectorized_lambdify((self.var_u, self.var_v), self.expr_x) fy = vectorized_lambdify((self.var_u, self.var_v), self.expr_y) fz = vectorized_lambdify((self.var_u, self.var_v), self.expr_z) mesh_x = fx(mesh_u, mesh_v) mesh_y = fy(mesh_u, mesh_v) mesh_z = fz(mesh_u, mesh_v) mesh_x = np.array(mesh_x, dtype=np.float64) mesh_y = np.array(mesh_y, dtype=np.float64) mesh_z = np.array(mesh_z, dtype=np.float64) mesh_x = np.ma.masked_invalid(mesh_x) mesh_y = np.ma.masked_invalid(mesh_y) mesh_z = np.ma.masked_invalid(mesh_z) self._xlim = (np.amin(mesh_x), np.amax(mesh_x)) self._ylim = (np.amin(mesh_y), np.amax(mesh_y)) self._zlim = (np.amin(mesh_z), np.amax(mesh_z)) return mesh_x, mesh_y, mesh_z ### Contours class ContourSeries(BaseSeries): # The code is mostly repetition of SurfaceOver2DRange. # Presently used in contour_plot function is_contour = True def __init__(self, expr, var_start_end_x, var_start_end_y): super().__init__() self.nb_of_points_x = 50 self.nb_of_points_y = 50 self.expr = sympify(expr) self.var_x = sympify(var_start_end_x[0]) self.start_x = float(var_start_end_x[1]) self.end_x = float(var_start_end_x[2]) self.var_y = sympify(var_start_end_y[0]) self.start_y = float(var_start_end_y[1]) self.end_y = float(var_start_end_y[2]) self.get_points = self.get_meshes self._xlim = (self.start_x, self.end_x) self._ylim = (self.start_y, self.end_y) def __str__(self): return ('contour: %s for ' '%s over %s and %s over %s') % ( str(self.expr), str(self.var_x), str((self.start_x, self.end_x)), str(self.var_y), str((self.start_y, self.end_y))) def get_meshes(self): np = import_module('numpy') mesh_x, mesh_y = np.meshgrid(np.linspace(self.start_x, self.end_x, num=self.nb_of_points_x), np.linspace(self.start_y, self.end_y, num=self.nb_of_points_y)) f = vectorized_lambdify((self.var_x, self.var_y), self.expr) return (mesh_x, mesh_y, f(mesh_x, mesh_y)) ############################################################################## # Backends ############################################################################## class BaseBackend: def __init__(self, parent): super().__init__() self.parent = parent def show(self): raise NotImplementedError def save(self, path): raise NotImplementedError def close(self): raise NotImplementedError # Don't have to check for the success of importing matplotlib in each case; class MatplotlibBackend(BaseBackend): def __init__(self, parent): super().__init__(parent) self.matplotlib = import_module('matplotlib', import_kwargs={'fromlist': ['pyplot', 'cm', 'collections']}, min_module_version='1.1.0', catch=(RuntimeError,)) self.plt = self.matplotlib.pyplot self.cm = self.matplotlib.cm self.LineCollection = self.matplotlib.collections.LineCollection aspect = getattr(self.parent, 'aspect_ratio', 'auto') if aspect != 'auto': aspect = float(aspect[1]) / aspect[0] if isinstance(self.parent, Plot): nrows, ncolumns = 1, 1 series_list = [self.parent._series] elif isinstance(self.parent, PlotGrid): nrows, ncolumns = self.parent.nrows, self.parent.ncolumns series_list = self.parent._series self.ax = [] self.fig = self.plt.figure(figsize=parent.size) for i, series in enumerate(series_list): are_3D = [s.is_3D for s in series] if any(are_3D) and not all(are_3D): raise ValueError('The matplotlib backend cannot mix 2D and 3D.') elif all(are_3D): mpl_toolkits = import_module('mpl_toolkits', import_kwargs={'fromlist': ['mplot3d']}) self.ax.append(self.fig.add_subplot(nrows, ncolumns, i + 1, projection='3d', aspect=aspect)) elif not any(are_3D): self.ax.append(self.fig.add_subplot(nrows, ncolumns, i + 1, aspect=aspect)) self.ax[i].spines['left'].set_position('zero') self.ax[i].spines['right'].set_color('none') self.ax[i].spines['bottom'].set_position('zero') self.ax[i].spines['top'].set_color('none') self.ax[i].xaxis.set_ticks_position('bottom') self.ax[i].yaxis.set_ticks_position('left') @staticmethod def get_segments(x, y, z=None): np = import_module('numpy') if z is not None: dim = 3 points = (x, y, z) else: dim = 2 points = (x, y) points = np.ma.array(points).T.reshape(-1, 1, dim) return np.ma.concatenate([points[:-1], points[1:]], axis=1) def _process_series(self, series, ax, parent): np = import_module('numpy') mpl_toolkits = import_module( 'mpl_toolkits', import_kwargs={'fromlist': ['mplot3d']}) xlims, ylims, zlims = [], [], [] for s in series: if s.is_2Dline: x, y = s.get_data() if (isinstance(s.line_color, (int, float)) or callable(s.line_color)): segments = self.get_segments(x, y) collection = self.LineCollection(segments) collection.set_array(s.get_color_array()) ax.add_collection(collection) else: lbl = _str_or_latex(s.label) line, = ax.plot(x, y, label=lbl, color=s.line_color) elif s.is_contour: ax.contour(*s.get_meshes()) elif s.is_3Dline: x, y, z = s.get_data() if (isinstance(s.line_color, (int, float)) or callable(s.line_color)): art3d = mpl_toolkits.mplot3d.art3d segments = self.get_segments(x, y, z) collection = art3d.Line3DCollection(segments) collection.set_array(s.get_color_array()) ax.add_collection(collection) else: lbl = _str_or_latex(s.label) ax.plot(x, y, z, label=lbl, color=s.line_color) xlims.append(s._xlim) ylims.append(s._ylim) zlims.append(s._zlim) elif s.is_3Dsurface: x, y, z = s.get_meshes() collection = ax.plot_surface(x, y, z, cmap=getattr(self.cm, 'viridis', self.cm.jet), rstride=1, cstride=1, linewidth=0.1) if isinstance(s.surface_color, (float, int, Callable)): color_array = s.get_color_array() color_array = color_array.reshape(color_array.size) collection.set_array(color_array) else: collection.set_color(s.surface_color) xlims.append(s._xlim) ylims.append(s._ylim) zlims.append(s._zlim) elif s.is_implicit: points = s.get_raster() if len(points) == 2: x, y = _matplotlib_list(points[0]) ax.fill(x, y, facecolor=s.line_color, edgecolor='None') else: ListedColormap = self.matplotlib.colors.ListedColormap colormap = ListedColormap(["white", s.line_color]) xarray, yarray, zarray, plot_type = points if plot_type == 'contour': ax.contour(xarray, yarray, zarray, cmap=colormap, label=_str_or_latex(s.label)) else: ax.contourf(xarray, yarray, zarray, cmap=colormap, label=_str_or_latex(s.label)) else: raise NotImplementedError( '{} is not supported in the SymPy plotting module ' 'with matplotlib backend. Please report this issue.' .format(ax)) Axes3D = mpl_toolkits.mplot3d.Axes3D if not isinstance(ax, Axes3D): ax.autoscale_view( scalex=ax.get_autoscalex_on(), scaley=ax.get_autoscaley_on()) else: if xlims: xlims = np.array(xlims) xlim = (np.amin(xlims[:, 0]), np.amax(xlims[:, 1])) ax.set_xlim(xlim) else: ax.set_xlim([0, 1]) if ylims: ylims = np.array(ylims) ylim = (np.amin(ylims[:, 0]), np.amax(ylims[:, 1])) ax.set_ylim(ylim) else: ax.set_ylim([0, 1]) if zlims: zlims = np.array(zlims) zlim = (np.amin(zlims[:, 0]), np.amax(zlims[:, 1])) ax.set_zlim(zlim) else: ax.set_zlim([0, 1]) if parent.xscale and not isinstance(ax, Axes3D): ax.set_xscale(parent.xscale) if parent.yscale and not isinstance(ax, Axes3D): ax.set_yscale(parent.yscale) if not isinstance(ax, Axes3D) or self.matplotlib.__version__ >= '1.2.0': ax.set_autoscale_on(parent.autoscale) if parent.axis_center: val = parent.axis_center if isinstance(ax, Axes3D): pass elif val == 'center': ax.spines['left'].set_position('center') ax.spines['bottom'].set_position('center') elif val == 'auto': xl, xh = ax.get_xlim() yl, yh = ax.get_ylim() pos_left = ('data', 0) if xl*xh <= 0 else 'center' pos_bottom = ('data', 0) if yl*yh <= 0 else 'center' ax.spines['left'].set_position(pos_left) ax.spines['bottom'].set_position(pos_bottom) else: ax.spines['left'].set_position(('data', val[0])) ax.spines['bottom'].set_position(('data', val[1])) if not parent.axis: ax.set_axis_off() if parent.legend: if ax.legend(): ax.legend_.set_visible(parent.legend) if parent.margin: ax.set_xmargin(parent.margin) ax.set_ymargin(parent.margin) if parent.title: ax.set_title(parent.title) if parent.xlabel: xlbl = _str_or_latex(parent.xlabel) ax.set_xlabel(xlbl, position=(1, 0)) if parent.ylabel: ylbl = _str_or_latex(parent.ylabel) ax.set_ylabel(ylbl, position=(0, 1)) if isinstance(ax, Axes3D) and parent.zlabel: zlbl = _str_or_latex(parent.zlabel) ax.set_zlabel(zlbl, position=(0, 1)) if parent.annotations: for a in parent.annotations: ax.annotate(**a) if parent.markers: for marker in parent.markers: m = marker.copy() args = m.pop('args') ax.plot(*args, **m) if parent.rectangles: for r in parent.rectangles: rect = self.matplotlib.patches.Rectangle(**r) ax.add_patch(rect) if parent.fill: ax.fill_between(**parent.fill) # xlim and ylim shoulld always be set at last so that plot limits # doesn't get altered during the process. if parent.xlim: ax.set_xlim(parent.xlim) if parent.ylim: ax.set_ylim(parent.ylim) def process_series(self): parent = self.parent if isinstance(parent, Plot): series_list = [parent._series] else: series_list = parent._series for i, (series, ax) in enumerate(zip(series_list, self.ax)): if isinstance(self.parent, PlotGrid): parent = self.parent.args[i] self._process_series(series, ax, parent) def show(self): self.process_series() if _show: self.fig.tight_layout() self.plt.show() else: self.close() def save(self, path): self.process_series() self.fig.savefig(path) def close(self): self.plt.close(self.fig) class TextBackend(BaseBackend): def __init__(self, parent): super().__init__(parent) def show(self): if not _show: return if len(self.parent._series) != 1: raise ValueError( 'The TextBackend supports only one graph per Plot.') elif not isinstance(self.parent._series[0], LineOver1DRangeSeries): raise ValueError( 'The TextBackend supports only expressions over a 1D range') else: ser = self.parent._series[0] textplot(ser.expr, ser.start, ser.end) def close(self): pass class DefaultBackend(BaseBackend): def __new__(cls, parent): matplotlib = import_module('matplotlib', min_module_version='1.1.0', catch=(RuntimeError,)) if matplotlib: return MatplotlibBackend(parent) else: return TextBackend(parent) plot_backends = { 'matplotlib': MatplotlibBackend, 'text': TextBackend, 'default': DefaultBackend } d" % nb_of_free_symbols) if len(args) == i + nb_of_free_symbols and isinstance(args[i], Tuple): ranges = Tuple(*[range_expr for range_expr in args[ i:i + nb_of_free_symbols]]) plots = [expr + ranges for expr in exprs] return plots else: default_range = Tuple(-10, 10) ranges = [] for symbol in free_symbols: ranges.append(Tuple(symbol) + default_range) for i in range(nb_of_free_symbols - len(free_symbols)): ranges.append(Tuple(Dummy()) + default_range) ranges = Tuple(*ranges) plots = [expr + ranges for expr in exprs] return plots elif isinstance(args[0], Tuple) and len(args[0]) == expr_len + nb_of_free_symbols: for arg in args: for i in range(expr_len): if not isinstance(arg[i], Expr): raise ValueError("Expected an expression, given %s" % str(arg[i])) for i in range(nb_of_free_symbols): if not len(arg[i + expr_len]) == 3: raise ValueError("The ranges should be a tuple of " "length 3, got %s" % str(arg[i + expr_len])) return args
true
true
f70e9f82018e07c830743995d2a988adcd978449
53,574
py
Python
qutip/qip/circuit.py
lebinyu/qutip
19f172cbb1689f3295dd2561057543da342fb646
[ "BSD-3-Clause" ]
null
null
null
qutip/qip/circuit.py
lebinyu/qutip
19f172cbb1689f3295dd2561057543da342fb646
[ "BSD-3-Clause" ]
null
null
null
qutip/qip/circuit.py
lebinyu/qutip
19f172cbb1689f3295dd2561057543da342fb646
[ "BSD-3-Clause" ]
null
null
null
# This file is part of QuTiP: Quantum Toolbox in Python. # # Copyright (c) 2011 and later, Paul D. Nation and Robert J. Johansson. # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are # met: # # 1. Redistributions of source code must retain the above copyright notice, # this list of conditions and the following disclaimer. # # 2. Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # 3. Neither the name of the QuTiP: Quantum Toolbox in Python nor the names # of its contributors may be used to endorse or promote products derived # from this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A # PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT # HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, # SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT # LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. ############################################################################### from collections.abc import Iterable import warnings import inspect import numpy as np import numpy as np import numpy as np from qutip.qip.circuit_latex import _latex_compile from qutip.qip.gates import * from qutip.qip.qubits import qubit_states __all__ = ['Gate', 'QubitCircuit'] class Gate(object): """ Representation of a quantum gate, with its required parametrs, and target and control qubits. Parameters ---------- name : string Gate name. targets : list or int Gate targets. controls : list or int Gate controls. arg_value : float Argument value(phi). arg_label : string Label for gate representation. """ def __init__(self, name, targets=None, controls=None, arg_value=None, arg_label=None): """ Create a gate with specified parameters. """ self.name = name self.targets = None self.controls = None if not isinstance(targets, Iterable) and targets is not None: self.targets = [targets] else: self.targets = targets if not isinstance(controls, Iterable) and controls is not None: self.controls = [controls] else: self.controls = controls for ind_list in [self.targets, self.controls]: if isinstance(ind_list, Iterable): all_integer = all( [isinstance(ind, np.int) for ind in ind_list]) if not all_integer: raise ValueError("Index of a qubit must be an integer") if name in ["SWAP", "ISWAP", "SQRTISWAP", "SQRTSWAP", "BERKELEY", "SWAPalpha"]: if (self.targets is None) or (len(self.targets) != 2): raise ValueError("Gate %s requires two targets" % name) if self.controls is not None: raise ValueError("Gate %s cannot have a control" % name) elif name in ["CNOT", "CSIGN", "CRX", "CRY", "CRZ"]: if self.targets is None or len(self.targets) != 1: raise ValueError("Gate %s requires one target" % name) if self.controls is None or len(self.controls) != 1: raise ValueError("Gate %s requires one control" % name) elif name in ["SNOT", "RX", "RY", "RZ", "PHASEGATE"]: if self.controls is not None: raise ValueError("Gate %s does not take controls" % name) elif name in ["RX", "RY", "RZ", "CPHASE", "SWAPalpha", "PHASEGATE", "GLOBALPHASE", "CRX", "CRY", "CRZ"]: if arg_value is None: raise ValueError("Gate %s requires an argument value" % name) self.arg_value = arg_value self.arg_label = arg_label def __str__(self): s = "Gate(%s, targets=%s, controls=%s)" % (self.name, self.targets, self.controls) return s def __repr__(self): return str(self) def _repr_latex_(self): return str(self) _gate_name_to_label = { 'RX': r'R_x', 'RY': r'R_y', 'RZ': r'R_z', 'CRX': r'R_x', 'CRY': r'R_y', 'CRZ': r'R_z', 'SQRTNOT': r'\sqrt{\rm NOT}', 'SNOT': r'{\rm H}', 'PHASEGATE': r'{\rm PHASE}', 'CPHASE': r'{\rm R}', 'CNOT': r'{\rm CNOT}', 'CSIGN': r'{\rm Z}', 'BERKELEY': r'{\rm BERKELEY}', 'SWAPalpha': r'{\rm SWAPalpha}', 'SWAP': r'{\rm SWAP}', 'ISWAP': r'{i}{\rm SWAP}', 'SQRTSWAP': r'\sqrt{\rm SWAP}', 'SQRTISWAP': r'\sqrt{{i}\rm SWAP}', 'FREDKIN': r'{\rm FREDKIN}', 'TOFFOLI': r'{\rm TOFFOLI}', 'GLOBALPHASE': r'{\rm Ph}', } def _gate_label(name, arg_label): if name in _gate_name_to_label: gate_label = _gate_name_to_label[name] else: warnings.warn("Unknown gate %s" % name) gate_label = name if arg_label: return r'%s(%s)' % (gate_label, arg_label) else: return r'%s' % gate_label class QubitCircuit(object): """ Representation of a quantum program/algorithm, maintaining a sequence of gates. Parameters ---------- N : int Number of qubits in the system. user_gates : dict Define a dictionary of the custom gates. See examples for detail. input_states : list A list of string such as `0`,'+', "A", "Y". Only used for latex. Examples -------- >>> def user_gate(): ... mat = np.array([[1., 0], ... [0., 1.j]]) ... return Qobj(mat, dims=[[2], [2]]) >>> qc.QubitCircuit(2, user_gates={"T":user_gate}) >>> qc.add_gate("T", targets=[0]) """ def __init__(self, N, input_states=None, output_states=None, reverse_states=True, user_gates=None): # number of qubits in the register self.N = N self.reverse_states = reverse_states self.gates = [] self.U_list = [] self.input_states = [None for i in range(N)] self.output_states = [None for i in range(N)] if user_gates is None: self.user_gates = {} else: if isinstance(user_gates, dict): self.user_gates = user_gates else: raise ValueError( "`user_gate` takes a python dictionary of the form" "{{str: gate_function}}, not {}".format(user_gates)) def add_state(self, state, targets=None, state_type="input"): """ Add an input or ouput state to the circuit. By default all the input and output states will be initialized to `None`. A particular state can be added by specifying the state and the qubit where it has to be added along with the type as input or output. Parameters ---------- state: str The state that has to be added. It can be any string such as `0`, '+', "A", "Y" targets: list A list of qubit positions where the given state has to be added. state_type: str One of either "input" or "output". This specifies whether the state to be added is an input or output. default: "input" """ if state_type == "input": for i in targets: self.input_states[i] = state if state_type == "output": for i in targets: self.output_states[i] = state def add_gate(self, gate, targets=None, controls=None, arg_value=None, arg_label=None, index=None): """ Adds a gate with specified parameters to the circuit. Parameters ---------- gate: string or `Gate` Gate name. If gate is an instance of `Gate`, parameters are unpacked and added. targets: list Gate targets. controls: list Gate controls. arg_value: float Argument value(phi). arg_label: string Label for gate representation. index : list Positions to add the gate. """ if isinstance(gate, Gate): name = gate.name targets = gate.targets controls = gate.controls arg_value = gate.arg_value arg_label = gate.arg_label else: name = gate if index is None: self.gates.append(Gate(name, targets=targets, controls=controls, arg_value=arg_value, arg_label=arg_label)) else: for position in index: self.gates.insert(position, Gate(name, targets=targets, controls=controls, arg_value=arg_value, arg_label=arg_label)) def add_1q_gate(self, name, start=0, end=None, qubits=None, arg_value=None, arg_label=None): """ Adds a single qubit gate with specified parameters on a variable number of qubits in the circuit. By default, it applies the given gate to all the qubits in the register. Parameters ---------- name : string Gate name. start : int Starting location of qubits. end : int Last qubit for the gate. qubits : list Specific qubits for applying gates. arg_value : float Argument value(phi). arg_label : string Label for gate representation. """ if name not in ["RX", "RY", "RZ", "SNOT", "SQRTNOT", "PHASEGATE"]: raise ValueError("%s is not a single qubit gate" % name) if qubits is not None: for i in range(len(qubits)): self.gates.append(Gate(name, targets=qubits[i], controls=None, arg_value=arg_value, arg_label=arg_label)) else: if end is None: end = self.N - 1 for i in range(start, end): self.gates.append(Gate(name, targets=i, controls=None, arg_value=arg_value, arg_label=arg_label)) def add_circuit(self, qc, start=0): """ Adds a block of a qubit circuit to the main circuit. Globalphase gates are not added. Parameters ---------- qc : QubitCircuit The circuit block to be added to the main circuit. start : int The qubit on which the first gate is applied. """ if self.N - start < qc.N: raise NotImplementedError("Targets exceed number of qubits.") for gate in qc.gates: if gate.name in ["RX", "RY", "RZ", "SNOT", "SQRTNOT", "PHASEGATE"]: self.add_gate(gate.name, gate.targets[0] + start, None, gate.arg_value, gate.arg_label) elif gate.name in ["CPHASE", "CNOT", "CSIGN", "CRX", "CRY", "CRZ"]: self.add_gate(gate.name, gate.targets[0] + start, gate.controls[0] + start, gate.arg_value, gate.arg_label) elif gate.name in ["BERKELEY", "SWAPalpha", "SWAP", "ISWAP", "SQRTSWAP", "SQRTISWAP"]: self.add_gate(gate.name, None, [gate.controls[0] + start, gate.controls[1] + start], None, None) elif gate.name in ["TOFFOLI"]: self.add_gate(gate.name, gate.targets[0] + start, [gate.controls[0] + start, gate.controls[1] + start], None, None) elif gate.name in ["FREDKIN"]: self.add_gate(gate.name, [gate.targets[0] + start, gate.targets[1] + start], gate.controls + start, None, None) elif gate.name in self.user_gates: self.add_gate( gate.name, targets=gate.targets, arg_value=gate.arg_value) def remove_gate(self, index=None, end=None, name=None, remove="first"): """ Remove a gate from a specific index or between two indexes or the first, last or all instances of a particular gate. Parameters ---------- index : int Location of gate to be removed. name : string Gate name to be removed. remove : string If first or all gate are to be removed. """ if index is not None and index <= self.N: if end is not None and end <= self.N: for i in range(end - index): self.gates.pop(index + i) elif end is not None and end > self.N: raise ValueError("End target exceeds number of gates.") else: self.gates.pop(index) elif name is not None and remove == "first": for gate in self.gates: if name == gate.name: self.gates.remove(gate) break elif name is not None and remove == "last": for i in range(self.N + 1): if name == self.gates[self.N - i].name: self.gates.remove(self.gates[self.N - i]) break elif name is not None and remove == "all": for j in range(self.N + 1): if name == self.gates[self.N - j].name: self.gates.remove(self.gates[self.N - j]) else: self.gates.pop() def reverse_circuit(self): """ Reverse an entire circuit of unitary gates. Returns ---------- qc : QubitCircuit Return QubitCircuit of resolved gates for the qubit circuit in the reverse order. """ temp = QubitCircuit(self.N, self.reverse_states) for gate in reversed(self.gates): temp.add_gate(gate) return temp def resolve_gates(self, basis=["CNOT", "RX", "RY", "RZ"]): """ Unitary matrix calculator for N qubits returning the individual steps as unitary matrices operating from left to right in the specified basis. Parameters ---------- basis : list. Basis of the resolved circuit. Returns ------- qc : QubitCircuit Return QubitCircuit of resolved gates for the qubit circuit in the desired basis. """ qc_temp = QubitCircuit(self.N, self.reverse_states) temp_resolved = [] basis_1q_valid = ["RX", "RY", "RZ"] basis_2q_valid = ["CNOT", "CSIGN", "ISWAP", "SQRTSWAP", "SQRTISWAP"] if isinstance(basis, list): basis_1q = [] basis_2q = [] for gate in basis: if gate in basis_2q_valid: basis_2q.append(gate) elif gate in basis_1q_valid: basis_1q.append(gate) else: raise NotImplementedError( "%s is not a valid basis gate" % gate) if len(basis_1q) == 1: raise ValueError("Not sufficient single-qubit gates in basis") elif len(basis_1q) == 0: basis_1q = ["RX", "RY", "RZ"] else: # only one 2q gate is given as basis basis_1q = ["RX", "RY", "RZ"] if basis in basis_2q_valid: basis_2q = [basis] else: raise ValueError("%s is not a valid two-qubit basis gate" % basis) for gate in self.gates: if gate.name == "RX": temp_resolved.append(gate) elif gate.name == "RY": temp_resolved.append(gate) elif gate.name == "RZ": temp_resolved.append(gate) elif gate.name == "SQRTNOT": temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "SNOT": temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) elif gate.name == "PHASEGATE": temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=gate.arg_value / 2, arg_label=gate.arg_label)) temp_resolved.append(Gate("RZ", gate.targets, None, gate.arg_value, gate.arg_label)) elif gate.name in basis_2q: # ignore all gate in 2q basis temp_resolved.append(gate) elif gate.name == "CPHASE": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "CNOT": temp_resolved.append(gate) elif gate.name == "CSIGN": temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls)) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi, arg_label=r"\pi")) elif gate.name == "BERKELEY": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "SWAPalpha": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "SWAP": if "ISWAP" in basis_2q: # dealed with separately temp_resolved.append(gate) else: temp_resolved.append( Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append( Gate("CNOT", gate.targets[1], gate.targets[0])) temp_resolved.append( Gate("CNOT", gate.targets[0], gate.targets[1])) elif gate.name == "ISWAP": temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.targets[0])) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append(Gate("RZ", gate.targets[0], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets[0], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append(Gate("RY", gate.targets[0], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "SQRTSWAP": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "SQRTISWAP": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "FREDKIN": temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.controls)) temp_resolved.append(Gate("RZ", gate.controls, None, arg_value=np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("RZ", [gate.targets[0]], None, arg_value=-np.pi / 8, arg_label=r"-\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.controls)) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets[1], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RZ", gate.targets[0], None, arg_value=np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.controls)) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=-np.pi / 8, arg_label=r"-\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.targets[0])) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.controls)) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=-np.pi / 8, arg_label=r"-\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.targets[0])) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets[1], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) elif gate.name == "TOFFOLI": temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=1 * np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("RZ", gate.controls[1], None, arg_value=np.pi/2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RZ", gate.controls[0], None, arg_value=np.pi / 4, arg_label=r"\pi/4")) temp_resolved.append(Gate("CNOT", gate.controls[1], gate.controls[0])) temp_resolved.append(Gate("RZ", gate.controls[1], None, arg_value=-np.pi / 4, arg_label=r"-\pi/4")) temp_resolved.append(Gate("CNOT", gate.controls[1], gate.controls[0])) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("RZ", gate.controls[1], None, arg_value=-np.pi / 4, arg_label=r"-\pi/4")) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls[0])) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 4, arg_label=r"-\pi/4")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls[1])) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls[0])) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 4, arg_label=r"-\pi/4")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls[1])) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) elif gate.name == "GLOBALPHASE": temp_resolved.append(Gate(gate.name, gate.targets, gate.controls, gate.arg_value, gate.arg_label)) else: raise NotImplementedError( "Gate {} " "cannot be resolved.".format(gate.name)) if "CSIGN" in basis_2q: for gate in temp_resolved: if gate.name == "CNOT": qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("CSIGN", gate.targets, gate.controls)) qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) else: qc_temp.gates.append(gate) elif "ISWAP" in basis_2q: for gate in temp_resolved: if gate.name == "CNOT": qc_temp.gates.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) qc_temp.gates.append(Gate("ISWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RY", gate.controls, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RZ", gate.controls, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) qc_temp.gates.append(Gate("ISWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "SWAP": qc_temp.gates.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) qc_temp.gates.append(Gate("ISWAP", gate.targets, None)) qc_temp.gates.append(Gate("RX", gate.targets[0], None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("ISWAP", gate.targets, None)) qc_temp.gates.append(Gate("RX", gate.targets[1], None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("ISWAP", [gate.targets[1], gate.targets[0]], None)) qc_temp.gates.append(Gate("RX", gate.targets[0], None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) else: qc_temp.gates.append(gate) elif "SQRTSWAP" in basis_2q: for gate in temp_resolved: if gate.name == "CNOT": qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) qc_temp.gates.append(Gate("SQRTSWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RZ", gate.controls, None, arg_value=np.pi, arg_label=r"\pi")) qc_temp.gates.append(Gate("SQRTSWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RZ", gate.controls, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) else: qc_temp.gates.append(gate) elif "SQRTISWAP" in basis_2q: for gate in temp_resolved: if gate.name == "CNOT": qc_temp.gates.append(Gate("RY", gate.controls, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RX", gate.controls, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) qc_temp.gates.append(Gate("RX", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("SQRTISWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RX", gate.controls, None, arg_value=np.pi, arg_label=r"\pi")) qc_temp.gates.append(Gate("SQRTISWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RY", gate.controls, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) qc_temp.gates.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) qc_temp.gates.append(Gate("RZ", gate.controls, None, arg_value=np.pi, arg_label=r"\pi")) qc_temp.gates.append(Gate("GLOBALPHASE", None, None, arg_value=3 * np.pi / 2, arg_label=r"3\pi/2")) else: qc_temp.gates.append(gate) else: qc_temp.gates = temp_resolved if len(basis_1q) == 2: temp_resolved = qc_temp.gates qc_temp.gates = [] for gate in temp_resolved: if gate.name == "RX" and "RX" not in basis_1q: qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RZ", gate.targets, None, gate.arg_value, gate.arg_label)) qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "RY" and "RY" not in basis_1q: qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RX", gate.targets, None, gate.arg_value, gate.arg_label)) qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "RZ" and "RZ" not in basis_1q: qc_temp.gates.append(Gate("RX", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RY", gate.targets, None, gate.arg_value, gate.arg_label)) qc_temp.gates.append(Gate("RX", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) else: qc_temp.gates.append(gate) return qc_temp def adjacent_gates(self): """ Method to resolve two qubit gates with non-adjacent control/s or target/s in terms of gates with adjacent interactions. Returns ------- qc : QubitCircuit Return QubitCircuit of the gates for the qubit circuit with the resolved non-adjacent gates. """ temp = QubitCircuit(self.N, reverse_states=self.reverse_states) swap_gates = ["SWAP", "ISWAP", "SQRTISWAP", "SQRTSWAP", "BERKELEY", "SWAPalpha"] for gate in self.gates: if gate.name == "CNOT" or gate.name == "CSIGN": start = min([gate.targets[0], gate.controls[0]]) end = max([gate.targets[0], gate.controls[0]]) i = start while i < end: if start + end - i - i == 1 and (end - start + 1) % 2 == 0: # Apply required gate if control, target are adjacent # to each other, provided |control-target| is even. if end == gate.controls[0]: temp.gates.append(Gate(gate.name, targets=[i], controls=[i + 1])) else: temp.gates.append(Gate(gate.name, targets=[i + 1], controls=[i])) elif (start + end - i - i == 2 and (end - start + 1) % 2 == 1): # Apply a swap between i and its adjacent gate, then # the required gate if and then another swap if control # and target have one qubit between them, provided # |control-target| is odd. temp.gates.append(Gate("SWAP", targets=[i, i + 1])) if end == gate.controls[0]: temp.gates.append(Gate(gate.name, targets=[i + 1], controls=[i + 2])) else: temp.gates.append(Gate(gate.name, targets=[i + 2], controls=[i + 1])) temp.gates.append(Gate("SWAP", targets=[i, i + 1])) i += 1 else: # Swap the target/s and/or control with their adjacent # qubit to bring them closer. temp.gates.append(Gate("SWAP", targets=[i, i + 1])) temp.gates.append(Gate("SWAP", targets=[start + end - i - 1, start + end - i])) i += 1 elif gate.name in swap_gates: start = min([gate.targets[0], gate.targets[1]]) end = max([gate.targets[0], gate.targets[1]]) i = start while i < end: if start + end - i - i == 1 and (end - start + 1) % 2 == 0: temp.gates.append(Gate(gate.name, targets=[i, i + 1])) elif ((start + end - i - i) == 2 and (end - start + 1) % 2 == 1): temp.gates.append(Gate("SWAP", targets=[i, i + 1])) temp.gates.append( Gate(gate.name, targets=[i + 1, i + 2])) temp.gates.append(Gate("SWAP", targets=[i, i + 1])) i += 1 else: temp.gates.append(Gate("SWAP", targets=[i, i + 1])) temp.gates.append(Gate("SWAP", targets=[start + end - i - 1, start + end - i])) i += 1 else: raise NotImplementedError( "`adjacent_gates` is not defined for " "gate {}.".format(gate.name)) return temp def propagators(self): """ Propagator matrix calculator for N qubits returning the individual steps as unitary matrices operating from left to right. Returns ------- U_list : list Return list of unitary matrices for the qubit circuit. """ self.U_list = [] for gate in self.gates: if gate.name == "RX": self.U_list.append(rx(gate.arg_value, self.N, gate.targets[0])) elif gate.name == "RY": self.U_list.append(ry(gate.arg_value, self.N, gate.targets[0])) elif gate.name == "RZ": self.U_list.append(rz(gate.arg_value, self.N, gate.targets[0])) elif gate.name == "SQRTNOT": self.U_list.append(sqrtnot(self.N, gate.targets[0])) elif gate.name == "SNOT": self.U_list.append(snot(self.N, gate.targets[0])) elif gate.name == "PHASEGATE": self.U_list.append(phasegate(gate.arg_value, self.N, gate.targets[0])) elif gate.name == "CRX": self.U_list.append(controlled_gate(rx(gate.arg_value), N=self.N, control=gate.controls[0], target=gate.targets[0])) elif gate.name == "CRY": self.U_list.append(controlled_gate(ry(gate.arg_value), N=self.N, control=gate.controls[0], target=gate.targets[0])) elif gate.name == "CRZ": self.U_list.append(controlled_gate(rz(gate.arg_value), N=self.N, control=gate.controls[0], target=gate.targets[0])) elif gate.name == "CPHASE": self.U_list.append(cphase(gate.arg_value, self.N, gate.controls[0], gate.targets[0])) elif gate.name == "CNOT": self.U_list.append(cnot(self.N, gate.controls[0], gate.targets[0])) elif gate.name == "CSIGN": self.U_list.append(csign(self.N, gate.controls[0], gate.targets[0])) elif gate.name == "BERKELEY": self.U_list.append(berkeley(self.N, gate.targets)) elif gate.name == "SWAPalpha": self.U_list.append(swapalpha(gate.arg_value, self.N, gate.targets)) elif gate.name == "SWAP": self.U_list.append(swap(self.N, gate.targets)) elif gate.name == "ISWAP": self.U_list.append(iswap(self.N, gate.targets)) elif gate.name == "SQRTSWAP": self.U_list.append(sqrtswap(self.N, gate.targets)) elif gate.name == "SQRTISWAP": self.U_list.append(sqrtiswap(self.N, gate.targets)) elif gate.name == "FREDKIN": self.U_list.append(fredkin(self.N, gate.controls[0], gate.targets)) elif gate.name == "TOFFOLI": self.U_list.append(toffoli(self.N, gate.controls, gate.targets[0])) elif gate.name == "GLOBALPHASE": self.U_list.append(globalphase(gate.arg_value, self.N)) elif gate.name in self.user_gates: if gate.controls is not None: raise ValueError( "A user defined gate {} takes only " "`targets` variable.".format(gate.name)) func = self.user_gates[gate.name] para_num = len(inspect.getfullargspec(func)[0]) if para_num == 0: oper = func() elif para_num == 1: oper = func(gate.arg_value) else: raise ValueError( "gate function takes at most one parameters.") self.U_list.append(expand_oper(oper, self.N, gate.targets)) else: raise NotImplementedError( "{} gate is an unknown gate.".format(gate.name)) return self.U_list def latex_code(self): rows = [] gates = self.gates for gate in gates: col = [] for n in range(self.N): if gate.targets and n in gate.targets: if len(gate.targets) > 1: if gate.name == "SWAP": col.append(r" \qswap \qwx ") elif ((self.reverse_states and n == max(gate.targets)) or (not self.reverse_states and n == min(gate.targets))): col.append(r" \multigate{%d}{%s} " % (len(gate.targets) - 1, _gate_label(gate.name, gate.arg_label))) else: col.append(r" \ghost{%s} " % (_gate_label(gate.name, gate.arg_label))) elif gate.name == "CNOT": col.append(r" \targ ") elif gate.name == "TOFFOLI": col.append(r" \targ ") else: col.append(r" \gate{%s} " % _gate_label(gate.name, gate.arg_label)) elif gate.controls and n in gate.controls: m = (gate.targets[0] - n) * (-1 if self.reverse_states else 1) col.append(r" \ctrl{%d} " % m) elif (not gate.controls and not gate.targets): # global gate if ((self.reverse_states and n == self.N - 1) or (not self.reverse_states and n == 0)): col.append(r" \multigate{%d}{%s} " % (self.N - 1, _gate_label(gate.name, gate.arg_label))) else: col.append(r" \ghost{%s} " % (_gate_label(gate.name, gate.arg_label))) else: col.append(r" \qw ") col.append(r" \qw ") rows.append(col) input_states = ["\lstick{\ket{" + x + "}}" if x is not None else "" for x in self.input_states] code = "" n_iter = (reversed(range(self.N)) if self.reverse_states else range(self.N)) for n in n_iter: code += r" & %s" % input_states[n] for m in range(len(gates)): code += r" & %s" % rows[m][n] code += r" & \qw \\ " + "\n" return code def _repr_png_(self): return _latex_compile(self.latex_code(), format="png") def _repr_svg_(self): return _latex_compile(self.latex_code(), format="svg") @property def png(self): from IPython.display import Image return Image(self._repr_png_(), embed=True) @property def svg(self): from IPython.display import SVG return SVG(self._repr_svg_()) def qasm(self): code = "# qasm code generated by QuTiP\n\n" for n in range(self.N): code += "\tqubit\tq%d\n" % n code += "\n" for gate in self.gates: code += "\t%s\t" % gate.name qtargets = ["q%d" % t for t in gate.targets] if gate.targets else [] qcontrols = (["q%d" % c for c in gate.controls] if gate.controls else []) code += ",".join(qtargets + qcontrols) code += "\n" return code
46.707934
79
0.434987
, 'ISWAP': r'{i}{\rm SWAP}', 'SQRTSWAP': r'\sqrt{\rm SWAP}', 'SQRTISWAP': r'\sqrt{{i}\rm SWAP}', 'FREDKIN': r'{\rm FREDKIN}', 'TOFFOLI': r'{\rm TOFFOLI}', 'GLOBALPHASE': r'{\rm Ph}', } def _gate_label(name, arg_label): if name in _gate_name_to_label: gate_label = _gate_name_to_label[name] else: warnings.warn("Unknown gate %s" % name) gate_label = name if arg_label: return r'%s(%s)' % (gate_label, arg_label) else: return r'%s' % gate_label class QubitCircuit(object): def __init__(self, N, input_states=None, output_states=None, reverse_states=True, user_gates=None): self.N = N self.reverse_states = reverse_states self.gates = [] self.U_list = [] self.input_states = [None for i in range(N)] self.output_states = [None for i in range(N)] if user_gates is None: self.user_gates = {} else: if isinstance(user_gates, dict): self.user_gates = user_gates else: raise ValueError( "`user_gate` takes a python dictionary of the form" "{{str: gate_function}}, not {}".format(user_gates)) def add_state(self, state, targets=None, state_type="input"): if state_type == "input": for i in targets: self.input_states[i] = state if state_type == "output": for i in targets: self.output_states[i] = state def add_gate(self, gate, targets=None, controls=None, arg_value=None, arg_label=None, index=None): if isinstance(gate, Gate): name = gate.name targets = gate.targets controls = gate.controls arg_value = gate.arg_value arg_label = gate.arg_label else: name = gate if index is None: self.gates.append(Gate(name, targets=targets, controls=controls, arg_value=arg_value, arg_label=arg_label)) else: for position in index: self.gates.insert(position, Gate(name, targets=targets, controls=controls, arg_value=arg_value, arg_label=arg_label)) def add_1q_gate(self, name, start=0, end=None, qubits=None, arg_value=None, arg_label=None): if name not in ["RX", "RY", "RZ", "SNOT", "SQRTNOT", "PHASEGATE"]: raise ValueError("%s is not a single qubit gate" % name) if qubits is not None: for i in range(len(qubits)): self.gates.append(Gate(name, targets=qubits[i], controls=None, arg_value=arg_value, arg_label=arg_label)) else: if end is None: end = self.N - 1 for i in range(start, end): self.gates.append(Gate(name, targets=i, controls=None, arg_value=arg_value, arg_label=arg_label)) def add_circuit(self, qc, start=0): if self.N - start < qc.N: raise NotImplementedError("Targets exceed number of qubits.") for gate in qc.gates: if gate.name in ["RX", "RY", "RZ", "SNOT", "SQRTNOT", "PHASEGATE"]: self.add_gate(gate.name, gate.targets[0] + start, None, gate.arg_value, gate.arg_label) elif gate.name in ["CPHASE", "CNOT", "CSIGN", "CRX", "CRY", "CRZ"]: self.add_gate(gate.name, gate.targets[0] + start, gate.controls[0] + start, gate.arg_value, gate.arg_label) elif gate.name in ["BERKELEY", "SWAPalpha", "SWAP", "ISWAP", "SQRTSWAP", "SQRTISWAP"]: self.add_gate(gate.name, None, [gate.controls[0] + start, gate.controls[1] + start], None, None) elif gate.name in ["TOFFOLI"]: self.add_gate(gate.name, gate.targets[0] + start, [gate.controls[0] + start, gate.controls[1] + start], None, None) elif gate.name in ["FREDKIN"]: self.add_gate(gate.name, [gate.targets[0] + start, gate.targets[1] + start], gate.controls + start, None, None) elif gate.name in self.user_gates: self.add_gate( gate.name, targets=gate.targets, arg_value=gate.arg_value) def remove_gate(self, index=None, end=None, name=None, remove="first"): if index is not None and index <= self.N: if end is not None and end <= self.N: for i in range(end - index): self.gates.pop(index + i) elif end is not None and end > self.N: raise ValueError("End target exceeds number of gates.") else: self.gates.pop(index) elif name is not None and remove == "first": for gate in self.gates: if name == gate.name: self.gates.remove(gate) break elif name is not None and remove == "last": for i in range(self.N + 1): if name == self.gates[self.N - i].name: self.gates.remove(self.gates[self.N - i]) break elif name is not None and remove == "all": for j in range(self.N + 1): if name == self.gates[self.N - j].name: self.gates.remove(self.gates[self.N - j]) else: self.gates.pop() def reverse_circuit(self): temp = QubitCircuit(self.N, self.reverse_states) for gate in reversed(self.gates): temp.add_gate(gate) return temp def resolve_gates(self, basis=["CNOT", "RX", "RY", "RZ"]): qc_temp = QubitCircuit(self.N, self.reverse_states) temp_resolved = [] basis_1q_valid = ["RX", "RY", "RZ"] basis_2q_valid = ["CNOT", "CSIGN", "ISWAP", "SQRTSWAP", "SQRTISWAP"] if isinstance(basis, list): basis_1q = [] basis_2q = [] for gate in basis: if gate in basis_2q_valid: basis_2q.append(gate) elif gate in basis_1q_valid: basis_1q.append(gate) else: raise NotImplementedError( "%s is not a valid basis gate" % gate) if len(basis_1q) == 1: raise ValueError("Not sufficient single-qubit gates in basis") elif len(basis_1q) == 0: basis_1q = ["RX", "RY", "RZ"] else: basis_1q = ["RX", "RY", "RZ"] if basis in basis_2q_valid: basis_2q = [basis] else: raise ValueError("%s is not a valid two-qubit basis gate" % basis) for gate in self.gates: if gate.name == "RX": temp_resolved.append(gate) elif gate.name == "RY": temp_resolved.append(gate) elif gate.name == "RZ": temp_resolved.append(gate) elif gate.name == "SQRTNOT": temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "SNOT": temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) elif gate.name == "PHASEGATE": temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=gate.arg_value / 2, arg_label=gate.arg_label)) temp_resolved.append(Gate("RZ", gate.targets, None, gate.arg_value, gate.arg_label)) elif gate.name in basis_2q: temp_resolved.append(gate) elif gate.name == "CPHASE": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "CNOT": temp_resolved.append(gate) elif gate.name == "CSIGN": temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls)) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi, arg_label=r"\pi")) elif gate.name == "BERKELEY": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "SWAPalpha": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "SWAP": if "ISWAP" in basis_2q: temp_resolved.append(gate) else: temp_resolved.append( Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append( Gate("CNOT", gate.targets[1], gate.targets[0])) temp_resolved.append( Gate("CNOT", gate.targets[0], gate.targets[1])) elif gate.name == "ISWAP": temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.targets[0])) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append(Gate("RZ", gate.targets[0], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets[0], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append(Gate("RY", gate.targets[0], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "SQRTSWAP": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "SQRTISWAP": raise NotImplementedError("Cannot be resolved in this basis") elif gate.name == "FREDKIN": temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.controls)) temp_resolved.append(Gate("RZ", gate.controls, None, arg_value=np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("RZ", [gate.targets[0]], None, arg_value=-np.pi / 8, arg_label=r"-\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.controls)) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets[1], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RZ", gate.targets[0], None, arg_value=np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.controls)) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=-np.pi / 8, arg_label=r"-\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.targets[0])) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.controls)) temp_resolved.append(Gate("RZ", gate.targets[1], None, arg_value=-np.pi / 8, arg_label=r"-\pi/8")) temp_resolved.append(Gate("CNOT", gate.targets[1], gate.targets[0])) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets[1], None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("CNOT", gate.targets[0], gate.targets[1])) elif gate.name == "TOFFOLI": temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=1 * np.pi / 8, arg_label=r"\pi/8")) temp_resolved.append(Gate("RZ", gate.controls[1], None, arg_value=np.pi/2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RZ", gate.controls[0], None, arg_value=np.pi / 4, arg_label=r"\pi/4")) temp_resolved.append(Gate("CNOT", gate.controls[1], gate.controls[0])) temp_resolved.append(Gate("RZ", gate.controls[1], None, arg_value=-np.pi / 4, arg_label=r"-\pi/4")) temp_resolved.append(Gate("CNOT", gate.controls[1], gate.controls[0])) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) temp_resolved.append(Gate("RZ", gate.controls[1], None, arg_value=-np.pi / 4, arg_label=r"-\pi/4")) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls[0])) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 4, arg_label=r"-\pi/4")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls[1])) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls[0])) temp_resolved.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 4, arg_label=r"-\pi/4")) temp_resolved.append(Gate("CNOT", gate.targets, gate.controls[1])) temp_resolved.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) temp_resolved.append(Gate("RX", gate.targets, None, arg_value=np.pi, arg_label=r"\pi")) elif gate.name == "GLOBALPHASE": temp_resolved.append(Gate(gate.name, gate.targets, gate.controls, gate.arg_value, gate.arg_label)) else: raise NotImplementedError( "Gate {} " "cannot be resolved.".format(gate.name)) if "CSIGN" in basis_2q: for gate in temp_resolved: if gate.name == "CNOT": qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("CSIGN", gate.targets, gate.controls)) qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) else: qc_temp.gates.append(gate) elif "ISWAP" in basis_2q: for gate in temp_resolved: if gate.name == "CNOT": qc_temp.gates.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) qc_temp.gates.append(Gate("ISWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RY", gate.controls, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RZ", gate.controls, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) qc_temp.gates.append(Gate("ISWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "SWAP": qc_temp.gates.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) qc_temp.gates.append(Gate("ISWAP", gate.targets, None)) qc_temp.gates.append(Gate("RX", gate.targets[0], None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("ISWAP", gate.targets, None)) qc_temp.gates.append(Gate("RX", gate.targets[1], None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("ISWAP", [gate.targets[1], gate.targets[0]], None)) qc_temp.gates.append(Gate("RX", gate.targets[0], None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) else: qc_temp.gates.append(gate) elif "SQRTSWAP" in basis_2q: for gate in temp_resolved: if gate.name == "CNOT": qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) qc_temp.gates.append(Gate("SQRTSWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RZ", gate.controls, None, arg_value=np.pi, arg_label=r"\pi")) qc_temp.gates.append(Gate("SQRTSWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RZ", gate.controls, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) else: qc_temp.gates.append(gate) elif "SQRTISWAP" in basis_2q: for gate in temp_resolved: if gate.name == "CNOT": qc_temp.gates.append(Gate("RY", gate.controls, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RX", gate.controls, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) qc_temp.gates.append(Gate("RX", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("SQRTISWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RX", gate.controls, None, arg_value=np.pi, arg_label=r"\pi")) qc_temp.gates.append(Gate("SQRTISWAP", [gate.controls[0], gate.targets[0]], None)) qc_temp.gates.append(Gate("RY", gate.controls, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) qc_temp.gates.append(Gate("GLOBALPHASE", None, None, arg_value=np.pi / 4, arg_label=r"\pi/4")) qc_temp.gates.append(Gate("RZ", gate.controls, None, arg_value=np.pi, arg_label=r"\pi")) qc_temp.gates.append(Gate("GLOBALPHASE", None, None, arg_value=3 * np.pi / 2, arg_label=r"3\pi/2")) else: qc_temp.gates.append(gate) else: qc_temp.gates = temp_resolved if len(basis_1q) == 2: temp_resolved = qc_temp.gates qc_temp.gates = [] for gate in temp_resolved: if gate.name == "RX" and "RX" not in basis_1q: qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RZ", gate.targets, None, gate.arg_value, gate.arg_label)) qc_temp.gates.append(Gate("RY", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "RY" and "RY" not in basis_1q: qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RX", gate.targets, None, gate.arg_value, gate.arg_label)) qc_temp.gates.append(Gate("RZ", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) elif gate.name == "RZ" and "RZ" not in basis_1q: qc_temp.gates.append(Gate("RX", gate.targets, None, arg_value=-np.pi / 2, arg_label=r"-\pi/2")) qc_temp.gates.append(Gate("RY", gate.targets, None, gate.arg_value, gate.arg_label)) qc_temp.gates.append(Gate("RX", gate.targets, None, arg_value=np.pi / 2, arg_label=r"\pi/2")) else: qc_temp.gates.append(gate) return qc_temp def adjacent_gates(self): temp = QubitCircuit(self.N, reverse_states=self.reverse_states) swap_gates = ["SWAP", "ISWAP", "SQRTISWAP", "SQRTSWAP", "BERKELEY", "SWAPalpha"] for gate in self.gates: if gate.name == "CNOT" or gate.name == "CSIGN": start = min([gate.targets[0], gate.controls[0]]) end = max([gate.targets[0], gate.controls[0]]) i = start while i < end: if start + end - i - i == 1 and (end - start + 1) % 2 == 0: if end == gate.controls[0]: temp.gates.append(Gate(gate.name, targets=[i], controls=[i + 1])) else: temp.gates.append(Gate(gate.name, targets=[i + 1], controls=[i])) elif (start + end - i - i == 2 and (end - start + 1) % 2 == 1): temp.gates.append(Gate("SWAP", targets=[i, i + 1])) if end == gate.controls[0]: temp.gates.append(Gate(gate.name, targets=[i + 1], controls=[i + 2])) else: temp.gates.append(Gate(gate.name, targets=[i + 2], controls=[i + 1])) temp.gates.append(Gate("SWAP", targets=[i, i + 1])) i += 1 else: temp.gates.append(Gate("SWAP", targets=[i, i + 1])) temp.gates.append(Gate("SWAP", targets=[start + end - i - 1, start + end - i])) i += 1 elif gate.name in swap_gates: start = min([gate.targets[0], gate.targets[1]]) end = max([gate.targets[0], gate.targets[1]]) i = start while i < end: if start + end - i - i == 1 and (end - start + 1) % 2 == 0: temp.gates.append(Gate(gate.name, targets=[i, i + 1])) elif ((start + end - i - i) == 2 and (end - start + 1) % 2 == 1): temp.gates.append(Gate("SWAP", targets=[i, i + 1])) temp.gates.append( Gate(gate.name, targets=[i + 1, i + 2])) temp.gates.append(Gate("SWAP", targets=[i, i + 1])) i += 1 else: temp.gates.append(Gate("SWAP", targets=[i, i + 1])) temp.gates.append(Gate("SWAP", targets=[start + end - i - 1, start + end - i])) i += 1 else: raise NotImplementedError( "`adjacent_gates` is not defined for " "gate {}.".format(gate.name)) return temp def propagators(self): self.U_list = [] for gate in self.gates: if gate.name == "RX": self.U_list.append(rx(gate.arg_value, self.N, gate.targets[0])) elif gate.name == "RY": self.U_list.append(ry(gate.arg_value, self.N, gate.targets[0])) elif gate.name == "RZ": self.U_list.append(rz(gate.arg_value, self.N, gate.targets[0])) elif gate.name == "SQRTNOT": self.U_list.append(sqrtnot(self.N, gate.targets[0])) elif gate.name == "SNOT": self.U_list.append(snot(self.N, gate.targets[0])) elif gate.name == "PHASEGATE": self.U_list.append(phasegate(gate.arg_value, self.N, gate.targets[0])) elif gate.name == "CRX": self.U_list.append(controlled_gate(rx(gate.arg_value), N=self.N, control=gate.controls[0], target=gate.targets[0])) elif gate.name == "CRY": self.U_list.append(controlled_gate(ry(gate.arg_value), N=self.N, control=gate.controls[0], target=gate.targets[0])) elif gate.name == "CRZ": self.U_list.append(controlled_gate(rz(gate.arg_value), N=self.N, control=gate.controls[0], target=gate.targets[0])) elif gate.name == "CPHASE": self.U_list.append(cphase(gate.arg_value, self.N, gate.controls[0], gate.targets[0])) elif gate.name == "CNOT": self.U_list.append(cnot(self.N, gate.controls[0], gate.targets[0])) elif gate.name == "CSIGN": self.U_list.append(csign(self.N, gate.controls[0], gate.targets[0])) elif gate.name == "BERKELEY": self.U_list.append(berkeley(self.N, gate.targets)) elif gate.name == "SWAPalpha": self.U_list.append(swapalpha(gate.arg_value, self.N, gate.targets)) elif gate.name == "SWAP": self.U_list.append(swap(self.N, gate.targets)) elif gate.name == "ISWAP": self.U_list.append(iswap(self.N, gate.targets)) elif gate.name == "SQRTSWAP": self.U_list.append(sqrtswap(self.N, gate.targets)) elif gate.name == "SQRTISWAP": self.U_list.append(sqrtiswap(self.N, gate.targets)) elif gate.name == "FREDKIN": self.U_list.append(fredkin(self.N, gate.controls[0], gate.targets)) elif gate.name == "TOFFOLI": self.U_list.append(toffoli(self.N, gate.controls, gate.targets[0])) elif gate.name == "GLOBALPHASE": self.U_list.append(globalphase(gate.arg_value, self.N)) elif gate.name in self.user_gates: if gate.controls is not None: raise ValueError( "A user defined gate {} takes only " "`targets` variable.".format(gate.name)) func = self.user_gates[gate.name] para_num = len(inspect.getfullargspec(func)[0]) if para_num == 0: oper = func() elif para_num == 1: oper = func(gate.arg_value) else: raise ValueError( "gate function takes at most one parameters.") self.U_list.append(expand_oper(oper, self.N, gate.targets)) else: raise NotImplementedError( "{} gate is an unknown gate.".format(gate.name)) return self.U_list def latex_code(self): rows = [] gates = self.gates for gate in gates: col = [] for n in range(self.N): if gate.targets and n in gate.targets: if len(gate.targets) > 1: if gate.name == "SWAP": col.append(r" \qswap \qwx ") elif ((self.reverse_states and n == max(gate.targets)) or (not self.reverse_states and n == min(gate.targets))): col.append(r" \multigate{%d}{%s} " % (len(gate.targets) - 1, _gate_label(gate.name, gate.arg_label))) else: col.append(r" \ghost{%s} " % (_gate_label(gate.name, gate.arg_label))) elif gate.name == "CNOT": col.append(r" \targ ") elif gate.name == "TOFFOLI": col.append(r" \targ ") else: col.append(r" \gate{%s} " % _gate_label(gate.name, gate.arg_label)) elif gate.controls and n in gate.controls: m = (gate.targets[0] - n) * (-1 if self.reverse_states else 1) col.append(r" \ctrl{%d} " % m) elif (not gate.controls and not gate.targets): if ((self.reverse_states and n == self.N - 1) or (not self.reverse_states and n == 0)): col.append(r" \multigate{%d}{%s} " % (self.N - 1, _gate_label(gate.name, gate.arg_label))) else: col.append(r" \ghost{%s} " % (_gate_label(gate.name, gate.arg_label))) else: col.append(r" \qw ") col.append(r" \qw ") rows.append(col) input_states = ["\lstick{\ket{" + x + "}}" if x is not None else "" for x in self.input_states] code = "" n_iter = (reversed(range(self.N)) if self.reverse_states else range(self.N)) for n in n_iter: code += r" & %s" % input_states[n] for m in range(len(gates)): code += r" & %s" % rows[m][n] code += r" & \qw \\ " + "\n" return code def _repr_png_(self): return _latex_compile(self.latex_code(), format="png") def _repr_svg_(self): return _latex_compile(self.latex_code(), format="svg") @property def png(self): from IPython.display import Image return Image(self._repr_png_(), embed=True) @property def svg(self): from IPython.display import SVG return SVG(self._repr_svg_()) def qasm(self): code = "# qasm code generated by QuTiP\n\n" for n in range(self.N): code += "\tqubit\tq%d\n" % n code += "\n" for gate in self.gates: code += "\t%s\t" % gate.name qtargets = ["q%d" % t for t in gate.targets] if gate.targets else [] qcontrols = (["q%d" % c for c in gate.controls] if gate.controls else []) code += ",".join(qtargets + qcontrols) code += "\n" return code
true
true
f70ea0db8a6447a49d3bf80ab3739ca2d79b0877
5,469
py
Python
docs/source/conf.py
holavpv/osmnx
df1feb06f2c2be3c042f524428e3993f22560ee6
[ "MIT" ]
1
2020-07-09T22:11:52.000Z
2020-07-09T22:11:52.000Z
docs/source/conf.py
holavpv/osmnx
df1feb06f2c2be3c042f524428e3993f22560ee6
[ "MIT" ]
null
null
null
docs/source/conf.py
holavpv/osmnx
df1feb06f2c2be3c042f524428e3993f22560ee6
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ OSMnx documentation build configuration file. Created by sphinx-quickstart on Sun Feb 4 13:53:34 2018. This file is execfile()d with the current directory set to its containing dir. Note that not all possible configuration values are present in this autogenerated file. All configuration values have a default; values that are commented out serve to show the default. If extensions (or modules to document with autodoc) are in another directory, add these directories to sys.path here. If the directory is relative to the documentation root, use os.path.abspath to make it absolute, like shown here. """ import os import sys # go up two levels from /docs/source to the package root sys.path.insert(0, os.path.abspath("../..")) # mock import these packages because readthedocs doesn't have them installed autodoc_mock_imports = [ "dateutil", "geopandas", "matplotlib", "matplotlib.cm", "matplotlib.colors", "matplotlib.pyplot", "networkx", "numpy", "pandas", "pyproj", "requests", "scipy", "scipy.spatial", "shapely", "shapely.geometry", "shapely.ops", "sklearn", "sklearn.neighbors", ] # -- General configuration ------------------------------------------------ # General information about the project. project = "OSMnx" copyright = "2020, Geoff Boeing" author = "Geoff Boeing" # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. version = release = "0.15.1" # If your documentation needs a minimal Sphinx version, state it here. # # needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = ["sphinx.ext.autodoc", "sphinx.ext.napoleon"] # Add any paths that contain templates here, relative to this directory. templates_path = ["_templates"] # The suffix(es) of source filenames. # You can specify multiple suffix as a list of string: # # source_suffix = ['.rst', '.md'] source_suffix = ".rst" # The master toctree document. master_doc = "index" # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. language = None # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. # This patterns also effect to html_static_path and html_extra_path exclude_patterns = [] # The name of the Pygments (syntax highlighting) style to use. pygments_style = "sphinx" # If true, `todo` and `todoList` produce output, else they produce nothing. todo_include_todos = False # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. # html_theme = "default" # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. # # html_theme_options = {} # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = [] # Custom sidebar templates, must be a dictionary that maps document names # to template names. # # This is required for the alabaster theme # refs: http://alabaster.readthedocs.io/en/latest/installation.html#sidebars html_sidebars = { "**": [ "relations.html", # needs 'show_related': True theme option to display "searchbox.html", ] } # -- Options for HTMLHelp output ------------------------------------------ # Output file base name for HTML help builder. htmlhelp_basename = "OSMnxdoc" # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). # # 'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). # # 'pointsize': '10pt', # Additional stuff for the LaTeX preamble. # # 'preamble': '', # Latex figure (float) alignment # # 'figure_align': 'htbp', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ (master_doc, "OSMnx.tex", "OSMnx Documentation", "Geoff Boeing", "manual"), ] # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [(master_doc, "osmnx", "OSMnx Documentation", [author], 1)] # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ( master_doc, "OSMnx", "OSMnx Documentation", author, "OSMnx", "Python for street networks.", "Miscellaneous", ), ]
29.403226
79
0.672518
import os import sys sys.path.insert(0, os.path.abspath("../..")) autodoc_mock_imports = [ "dateutil", "geopandas", "matplotlib", "matplotlib.cm", "matplotlib.colors", "matplotlib.pyplot", "networkx", "numpy", "pandas", "pyproj", "requests", "scipy", "scipy.spatial", "shapely", "shapely.geometry", "shapely.ops", "sklearn", "sklearn.neighbors", ] # -- General configuration ------------------------------------------------ # General information about the project. project = "OSMnx" copyright = "2020, Geoff Boeing" author = "Geoff Boeing" # The version info for the project you're documenting, acts as replacement for version = release = "0.15.1" extensions = ["sphinx.ext.autodoc", "sphinx.ext.napoleon"] templates_path = ["_templates"] source_suffix = ".rst" master_doc = "index" language = None exclude_patterns = [] pygments_style = "sphinx" todo_include_todos = False html_theme = "default" html_static_path = [] ebars = { "**": [ "relations.html", "searchbox.html", ] } htmlhelp_basename = "OSMnxdoc" latex_elements = { } latex_documents = [ (master_doc, "OSMnx.tex", "OSMnx Documentation", "Geoff Boeing", "manual"), ] man_pages = [(master_doc, "osmnx", "OSMnx Documentation", [author], 1)] texinfo_documents = [ ( master_doc, "OSMnx", "OSMnx Documentation", author, "OSMnx", "Python for street networks.", "Miscellaneous", ), ]
true
true
f70ea0e76e8e27653ad98588d7b83e00bb2a047a
5,613
py
Python
tests/unit/test_config.py
nickatnight/praw
6ba5c92e5d5210338c0a2a2755a5e5e226a002fa
[ "BSD-2-Clause" ]
2,360
2015-01-03T18:27:44.000Z
2022-03-26T23:24:49.000Z
tests/unit/test_config.py
nickatnight/praw
6ba5c92e5d5210338c0a2a2755a5e5e226a002fa
[ "BSD-2-Clause" ]
1,187
2015-01-04T18:42:10.000Z
2022-03-28T13:46:33.000Z
tests/unit/test_config.py
nickatnight/praw
6ba5c92e5d5210338c0a2a2755a5e5e226a002fa
[ "BSD-2-Clause" ]
591
2015-01-04T17:33:34.000Z
2022-03-27T20:28:26.000Z
import os import sys from unittest import mock import pytest from praw.config import Config from praw.exceptions import ClientException class TestConfig: @staticmethod def _assert_config_read(environment, mock_config): mock_instance = mock_config.return_value Config.CONFIG = None # Force config file reload prev_environment = {environment: None} for env_name in ["APPDATA", "HOME", "XDG_CONFIG_HOME"]: if env_name in os.environ: prev_environment[env_name] = os.environ[env_name] del os.environ[env_name] os.environ[environment] = "/MOCK" module_dir = os.path.dirname(sys.modules["praw"].__file__) environ_path = os.path.join( "/MOCK", ".config" if environment == "HOME" else "", "praw.ini" ) locations = [ os.path.join(module_dir, "praw.ini"), environ_path, "praw.ini", ] try: Config._load_config() mock_instance.read.assert_called_with(locations) finally: Config.CONFIG = None # Force config file reload for env_name in prev_environment: if prev_environment[env_name] is None: del os.environ[env_name] else: os.environ[env_name] = prev_environment[env_name] def test_check_for_updates__false(self): for value in [False, "False", "other"]: config = Config("DEFAULT", check_for_updates=value) assert config.check_for_updates is False def test_custom__extra_values_set(self): config = Config("DEFAULT", user1="foo", user2="bar") assert config.custom == {"user1": "foo", "user2": "bar"} def test_custom__no_extra_values_set(self): config = Config("DEFAULT") assert config.custom == {} def test_check_for_updates__true(self): for value in [True, "1", "true", "YES", "on"]: config = Config("DEFAULT", check_for_updates=value) assert config.check_for_updates is True @mock.patch("configparser.ConfigParser") def test_load_ini_from_appdata(self, mock_config): self._assert_config_read("APPDATA", mock_config) @mock.patch("configparser.ConfigParser") def test_load_ini_from_home(self, mock_config): self._assert_config_read("HOME", mock_config) @mock.patch("configparser.ConfigParser") def test_load_ini_from_xdg_config_home(self, mock_config): self._assert_config_read("XDG_CONFIG_HOME", mock_config) @mock.patch("configparser.ConfigParser") def test_load_ini_with_no_config_directory(self, mock_config): mock_instance = mock_config.return_value Config.CONFIG = None # Force config file reload prev_environment = {} for key in ["APPDATA", "HOME", "XDG_CONFIG_HOME"]: if key in os.environ: prev_environment[key] = os.environ[key] del os.environ[key] module_dir = os.path.dirname(sys.modules["praw"].__file__) locations = [os.path.join(module_dir, "praw.ini"), "praw.ini"] try: Config._load_config() mock_instance.read.assert_called_with(locations) finally: Config.CONFIG = None # Force config file reload for key, value in prev_environment.items(): os.environ[key] = value def test_short_url(self): config = Config("DEFAULT") assert config.short_url == "https://redd.it" def test_short_url_not_defined(self): config = Config("DEFAULT", short_url=None) with pytest.raises(ClientException) as excinfo: config.short_url assert str(excinfo.value) == "No short domain specified." def test_unset_value_has_useful_string_representation(self): config = Config("DEFAULT", password=Config.CONFIG_NOT_SET) assert str(config.password) == "NotSet" class TestConfigInterpolation: def test_no_interpolation(self): Config.CONFIG = None # Force config file reload with mock.patch.dict( "os.environ", { "APPDATA": os.path.dirname(__file__), "XDG_CONFIG_HOME": os.path.dirname(__file__), }, ): config = Config("INTERPOLATION") assert config.custom["basic_interpolation"] == "%(reddit_url)s" assert config.custom["extended_interpolation"] == "${reddit_url}" def test_basic_interpolation(self): Config.CONFIG = None # Force config file reload with mock.patch.dict( "os.environ", { "APPDATA": os.path.dirname(__file__), "XDG_CONFIG_HOME": os.path.dirname(__file__), }, ): config = Config("INTERPOLATION", config_interpolation="basic") assert config.custom["basic_interpolation"] == config.reddit_url assert config.custom["extended_interpolation"] == "${reddit_url}" def test_extended_interpolation(self): Config.CONFIG = None # Force config file reload with mock.patch.dict( "os.environ", { "APPDATA": os.path.dirname(__file__), "XDG_CONFIG_HOME": os.path.dirname(__file__), }, ): config = Config("INTERPOLATION", config_interpolation="extended") assert config.custom["basic_interpolation"] == "%(reddit_url)s" assert config.custom["extended_interpolation"] == config.reddit_url
37.42
79
0.618742
import os import sys from unittest import mock import pytest from praw.config import Config from praw.exceptions import ClientException class TestConfig: @staticmethod def _assert_config_read(environment, mock_config): mock_instance = mock_config.return_value Config.CONFIG = None prev_environment = {environment: None} for env_name in ["APPDATA", "HOME", "XDG_CONFIG_HOME"]: if env_name in os.environ: prev_environment[env_name] = os.environ[env_name] del os.environ[env_name] os.environ[environment] = "/MOCK" module_dir = os.path.dirname(sys.modules["praw"].__file__) environ_path = os.path.join( "/MOCK", ".config" if environment == "HOME" else "", "praw.ini" ) locations = [ os.path.join(module_dir, "praw.ini"), environ_path, "praw.ini", ] try: Config._load_config() mock_instance.read.assert_called_with(locations) finally: Config.CONFIG = None for env_name in prev_environment: if prev_environment[env_name] is None: del os.environ[env_name] else: os.environ[env_name] = prev_environment[env_name] def test_check_for_updates__false(self): for value in [False, "False", "other"]: config = Config("DEFAULT", check_for_updates=value) assert config.check_for_updates is False def test_custom__extra_values_set(self): config = Config("DEFAULT", user1="foo", user2="bar") assert config.custom == {"user1": "foo", "user2": "bar"} def test_custom__no_extra_values_set(self): config = Config("DEFAULT") assert config.custom == {} def test_check_for_updates__true(self): for value in [True, "1", "true", "YES", "on"]: config = Config("DEFAULT", check_for_updates=value) assert config.check_for_updates is True @mock.patch("configparser.ConfigParser") def test_load_ini_from_appdata(self, mock_config): self._assert_config_read("APPDATA", mock_config) @mock.patch("configparser.ConfigParser") def test_load_ini_from_home(self, mock_config): self._assert_config_read("HOME", mock_config) @mock.patch("configparser.ConfigParser") def test_load_ini_from_xdg_config_home(self, mock_config): self._assert_config_read("XDG_CONFIG_HOME", mock_config) @mock.patch("configparser.ConfigParser") def test_load_ini_with_no_config_directory(self, mock_config): mock_instance = mock_config.return_value Config.CONFIG = None prev_environment = {} for key in ["APPDATA", "HOME", "XDG_CONFIG_HOME"]: if key in os.environ: prev_environment[key] = os.environ[key] del os.environ[key] module_dir = os.path.dirname(sys.modules["praw"].__file__) locations = [os.path.join(module_dir, "praw.ini"), "praw.ini"] try: Config._load_config() mock_instance.read.assert_called_with(locations) finally: Config.CONFIG = None for key, value in prev_environment.items(): os.environ[key] = value def test_short_url(self): config = Config("DEFAULT") assert config.short_url == "https://redd.it" def test_short_url_not_defined(self): config = Config("DEFAULT", short_url=None) with pytest.raises(ClientException) as excinfo: config.short_url assert str(excinfo.value) == "No short domain specified." def test_unset_value_has_useful_string_representation(self): config = Config("DEFAULT", password=Config.CONFIG_NOT_SET) assert str(config.password) == "NotSet" class TestConfigInterpolation: def test_no_interpolation(self): Config.CONFIG = None with mock.patch.dict( "os.environ", { "APPDATA": os.path.dirname(__file__), "XDG_CONFIG_HOME": os.path.dirname(__file__), }, ): config = Config("INTERPOLATION") assert config.custom["basic_interpolation"] == "%(reddit_url)s" assert config.custom["extended_interpolation"] == "${reddit_url}" def test_basic_interpolation(self): Config.CONFIG = None with mock.patch.dict( "os.environ", { "APPDATA": os.path.dirname(__file__), "XDG_CONFIG_HOME": os.path.dirname(__file__), }, ): config = Config("INTERPOLATION", config_interpolation="basic") assert config.custom["basic_interpolation"] == config.reddit_url assert config.custom["extended_interpolation"] == "${reddit_url}" def test_extended_interpolation(self): Config.CONFIG = None with mock.patch.dict( "os.environ", { "APPDATA": os.path.dirname(__file__), "XDG_CONFIG_HOME": os.path.dirname(__file__), }, ): config = Config("INTERPOLATION", config_interpolation="extended") assert config.custom["basic_interpolation"] == "%(reddit_url)s" assert config.custom["extended_interpolation"] == config.reddit_url
true
true
f70ea1c658fc200cff6ab95f636b343fc4f6d6b7
2,057
py
Python
Validation/Performance/scripts/cmsScimarkStop.py
ckamtsikis/cmssw
ea19fe642bb7537cbf58451dcf73aa5fd1b66250
[ "Apache-2.0" ]
13
2015-11-30T15:49:45.000Z
2022-02-08T16:11:30.000Z
Validation/Performance/scripts/cmsScimarkStop.py
ckamtsikis/cmssw
ea19fe642bb7537cbf58451dcf73aa5fd1b66250
[ "Apache-2.0" ]
640
2015-02-11T18:55:47.000Z
2022-03-31T14:12:23.000Z
Validation/Performance/scripts/cmsScimarkStop.py
ckamtsikis/cmssw
ea19fe642bb7537cbf58451dcf73aa5fd1b66250
[ "Apache-2.0" ]
51
2015-08-11T21:01:40.000Z
2022-03-30T07:31:34.000Z
#! /usr/bin/env python #Script to #1-check for cmsScimarkLaunch (infinite loop) scripts #2-kill them #3-report their results using cmsScimarkParser.py from __future__ import print_function import subprocess,os,sys def main(): #Use ps -ef to look for cmsScimarkLaunch processes ps_stdouterr=subprocess.Popen("ps -efww|grep cmsScimarkLaunch|grep -v grep|grep -v 'sh -c'",shell=True,stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout if ps_stdouterr: ps_lines=ps_stdouterr.readlines() #print ps_lines if ps_lines: for line in ps_lines: tokens=line.split() #Look up the PID PID=tokens[1] #Look up the cpu core core=tokens[9] print("Found process:\n%s"%line[:-1]) #to eliminate the extra \n #Kill the PID print("Killing process with PID %s"%PID) kill_stdouterr=subprocess.Popen("kill %s"%PID,shell=True,stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout.read() print(kill_stdouterr) #Harvest the cmsScimark scores #Look for the cmsScimark log: if os.path.exists("cmsScimark_%s.log"%core): #Create the results dir mkdir_stdouterr=subprocess.Popen("mkdir cmsScimarkResults_cpu%s"%core,shell=True,stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout.read() print(mkdir_stdouterr) #Execute the harvesting scrip cmsScimarkParser.py (it is in the release) harvest_stdouterr=subprocess.Popen("cmsScimarkParser.py -i cmsScimark_%s.log -o cmsScimarkResults_cpu%s"%(core,core),shell=True,stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout.read() print(harvest_stdouterr) else: print("No cmsScimark_%s.log file was found for cpu%s, log might be in another directory!"%(core,core)) else: print("No cmsScimarkLaunch processes found in the ps -ef output") return 0 if __name__ == "__main__": sys.exit(main())
45.711111
207
0.656782
from __future__ import print_function import subprocess,os,sys def main(): ps_stdouterr=subprocess.Popen("ps -efww|grep cmsScimarkLaunch|grep -v grep|grep -v 'sh -c'",shell=True,stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout if ps_stdouterr: ps_lines=ps_stdouterr.readlines() if ps_lines: for line in ps_lines: tokens=line.split() PID=tokens[1] core=tokens[9] print("Found process:\n%s"%line[:-1]) print("Killing process with PID %s"%PID) kill_stdouterr=subprocess.Popen("kill %s"%PID,shell=True,stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout.read() print(kill_stdouterr) if os.path.exists("cmsScimark_%s.log"%core): mkdir_stdouterr=subprocess.Popen("mkdir cmsScimarkResults_cpu%s"%core,shell=True,stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout.read() print(mkdir_stdouterr) harvest_stdouterr=subprocess.Popen("cmsScimarkParser.py -i cmsScimark_%s.log -o cmsScimarkResults_cpu%s"%(core,core),shell=True,stdout=subprocess.PIPE, stderr=subprocess.STDOUT).stdout.read() print(harvest_stdouterr) else: print("No cmsScimark_%s.log file was found for cpu%s, log might be in another directory!"%(core,core)) else: print("No cmsScimarkLaunch processes found in the ps -ef output") return 0 if __name__ == "__main__": sys.exit(main())
true
true
f70ea28bcf4185dc675728f2549fd2e9447346be
77,417
py
Python
Graphing_Summaries.py
GrantRoss-Tenki/Malawi-CQC-CSC-OSU-Work
a720e0451579945ba10eafdafe2e0d59a86d5cfb
[ "MIT" ]
null
null
null
Graphing_Summaries.py
GrantRoss-Tenki/Malawi-CQC-CSC-OSU-Work
a720e0451579945ba10eafdafe2e0d59a86d5cfb
[ "MIT" ]
null
null
null
Graphing_Summaries.py
GrantRoss-Tenki/Malawi-CQC-CSC-OSU-Work
a720e0451579945ba10eafdafe2e0d59a86d5cfb
[ "MIT" ]
null
null
null
import matplotlib.pyplot as plt import numpy as np import pandas as pd #from pylab import plot, show, xlim,figure,hold, ylim,legend, boxplot, setup, axes import seaborn as sns # Is this a personal or work computer # Are you graphing for hood or no hood Computer = 'personal' #or 'personal' or 'work' Hood_or_no = 'no_hood' # 'no_hood' or 'hood' #what household do you want to remove make sure it is in ascending order # if there is nothing, then put a placeholder of 1045 or higher Household_removal = [1045] #Household_removal = Household_removal.sort(reverse=False) Household_removal_NO_Hood_fuel_day_adult = [1045] Household_removal_Hood_fuel_day_adult = [2020] Household_removal_NO_Hood_PM = [1045] Household_removal_Hood_PM = [2020] pd.set_option('display.max_rows', 500) pd.set_option('display.max_columns', 500) pd.set_option('display.width', 1000) if Hood_or_no == 'hood': C_Place_holder = 2001 else: C_Place_holder = 1001 if Computer == 'personal' and Hood_or_no == 'no_hood': # 1N datafile_path_day_1N ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_Summary_Day_1_exact.csv" Day_1N = pd.read_csv(datafile_path_day_1N, skiprows=2) datafile_path_event_1N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_Summary_Event_1_exact.csv" Event_1N = pd.read_csv(datafile_path_event_1N, skiprows=2) # there is no second exact in phase 1N #1N Survey datafile_path_survey_1N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_1H_Survey_summary_.csv" Filter_1n_survey = pd.read_csv(datafile_path_survey_1N, skiprows=0) #print(Filter_1n_survey.iloc[0:40, :]) Survey_1N = Filter_1n_survey.iloc[0:40,:] #24 hour Kitchen pm breakdown data_file_path_24_PM_1N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_24_hour_Kitchen_PM.csv" Kit_PM_1N_24hr = pd.read_csv(data_file_path_24_PM_1N, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_Fuel_1N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_24_hour_Fuel_removal.csv" Fuel_remove_1N_24hr = pd.read_csv(data_file_path_24_Fuel_1N, skiprows=0) #2N datafile_path_day_2N ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_Summary_Day_1_exact.csv" Day_2N = pd.read_csv(datafile_path_day_2N, skiprows=2) datafile_path_event_2N_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_Summary_Event_1_exact.csv" Event_2N_1 = pd.read_csv(datafile_path_event_2N_1, skiprows=2) #2N second Exact datafile_path_event_2N_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_Summary_Event_2_exact.csv" Event_2N_2 = pd.read_csv(datafile_path_event_2N_2, skiprows=2) #2N Survey datafile_path_survey_2N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_Survey_summary_.csv" Survey_2N = pd.read_csv(datafile_path_survey_2N, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_2N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_24_hour_Kitchen_PM.csv" Kit_PM_2N_24hr = pd.read_csv(data_file_path_24_PM_2N, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_Fuel_2N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_24_hour_Fuel_removal.csv" Fuel_remove_2N_24hr = pd.read_csv(data_file_path_24_Fuel_2N, skiprows=0) #3N datafile_path_day_3N ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_Summary_Day_1_exact.csv" Day_3N = pd.read_csv(datafile_path_day_3N, skiprows=2) datafile_path_event_3N_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_Summary_Event_1_exact.csv" Event_3N_1 = pd.read_csv(datafile_path_event_3N_1, skiprows=2) #3N second Exact datafile_path_event_3N_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_Summary_Event_2_exact.csv" Event_3N_2 = pd.read_csv(datafile_path_event_3N_2, skiprows=2) #3N Survey datafile_path_survey_3N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_Survey_summary_.csv" Survey_3N = pd.read_csv(datafile_path_survey_3N, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_3N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_24_hour_Kitchen_PM.csv" Kit_PM_3N_24hr = pd.read_csv(data_file_path_24_PM_3N, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_Fuel_3N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_24_hour_Fuel_removal.csv" Fuel_remove_3N_24hr = pd.read_csv(data_file_path_24_Fuel_3N, skiprows=0) #4N datafile_path_day_4N ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_Summary_Day_1_exact.csv" Day_4N = pd.read_csv(datafile_path_day_4N, skiprows=2) datafile_path_event_4N_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_Summary_Event_1_exact.csv" Event_4N_1 = pd.read_csv(datafile_path_event_4N_1, skiprows=2) #4N second Exact datafile_path_event_4N_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_Summary_Event_2_exact.csv" Event_4N_2 = pd.read_csv(datafile_path_event_4N_2, skiprows=2) #4N Survey datafile_path_survey_4N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_Survey_summary_.csv" Survey_4N = pd.read_csv(datafile_path_survey_4N, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_4N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_24_hour_Kitchen_PM.csv" Kit_PM_4N_24hr = pd.read_csv(data_file_path_24_PM_4N, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_Fuel_4N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_24_hour_Fuel_removal.csv" Fuel_remove_4N_24hr = pd.read_csv(data_file_path_24_Fuel_4N, skiprows=0) elif Computer == 'personal' and Hood_or_no == 'hood': #1H datafile_path_day_1H ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1H/1H_Summary_Day_1_exact.csv" Day_1H = pd.read_csv(datafile_path_day_1H, skiprows=2) datafile_path_event_1H ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1H/1H_Summary_Event_1_exact.csv" Event_1H = pd.read_csv(datafile_path_event_1H, skiprows=2) #there is no second exact in phase 1H #1H Survey (row 40 or so afterward is Hood portion column 1 is houshold number) datafile_path_survey_1H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_1H_Survey_summary_.csv" Survey_1H = pd.read_csv(datafile_path_survey_1H, skiprows=40) #24 hour Kitchen pm breakdown data_file_path_24_PM_1H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1H/1H_24_hour_Kitchen_PM.csv" Kit_PM_1H_24hr = pd.read_csv(data_file_path_24_PM_1H, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_fuel_1H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1H/1H_24_hour_Fuel_removal.csv" Fuel_remove_1H_24hr = pd.read_csv(data_file_path_24_fuel_1H, skiprows=0) #2H datafile_path_day_2H ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_Summary_Day_1_exact.csv" Day_2H = pd.read_csv(datafile_path_day_2H, skiprows=2) datafile_path_event_2H_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_Summary_Event_1_exact.csv" Event_2H_1 = pd.read_csv(datafile_path_event_2H_1, skiprows=2) #2H second Exact datafile_path_event_2H_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_Summary_Event_2_exact.csv" Event_2H_2 = pd.read_csv(datafile_path_event_2H_2, skiprows=2) #2H survey datafile_path_survey_2H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_Survey_summary_.csv" Survey_2H = pd.read_csv(datafile_path_survey_2H, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_2H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_24_hour_Kitchen_PM.csv" Kit_PM_2H_24hr = pd.read_csv(data_file_path_24_PM_2H, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_fuel_2H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_24_hour_Fuel_removal.csv" Fuel_remove_2H_24hr = pd.read_csv(data_file_path_24_fuel_2H, skiprows=0) #3H datafile_path_day_3H ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_Summary_Day_1_exact.csv" Day_3H = pd.read_csv(datafile_path_day_3H, skiprows=2) datafile_path_event_3N_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_Summary_Event_1_exact.csv" Event_3H_1 = pd.read_csv(datafile_path_event_3N_1, skiprows=2) #3H second Exact datafile_path_event_3H_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_Summary_Event_2_exact.csv" Event_3H_2 = pd.read_csv(datafile_path_event_3H_2, skiprows=2) #3H survey datafile_path_survey_3H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_Survey_summary_.csv" Survey_3H = pd.read_csv(datafile_path_survey_3H, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_3H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_24_hour_Kitchen_PM.csv" Kit_PM_3H_24hr = pd.read_csv(data_file_path_24_PM_3H, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_fuel_3H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_24_hour_Fuel_removal.csv" Fuel_remove_3H_24hr = pd.read_csv(data_file_path_24_fuel_3H, skiprows=0) #work uses box information and not local data elif Computer == 'work' and Hood_or_no == 'no_hood': # 1N for box file system datafile_path_day_1N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_Summary_Day_1_exact.csv" Day_1N = pd.read_csv(datafile_path_day_1N, skiprows=2) datafile_path_event_1N ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_Summary_Event_1_exact.csv" Event_1N = pd.read_csv(datafile_path_event_1N, skiprows=2) # there is no second exact in phase 1N #1N Survey datafile_path_survey_1N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_1H_Survey_summary_.csv" Filter_1n_survey = pd.read_csv(datafile_path_survey_1N, skiprows=0) #print(Filter_1n_survey.iloc[0:40, :]) Survey_1N = Filter_1n_survey.iloc[0:40,:] #24 hour Kitchen pm breakdown data_file_path_24_PM_1N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_24_hour_Kitchen_PM.csv" Kit_PM_1N_24hr = pd.read_csv(data_file_path_24_PM_1N, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_Fuel_1N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_24_hour_Fuel_removal.csv" Fuel_remove_1N_24hr = pd.read_csv(data_file_path_24_Fuel_1N, skiprows=0) #2N datafile_path_day_2N ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_Summary_Day_1_exact.csv" Day_2N = pd.read_csv(datafile_path_day_2N, skiprows=2) datafile_path_event_2N_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_Summary_Event_1_exact.csv" Event_2N_1 = pd.read_csv(datafile_path_event_2N_1, skiprows=2) #2N second Exact datafile_path_event_2N_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_Summary_Event_2_exact.csv" Event_2N_2 = pd.read_csv(datafile_path_event_2N_2, skiprows=2) #2N Survey datafile_path_survey_2N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_Survey_summary_.csv" Survey_2N = pd.read_csv(datafile_path_survey_2N, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_2N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_24_hour_Kitchen_PM.csv" Kit_PM_2N_24hr = pd.read_csv(data_file_path_24_PM_2N, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_Fuel_2N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_24_hour_Fuel_removal.csv" Fuel_remove_2N_24hr = pd.read_csv(data_file_path_24_Fuel_2N, skiprows=0) #3N datafile_path_day_3N ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_Summary_Day_1_exact.csv" Day_3N = pd.read_csv(datafile_path_day_3N, skiprows=2) datafile_path_event_3N_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_Summary_Event_1_exact.csv" Event_3N_1 = pd.read_csv(datafile_path_event_3N_1, skiprows=2) #3N second Exact datafile_path_event_3N_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_Summary_Event_2_exact.csv" Event_3N_2 = pd.read_csv(datafile_path_event_3N_2, skiprows=2) #3N survey datafile_path_survey_3N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_Survey_summary_.csv" Survey_3N = pd.read_csv(datafile_path_survey_3N, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_3N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_24_hour_Kitchen_PM.csv" Kit_PM_3N_24hr = pd.read_csv(data_file_path_24_PM_3N, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_Fuel_3N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_24_hour_Fuel_removal.csv" Fuel_remove_3N_24hr = pd.read_csv(data_file_path_24_Fuel_3N, skiprows=0) #4N datafile_path_day_4N ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_Summary_Day_1_exact.csv" Day_4N = pd.read_csv(datafile_path_day_4N, skiprows=2) datafile_path_event_4N_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_Summary_Event_1_exact.csv" Event_4N_1 = pd.read_csv(datafile_path_event_4N_1, skiprows=2) #4N second Exact datafile_path_event_4N_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_Summary_Event_2_exact.csv" Event_4N_2 = pd.read_csv(datafile_path_event_4N_2, skiprows=2) #4N Survey datafile_path_survey_4N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_Survey_summary_.csv" Survey_4N = pd.read_csv(datafile_path_survey_4N, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_4N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_24_hour_Kitchen_PM.csv" Kit_PM_4N_24hr = pd.read_csv(data_file_path_24_PM_4N, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_Fuel_4N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_24_hour_Fuel_removal.csv" Fuel_remove_4N_24hr = pd.read_csv(data_file_path_24_Fuel_4N, skiprows=0) else: #1H datafile_path_day_1H ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1H/1H_Summary_Day_1_exact.csv" Day_1H = pd.read_csv(datafile_path_day_1H, skiprows=2) datafile_path_event_1H ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1H/1H_Summary_Event_1_exact.csv" Event_1H = pd.read_csv(datafile_path_event_1H, skiprows=2) #there is no second exact in phase 1H #1H Survey (row 40 or so afterward is Hood portion column 1 is houshold number) datafile_path_survey_1H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_1H_Survey_summary_.csv" Survey_1H = pd.read_csv(datafile_path_survey_1H, skiprows=40) #24 hour Kitchen pm breakdown data_file_path_24_PM_1H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1H/1H_24_hour_Kitchen_PM.csv" Kit_PM_1H_24hr = pd.read_csv(data_file_path_24_PM_1H, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_fuel_1H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1H/1H_24_hour_Fuel_removal.csv" Fuel_remove_1H_24hr = pd.read_csv(data_file_path_24_fuel_1H, skiprows=0) #2H datafile_path_day_2H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_Summary_Day_1_exact.csv" Day_2H = pd.read_csv(datafile_path_day_2H, skiprows=2) datafile_path_event_2H_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_Summary_Event_1_exact.csv" Event_2H_1 = pd.read_csv(datafile_path_event_2H_1, skiprows=2) #2H second Exact datafile_path_event_2H_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_Summary_Event_2_exact.csv" Event_2H_2 = pd.read_csv(datafile_path_event_2H_2, skiprows=2) #2H survey datafile_path_survey_2H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_Survey_summary_.csv" Survey_2H = pd.read_csv(datafile_path_survey_2H, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_2H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_24_hour_Kitchen_PM.csv" Kit_PM_2H_24hr = pd.read_csv(data_file_path_24_PM_2H, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_fuel_2H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_24_hour_Fuel_removal.csv" Fuel_remove_2H_24hr = pd.read_csv(data_file_path_24_fuel_2H, skiprows=0) #3H datafile_path_day_3H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_Summary_Day_1_exact.csv" Day_3H = pd.read_csv(datafile_path_day_3H, skiprows=2) datafile_path_event_3N_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_Summary_Event_1_exact.csv" Event_3H_1 = pd.read_csv(datafile_path_event_3N_1, skiprows=2) #3H second Exact datafile_path_event_3H_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_Summary_Event_2_exact.csv" Event_3H_2 = pd.read_csv(datafile_path_event_3H_2, skiprows=2) #3H survey datafile_path_survey_3H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_Survey_summary_.csv" Survey_3H = pd.read_csv(datafile_path_survey_3H, skiprows=0) #24 hour Kitchen pm breakdown data_file_path_24_PM_3H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_24_hour_Kitchen_PM.csv" Kit_PM_3H_24hr = pd.read_csv(data_file_path_24_PM_3H, skiprows=0) #24 hour Fuel Removal breakdown data_file_path_24_fuel_3H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_24_hour_Fuel_removal.csv" Fuel_remove_3H_24hr = pd.read_csv(data_file_path_24_fuel_3H, skiprows=0) #time to start ploting fun things #1st starting with the fuel per day per adult histogram and box plot NO_hood_counter = np.arange(0,39) hood_counter = np.arange(0,14) #what household do you want to remove from the graphs (1046 is a dummy spacer) print('---------------Fuel per Day per Adult No-Hood Phase---------------------') if Hood_or_no == 'no_hood': Fuel_per_day_per_adult_1N = [] f_d_a_1N = [] Fuel_per_day_per_adult_2N = [] f_d_a_2N = [] Fuel_per_day_per_adult_3N = [] f_d_a_3N = [] Fuel_per_day_per_adult_4N = [] f_d_a_4N =[] count_t = 0 count_f = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_NO_Hood_fuel_day_adult[count_f] - C_Place_holder): count_f = count_f + 1 if count_f == len(Household_removal_NO_Hood_fuel_day_adult): count_f = 0 continue if Fuel_remove_1N_24hr.iloc[c,6]!= -1.00: Fuel_per_day_per_adult_1N.append(Fuel_remove_1N_24hr.iloc[c,6]/Survey_1N.iloc[c,7]) f_d_a_1N.append(Day_1N.iloc[c,0]) if Fuel_remove_2N_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_2N.append(Fuel_remove_2N_24hr.iloc[c,6] / Survey_2N.iloc[c, 7]) f_d_a_2N.append(Day_2N.iloc[c,0]) if Fuel_remove_3N_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_3N.append(Fuel_remove_3N_24hr.iloc[c,6]/ Survey_3N.iloc[c, 7]) f_d_a_3N.append(Day_3N.iloc[c, 0]) if Fuel_remove_4N_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_4N.append(Fuel_remove_4N_24hr.iloc[c,6] / Survey_4N.iloc[c, 7]) f_d_a_4N.append(Day_3N.iloc[c, 0]) # percentage Change of Fuel per day between the phases Fuel_per_day_per_adult_2N_1N = [] f_d_a_2N_1N = [] Fuel_per_day_per_adult_3N_1N = [] f_d_a_3N_1N = [] Fuel_per_day_per_adult_4N_1N = [] f_d_a_4N_1N = [] Fuel_per_day_per_adult_3N_2N = [] f_d_a_3N_2N = [] Fuel_per_day_per_adult_4N_3N = [] f_d_a_4N_3N = [] Fuel_per_day_per_adult_4N_2N = [] f_d_a_4N_2N = [] count_t = 0 count_f = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_NO_Hood_fuel_day_adult[count_f] - C_Place_holder): count_f = count_f + 1 if count_f == len(Household_removal_NO_Hood_fuel_day_adult): count_f = 0 continue if (len(Fuel_per_day_per_adult_2N)-1) >= c and (len(Fuel_per_day_per_adult_1N)-1) >= c: if Day_1N.iloc[c,13] > 0 and Day_2N.iloc[c,13] > 0 and Day_1N.iloc[c,0] == Day_2N.iloc[c,0]: Fuel_per_day_per_adult_2N_1N.append(Fuel_per_day_per_adult_2N[c]/Fuel_per_day_per_adult_1N[c]) f_d_a_2N_1N.append(Day_1N.iloc[c,0]) if (len(Fuel_per_day_per_adult_3N)-1) >= c and (len(Fuel_per_day_per_adult_1N)-1) >= c: if Day_3N.iloc[c,13] > 0 and Day_1N.iloc[c,13] > 0 and Day_3N.iloc[c,0] == Day_1N.iloc[c,0]: Fuel_per_day_per_adult_3N_1N.append(Fuel_per_day_per_adult_3N[c]/Fuel_per_day_per_adult_1N[c]) f_d_a_3N_1N.append(Day_1N.iloc[c,0]) if (len(Fuel_per_day_per_adult_4N)-1) >= c and (len(Fuel_per_day_per_adult_1N)-1) >= c: if Day_4N.iloc[c,13] > 0 and Day_1N.iloc[c,13] > 0 and Day_4N.iloc[c,0] == Day_1N.iloc[c,0]: Fuel_per_day_per_adult_4N_1N.append(Fuel_per_day_per_adult_4N[c]/Fuel_per_day_per_adult_1N[c]) f_d_a_4N_1N.append(Day_1N.iloc[c,0]) if (len(Fuel_per_day_per_adult_3N)-1) >= c and (len(Fuel_per_day_per_adult_2N)-1) >= c: if Day_3N.iloc[c,13] > 0 and Day_2N.iloc[c,13] > 0 and Day_3N.iloc[c,0] == Day_2N.iloc[c,0]: Fuel_per_day_per_adult_3N_2N.append(Fuel_per_day_per_adult_3N[c]/Fuel_per_day_per_adult_2N[c]) f_d_a_3N_2N.append(Day_2N.iloc[c,0]) if (len(Fuel_per_day_per_adult_4N)-1) >= c and (len(Fuel_per_day_per_adult_3N)-1) >= c: if Day_4N.iloc[c,13] > 0 and Day_3N.iloc[c,13] > 0 and Day_4N.iloc[c,0] == Day_3N.iloc[c,0]: Fuel_per_day_per_adult_4N_3N.append(Fuel_per_day_per_adult_4N[c]/Fuel_per_day_per_adult_3N[c]) f_d_a_4N_3N.append(Day_3N.iloc[c,0]) if (len(Fuel_per_day_per_adult_4N)-1) >= c and (len(Fuel_per_day_per_adult_2N)-1) >= c: if Day_4N.iloc[c,13] > 0 and Day_2N.iloc[c,13] > 0 and Day_4N.iloc[c,0] == Day_2N.iloc[c,0]: Fuel_per_day_per_adult_4N_2N.append(Fuel_per_day_per_adult_4N[c]/Fuel_per_day_per_adult_2N[c]) f_d_a_4N_2N.append(Day_4N.iloc[c,0]) # now for box plotting for Fuel per day beteen Phases #1N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_1N, ax=ax_box, color='b') sns.distplot(Fuel_per_day_per_adult_1N, ax=ax_hist, color='b') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('1N Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) #2N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_2N, ax=ax_box, color='g') sns.distplot(Fuel_per_day_per_adult_2N, ax=ax_hist, color='g') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('2N Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) #3N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_3N, ax=ax_box, color='r') sns.distplot(Fuel_per_day_per_adult_3N, ax=ax_hist, color='r') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('3N Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) #4N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_4N, ax=ax_box, color='y') sns.distplot(Fuel_per_day_per_adult_4N, ax=ax_hist, color='y') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('4N Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) #Plotting on the same graph fig, ax = plt.subplots() plt.title('No-Hood Fuel per Day per Adult') #plt.hold(True) #1N quant_1_1N = np.percentile(Fuel_per_day_per_adult_1N, [25,50,75]) Top_lim_1_1N = quant_1_1N[2] + 1.5*(quant_1_1N[2] - quant_1_1N[0]) Low_lim_1_1N = quant_1_1N[0] - 1.5*(quant_1_1N[2] - quant_1_1N[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_1N, positions = [1], widths = 0.6) Fuel_D_A_1N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_1N): if a > Top_lim_1_1N or a < Low_lim_1_1N: Fuel_D_A_1N_outlier.append(f_d_a_1N[v]) plt.text(1,a,f_d_a_1N[v]) plt.text(1,0.1,'1N',color='b') #2N quant_1_2N = np.percentile(Fuel_per_day_per_adult_2N, [25,50,75]) Top_lim_1_2N = quant_1_2N[2] + 1.5*(quant_1_2N[2] - quant_1_2N[0]) Low_lim_1_2N = quant_1_2N[0] - 1.5*(quant_1_2N[2] - quant_1_2N[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_2N,positions = [2], widths = 0.6) Fuel_D_A_2N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_2N): if a > Top_lim_1_2N or a < Low_lim_1_2N: Fuel_D_A_2N_outlier.append(f_d_a_2N[v]) plt.text(2,a,f_d_a_2N[v]) plt.text(2,0.1,'2N', color= 'g') #3N quant_1_3N = np.percentile(Fuel_per_day_per_adult_3N, [25,50,75]) Top_lim_1_3N = quant_1_3N[2] + 1.5*(quant_1_3N[2] - quant_1_3N[0]) Low_lim_1_3N = quant_1_3N[0] - 1.5*(quant_1_3N[2] - quant_1_3N[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_3N,positions = [3], widths = 0.6) count = 0 Fuel_D_A_3N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3N): if a > Top_lim_1_3N or a < Low_lim_1_3N: Fuel_D_A_3N_outlier.append(f_d_a_3N[v]) count = count + 1 if count == 2: plt.text(3,a,f_d_a_3N[v],ha='left',va='bottom') elif count != 2: plt.text(3,a,f_d_a_3N[v],ha='right',va='bottom') plt.text(3,0.1,'3N', color='r') #4N quant_1_4N = np.percentile(Fuel_per_day_per_adult_4N, [25,50,75]) Top_lim_1_4N = quant_1_4N[2] + 1.5*(quant_1_4N[2] - quant_1_4N[0]) Low_lim_1_4N = quant_1_4N[0] - 1.5*(quant_1_4N[2] - quant_1_4N[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_4N,positions = [4], widths = 0.6) Fuel_D_A_4N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_4N): if a > Top_lim_1_4N or a < Low_lim_1_4N: Fuel_D_A_4N_outlier.append(f_d_a_4N[v]) plt.text(4,a,f_d_a_4N[v]) plt.text(4,0.1,'4N', color='y') plt.xlim(0,5) plt.ylim(0,2.3) print('Fuel/Day/Adult 1N had these values as outliers ', Fuel_D_A_1N_outlier) print('Fuel/Day/Adult 2N had these values as outliers ', Fuel_D_A_2N_outlier) print('Fuel/Day/Adult 3N had these values as outliers ', Fuel_D_A_3N_outlier) print('Fuel/Day/Adult 4N had these values as outliers ', Fuel_D_A_4N_outlier) plt.show() # % change of fuel per day per adult between each phase fig_2, ax2 = plt.subplots() plt.title('% No_hood Change from Fuel per Day per Adult' ) #plt.hold(True) #2N to 1N quant_1_2N_1N = np.percentile(Fuel_per_day_per_adult_2N_1N, [25,50,75]) Top_lim_1_2N_1N = quant_1_2N_1N[2] + 1.5*(quant_1_2N_1N[2]-quant_1_2N_1N[0]) Low_lim_1_2N_1N = quant_1_2N_1N[0] - 1.5*(quant_1_2N_1N[2]-quant_1_2N_1N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_2N_1N, positions=[1], widths= 0.6) Fuel_D_A_2N_1N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_2N_1N): if a > Top_lim_1_2N_1N or a < Low_lim_1_2N_1N: Fuel_D_A_2N_1N_outlier.append(f_d_a_2N_1N[v]) plt.text(1, a, f_d_a_2N_1N[v]) plt.text(0.5, 0, '2N / 1N', color= 'g') #3N to 1N quant_1_3N_1N = np.percentile(Fuel_per_day_per_adult_3N_1N, [25,50,75]) Top_lim_1_3N_1N = quant_1_3N_1N[2] + 1.5*(quant_1_3N_1N[2]-quant_1_3N_1N[0]) Low_lim_1_3N_1N = quant_1_3N_1N[0] - 1.5*(quant_1_3N_1N[2]-quant_1_3N_1N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_3N_1N, positions=[2], widths= 0.6) Fuel_D_A_3N_1N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3N_1N): if a > Top_lim_1_3N_1N or a < Low_lim_1_3N_1N: Fuel_D_A_3N_1N_outlier.append(f_d_a_3N_1N[v]) plt.text(2, a, f_d_a_3N_1N[v]) plt.text(1.5, 0, '3N / 1N', color= 'r') #4N to 1N quant_1_4N_1N = np.percentile(Fuel_per_day_per_adult_4N_1N, [25,50,75]) Top_lim_1_4N_1N = quant_1_4N_1N[2] + 1.5*(quant_1_4N_1N[2]-quant_1_4N_1N[0]) Low_lim_1_4N_1N = quant_1_4N_1N[0] - 1.5*(quant_1_4N_1N[2]-quant_1_4N_1N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_4N_1N, positions=[3], widths= 0.6) Fuel_D_A_4N_1N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_4N_1N): if a > Top_lim_1_4N_1N or a < Low_lim_1_4N_1N: Fuel_D_A_4N_1N_outlier.append(f_d_a_4N_1N[v]) plt.text(3, a, f_d_a_4N_1N[v]) plt.text(2.5, 0, '4N / 1N', color= 'y') #3N to 2N quant_1_3N_2N = np.percentile(Fuel_per_day_per_adult_3N_2N, [25,50,75]) Top_lim_1_3N_2N = quant_1_3N_2N[2] + 1.5*(quant_1_3N_2N[2]-quant_1_3N_2N[0]) Low_lim_1_3N_2N = quant_1_3N_2N[0] - 1.5*(quant_1_3N_2N[2]-quant_1_3N_2N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_3N_2N, positions=[4], widths= 0.6) Fuel_D_A_3N_2N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3N_2N): if a > Top_lim_1_3N_2N or a < Low_lim_1_3N_2N: Fuel_D_A_3N_2N_outlier.append(f_d_a_3N_2N[v]) plt.text(4, a, f_d_a_3N_2N[v]) plt.text(3.5, 0, '3N / 2N', color= 'm') #4N to 3N quant_1_4N_3N = np.percentile(Fuel_per_day_per_adult_4N_3N, [25,50,75]) Top_lim_1_4N_3N = quant_1_4N_3N[2] + 1.5*(quant_1_4N_3N[2]-quant_1_4N_3N[0]) Low_lim_1_4N_3N = quant_1_4N_3N[0] - 1.5*(quant_1_4N_3N[2]-quant_1_4N_3N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_4N_3N, positions=[5], widths= 0.6) Fuel_D_A_4N_3N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_4N_3N): if a > Top_lim_1_4N_3N or a < Low_lim_1_4N_3N: Fuel_D_A_4N_3N_outlier.append(f_d_a_4N_3N[v]) plt.text(5, a, f_d_a_4N_3N[v]) plt.text(4.5, 0, '4N / 3N', color= 'k') #4N to 2N quant_1_4N_2N = np.percentile(Fuel_per_day_per_adult_4N_2N, [25,50,75]) Top_lim_1_4N_2N = quant_1_4N_2N[2] + 1.5*(quant_1_4N_2N[2]-quant_1_4N_2N[0]) Low_lim_1_4N_2N = quant_1_4N_2N[0] - 1.5*(quant_1_4N_2N[2]-quant_1_4N_2N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_4N_2N, positions=[6], widths= 0.6) Fuel_D_A_4N_2N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_4N_2N): if a > Top_lim_1_4N_2N or a < Low_lim_1_4N_2N: Fuel_D_A_4N_2N_outlier.append(f_d_a_4N_2N[v]) plt.text(6, a, f_d_a_4N_2N[v]) plt.text(5.5, 0, '4N / 2N', color= 'tab:orange') plt.xlim(0,7) plt.ylim(-0.5,4) print('Fuel/Day/Adult 2N/1N had these values as outliers ', Fuel_D_A_2N_1N_outlier) print('Fuel/Day/Adult 3N/1N had these values as outliers ', Fuel_D_A_3N_1N_outlier) print('Fuel/Day/Adult 4N/1N had these values as outliers ', Fuel_D_A_4N_1N_outlier) print('Fuel/Day/Adult 3N/2N had these values as outliers ', Fuel_D_A_3N_2N_outlier) print('Fuel/Day/Adult 4N/3N had these values as outliers ', Fuel_D_A_4N_3N_outlier) print('Fuel/Day/Adult 4N/2N had these values as outliers ', Fuel_D_A_4N_2N_outlier) plt.show() #adding averages to the tables quant_1_1N = np.append(quant_1_1N, np.average(Fuel_per_day_per_adult_1N)) quant_1_2N = np.append(quant_1_2N, np.average(Fuel_per_day_per_adult_2N)) quant_1_3N = np.append(quant_1_3N, np.average(Fuel_per_day_per_adult_3N)) quant_1_4N = np.append(quant_1_4N, np.average(Fuel_per_day_per_adult_4N)) D_50_quant_phase_f_d_a = {'Percentile %': ['25','50','75', 'Avg'], '1N': quant_1_1N, '2N': quant_1_2N,'3N' : quant_1_3N,'4N': quant_1_4N} F_D_A_50_phase_no_hood = pd.DataFrame(data=D_50_quant_phase_f_d_a, columns=['Percentile %','1N', '2N', '3N','4N']) quant_1_2N_1N = np.append(quant_1_2N_1N , np.average(Fuel_per_day_per_adult_2N_1N)) quant_1_3N_1N = np.append(quant_1_3N_1N , np.average(Fuel_per_day_per_adult_3N_1N)) quant_1_4N_1N = np.append(quant_1_4N_1N , np.average(Fuel_per_day_per_adult_4N_1N)) quant_1_3N_2N = np.append(quant_1_3N_2N , np.average(Fuel_per_day_per_adult_3N_2N)) quant_1_4N_3N = np.append(quant_1_4N_3N , np.average(Fuel_per_day_per_adult_4N_3N)) quant_1_4N_2N = np.append(quant_1_4N_2N , np.average(Fuel_per_day_per_adult_4N_2N)) D_50_quant_percent_f_d_a ={'Percentile %': ['25','50','75', 'Avg'],'2N / 1N': quant_1_2N_1N,'3N / 1N': quant_1_3N_1N,'4N / 1N': quant_1_4N_1N, '3N / 2N': quant_1_3N_2N,'4N / 3N': quant_1_4N_3N,'4N / 2N': quant_1_4N_2N} F_D_A_50_percent_change_no_hood = pd.DataFrame(data=D_50_quant_percent_f_d_a, columns=['Percentile %','2N / 1N','3N / 1N', '4N / 1N' ,'3N / 2N','4N / 3N','4N / 2N']) print(F_D_A_50_phase_no_hood) print(F_D_A_50_percent_change_no_hood) # add more print ('-------------------Fuel per Day per Adult Hood Phase -------------------') if Hood_or_no == 'hood': Fuel_per_day_per_adult_1H = [] f_d_a_1H = [] Fuel_per_day_per_adult_2H = [] f_d_a_2H = [] Fuel_per_day_per_adult_3H = [] f_d_a_3H = [] count_t = 0 count_f = 0 for c in hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_Hood_fuel_day_adult[count_f] - C_Place_holder): count_f = count_f + 1 if count_f == len(Household_removal_Hood_fuel_day_adult): count_f = 0 continue if Fuel_remove_1H_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_1H.append(Fuel_remove_1H_24hr.iloc[c,6]/Survey_1H.iloc[c,7]) f_d_a_1H.append(Day_1H.iloc[c,0]) if Fuel_remove_2H_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_2H.append(Fuel_remove_2H_24hr.iloc[c,6] / Survey_2H.iloc[c, 7]) f_d_a_2H.append(Day_2H.iloc[c,0]) if Fuel_remove_3H_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_3H.append(Fuel_remove_3H_24hr.iloc[c,6]/ Survey_3H.iloc[c, 7]) f_d_a_3H.append(Day_3H.iloc[c, 0]) # percentage Change of Fuel per day between the phases Fuel_per_day_per_adult_2H_1H = [] f_d_a_2H_1H = [] Fuel_per_day_per_adult_3H_1H = [] f_d_a_3H_1H = [] Fuel_per_day_per_adult_3H_2H = [] f_d_a_3H_2H = [] count_t = 0 count_f = 0 for c in hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_Hood_fuel_day_adult[count_f] - C_Place_holder): count_f = count_f + 1 if count_f == len(Household_removal_Hood_fuel_day_adult): count_f = 0 continue if (len(Fuel_per_day_per_adult_2H)-1) >= c and (len(Fuel_per_day_per_adult_1H)-1) >= c: if Day_1H.iloc[c,13] > 0 and Day_2H.iloc[c,13] > 0 and Day_1H.iloc[c,0] == Day_2H.iloc[c,0]: Fuel_per_day_per_adult_2H_1H.append(Fuel_per_day_per_adult_2H[c]/Fuel_per_day_per_adult_1H[c]) f_d_a_2H_1H.append(Day_1H.iloc[c,0]) if (len(Fuel_per_day_per_adult_3H)-1) >= c and (len(Fuel_per_day_per_adult_1H)-1) >= c: if Day_3H.iloc[c,13] > 0 and Day_1H.iloc[c,13] > 0 and Day_3H.iloc[c,0] == Day_1H.iloc[c,0]: Fuel_per_day_per_adult_3H_1H.append(Fuel_per_day_per_adult_3H[c]/Fuel_per_day_per_adult_1H[c]) f_d_a_3H_1H.append(Day_1H.iloc[c,0]) if (len(Fuel_per_day_per_adult_3H)-1) >= c and (len(Fuel_per_day_per_adult_2H)-1) >= c: if Day_3H.iloc[c,13] > 0 and Day_2H.iloc[c,13] > 0 and Day_3H.iloc[c,0] == Day_2H.iloc[c,0]: Fuel_per_day_per_adult_3H_2H.append(Fuel_per_day_per_adult_3H[c]/Fuel_per_day_per_adult_2H[c]) f_d_a_3H_2H.append(Day_1H.iloc[c,0]) # now for plotting #1H sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_1H, ax=ax_box, color='b') sns.distplot(Fuel_per_day_per_adult_1H, ax=ax_hist, color='b') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('1H Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) #2H sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_2H, ax=ax_box, color='g') sns.distplot(Fuel_per_day_per_adult_2H, ax=ax_hist, color='g') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('2H Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) #3H sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_3H, ax=ax_box, color='r') sns.distplot(Fuel_per_day_per_adult_3H, ax=ax_hist, color='r') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('3H Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) fig_2, ax_2 = plt.subplots() plt.title('Hood Fuel per Day per Adult') #plt.hold(True) quant_1_1H = np.percentile(Fuel_per_day_per_adult_1H, [25,50,75]) Top_lim_1_1H = quant_1_1H[2] + 1.5*(quant_1_1H[2] - quant_1_1H[0]) Low_lim_1_1H = quant_1_1H[0] - 1.5*(quant_1_1H[2] - quant_1_1H[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_1H, positions = [1], widths = 0.6) Fuel_D_A_1H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_1H): if a > Top_lim_1_1H or a < Low_lim_1_1H: Fuel_D_A_1H_outlier.append(f_d_a_1H[v]) plt.text(1,a,f_d_a_1H[v]) plt.text(1,0,'1H',color='b') quant_1_2H = np.percentile(Fuel_per_day_per_adult_2H, [25,50,75]) Top_lim_1_2H = quant_1_2H[2] + 1.5*(quant_1_2H[2] - quant_1_2H[0]) Low_lim_1_2H = quant_1_2H[0] - 1.5*(quant_1_2H[2] - quant_1_2H[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_2H,positions = [2], widths = 0.6) count = 0 Fuel_D_A_2H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_2H): if a > Top_lim_1_2H or a < Low_lim_1_2H: Fuel_D_A_2H_outlier.append(f_d_a_2H[v]) count = count + 1 if count == 1: plt.text(2,a,f_d_a_2H[v],ha='left',va='bottom') elif count !=1: plt.text(2,a,f_d_a_2H[v],ha='right',va='bottom') plt.text(2,0,'2H', color= 'g') quant_1_3H = np.percentile(Fuel_per_day_per_adult_3H, [25,50,75]) Top_lim_1_3H = quant_1_3H[2] + 1.5*(quant_1_3H[2] - quant_1_3H[0]) Low_lim_1_3H = quant_1_3H[0] - 1.5*(quant_1_3H[2] - quant_1_3H[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_3H,positions = [3], widths = 0.6) count = 0 Fuel_D_A_3H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3H): if a > Top_lim_1_3H or a < Low_lim_1_3H: Fuel_D_A_3H_outlier.append(f_d_a_3H[v]) count = count + 1 if count == 3: plt.text(3,a,f_d_a_3H[v],ha='left',va='bottom') elif count != 1: plt.text(3,a,f_d_a_3H[v],ha='right',va='bottom') plt.text(3,0,'3H', color='r') plt.xlim(-0,4) plt.ylim(-0.25,2.5) print('Fuel/Day/Adult 1H had these values as outliers ', Fuel_D_A_1H_outlier) print('Fuel/Day/Adult 2H had these values as outliers ', Fuel_D_A_2H_outlier) print('Fuel/Day/Adult 3H had these values as outliers ', Fuel_D_A_3H_outlier) plt.show() #% change of fuel perday per adult between each phase fig_2, ax2 = plt.subplots() plt.title('% No_hood Change from Fuel per Day per Adult' ) #plt.hold(True) #2H to 1H quant_1_2H_1H = np.percentile(Fuel_per_day_per_adult_2H_1H, [25,50,75]) Top_lim_1_2H_1H = quant_1_2H_1H[2] + 1.5*(quant_1_2H_1H[2]-quant_1_2H_1H[0]) Low_lim_1_2H_1H = quant_1_2H_1H[0] - 1.5*(quant_1_2H_1H[2]-quant_1_2H_1H[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_2H_1H, positions=[1], widths= 0.6) Fuel_D_A_2H_1H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_2H_1H): if a > Top_lim_1_2H_1H or a < Low_lim_1_2H_1H: Fuel_D_A_2H_1H_outlier.append(f_d_a_2H_1H[v]) plt.text(1, a, f_d_a_2H_1H[v]) plt.text(0.75, -0.25, '2H / 1H', color= 'g') #3H to 1H quant_1_3H_1H = np.percentile(Fuel_per_day_per_adult_3H_1H, [25,50,75]) Top_lim_1_3H_1H = quant_1_3H_1H[2] + 1.5*(quant_1_3H_1H[2]-quant_1_3H_1H[0]) Low_lim_1_3H_1H = quant_1_3H_1H[0] - 1.5*(quant_1_3H_1H[2]-quant_1_3H_1H[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_3H_1H, positions=[2], widths= 0.6) Fuel_D_A_3H_1H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3H_1H): if a > Top_lim_1_3H_1H or a < Low_lim_1_3H_1H: Fuel_D_A_3H_1H_outlier.append(f_d_a_3H_1H[v]) plt.text(2, a, f_d_a_3H_1H[v]) plt.text(1.75, -0.25, '3H / 1H', color= 'r') #3H to 2H quant_1_3H_2H = np.percentile(Fuel_per_day_per_adult_3H_2H, [25,50,75]) Top_lim_1_3H_2H = quant_1_3H_2H[2] + 1.5*(quant_1_3H_2H[2]-quant_1_3H_2H[0]) Low_lim_1_3H_2H = quant_1_3H_2H[0] - 1.5*(quant_1_3H_2H[2]-quant_1_3H_2H[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_3H_2H, positions=[3], widths= 0.6) Fuel_D_A_3H_2H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3H_2H): if a > Top_lim_1_3H_2H or a < Low_lim_1_3H_2H: Fuel_D_A_3H_2H_outlier.append(f_d_a_3H_2H[v]) plt.text(3, a, f_d_a_3H_2H[v]) plt.text(2.75, -0.25, '2H / 1H', color= 'm') plt.xlim(-0,4) plt.ylim(-0.25,6) print('Fuel/Day/Adult 2H/1H had these values as outliers ', Fuel_D_A_2H_1H_outlier) print('Fuel/Day/Adult 3H/1H had these values as outliers ', Fuel_D_A_3H_1H_outlier) print('Fuel/Day/Adult 3H/2H had these values as outliers ', Fuel_D_A_3H_2H_outlier) plt.show() quant_1_1H = np.append(quant_1_1H, np.average(Fuel_per_day_per_adult_1H)) quant_1_2H = np.append(quant_1_2H, np.average(Fuel_per_day_per_adult_2H)) quant_1_3H = np.append(quant_1_3H, np.average(Fuel_per_day_per_adult_3H)) D_50_quant_phase_f_d_a_hood = {'Percentile %': ['25','50','75', 'Avg'], '1H': quant_1_1H, '2H': quant_1_2H,'3H' : quant_1_3H} F_D_A_50_phase_hood = pd.DataFrame(data=D_50_quant_phase_f_d_a_hood, columns=['Percentile %','1H', '2H','3H'] ) quant_1_2H_1H = np.append(quant_1_2H_1H , np.average(Fuel_per_day_per_adult_2H_1H)) quant_1_3H_1H = np.append(quant_1_3H_1H , np.average(Fuel_per_day_per_adult_3H_1H)) quant_1_3H_2H = np.append(quant_1_3H_2H , np.average(Fuel_per_day_per_adult_3H_2H)) D_50_quant_percent_f_d_a_hood ={'Percentile %': ['25','50','75', 'Avg'],'2H / 1H': quant_1_2H_1H,'3H / 1H': quant_1_3H_1H,'3H / 2H': quant_1_3H_2H} F_D_A_50_percent_change_hood = pd.DataFrame(data=D_50_quant_percent_f_d_a_hood, columns=['Percentile %','2H / 1H','3H / 1H','3H / 2H']) print(F_D_A_50_phase_hood) print(F_D_A_50_percent_change_hood) print('----------------------- Kitchen PM per Day -----------------------------') if Hood_or_no == 'no_hood': Kit_PM_per_day_1N = [] K_PM_D_1N = [] Kit_PM_per_day_2N = [] K_PM_D_2N = [] Kit_PM_per_day_3N = [] K_PM_D_3N = [] Kit_PM_per_day_4N = [] K_PM_D_4N = [] count_t = 0 count_pm = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_NO_Hood_PM[count_pm] - C_Place_holder): count_pm = count_pm + 1 if count_pm == len(Household_removal_NO_Hood_PM): count_pm = 0 continue # if Day_1N.iloc[c,7] != -1.00: # Kit_PM_per_day_1N.append(Day_1N.iloc[c,7]/Day_1N.iloc[c,1]) # K_PM_D_1N.append(Day_1N.iloc[c,0]) if Kit_PM_1N_24hr.iloc[c,6] != -1.00: Kit_PM_per_day_1N.append(Kit_PM_1N_24hr.iloc[c,6]) K_PM_D_1N.append(Kit_PM_1N_24hr.iloc[c, 0]) #if Day_2N.iloc[c, 7] != -1.00: # Kit_PM_per_day_2N.append(Day_2N.iloc[c,7]/Day_2N.iloc[c,1]) # K_PM_D_2N.append(Day_2N.iloc[c,0]) if Kit_PM_2N_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_2N.append(Kit_PM_2N_24hr.iloc[c, 6]) K_PM_D_2N.append(Kit_PM_2N_24hr.iloc[c, 0]) # if Day_3N.iloc[c, 7] != -1.00: # Kit_PM_per_day_3N.append(Day_3N.iloc[c,7]/Day_3N.iloc[c,1]) # K_PM_D_3N.append(Day_3N.iloc[c, 0]) if Kit_PM_3N_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_3N.append(Kit_PM_3N_24hr.iloc[c, 6]) K_PM_D_3N.append(Kit_PM_3N_24hr.iloc[c, 0]) # if Day_4N.iloc[c, 7] != -1.00: # Kit_PM_per_day_4N.append(Day_4N.iloc[c,7]/Day_4N.iloc[c,1]) # K_PM_D_4N.append(Day_4N.iloc[c, 0]) if Kit_PM_4N_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_4N.append(Kit_PM_4N_24hr.iloc[c, 6]) K_PM_D_4N.append(Kit_PM_4N_24hr.iloc[c, 0]) # percentages Between Phases of kitchen PM per day Kit_per_day_2N_1N = [] K_PM_D_2N_1N = [] Kit_per_day_3N_1N = [] K_PM_D_3N_1N = [] Kit_per_day_4N_1N = [] K_PM_D_4N_1N = [] Kit_per_day_3N_2N = [] K_PM_D_3N_2N = [] Kit_per_day_4N_3N = [] K_PM_D_4N_3N = [] Kit_per_day_4N_2N = [] K_PM_D_4N_2N = [] count_t = 0 count_pm = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_NO_Hood_PM[count_pm] - C_Place_holder): count_pm = count_pm + 1 if count_pm == len(Household_removal_NO_Hood_PM): count_pm = 0 continue if (len(Kit_PM_per_day_2N)-1) >= c and (len(Kit_PM_per_day_1N)-1) >= c: #if Day_1N.iloc[c,7] > 0 and Day_2N.iloc[c,7] > 0 and Day_1N.iloc[c,0] == Day_2N.iloc[c,0]: if Kit_PM_1N_24hr.iloc[c,6] > 0 and Kit_PM_2N_24hr.iloc[c,6] > 0 and Kit_PM_1N_24hr.iloc[c,0] == Kit_PM_2N_24hr.iloc[c,0]: Kit_per_day_2N_1N.append(Kit_PM_per_day_2N[c]/Kit_PM_per_day_1N[c]) K_PM_D_2N_1N.append(Day_1N.iloc[c,0]) if (len(Kit_PM_per_day_3N)-1) >= c and (len(Kit_PM_per_day_1N)-1) >= c: #if Day_3N.iloc[c,7] > 0 and Day_1N.iloc[c,7] > 0 and Day_3N.iloc[c,0] == Day_1N.iloc[c,0]: if Kit_PM_3N_24hr.iloc[c, 6] > 0 and Kit_PM_1N_24hr.iloc[c, 6] > 0 and Kit_PM_3N_24hr.iloc[c, 0] == \ Kit_PM_1N_24hr.iloc[c, 0]: Kit_per_day_3N_1N.append(Kit_PM_per_day_3N[c]/Kit_PM_per_day_1N[c]) K_PM_D_3N_1N.append(Day_1N.iloc[c,0]) if (len(Kit_PM_per_day_4N)-1) >= c and (len(Kit_PM_per_day_1N)-1) >= c: #if Day_4N.iloc[c,7] > 0 and Day_1N.iloc[c,7] > 0 and Day_4N.iloc[c,0] == Day_1N.iloc[c,0]: if Kit_PM_4N_24hr.iloc[c, 6] > 0 and Kit_PM_1N_24hr.iloc[c, 6] > 0 and Kit_PM_4N_24hr.iloc[c, 0] == \ Kit_PM_1N_24hr.iloc[c, 0]: Kit_per_day_4N_1N.append(Kit_PM_per_day_4N[c]/Kit_PM_per_day_1N[c]) K_PM_D_4N_1N.append(Day_1N.iloc[c,0]) if (len(Kit_PM_per_day_3N)-1) >= c and (len(Kit_PM_per_day_2N)-1) >= c: #if Day_3N.iloc[c,7] > 0 and Day_2N.iloc[c,7] > 0 and Day_3N.iloc[c,0] == Day_2N.iloc[c,0]: if Kit_PM_3N_24hr.iloc[c, 6] > 0 and Kit_PM_2N_24hr.iloc[c, 6] > 0 and Kit_PM_3N_24hr.iloc[c, 0] == \ Kit_PM_2N_24hr.iloc[c, 0]: Kit_per_day_3N_2N.append(Kit_PM_per_day_3N[c]/Kit_PM_per_day_2N[c]) K_PM_D_3N_2N.append(Day_2N.iloc[c,0]) if (len(Kit_PM_per_day_4N)-1) >= c and (len(Kit_PM_per_day_3N)-1) >= c: #if Day_4N.iloc[c,7] > 0 and Day_3N.iloc[c,7] > 0 and Day_4N.iloc[c,0] == Day_3N.iloc[c,0]: if Kit_PM_4N_24hr.iloc[c, 6] > 0 and Kit_PM_3N_24hr.iloc[c, 6] > 0 and Kit_PM_3N_24hr.iloc[c, 0] == \ Kit_PM_4N_24hr.iloc[c, 0]: Kit_per_day_4N_3N.append(Kit_PM_per_day_4N[c]/Kit_PM_per_day_3N[c]) K_PM_D_4N_3N.append(Day_3N.iloc[c,0]) if (len(Kit_PM_per_day_4N)-1) >= c and (len(Kit_PM_per_day_2N)-1) >= c: #if Day_4N.iloc[c,7] > 0 and Day_2N.iloc[c,7] > 0 and Day_4N.iloc[c,0] == Day_2N.iloc[c,0]: if Kit_PM_4N_24hr.iloc[c, 6] > 0 and Kit_PM_4N_24hr.iloc[c, 6] > 0 and Kit_PM_4N_24hr.iloc[c, 0] == \ Kit_PM_2N_24hr.iloc[c, 0]: Kit_per_day_4N_2N.append(Kit_PM_per_day_4N[c]/Kit_PM_per_day_2N[c]) K_PM_D_4N_2N.append(Day_4N.iloc[c,0]) # now for box plotting for Kitchen PM per day percent changes #2N to 1N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_2N_1N, ax=ax_box, color='g') sns.distplot(Kit_per_day_2N_1N, ax=ax_hist, color='g') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 2N/1N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) #3N to 1N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_3N_1N, ax=ax_box, color='r') sns.distplot(Kit_per_day_3N_1N, ax=ax_hist, color='r') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 3N/1N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) #4N to 1N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_4N_1N, ax=ax_box, color='y') sns.distplot(Kit_per_day_4N_1N, ax=ax_hist, color='y') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 4N/1N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) #3N to 2N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_3N_2N, ax=ax_box, color='m') sns.distplot(Kit_per_day_3N_2N, ax=ax_hist, color='m') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 3N/2N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) #4N to 3N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_4N_3N, ax=ax_box, color='k') sns.distplot(Kit_per_day_4N_3N, ax=ax_hist, color='k') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 4N/3N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) #4N to 2N sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_4N_2N, ax=ax_box, color='tab:orange') sns.distplot(Kit_per_day_4N_2N, ax=ax_hist, color='tab:orange') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 4N/2N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) #Plotting on the same graph fig, ax = plt.subplots() plt.title('No-Hood Kitchen PM per day') #plt.hold() #1N quant_1_1N = np.percentile(Kit_PM_per_day_1N, [25,50,75]) Top_lim_1_1N = quant_1_1N[2] + 1.5*(quant_1_1N[2] - quant_1_1N[0]) Low_lim_1_1N = quant_1_1N[0] - 1.5*(quant_1_1N[2] - quant_1_1N[0]) bp_1 = plt.boxplot(Kit_PM_per_day_1N, positions = [1], widths = 0.6) kitchen_pm_1N_outlier = [] for v,a in enumerate(Kit_PM_per_day_1N): if a > Top_lim_1_1N or a < Low_lim_1_1N: kitchen_pm_1N_outlier.append(K_PM_D_1N[v]) plt.text(1,a,K_PM_D_1N[v]) plt.text(1,0.1,'1N',color='b') #2N quant_1_2N = np.percentile(Kit_PM_per_day_2N, [25,50,75]) Top_lim_1_2N = quant_1_2N[2] + 1.5*(quant_1_2N[2] - quant_1_2N[0]) Low_lim_1_2N = quant_1_2N[0] - 1.5*(quant_1_2N[2] - quant_1_2N[0]) bp_1 = plt.boxplot(Kit_PM_per_day_2N,positions = [2], widths = 0.6) kitchen_pm_2N_outlier = [] for v,a in enumerate(Kit_PM_per_day_2N): if a > Top_lim_1_2N or a < Low_lim_1_2N: kitchen_pm_2N_outlier.append(K_PM_D_2N[v]) plt.text(2,a,K_PM_D_2N[v]) plt.text(2,0.1,'2N', color= 'g') #3N quant_1_3N = np.percentile(Kit_PM_per_day_3N, [25,50,75]) Top_lim_1_3N = quant_1_3N[2] + 1.5*(quant_1_3N[2] - quant_1_3N[0]) Low_lim_1_3N = quant_1_3N[0] - 1.5*(quant_1_3N[2] - quant_1_3N[0]) kitchen_pm_3N_outlier = [] bp_1 = plt.boxplot(Kit_PM_per_day_3N,positions = [3], widths = 0.6) count = 0 for v,a in enumerate(Kit_PM_per_day_3N): if a > Top_lim_1_3N or a < Low_lim_1_3N: kitchen_pm_3N_outlier.append(K_PM_D_3N[v]) count = count + 1 if count == (3): plt.text(3,a,K_PM_D_3N[v],ha='left', va='bottom') if count == (1): plt.text(3,a,K_PM_D_3N[v],ha='left', va='top') else: plt.text(3,a,K_PM_D_3N[v],ha='right', va='bottom') plt.text(3,0.1,'3N', color='r') #4N quant_1_4N = np.percentile(Kit_PM_per_day_4N, [25,50,75]) Top_lim_1_4N = quant_1_4N[2] + 1.5*(quant_1_4N[2] - quant_1_4N[0]) Low_lim_1_4N = quant_1_4N[0] - 1.5*(quant_1_4N[2] - quant_1_4N[0]) bp_1 = plt.boxplot(Kit_PM_per_day_4N,positions = [4], widths = 0.6) kitchen_pm_4N_outlier = [] for v,a in enumerate(Kit_PM_per_day_4N): if a > Top_lim_1_4N or a < Low_lim_1_4N: kitchen_pm_4N_outlier.append(K_PM_D_4N[v]) plt.text(4,a,K_PM_D_4N[v]) plt.text(4,0.1,'4N', color='y') plt.xlim(0,5) plt.ylim(0,1200) print('Kitchen PM 1N had these values as outliers ', kitchen_pm_1N_outlier) print('Kitchen PM 2N had these values as outliers ', kitchen_pm_2N_outlier) print('Kitchen PM 3N had these values as outliers ', kitchen_pm_3N_outlier) print('Kitchen PM 4N had these values as outliers ', kitchen_pm_4N_outlier) plt.show() # % change of PM per day fig_2, ax2 = plt.subplots() plt.title('% No_hood PM per Day Change' ) #plt.hold(True) #2N to 1N quant_1_2N_1N = np.percentile(Kit_per_day_2N_1N, [25,50,75]) Top_lim_1_2N_1N = quant_1_2N_1N[2] + 1.5*(quant_1_2N_1N[2]-quant_1_2N_1N[0]) Low_lim_1_2N_1N = quant_1_2N_1N[0] - 1.5*(quant_1_2N_1N[2]-quant_1_2N_1N[0]) bp_1_1 = plt.boxplot(Kit_per_day_2N_1N, positions=[1], widths= 0.6) kitchen_pm_2N_1N_outlier = [] for v,a in enumerate(Kit_per_day_2N_1N): if a > Top_lim_1_2N_1N or a < Low_lim_1_2N_1N: kitchen_pm_2N_1N_outlier.append(K_PM_D_2N_1N[v]) plt.text(1, a, K_PM_D_2N_1N[v]) plt.text(0.5, -0.25, '2N / 1N', color= 'g') #3N to 1N quant_1_3N_1N = np.percentile(Kit_per_day_3N_1N, [25,50,75]) Top_lim_1_3N_1N = quant_1_3N_1N[2] + 1.5*(quant_1_3N_1N[2]-quant_1_3N_1N[0]) Low_lim_1_3N_1N = quant_1_3N_1N[0] - 1.5*(quant_1_3N_1N[2]-quant_1_3N_1N[0]) bp_1_1 = plt.boxplot(Kit_per_day_3N_1N, positions=[2], widths= 0.6) kitchen_pm_3N_1N_outlier = [] for v,a in enumerate(Kit_per_day_3N_1N): if a > Top_lim_1_3N_1N or a < Low_lim_1_3N_1N: kitchen_pm_3N_1N_outlier.append(K_PM_D_3N_1N[v]) plt.text(2, a, K_PM_D_3N_1N[v]) plt.text(1.5, -0.25, '3N / 1N', color= 'r') #4N to 1N quant_1_4N_1N = np.percentile(Kit_per_day_4N_1N, [25,50,75]) Top_lim_1_4N_1N = quant_1_4N_1N[2] + 1.5*(quant_1_4N_1N[2]-quant_1_4N_1N[0]) Low_lim_1_4N_1N = quant_1_4N_1N[0] - 1.5*(quant_1_4N_1N[2]-quant_1_4N_1N[0]) bp_1_1 = plt.boxplot(Kit_per_day_4N_1N, positions=[3], widths= 0.6) kitchen_pm_4N_1N_outlier = [] for v,a in enumerate(Kit_per_day_4N_1N): if a > Top_lim_1_4N_1N or a < Low_lim_1_4N_1N: kitchen_pm_4N_1N_outlier.append(K_PM_D_4N_1N[v]) plt.text(3, a, K_PM_D_4N_1N[v]) plt.text(2.5, -0.25, '4N / 1N', color= 'y') #3N to 2N quant_1_3N_2N = np.percentile(Kit_per_day_3N_2N, [25,50,75]) Top_lim_1_3N_2N = quant_1_3N_2N[2] + 1.5*(quant_1_3N_2N[2]-quant_1_3N_2N[0]) Low_lim_1_3N_2N = quant_1_3N_2N[0] - 1.5*(quant_1_3N_2N[2]-quant_1_3N_2N[0]) bp_1_1 = plt.boxplot(Kit_per_day_3N_2N, positions=[4], widths= 0.6) kitchen_pm_3N_2N_outlier = [] for v,a in enumerate(Kit_per_day_3N_2N): if a > Top_lim_1_3N_2N or a < Low_lim_1_3N_2N: kitchen_pm_3N_2N_outlier.append(K_PM_D_3N_2N[v]) plt.text(4, a, K_PM_D_3N_2N[v]) plt.text(3.5, -0.25, '3N / 2N', color= 'm') #4N to 3N quant_1_4N_3N = np.percentile(Kit_per_day_4N_3N, [25,50,75]) Top_lim_1_4N_3N = quant_1_4N_3N[2] + 1.5*(quant_1_4N_3N[2]-quant_1_4N_3N[0]) Low_lim_1_4N_3N = quant_1_4N_3N[0] - 1.5*(quant_1_4N_3N[2]-quant_1_4N_3N[0]) bp_1_1 = plt.boxplot(Kit_per_day_4N_3N, positions=[5], widths= 0.6) kitchen_pm_4N_3N_outlier = [] for v,a in enumerate(Kit_per_day_4N_3N): if a > Top_lim_1_4N_3N or a < Low_lim_1_4N_3N: kitchen_pm_4N_3N_outlier.append(K_PM_D_4N_3N[v]) plt.text(5, a, K_PM_D_4N_3N[v]) plt.text(4.5, -0.25, '4N / 3N', color= 'k') #4N to 2N quant_1_4N_2N = np.percentile(Kit_per_day_4N_2N, [25,50,75]) Top_lim_1_4N_2N = quant_1_4N_2N[2] + 1.5*(quant_1_4N_2N[2]-quant_1_4N_2N[0]) Low_lim_1_4N_2N = quant_1_4N_2N[0] - 1.5*(quant_1_4N_2N[2]-quant_1_4N_2N[0]) bp_1_1 = plt.boxplot(Kit_per_day_4N_2N, positions=[6], widths= 0.6) kitchen_pm_4N_2N_outlier = [] for v,a in enumerate(Kit_per_day_4N_2N): if a > Top_lim_1_4N_2N or a < Low_lim_1_4N_2N: kitchen_pm_4N_2N_outlier.append(K_PM_D_4N_2N[v]) plt.text(6, a, K_PM_D_4N_2N[v]) plt.text(5.5, -0.25, '4N / 2N', color= 'tab:orange') plt.xlim(0,7) plt.ylim(-0.5,5) print('Kitchen PM 2N/1N had these values as outliers ', kitchen_pm_2N_1N_outlier) print('Kitchen PM 3N/1N had these values as outliers ', kitchen_pm_3N_1N_outlier) print('Kitchen PM 4N/1N had these values as outliers ', kitchen_pm_4N_1N_outlier) print('Kitchen PM 3N/2N had these values as outliers ', kitchen_pm_3N_2N_outlier) print('Kitchen PM 4N/3N had these values as outliers ', kitchen_pm_4N_3N_outlier) print('Kitchen PM 4N/2N had these values as outliers ', kitchen_pm_4N_2N_outlier) plt.show() #adding averages to the tables quant_1_1N = np.append(quant_1_1N, np.average(Kit_PM_per_day_1N)) quant_1_2N = np.append(quant_1_2N, np.average(Kit_PM_per_day_2N)) quant_1_3N = np.append(quant_1_3N, np.average(Kit_PM_per_day_3N)) quant_1_4N = np.append(quant_1_4N, np.average(Kit_PM_per_day_4N)) D_50_quant_phase_PM_d = {'Percentile %': ['25','50','75', 'Avg'], '1N': quant_1_1N, '2N': quant_1_2N,'3N' : quant_1_3N,'4N': quant_1_4N} PM_D_50_phase_no_hood = pd.DataFrame(data=D_50_quant_phase_PM_d,columns=['Percentile %','1N', '2N', '3N','4N']) quant_1_2N_1N = np.append(quant_1_2N_1N , np.average(Kit_per_day_2N_1N)) quant_1_3N_1N = np.append(quant_1_3N_1N , np.average(Kit_per_day_3N_1N)) quant_1_4N_1N = np.append(quant_1_4N_1N , np.average(Kit_per_day_4N_1N)) quant_1_3N_2N = np.append(quant_1_3N_2N , np.average(Kit_per_day_3N_2N)) quant_1_4N_3N = np.append(quant_1_4N_3N , np.average(Kit_per_day_4N_3N)) quant_1_4N_2N = np.append(quant_1_4N_2N , np.average(Kit_per_day_4N_2N)) D_50_quant_percent_PM_d ={'Percentile %': ['25','50','75', 'Avg'],'2N / 1N': quant_1_2N_1N,'3N / 1N': quant_1_3N_1N,'4N / 1N': quant_1_4N_1N, '3N / 2N': quant_1_3N_2N,'4N / 3N': quant_1_4N_3N,'4N / 2N': quant_1_4N_2N} PM_D_50_percent_change_no_hood = pd.DataFrame(data=D_50_quant_percent_PM_d, columns=['Percentile %','2N / 1N','3N / 1N', '4N / 1N' ,'3N / 2N','4N / 3N','4N / 2N']) print(PM_D_50_phase_no_hood) print(PM_D_50_percent_change_no_hood) # hood Pm per day if Hood_or_no == 'hood': Kit_PM_per_day_1H = [] K_PM_D_1H = [] Kit_PM_per_day_2H = [] K_PM_D_2H = [] Kit_PM_per_day_3H = [] K_PM_D_3H = [] count_t = 0 count_pm = 0 for c in hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_Hood_PM[count_pm] - C_Place_holder): count_pm = count_pm + 1 if count_pm == len(Household_removal_Hood_PM): count_pm = 0 continue # if Day_1H.iloc[c,7] != -1.00: # Kit_PM_per_day_1H.append(Day_1H.iloc[c,7]/Day_1H.iloc[c,1]) # K_PM_D_1H.append(Day_1H.iloc[c,0]) if Kit_PM_1H_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_1H.append(Kit_PM_1H_24hr.iloc[c,6]) K_PM_D_1H.append(Kit_PM_1H_24hr.iloc[c,0]) # if Day_2H.iloc[c, 7] != -1.00: # Kit_PM_per_day_2H.append(Day_2H.iloc[c,7]/Day_2H.iloc[c,1]) # K_PM_D_2H.append(Day_2H.iloc[c,0]) if Kit_PM_2H_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_2H.append(Kit_PM_2H_24hr.iloc[c,6]) K_PM_D_2H.append(Kit_PM_2H_24hr.iloc[c,0]) # if Day_3H.iloc[c, 7] != -1.00: # Kit_PM_per_day_3H.append(Day_3H.iloc[c,7]/Day_3H.iloc[c,1]) # K_PM_D_3H.append(Day_3H.iloc[c, 0]) if Kit_PM_3H_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_3H.append(Kit_PM_3H_24hr.iloc[c,6]) K_PM_D_3H.append(Kit_PM_3H_24hr.iloc[c,0]) # percentages Between Phases of kitchen PM per day Kit_per_day_2H_1H = [] K_PM_D_2H_1H = [] Kit_per_day_3H_1H = [] K_PM_D_3H_1H = [] Kit_per_day_3H_2H = [] K_PM_D_3H_2H = [] count_t = 0 count_pm = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_Hood_PM[count_pm] - C_Place_holder): count_pm = count_pm + 1 if count_pm == len(Household_removal_Hood_PM): count_pm = 0 continue if (len(Kit_PM_per_day_2H)-1) >= c and (len(Kit_PM_per_day_1H)-1) >= c: #if Day_1H.iloc[c,7] > 0 and Day_2H.iloc[c,7] > 0 and Day_1H.iloc[c,0] == Day_2H.iloc[c,0]: if Kit_PM_1H_24hr.iloc[c, 6] > 0 and Kit_PM_2H_24hr.iloc[c, 6] > 0 and Kit_PM_1H_24hr.iloc[c, 0] == Kit_PM_2H_24hr.iloc[c, 0]: Kit_per_day_2H_1H.append(Kit_PM_per_day_2H[c]/Kit_PM_per_day_1H[c]) K_PM_D_2H_1H.append(Day_1H.iloc[c,0]) if (len(Kit_PM_per_day_3H)-1) >= c and (len(Kit_PM_per_day_1H)-1) >= c: #if Day_3H.iloc[c,7] > 0 and Day_1H.iloc[c,7] > 0 and Day_3H.iloc[c,0] == Day_1H.iloc[c,0]: if Kit_PM_3H_24hr.iloc[c, 6] > 0 and Kit_PM_1H_24hr.iloc[c, 6] > 0 and Kit_PM_1H_24hr.iloc[c, 0] == \ Kit_PM_3H_24hr.iloc[c, 0]: Kit_per_day_3H_1H.append(Kit_PM_per_day_3H[c]/Kit_PM_per_day_1H[c]) K_PM_D_3H_1H.append(Day_1H.iloc[c,0]) if (len(Kit_PM_per_day_3H)-1) >= c and (len(Kit_PM_per_day_2H)-1) >= c: #if Day_3H.iloc[c,7] > 0 and Day_2H.iloc[c,7] > 0 and Day_3H.iloc[c,0] == Day_2H.iloc[c,0]: if Kit_PM_3H_24hr.iloc[c, 6] > 0 and Kit_PM_2H_24hr.iloc[c, 6] > 0 and Kit_PM_3H_24hr.iloc[c, 0] == \ Kit_PM_2H_24hr.iloc[c, 0]: Kit_per_day_3H_2H.append(Kit_PM_per_day_3H[c]/Kit_PM_per_day_2H[c]) K_PM_D_3H_2H.append(Day_2H.iloc[c,0]) # now for box plotting for Kitchen PM per day percent changes #2H to 1H sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_2H_1H, ax=ax_box, color='g') sns.distplot(Kit_per_day_2H_1H, ax=ax_hist, color='g') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 2H/1H (Kitchen PM per Day)') plt.ylim(top=1.5) plt.ylim(bottom = 0) #3H to 1H sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_3H_1H, ax=ax_box, color='r') sns.distplot(Kit_per_day_3H_1H, ax=ax_hist, color='r') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 3H/1H (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) #3H to 2H sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_3H_2H, ax=ax_box, color='m') sns.distplot(Kit_per_day_3H_2H, ax=ax_hist, color='m') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 3H/2H (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) #Plotting on the same graph fig, ax = plt.subplots() plt.title('Hood Kitchen PM per day') #1H quant_1_1H = np.percentile(Kit_PM_per_day_1H, [25,50,75]) Top_lim_1_1H = quant_1_1H[2] + 1.5*(quant_1_1H[2] - quant_1_1H[0]) Low_lim_1_1H = quant_1_1H[0] - 1.5*(quant_1_1H[2] - quant_1_1H[0]) bp_1 = plt.boxplot(Kit_PM_per_day_1H, positions = [1], widths = 0.6) kitchen_pm_1H_outlier = [] for v,a in enumerate(Kit_PM_per_day_1H): if a > Top_lim_1_1H or a < Low_lim_1_1H: kitchen_pm_1H_outlier.append(K_PM_D_1H[v]) plt.text(1,a,K_PM_D_1H[v]) plt.text(0.5,0.1,'1H',color='b') #2N quant_1_2H = np.percentile(Kit_PM_per_day_2H, [25,50,75]) Top_lim_1_2N = quant_1_2H[2] + 1.5*(quant_1_2H[2] - quant_1_2H[0]) Low_lim_1_2N = quant_1_2H[0] - 1.5*(quant_1_2H[2] - quant_1_2H[0]) bp_1 = plt.boxplot(Kit_PM_per_day_2H,positions = [2], widths = 0.6) kitchen_pm_2H_outlier = [] for v,a in enumerate(Kit_PM_per_day_2H): if a > Top_lim_1_2N or a < Low_lim_1_2N: kitchen_pm_2H_outlier.append(K_PM_D_2H[v]) plt.text(2,a,K_PM_D_2H[v]) plt.text(1.5,0.1,'2H', color= 'g') #3H quant_1_3H = np.percentile(Kit_PM_per_day_3H, [25,50,75]) Top_lim_1_3N = quant_1_3H[2] + 1.5*(quant_1_3H[2] - quant_1_3H[0]) Low_lim_1_3N = quant_1_3H[0] - 1.5*(quant_1_3H[2] - quant_1_3H[0]) kitchen_3H_outlier = [] bp_1 = plt.boxplot(Kit_PM_per_day_3H,positions = [3], widths = 0.6) count = 0 kitchen_pm_3H_outlier = [] for v,a in enumerate(Kit_PM_per_day_3H): if a > Top_lim_1_3N or a < Low_lim_1_3N: kitchen_pm_3H_outlier.append(K_PM_D_3H[v]) plt.text(3,a,K_PM_D_3H[v]) # kitchen_3N_outlier.append(K_PM_D_3N[v]) # count = count + 1 # if count == (3): # plt.text(3,a,K_PM_D_3N[v],ha='left', va='bottom') # if count == (1): # plt.text(3,a,K_PM_D_3N[v],ha='left', va='top') # else: # plt.text(3,a,K_PM_D_3N[v],ha='right', va='bottom') plt.text(2.5,0.1,'3H', color='r') plt.xlim(0,4) plt.ylim(0,1200) print('Kitchen PM 1H had these values as outliers ', kitchen_pm_1H_outlier) print('Kitchen PM 2H had these values as outliers ', kitchen_pm_2H_outlier) print('Kitchen PM 3H had these values as outliers ', kitchen_pm_3H_outlier) plt.show() #print('3N had these values as outliers ' , kitchen_3N_outlier) # % change of PM per day fig_2, ax2 = plt.subplots() plt.title('% hood PM per Day Change' ) #plt.hold(True) #2H to 1H quant_1_2H_1H = np.percentile(Kit_per_day_2H_1H, [25,50,75]) Top_lim_1_2N_1N = quant_1_2H_1H[2] + 1.5*(quant_1_2H_1H[2]-quant_1_2H_1H[0]) Low_lim_1_2N_1N = quant_1_2H_1H[0] - 1.5*(quant_1_2H_1H[2]-quant_1_2H_1H[0]) bp_1_1 = plt.boxplot(Kit_per_day_2H_1H, positions=[1], widths= 0.6) kitchen_pm_2H_1H_outlier = [] for v,a in enumerate(Kit_per_day_2H_1H): if a > Top_lim_1_2N_1N or a < Low_lim_1_2N_1N: kitchen_pm_2H_1H_outlier.append(K_PM_D_2H_1H[v]) plt.text(1, a, K_PM_D_2H_1H[v]) plt.text(0.75, -0.25, '2H / 1H', color= 'g') #3H to 1H quant_1_3H_1H = np.percentile(Kit_per_day_3H_1H, [25,50,75]) Top_lim_1_3N_1N = quant_1_3H_1H[2] + 1.5*(quant_1_3H_1H[2]-quant_1_3H_1H[0]) Low_lim_1_3N_1N = quant_1_3H_1H[0] - 1.5*(quant_1_3H_1H[2]-quant_1_3H_1H[0]) bp_1_1 = plt.boxplot(Kit_per_day_3H_1H, positions=[2], widths= 0.6) kitchen_pm_3H_1H_outlier = [] for v,a in enumerate(Kit_per_day_3H_1H): if a > Top_lim_1_3N_1N or a < Low_lim_1_3N_1N: kitchen_pm_3H_1H_outlier.append(K_PM_D_3H_1H[v]) plt.text(2, a, K_PM_D_3H_1H[v]) plt.text(1.75, -0.25, '3H / 1H', color= 'r') #3H to 2H quant_1_3H_2H = np.percentile(Kit_per_day_3H_2H, [25,50,75]) Top_lim_1_3N_2N = quant_1_3H_2H[2] + 1.5*(quant_1_3H_2H[2]-quant_1_3H_2H[0]) Low_lim_1_3N_2N = quant_1_3H_2H[0] - 1.5*(quant_1_3H_2H[2]-quant_1_3H_2H[0]) bp_1_1 = plt.boxplot(Kit_per_day_3H_2H, positions=[3], widths= 0.6) kitchen_pm_3H_2H_outlier = [] for v,a in enumerate(Kit_per_day_3H_2H): if a > Top_lim_1_3N_2N or a < Low_lim_1_3N_2N: kitchen_pm_3H_2H_outlier.append(K_PM_D_3H_2H[v]) plt.text(3, a, K_PM_D_3H_2H[v]) plt.text(2.75, -0.25, '3H / 2H', color= 'm') plt.xlim(0,4) plt.ylim(-0.5,5) print('Kitchen PM 2H/1H had these values as outliers ', kitchen_pm_2H_1H_outlier) print('Kitchen PM 3H/1H had these values as outliers ', kitchen_pm_3H_1H_outlier) print('Kitchen PM 3H/2H had these values as outliers ', kitchen_pm_3H_2H_outlier) plt.show() quant_1_1H = np.append(quant_1_1H, np.average(Kit_PM_per_day_1H)) quant_1_2H = np.append(quant_1_2H, np.average(Kit_PM_per_day_2H)) quant_1_3H = np.append(quant_1_3H, np.average(Kit_PM_per_day_3H)) D_50_quant_phase_PM_D_hood = {'Percentile %': ['25','50','75', 'Avg'], '1H': quant_1_1H, '2H': quant_1_2H,'3H' : quant_1_3H} PM_D_50_phase_hood = pd.DataFrame(data=D_50_quant_phase_PM_D_hood, columns= ['Percentile %','1H','2H','3H' ]) quant_1_2H_1H = np.append(quant_1_2H_1H , np.average(Kit_per_day_2H_1H)) quant_1_3H_1H = np.append(quant_1_3H_1H , np.average(Kit_per_day_3H_1H)) quant_1_3H_2H = np.append(quant_1_3H_2H , np.average(Kit_per_day_3H_2H)) D_50_quant_percent_PM_D_hood ={'Percentile %': ['25','50','75', 'Avg'],'2H / 1H': quant_1_2H_1H,'3H / 1H': quant_1_3H_1H,'3H / 2H': quant_1_3H_2H} PM_D_50_percent_change_hood = pd.DataFrame(data=D_50_quant_percent_PM_D_hood, columns=['Percentile %','2H / 1H','3H / 1H','3H / 2H']) print(PM_D_50_phase_hood) print(PM_D_50_percent_change_hood) # when i am ready to transfer to a data frame and get the differences #histograms for the comparison if Hood_or_no == 'no_hood': plt.title('Histogram of Fuel per 24 Hours per Person - No Hood' ) plt.hist([Fuel_per_day_per_adult_1N], color=['b'], alpha=0.5, label='1N') plt.hist([Fuel_per_day_per_adult_2N], color=['g'], alpha=0.5, label='2N') plt.hist([Fuel_per_day_per_adult_3N], color=['r'], alpha=0.5, label='3N') plt.hist([Fuel_per_day_per_adult_4N], color=['y'], alpha=0.5, label='4N') plt.legend(loc='upper right') plt.show() plt.title('Histogram of Kitchen PM 24 Hours - No Hood' ) plt.hist([Kit_PM_per_day_1N], color=['b'], alpha=0.5, label='1N') plt.hist([Kit_PM_per_day_2N], color=['g'], alpha=0.5, label='2N') plt.hist([Kit_PM_per_day_3N], color=['r'], alpha=0.5, label='3N') plt.hist([Kit_PM_per_day_4N], color=['y'], alpha=0.5, label='4N') plt.legend(loc='upper right') plt.show() if Hood_or_no == 'hood': plt.title('Histogram of Fuel per 24 Hours per Person - Hood' ) plt.hist([Fuel_per_day_per_adult_1H], color=['b'], alpha=0.5, label='1H') plt.hist([Fuel_per_day_per_adult_2H], color=['g'], alpha=0.5, label='2H') plt.hist([Fuel_per_day_per_adult_3H], color=['r'], alpha=0.5, label='3H') plt.legend(loc='upper right') plt.show() plt.title('Histogram of Kitchen PM 24 Hours - Hood' ) plt.hist([Kit_PM_per_day_1H], color=['b'], alpha=0.5, label='1H') plt.hist([Kit_PM_per_day_2H], color=['g'], alpha=0.5, label='2H') plt.hist([Kit_PM_per_day_3H], color=['r'], alpha=0.5, label='3H') plt.legend(loc='upper right') plt.show()
50.965767
152
0.657504
import matplotlib.pyplot as plt import numpy as np import pandas as pd import seaborn as sns Computer = 'personal' Hood_or_no = 'no_hood' Household_removal = [1045] Household_removal_NO_Hood_fuel_day_adult = [1045] Household_removal_Hood_fuel_day_adult = [2020] Household_removal_NO_Hood_PM = [1045] Household_removal_Hood_PM = [2020] pd.set_option('display.max_rows', 500) pd.set_option('display.max_columns', 500) pd.set_option('display.width', 1000) if Hood_or_no == 'hood': C_Place_holder = 2001 else: C_Place_holder = 1001 if Computer == 'personal' and Hood_or_no == 'no_hood': datafile_path_day_1N ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_Summary_Day_1_exact.csv" Day_1N = pd.read_csv(datafile_path_day_1N, skiprows=2) datafile_path_event_1N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_Summary_Event_1_exact.csv" Event_1N = pd.read_csv(datafile_path_event_1N, skiprows=2) datafile_path_survey_1N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_1H_Survey_summary_.csv" Filter_1n_survey = pd.read_csv(datafile_path_survey_1N, skiprows=0) Survey_1N = Filter_1n_survey.iloc[0:40,:] data_file_path_24_PM_1N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_24_hour_Kitchen_PM.csv" Kit_PM_1N_24hr = pd.read_csv(data_file_path_24_PM_1N, skiprows=0) data_file_path_24_Fuel_1N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_24_hour_Fuel_removal.csv" Fuel_remove_1N_24hr = pd.read_csv(data_file_path_24_Fuel_1N, skiprows=0) datafile_path_day_2N ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_Summary_Day_1_exact.csv" Day_2N = pd.read_csv(datafile_path_day_2N, skiprows=2) datafile_path_event_2N_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_Summary_Event_1_exact.csv" Event_2N_1 = pd.read_csv(datafile_path_event_2N_1, skiprows=2) datafile_path_event_2N_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_Summary_Event_2_exact.csv" Event_2N_2 = pd.read_csv(datafile_path_event_2N_2, skiprows=2) datafile_path_survey_2N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_Survey_summary_.csv" Survey_2N = pd.read_csv(datafile_path_survey_2N, skiprows=0) data_file_path_24_PM_2N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_24_hour_Kitchen_PM.csv" Kit_PM_2N_24hr = pd.read_csv(data_file_path_24_PM_2N, skiprows=0) data_file_path_24_Fuel_2N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2N/2N_24_hour_Fuel_removal.csv" Fuel_remove_2N_24hr = pd.read_csv(data_file_path_24_Fuel_2N, skiprows=0) datafile_path_day_3N ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_Summary_Day_1_exact.csv" Day_3N = pd.read_csv(datafile_path_day_3N, skiprows=2) datafile_path_event_3N_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_Summary_Event_1_exact.csv" Event_3N_1 = pd.read_csv(datafile_path_event_3N_1, skiprows=2) datafile_path_event_3N_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_Summary_Event_2_exact.csv" Event_3N_2 = pd.read_csv(datafile_path_event_3N_2, skiprows=2) datafile_path_survey_3N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_Survey_summary_.csv" Survey_3N = pd.read_csv(datafile_path_survey_3N, skiprows=0) data_file_path_24_PM_3N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_24_hour_Kitchen_PM.csv" Kit_PM_3N_24hr = pd.read_csv(data_file_path_24_PM_3N, skiprows=0) data_file_path_24_Fuel_3N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3N/3N_24_hour_Fuel_removal.csv" Fuel_remove_3N_24hr = pd.read_csv(data_file_path_24_Fuel_3N, skiprows=0) datafile_path_day_4N ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_Summary_Day_1_exact.csv" Day_4N = pd.read_csv(datafile_path_day_4N, skiprows=2) datafile_path_event_4N_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_Summary_Event_1_exact.csv" Event_4N_1 = pd.read_csv(datafile_path_event_4N_1, skiprows=2) datafile_path_event_4N_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_Summary_Event_2_exact.csv" Event_4N_2 = pd.read_csv(datafile_path_event_4N_2, skiprows=2) datafile_path_survey_4N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_Survey_summary_.csv" Survey_4N = pd.read_csv(datafile_path_survey_4N, skiprows=0) data_file_path_24_PM_4N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_24_hour_Kitchen_PM.csv" Kit_PM_4N_24hr = pd.read_csv(data_file_path_24_PM_4N, skiprows=0) data_file_path_24_Fuel_4N = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/4N/4N_24_hour_Fuel_removal.csv" Fuel_remove_4N_24hr = pd.read_csv(data_file_path_24_Fuel_4N, skiprows=0) elif Computer == 'personal' and Hood_or_no == 'hood': datafile_path_day_1H ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1H/1H_Summary_Day_1_exact.csv" Day_1H = pd.read_csv(datafile_path_day_1H, skiprows=2) datafile_path_event_1H ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1H/1H_Summary_Event_1_exact.csv" Event_1H = pd.read_csv(datafile_path_event_1H, skiprows=2) datafile_path_survey_1H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1N/1N_1H_Survey_summary_.csv" Survey_1H = pd.read_csv(datafile_path_survey_1H, skiprows=40) data_file_path_24_PM_1H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1H/1H_24_hour_Kitchen_PM.csv" Kit_PM_1H_24hr = pd.read_csv(data_file_path_24_PM_1H, skiprows=0) data_file_path_24_fuel_1H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/1H/1H_24_hour_Fuel_removal.csv" Fuel_remove_1H_24hr = pd.read_csv(data_file_path_24_fuel_1H, skiprows=0) datafile_path_day_2H ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_Summary_Day_1_exact.csv" Day_2H = pd.read_csv(datafile_path_day_2H, skiprows=2) datafile_path_event_2H_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_Summary_Event_1_exact.csv" Event_2H_1 = pd.read_csv(datafile_path_event_2H_1, skiprows=2) datafile_path_event_2H_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_Summary_Event_2_exact.csv" Event_2H_2 = pd.read_csv(datafile_path_event_2H_2, skiprows=2) datafile_path_survey_2H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_Survey_summary_.csv" Survey_2H = pd.read_csv(datafile_path_survey_2H, skiprows=0) data_file_path_24_PM_2H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_24_hour_Kitchen_PM.csv" Kit_PM_2H_24hr = pd.read_csv(data_file_path_24_PM_2H, skiprows=0) data_file_path_24_fuel_2H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/2H/2H_24_hour_Fuel_removal.csv" Fuel_remove_2H_24hr = pd.read_csv(data_file_path_24_fuel_2H, skiprows=0) datafile_path_day_3H ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_Summary_Day_1_exact.csv" Day_3H = pd.read_csv(datafile_path_day_3H, skiprows=2) datafile_path_event_3N_1 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_Summary_Event_1_exact.csv" Event_3H_1 = pd.read_csv(datafile_path_event_3N_1, skiprows=2) datafile_path_event_3H_2 ="C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_Summary_Event_2_exact.csv" Event_3H_2 = pd.read_csv(datafile_path_event_3H_2, skiprows=2) datafile_path_survey_3H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_Survey_summary_.csv" Survey_3H = pd.read_csv(datafile_path_survey_3H, skiprows=0) data_file_path_24_PM_3H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_24_hour_Kitchen_PM.csv" Kit_PM_3H_24hr = pd.read_csv(data_file_path_24_PM_3H, skiprows=0) data_file_path_24_fuel_3H = "C:/Users/gvros/Desktop/Oregon State Masters/Work/OSU, CSC, CQC Project files/3H/3H_24_hour_Fuel_removal.csv" Fuel_remove_3H_24hr = pd.read_csv(data_file_path_24_fuel_3H, skiprows=0) elif Computer == 'work' and Hood_or_no == 'no_hood': datafile_path_day_1N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_Summary_Day_1_exact.csv" Day_1N = pd.read_csv(datafile_path_day_1N, skiprows=2) datafile_path_event_1N ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_Summary_Event_1_exact.csv" Event_1N = pd.read_csv(datafile_path_event_1N, skiprows=2) datafile_path_survey_1N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_1H_Survey_summary_.csv" Filter_1n_survey = pd.read_csv(datafile_path_survey_1N, skiprows=0) Survey_1N = Filter_1n_survey.iloc[0:40,:] data_file_path_24_PM_1N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_24_hour_Kitchen_PM.csv" Kit_PM_1N_24hr = pd.read_csv(data_file_path_24_PM_1N, skiprows=0) data_file_path_24_Fuel_1N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_24_hour_Fuel_removal.csv" Fuel_remove_1N_24hr = pd.read_csv(data_file_path_24_Fuel_1N, skiprows=0) datafile_path_day_2N ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_Summary_Day_1_exact.csv" Day_2N = pd.read_csv(datafile_path_day_2N, skiprows=2) datafile_path_event_2N_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_Summary_Event_1_exact.csv" Event_2N_1 = pd.read_csv(datafile_path_event_2N_1, skiprows=2) datafile_path_event_2N_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_Summary_Event_2_exact.csv" Event_2N_2 = pd.read_csv(datafile_path_event_2N_2, skiprows=2) datafile_path_survey_2N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_Survey_summary_.csv" Survey_2N = pd.read_csv(datafile_path_survey_2N, skiprows=0) data_file_path_24_PM_2N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_24_hour_Kitchen_PM.csv" Kit_PM_2N_24hr = pd.read_csv(data_file_path_24_PM_2N, skiprows=0) data_file_path_24_Fuel_2N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2N/2N_24_hour_Fuel_removal.csv" Fuel_remove_2N_24hr = pd.read_csv(data_file_path_24_Fuel_2N, skiprows=0) datafile_path_day_3N ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_Summary_Day_1_exact.csv" Day_3N = pd.read_csv(datafile_path_day_3N, skiprows=2) datafile_path_event_3N_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_Summary_Event_1_exact.csv" Event_3N_1 = pd.read_csv(datafile_path_event_3N_1, skiprows=2) datafile_path_event_3N_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_Summary_Event_2_exact.csv" Event_3N_2 = pd.read_csv(datafile_path_event_3N_2, skiprows=2) datafile_path_survey_3N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_Survey_summary_.csv" Survey_3N = pd.read_csv(datafile_path_survey_3N, skiprows=0) data_file_path_24_PM_3N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_24_hour_Kitchen_PM.csv" Kit_PM_3N_24hr = pd.read_csv(data_file_path_24_PM_3N, skiprows=0) data_file_path_24_Fuel_3N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3N/3N_24_hour_Fuel_removal.csv" Fuel_remove_3N_24hr = pd.read_csv(data_file_path_24_Fuel_3N, skiprows=0) datafile_path_day_4N ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_Summary_Day_1_exact.csv" Day_4N = pd.read_csv(datafile_path_day_4N, skiprows=2) datafile_path_event_4N_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_Summary_Event_1_exact.csv" Event_4N_1 = pd.read_csv(datafile_path_event_4N_1, skiprows=2) datafile_path_event_4N_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_Summary_Event_2_exact.csv" Event_4N_2 = pd.read_csv(datafile_path_event_4N_2, skiprows=2) datafile_path_survey_4N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_Survey_summary_.csv" Survey_4N = pd.read_csv(datafile_path_survey_4N, skiprows=0) data_file_path_24_PM_4N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_24_hour_Kitchen_PM.csv" Kit_PM_4N_24hr = pd.read_csv(data_file_path_24_PM_4N, skiprows=0) data_file_path_24_Fuel_4N = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/4N/4N_24_hour_Fuel_removal.csv" Fuel_remove_4N_24hr = pd.read_csv(data_file_path_24_Fuel_4N, skiprows=0) else: datafile_path_day_1H ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1H/1H_Summary_Day_1_exact.csv" Day_1H = pd.read_csv(datafile_path_day_1H, skiprows=2) datafile_path_event_1H ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1H/1H_Summary_Event_1_exact.csv" Event_1H = pd.read_csv(datafile_path_event_1H, skiprows=2) datafile_path_survey_1H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1N/1N_1H_Survey_summary_.csv" Survey_1H = pd.read_csv(datafile_path_survey_1H, skiprows=40) data_file_path_24_PM_1H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1H/1H_24_hour_Kitchen_PM.csv" Kit_PM_1H_24hr = pd.read_csv(data_file_path_24_PM_1H, skiprows=0) data_file_path_24_fuel_1H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/1H/1H_24_hour_Fuel_removal.csv" Fuel_remove_1H_24hr = pd.read_csv(data_file_path_24_fuel_1H, skiprows=0) datafile_path_day_2H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_Summary_Day_1_exact.csv" Day_2H = pd.read_csv(datafile_path_day_2H, skiprows=2) datafile_path_event_2H_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_Summary_Event_1_exact.csv" Event_2H_1 = pd.read_csv(datafile_path_event_2H_1, skiprows=2) datafile_path_event_2H_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_Summary_Event_2_exact.csv" Event_2H_2 = pd.read_csv(datafile_path_event_2H_2, skiprows=2) datafile_path_survey_2H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_Survey_summary_.csv" Survey_2H = pd.read_csv(datafile_path_survey_2H, skiprows=0) data_file_path_24_PM_2H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_24_hour_Kitchen_PM.csv" Kit_PM_2H_24hr = pd.read_csv(data_file_path_24_PM_2H, skiprows=0) data_file_path_24_fuel_2H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/2H/2H_24_hour_Fuel_removal.csv" Fuel_remove_2H_24hr = pd.read_csv(data_file_path_24_fuel_2H, skiprows=0) datafile_path_day_3H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_Summary_Day_1_exact.csv" Day_3H = pd.read_csv(datafile_path_day_3H, skiprows=2) datafile_path_event_3N_1 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_Summary_Event_1_exact.csv" Event_3H_1 = pd.read_csv(datafile_path_event_3N_1, skiprows=2) datafile_path_event_3H_2 ="C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_Summary_Event_2_exact.csv" Event_3H_2 = pd.read_csv(datafile_path_event_3H_2, skiprows=2) datafile_path_survey_3H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_Survey_summary_.csv" Survey_3H = pd.read_csv(datafile_path_survey_3H, skiprows=0) data_file_path_24_PM_3H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_24_hour_Kitchen_PM.csv" Kit_PM_3H_24hr = pd.read_csv(data_file_path_24_PM_3H, skiprows=0) data_file_path_24_fuel_3H = "C:/Users/rossgra/Box/OSU, CSC, CQC Project files/3H/3H_24_hour_Fuel_removal.csv" Fuel_remove_3H_24hr = pd.read_csv(data_file_path_24_fuel_3H, skiprows=0) NO_hood_counter = np.arange(0,39) hood_counter = np.arange(0,14) print('---------------Fuel per Day per Adult No-Hood Phase---------------------') if Hood_or_no == 'no_hood': Fuel_per_day_per_adult_1N = [] f_d_a_1N = [] Fuel_per_day_per_adult_2N = [] f_d_a_2N = [] Fuel_per_day_per_adult_3N = [] f_d_a_3N = [] Fuel_per_day_per_adult_4N = [] f_d_a_4N =[] count_t = 0 count_f = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_NO_Hood_fuel_day_adult[count_f] - C_Place_holder): count_f = count_f + 1 if count_f == len(Household_removal_NO_Hood_fuel_day_adult): count_f = 0 continue if Fuel_remove_1N_24hr.iloc[c,6]!= -1.00: Fuel_per_day_per_adult_1N.append(Fuel_remove_1N_24hr.iloc[c,6]/Survey_1N.iloc[c,7]) f_d_a_1N.append(Day_1N.iloc[c,0]) if Fuel_remove_2N_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_2N.append(Fuel_remove_2N_24hr.iloc[c,6] / Survey_2N.iloc[c, 7]) f_d_a_2N.append(Day_2N.iloc[c,0]) if Fuel_remove_3N_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_3N.append(Fuel_remove_3N_24hr.iloc[c,6]/ Survey_3N.iloc[c, 7]) f_d_a_3N.append(Day_3N.iloc[c, 0]) if Fuel_remove_4N_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_4N.append(Fuel_remove_4N_24hr.iloc[c,6] / Survey_4N.iloc[c, 7]) f_d_a_4N.append(Day_3N.iloc[c, 0]) Fuel_per_day_per_adult_2N_1N = [] f_d_a_2N_1N = [] Fuel_per_day_per_adult_3N_1N = [] f_d_a_3N_1N = [] Fuel_per_day_per_adult_4N_1N = [] f_d_a_4N_1N = [] Fuel_per_day_per_adult_3N_2N = [] f_d_a_3N_2N = [] Fuel_per_day_per_adult_4N_3N = [] f_d_a_4N_3N = [] Fuel_per_day_per_adult_4N_2N = [] f_d_a_4N_2N = [] count_t = 0 count_f = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_NO_Hood_fuel_day_adult[count_f] - C_Place_holder): count_f = count_f + 1 if count_f == len(Household_removal_NO_Hood_fuel_day_adult): count_f = 0 continue if (len(Fuel_per_day_per_adult_2N)-1) >= c and (len(Fuel_per_day_per_adult_1N)-1) >= c: if Day_1N.iloc[c,13] > 0 and Day_2N.iloc[c,13] > 0 and Day_1N.iloc[c,0] == Day_2N.iloc[c,0]: Fuel_per_day_per_adult_2N_1N.append(Fuel_per_day_per_adult_2N[c]/Fuel_per_day_per_adult_1N[c]) f_d_a_2N_1N.append(Day_1N.iloc[c,0]) if (len(Fuel_per_day_per_adult_3N)-1) >= c and (len(Fuel_per_day_per_adult_1N)-1) >= c: if Day_3N.iloc[c,13] > 0 and Day_1N.iloc[c,13] > 0 and Day_3N.iloc[c,0] == Day_1N.iloc[c,0]: Fuel_per_day_per_adult_3N_1N.append(Fuel_per_day_per_adult_3N[c]/Fuel_per_day_per_adult_1N[c]) f_d_a_3N_1N.append(Day_1N.iloc[c,0]) if (len(Fuel_per_day_per_adult_4N)-1) >= c and (len(Fuel_per_day_per_adult_1N)-1) >= c: if Day_4N.iloc[c,13] > 0 and Day_1N.iloc[c,13] > 0 and Day_4N.iloc[c,0] == Day_1N.iloc[c,0]: Fuel_per_day_per_adult_4N_1N.append(Fuel_per_day_per_adult_4N[c]/Fuel_per_day_per_adult_1N[c]) f_d_a_4N_1N.append(Day_1N.iloc[c,0]) if (len(Fuel_per_day_per_adult_3N)-1) >= c and (len(Fuel_per_day_per_adult_2N)-1) >= c: if Day_3N.iloc[c,13] > 0 and Day_2N.iloc[c,13] > 0 and Day_3N.iloc[c,0] == Day_2N.iloc[c,0]: Fuel_per_day_per_adult_3N_2N.append(Fuel_per_day_per_adult_3N[c]/Fuel_per_day_per_adult_2N[c]) f_d_a_3N_2N.append(Day_2N.iloc[c,0]) if (len(Fuel_per_day_per_adult_4N)-1) >= c and (len(Fuel_per_day_per_adult_3N)-1) >= c: if Day_4N.iloc[c,13] > 0 and Day_3N.iloc[c,13] > 0 and Day_4N.iloc[c,0] == Day_3N.iloc[c,0]: Fuel_per_day_per_adult_4N_3N.append(Fuel_per_day_per_adult_4N[c]/Fuel_per_day_per_adult_3N[c]) f_d_a_4N_3N.append(Day_3N.iloc[c,0]) if (len(Fuel_per_day_per_adult_4N)-1) >= c and (len(Fuel_per_day_per_adult_2N)-1) >= c: if Day_4N.iloc[c,13] > 0 and Day_2N.iloc[c,13] > 0 and Day_4N.iloc[c,0] == Day_2N.iloc[c,0]: Fuel_per_day_per_adult_4N_2N.append(Fuel_per_day_per_adult_4N[c]/Fuel_per_day_per_adult_2N[c]) f_d_a_4N_2N.append(Day_4N.iloc[c,0]) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_1N, ax=ax_box, color='b') sns.distplot(Fuel_per_day_per_adult_1N, ax=ax_hist, color='b') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('1N Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_2N, ax=ax_box, color='g') sns.distplot(Fuel_per_day_per_adult_2N, ax=ax_hist, color='g') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('2N Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_3N, ax=ax_box, color='r') sns.distplot(Fuel_per_day_per_adult_3N, ax=ax_hist, color='r') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('3N Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_4N, ax=ax_box, color='y') sns.distplot(Fuel_per_day_per_adult_4N, ax=ax_hist, color='y') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('4N Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) fig, ax = plt.subplots() plt.title('No-Hood Fuel per Day per Adult') quant_1_1N = np.percentile(Fuel_per_day_per_adult_1N, [25,50,75]) Top_lim_1_1N = quant_1_1N[2] + 1.5*(quant_1_1N[2] - quant_1_1N[0]) Low_lim_1_1N = quant_1_1N[0] - 1.5*(quant_1_1N[2] - quant_1_1N[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_1N, positions = [1], widths = 0.6) Fuel_D_A_1N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_1N): if a > Top_lim_1_1N or a < Low_lim_1_1N: Fuel_D_A_1N_outlier.append(f_d_a_1N[v]) plt.text(1,a,f_d_a_1N[v]) plt.text(1,0.1,'1N',color='b') quant_1_2N = np.percentile(Fuel_per_day_per_adult_2N, [25,50,75]) Top_lim_1_2N = quant_1_2N[2] + 1.5*(quant_1_2N[2] - quant_1_2N[0]) Low_lim_1_2N = quant_1_2N[0] - 1.5*(quant_1_2N[2] - quant_1_2N[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_2N,positions = [2], widths = 0.6) Fuel_D_A_2N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_2N): if a > Top_lim_1_2N or a < Low_lim_1_2N: Fuel_D_A_2N_outlier.append(f_d_a_2N[v]) plt.text(2,a,f_d_a_2N[v]) plt.text(2,0.1,'2N', color= 'g') quant_1_3N = np.percentile(Fuel_per_day_per_adult_3N, [25,50,75]) Top_lim_1_3N = quant_1_3N[2] + 1.5*(quant_1_3N[2] - quant_1_3N[0]) Low_lim_1_3N = quant_1_3N[0] - 1.5*(quant_1_3N[2] - quant_1_3N[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_3N,positions = [3], widths = 0.6) count = 0 Fuel_D_A_3N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3N): if a > Top_lim_1_3N or a < Low_lim_1_3N: Fuel_D_A_3N_outlier.append(f_d_a_3N[v]) count = count + 1 if count == 2: plt.text(3,a,f_d_a_3N[v],ha='left',va='bottom') elif count != 2: plt.text(3,a,f_d_a_3N[v],ha='right',va='bottom') plt.text(3,0.1,'3N', color='r') quant_1_4N = np.percentile(Fuel_per_day_per_adult_4N, [25,50,75]) Top_lim_1_4N = quant_1_4N[2] + 1.5*(quant_1_4N[2] - quant_1_4N[0]) Low_lim_1_4N = quant_1_4N[0] - 1.5*(quant_1_4N[2] - quant_1_4N[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_4N,positions = [4], widths = 0.6) Fuel_D_A_4N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_4N): if a > Top_lim_1_4N or a < Low_lim_1_4N: Fuel_D_A_4N_outlier.append(f_d_a_4N[v]) plt.text(4,a,f_d_a_4N[v]) plt.text(4,0.1,'4N', color='y') plt.xlim(0,5) plt.ylim(0,2.3) print('Fuel/Day/Adult 1N had these values as outliers ', Fuel_D_A_1N_outlier) print('Fuel/Day/Adult 2N had these values as outliers ', Fuel_D_A_2N_outlier) print('Fuel/Day/Adult 3N had these values as outliers ', Fuel_D_A_3N_outlier) print('Fuel/Day/Adult 4N had these values as outliers ', Fuel_D_A_4N_outlier) plt.show() fig_2, ax2 = plt.subplots() plt.title('% No_hood Change from Fuel per Day per Adult' ) quant_1_2N_1N = np.percentile(Fuel_per_day_per_adult_2N_1N, [25,50,75]) Top_lim_1_2N_1N = quant_1_2N_1N[2] + 1.5*(quant_1_2N_1N[2]-quant_1_2N_1N[0]) Low_lim_1_2N_1N = quant_1_2N_1N[0] - 1.5*(quant_1_2N_1N[2]-quant_1_2N_1N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_2N_1N, positions=[1], widths= 0.6) Fuel_D_A_2N_1N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_2N_1N): if a > Top_lim_1_2N_1N or a < Low_lim_1_2N_1N: Fuel_D_A_2N_1N_outlier.append(f_d_a_2N_1N[v]) plt.text(1, a, f_d_a_2N_1N[v]) plt.text(0.5, 0, '2N / 1N', color= 'g') quant_1_3N_1N = np.percentile(Fuel_per_day_per_adult_3N_1N, [25,50,75]) Top_lim_1_3N_1N = quant_1_3N_1N[2] + 1.5*(quant_1_3N_1N[2]-quant_1_3N_1N[0]) Low_lim_1_3N_1N = quant_1_3N_1N[0] - 1.5*(quant_1_3N_1N[2]-quant_1_3N_1N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_3N_1N, positions=[2], widths= 0.6) Fuel_D_A_3N_1N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3N_1N): if a > Top_lim_1_3N_1N or a < Low_lim_1_3N_1N: Fuel_D_A_3N_1N_outlier.append(f_d_a_3N_1N[v]) plt.text(2, a, f_d_a_3N_1N[v]) plt.text(1.5, 0, '3N / 1N', color= 'r') quant_1_4N_1N = np.percentile(Fuel_per_day_per_adult_4N_1N, [25,50,75]) Top_lim_1_4N_1N = quant_1_4N_1N[2] + 1.5*(quant_1_4N_1N[2]-quant_1_4N_1N[0]) Low_lim_1_4N_1N = quant_1_4N_1N[0] - 1.5*(quant_1_4N_1N[2]-quant_1_4N_1N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_4N_1N, positions=[3], widths= 0.6) Fuel_D_A_4N_1N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_4N_1N): if a > Top_lim_1_4N_1N or a < Low_lim_1_4N_1N: Fuel_D_A_4N_1N_outlier.append(f_d_a_4N_1N[v]) plt.text(3, a, f_d_a_4N_1N[v]) plt.text(2.5, 0, '4N / 1N', color= 'y') quant_1_3N_2N = np.percentile(Fuel_per_day_per_adult_3N_2N, [25,50,75]) Top_lim_1_3N_2N = quant_1_3N_2N[2] + 1.5*(quant_1_3N_2N[2]-quant_1_3N_2N[0]) Low_lim_1_3N_2N = quant_1_3N_2N[0] - 1.5*(quant_1_3N_2N[2]-quant_1_3N_2N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_3N_2N, positions=[4], widths= 0.6) Fuel_D_A_3N_2N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3N_2N): if a > Top_lim_1_3N_2N or a < Low_lim_1_3N_2N: Fuel_D_A_3N_2N_outlier.append(f_d_a_3N_2N[v]) plt.text(4, a, f_d_a_3N_2N[v]) plt.text(3.5, 0, '3N / 2N', color= 'm') quant_1_4N_3N = np.percentile(Fuel_per_day_per_adult_4N_3N, [25,50,75]) Top_lim_1_4N_3N = quant_1_4N_3N[2] + 1.5*(quant_1_4N_3N[2]-quant_1_4N_3N[0]) Low_lim_1_4N_3N = quant_1_4N_3N[0] - 1.5*(quant_1_4N_3N[2]-quant_1_4N_3N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_4N_3N, positions=[5], widths= 0.6) Fuel_D_A_4N_3N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_4N_3N): if a > Top_lim_1_4N_3N or a < Low_lim_1_4N_3N: Fuel_D_A_4N_3N_outlier.append(f_d_a_4N_3N[v]) plt.text(5, a, f_d_a_4N_3N[v]) plt.text(4.5, 0, '4N / 3N', color= 'k') quant_1_4N_2N = np.percentile(Fuel_per_day_per_adult_4N_2N, [25,50,75]) Top_lim_1_4N_2N = quant_1_4N_2N[2] + 1.5*(quant_1_4N_2N[2]-quant_1_4N_2N[0]) Low_lim_1_4N_2N = quant_1_4N_2N[0] - 1.5*(quant_1_4N_2N[2]-quant_1_4N_2N[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_4N_2N, positions=[6], widths= 0.6) Fuel_D_A_4N_2N_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_4N_2N): if a > Top_lim_1_4N_2N or a < Low_lim_1_4N_2N: Fuel_D_A_4N_2N_outlier.append(f_d_a_4N_2N[v]) plt.text(6, a, f_d_a_4N_2N[v]) plt.text(5.5, 0, '4N / 2N', color= 'tab:orange') plt.xlim(0,7) plt.ylim(-0.5,4) print('Fuel/Day/Adult 2N/1N had these values as outliers ', Fuel_D_A_2N_1N_outlier) print('Fuel/Day/Adult 3N/1N had these values as outliers ', Fuel_D_A_3N_1N_outlier) print('Fuel/Day/Adult 4N/1N had these values as outliers ', Fuel_D_A_4N_1N_outlier) print('Fuel/Day/Adult 3N/2N had these values as outliers ', Fuel_D_A_3N_2N_outlier) print('Fuel/Day/Adult 4N/3N had these values as outliers ', Fuel_D_A_4N_3N_outlier) print('Fuel/Day/Adult 4N/2N had these values as outliers ', Fuel_D_A_4N_2N_outlier) plt.show() quant_1_1N = np.append(quant_1_1N, np.average(Fuel_per_day_per_adult_1N)) quant_1_2N = np.append(quant_1_2N, np.average(Fuel_per_day_per_adult_2N)) quant_1_3N = np.append(quant_1_3N, np.average(Fuel_per_day_per_adult_3N)) quant_1_4N = np.append(quant_1_4N, np.average(Fuel_per_day_per_adult_4N)) D_50_quant_phase_f_d_a = {'Percentile %': ['25','50','75', 'Avg'], '1N': quant_1_1N, '2N': quant_1_2N,'3N' : quant_1_3N,'4N': quant_1_4N} F_D_A_50_phase_no_hood = pd.DataFrame(data=D_50_quant_phase_f_d_a, columns=['Percentile %','1N', '2N', '3N','4N']) quant_1_2N_1N = np.append(quant_1_2N_1N , np.average(Fuel_per_day_per_adult_2N_1N)) quant_1_3N_1N = np.append(quant_1_3N_1N , np.average(Fuel_per_day_per_adult_3N_1N)) quant_1_4N_1N = np.append(quant_1_4N_1N , np.average(Fuel_per_day_per_adult_4N_1N)) quant_1_3N_2N = np.append(quant_1_3N_2N , np.average(Fuel_per_day_per_adult_3N_2N)) quant_1_4N_3N = np.append(quant_1_4N_3N , np.average(Fuel_per_day_per_adult_4N_3N)) quant_1_4N_2N = np.append(quant_1_4N_2N , np.average(Fuel_per_day_per_adult_4N_2N)) D_50_quant_percent_f_d_a ={'Percentile %': ['25','50','75', 'Avg'],'2N / 1N': quant_1_2N_1N,'3N / 1N': quant_1_3N_1N,'4N / 1N': quant_1_4N_1N, '3N / 2N': quant_1_3N_2N,'4N / 3N': quant_1_4N_3N,'4N / 2N': quant_1_4N_2N} F_D_A_50_percent_change_no_hood = pd.DataFrame(data=D_50_quant_percent_f_d_a, columns=['Percentile %','2N / 1N','3N / 1N', '4N / 1N' ,'3N / 2N','4N / 3N','4N / 2N']) print(F_D_A_50_phase_no_hood) print(F_D_A_50_percent_change_no_hood) print ('-------------------Fuel per Day per Adult Hood Phase -------------------') if Hood_or_no == 'hood': Fuel_per_day_per_adult_1H = [] f_d_a_1H = [] Fuel_per_day_per_adult_2H = [] f_d_a_2H = [] Fuel_per_day_per_adult_3H = [] f_d_a_3H = [] count_t = 0 count_f = 0 for c in hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_Hood_fuel_day_adult[count_f] - C_Place_holder): count_f = count_f + 1 if count_f == len(Household_removal_Hood_fuel_day_adult): count_f = 0 continue if Fuel_remove_1H_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_1H.append(Fuel_remove_1H_24hr.iloc[c,6]/Survey_1H.iloc[c,7]) f_d_a_1H.append(Day_1H.iloc[c,0]) if Fuel_remove_2H_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_2H.append(Fuel_remove_2H_24hr.iloc[c,6] / Survey_2H.iloc[c, 7]) f_d_a_2H.append(Day_2H.iloc[c,0]) if Fuel_remove_3H_24hr.iloc[c,6] != -1.00: Fuel_per_day_per_adult_3H.append(Fuel_remove_3H_24hr.iloc[c,6]/ Survey_3H.iloc[c, 7]) f_d_a_3H.append(Day_3H.iloc[c, 0]) Fuel_per_day_per_adult_2H_1H = [] f_d_a_2H_1H = [] Fuel_per_day_per_adult_3H_1H = [] f_d_a_3H_1H = [] Fuel_per_day_per_adult_3H_2H = [] f_d_a_3H_2H = [] count_t = 0 count_f = 0 for c in hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_Hood_fuel_day_adult[count_f] - C_Place_holder): count_f = count_f + 1 if count_f == len(Household_removal_Hood_fuel_day_adult): count_f = 0 continue if (len(Fuel_per_day_per_adult_2H)-1) >= c and (len(Fuel_per_day_per_adult_1H)-1) >= c: if Day_1H.iloc[c,13] > 0 and Day_2H.iloc[c,13] > 0 and Day_1H.iloc[c,0] == Day_2H.iloc[c,0]: Fuel_per_day_per_adult_2H_1H.append(Fuel_per_day_per_adult_2H[c]/Fuel_per_day_per_adult_1H[c]) f_d_a_2H_1H.append(Day_1H.iloc[c,0]) if (len(Fuel_per_day_per_adult_3H)-1) >= c and (len(Fuel_per_day_per_adult_1H)-1) >= c: if Day_3H.iloc[c,13] > 0 and Day_1H.iloc[c,13] > 0 and Day_3H.iloc[c,0] == Day_1H.iloc[c,0]: Fuel_per_day_per_adult_3H_1H.append(Fuel_per_day_per_adult_3H[c]/Fuel_per_day_per_adult_1H[c]) f_d_a_3H_1H.append(Day_1H.iloc[c,0]) if (len(Fuel_per_day_per_adult_3H)-1) >= c and (len(Fuel_per_day_per_adult_2H)-1) >= c: if Day_3H.iloc[c,13] > 0 and Day_2H.iloc[c,13] > 0 and Day_3H.iloc[c,0] == Day_2H.iloc[c,0]: Fuel_per_day_per_adult_3H_2H.append(Fuel_per_day_per_adult_3H[c]/Fuel_per_day_per_adult_2H[c]) f_d_a_3H_2H.append(Day_1H.iloc[c,0]) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_1H, ax=ax_box, color='b') sns.distplot(Fuel_per_day_per_adult_1H, ax=ax_hist, color='b') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('1H Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_2H, ax=ax_box, color='g') sns.distplot(Fuel_per_day_per_adult_2H, ax=ax_hist, color='g') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('2H Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Fuel_per_day_per_adult_3H, ax=ax_box, color='r') sns.distplot(Fuel_per_day_per_adult_3H, ax=ax_hist, color='r') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('3H Fuel per Day per Adult') plt.ylim(top=2) plt.ylim(bottom = 0) fig_2, ax_2 = plt.subplots() plt.title('Hood Fuel per Day per Adult') quant_1_1H = np.percentile(Fuel_per_day_per_adult_1H, [25,50,75]) Top_lim_1_1H = quant_1_1H[2] + 1.5*(quant_1_1H[2] - quant_1_1H[0]) Low_lim_1_1H = quant_1_1H[0] - 1.5*(quant_1_1H[2] - quant_1_1H[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_1H, positions = [1], widths = 0.6) Fuel_D_A_1H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_1H): if a > Top_lim_1_1H or a < Low_lim_1_1H: Fuel_D_A_1H_outlier.append(f_d_a_1H[v]) plt.text(1,a,f_d_a_1H[v]) plt.text(1,0,'1H',color='b') quant_1_2H = np.percentile(Fuel_per_day_per_adult_2H, [25,50,75]) Top_lim_1_2H = quant_1_2H[2] + 1.5*(quant_1_2H[2] - quant_1_2H[0]) Low_lim_1_2H = quant_1_2H[0] - 1.5*(quant_1_2H[2] - quant_1_2H[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_2H,positions = [2], widths = 0.6) count = 0 Fuel_D_A_2H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_2H): if a > Top_lim_1_2H or a < Low_lim_1_2H: Fuel_D_A_2H_outlier.append(f_d_a_2H[v]) count = count + 1 if count == 1: plt.text(2,a,f_d_a_2H[v],ha='left',va='bottom') elif count !=1: plt.text(2,a,f_d_a_2H[v],ha='right',va='bottom') plt.text(2,0,'2H', color= 'g') quant_1_3H = np.percentile(Fuel_per_day_per_adult_3H, [25,50,75]) Top_lim_1_3H = quant_1_3H[2] + 1.5*(quant_1_3H[2] - quant_1_3H[0]) Low_lim_1_3H = quant_1_3H[0] - 1.5*(quant_1_3H[2] - quant_1_3H[0]) bp_1 = plt.boxplot(Fuel_per_day_per_adult_3H,positions = [3], widths = 0.6) count = 0 Fuel_D_A_3H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3H): if a > Top_lim_1_3H or a < Low_lim_1_3H: Fuel_D_A_3H_outlier.append(f_d_a_3H[v]) count = count + 1 if count == 3: plt.text(3,a,f_d_a_3H[v],ha='left',va='bottom') elif count != 1: plt.text(3,a,f_d_a_3H[v],ha='right',va='bottom') plt.text(3,0,'3H', color='r') plt.xlim(-0,4) plt.ylim(-0.25,2.5) print('Fuel/Day/Adult 1H had these values as outliers ', Fuel_D_A_1H_outlier) print('Fuel/Day/Adult 2H had these values as outliers ', Fuel_D_A_2H_outlier) print('Fuel/Day/Adult 3H had these values as outliers ', Fuel_D_A_3H_outlier) plt.show() fig_2, ax2 = plt.subplots() plt.title('% No_hood Change from Fuel per Day per Adult' ) quant_1_2H_1H = np.percentile(Fuel_per_day_per_adult_2H_1H, [25,50,75]) Top_lim_1_2H_1H = quant_1_2H_1H[2] + 1.5*(quant_1_2H_1H[2]-quant_1_2H_1H[0]) Low_lim_1_2H_1H = quant_1_2H_1H[0] - 1.5*(quant_1_2H_1H[2]-quant_1_2H_1H[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_2H_1H, positions=[1], widths= 0.6) Fuel_D_A_2H_1H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_2H_1H): if a > Top_lim_1_2H_1H or a < Low_lim_1_2H_1H: Fuel_D_A_2H_1H_outlier.append(f_d_a_2H_1H[v]) plt.text(1, a, f_d_a_2H_1H[v]) plt.text(0.75, -0.25, '2H / 1H', color= 'g') quant_1_3H_1H = np.percentile(Fuel_per_day_per_adult_3H_1H, [25,50,75]) Top_lim_1_3H_1H = quant_1_3H_1H[2] + 1.5*(quant_1_3H_1H[2]-quant_1_3H_1H[0]) Low_lim_1_3H_1H = quant_1_3H_1H[0] - 1.5*(quant_1_3H_1H[2]-quant_1_3H_1H[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_3H_1H, positions=[2], widths= 0.6) Fuel_D_A_3H_1H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3H_1H): if a > Top_lim_1_3H_1H or a < Low_lim_1_3H_1H: Fuel_D_A_3H_1H_outlier.append(f_d_a_3H_1H[v]) plt.text(2, a, f_d_a_3H_1H[v]) plt.text(1.75, -0.25, '3H / 1H', color= 'r') quant_1_3H_2H = np.percentile(Fuel_per_day_per_adult_3H_2H, [25,50,75]) Top_lim_1_3H_2H = quant_1_3H_2H[2] + 1.5*(quant_1_3H_2H[2]-quant_1_3H_2H[0]) Low_lim_1_3H_2H = quant_1_3H_2H[0] - 1.5*(quant_1_3H_2H[2]-quant_1_3H_2H[0]) bp_1_1 = plt.boxplot(Fuel_per_day_per_adult_3H_2H, positions=[3], widths= 0.6) Fuel_D_A_3H_2H_outlier = [] for v,a in enumerate(Fuel_per_day_per_adult_3H_2H): if a > Top_lim_1_3H_2H or a < Low_lim_1_3H_2H: Fuel_D_A_3H_2H_outlier.append(f_d_a_3H_2H[v]) plt.text(3, a, f_d_a_3H_2H[v]) plt.text(2.75, -0.25, '2H / 1H', color= 'm') plt.xlim(-0,4) plt.ylim(-0.25,6) print('Fuel/Day/Adult 2H/1H had these values as outliers ', Fuel_D_A_2H_1H_outlier) print('Fuel/Day/Adult 3H/1H had these values as outliers ', Fuel_D_A_3H_1H_outlier) print('Fuel/Day/Adult 3H/2H had these values as outliers ', Fuel_D_A_3H_2H_outlier) plt.show() quant_1_1H = np.append(quant_1_1H, np.average(Fuel_per_day_per_adult_1H)) quant_1_2H = np.append(quant_1_2H, np.average(Fuel_per_day_per_adult_2H)) quant_1_3H = np.append(quant_1_3H, np.average(Fuel_per_day_per_adult_3H)) D_50_quant_phase_f_d_a_hood = {'Percentile %': ['25','50','75', 'Avg'], '1H': quant_1_1H, '2H': quant_1_2H,'3H' : quant_1_3H} F_D_A_50_phase_hood = pd.DataFrame(data=D_50_quant_phase_f_d_a_hood, columns=['Percentile %','1H', '2H','3H'] ) quant_1_2H_1H = np.append(quant_1_2H_1H , np.average(Fuel_per_day_per_adult_2H_1H)) quant_1_3H_1H = np.append(quant_1_3H_1H , np.average(Fuel_per_day_per_adult_3H_1H)) quant_1_3H_2H = np.append(quant_1_3H_2H , np.average(Fuel_per_day_per_adult_3H_2H)) D_50_quant_percent_f_d_a_hood ={'Percentile %': ['25','50','75', 'Avg'],'2H / 1H': quant_1_2H_1H,'3H / 1H': quant_1_3H_1H,'3H / 2H': quant_1_3H_2H} F_D_A_50_percent_change_hood = pd.DataFrame(data=D_50_quant_percent_f_d_a_hood, columns=['Percentile %','2H / 1H','3H / 1H','3H / 2H']) print(F_D_A_50_phase_hood) print(F_D_A_50_percent_change_hood) print('----------------------- Kitchen PM per Day -----------------------------') if Hood_or_no == 'no_hood': Kit_PM_per_day_1N = [] K_PM_D_1N = [] Kit_PM_per_day_2N = [] K_PM_D_2N = [] Kit_PM_per_day_3N = [] K_PM_D_3N = [] Kit_PM_per_day_4N = [] K_PM_D_4N = [] count_t = 0 count_pm = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_NO_Hood_PM[count_pm] - C_Place_holder): count_pm = count_pm + 1 if count_pm == len(Household_removal_NO_Hood_PM): count_pm = 0 continue if Kit_PM_1N_24hr.iloc[c,6] != -1.00: Kit_PM_per_day_1N.append(Kit_PM_1N_24hr.iloc[c,6]) K_PM_D_1N.append(Kit_PM_1N_24hr.iloc[c, 0]) if Kit_PM_2N_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_2N.append(Kit_PM_2N_24hr.iloc[c, 6]) K_PM_D_2N.append(Kit_PM_2N_24hr.iloc[c, 0]) if Kit_PM_3N_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_3N.append(Kit_PM_3N_24hr.iloc[c, 6]) K_PM_D_3N.append(Kit_PM_3N_24hr.iloc[c, 0]) if Kit_PM_4N_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_4N.append(Kit_PM_4N_24hr.iloc[c, 6]) K_PM_D_4N.append(Kit_PM_4N_24hr.iloc[c, 0]) Kit_per_day_2N_1N = [] K_PM_D_2N_1N = [] Kit_per_day_3N_1N = [] K_PM_D_3N_1N = [] Kit_per_day_4N_1N = [] K_PM_D_4N_1N = [] Kit_per_day_3N_2N = [] K_PM_D_3N_2N = [] Kit_per_day_4N_3N = [] K_PM_D_4N_3N = [] Kit_per_day_4N_2N = [] K_PM_D_4N_2N = [] count_t = 0 count_pm = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_NO_Hood_PM[count_pm] - C_Place_holder): count_pm = count_pm + 1 if count_pm == len(Household_removal_NO_Hood_PM): count_pm = 0 continue if (len(Kit_PM_per_day_2N)-1) >= c and (len(Kit_PM_per_day_1N)-1) >= c: if Kit_PM_1N_24hr.iloc[c,6] > 0 and Kit_PM_2N_24hr.iloc[c,6] > 0 and Kit_PM_1N_24hr.iloc[c,0] == Kit_PM_2N_24hr.iloc[c,0]: Kit_per_day_2N_1N.append(Kit_PM_per_day_2N[c]/Kit_PM_per_day_1N[c]) K_PM_D_2N_1N.append(Day_1N.iloc[c,0]) if (len(Kit_PM_per_day_3N)-1) >= c and (len(Kit_PM_per_day_1N)-1) >= c: if Kit_PM_3N_24hr.iloc[c, 6] > 0 and Kit_PM_1N_24hr.iloc[c, 6] > 0 and Kit_PM_3N_24hr.iloc[c, 0] == \ Kit_PM_1N_24hr.iloc[c, 0]: Kit_per_day_3N_1N.append(Kit_PM_per_day_3N[c]/Kit_PM_per_day_1N[c]) K_PM_D_3N_1N.append(Day_1N.iloc[c,0]) if (len(Kit_PM_per_day_4N)-1) >= c and (len(Kit_PM_per_day_1N)-1) >= c: if Kit_PM_4N_24hr.iloc[c, 6] > 0 and Kit_PM_1N_24hr.iloc[c, 6] > 0 and Kit_PM_4N_24hr.iloc[c, 0] == \ Kit_PM_1N_24hr.iloc[c, 0]: Kit_per_day_4N_1N.append(Kit_PM_per_day_4N[c]/Kit_PM_per_day_1N[c]) K_PM_D_4N_1N.append(Day_1N.iloc[c,0]) if (len(Kit_PM_per_day_3N)-1) >= c and (len(Kit_PM_per_day_2N)-1) >= c: if Kit_PM_3N_24hr.iloc[c, 6] > 0 and Kit_PM_2N_24hr.iloc[c, 6] > 0 and Kit_PM_3N_24hr.iloc[c, 0] == \ Kit_PM_2N_24hr.iloc[c, 0]: Kit_per_day_3N_2N.append(Kit_PM_per_day_3N[c]/Kit_PM_per_day_2N[c]) K_PM_D_3N_2N.append(Day_2N.iloc[c,0]) if (len(Kit_PM_per_day_4N)-1) >= c and (len(Kit_PM_per_day_3N)-1) >= c: if Kit_PM_4N_24hr.iloc[c, 6] > 0 and Kit_PM_3N_24hr.iloc[c, 6] > 0 and Kit_PM_3N_24hr.iloc[c, 0] == \ Kit_PM_4N_24hr.iloc[c, 0]: Kit_per_day_4N_3N.append(Kit_PM_per_day_4N[c]/Kit_PM_per_day_3N[c]) K_PM_D_4N_3N.append(Day_3N.iloc[c,0]) if (len(Kit_PM_per_day_4N)-1) >= c and (len(Kit_PM_per_day_2N)-1) >= c: if Kit_PM_4N_24hr.iloc[c, 6] > 0 and Kit_PM_4N_24hr.iloc[c, 6] > 0 and Kit_PM_4N_24hr.iloc[c, 0] == \ Kit_PM_2N_24hr.iloc[c, 0]: Kit_per_day_4N_2N.append(Kit_PM_per_day_4N[c]/Kit_PM_per_day_2N[c]) K_PM_D_4N_2N.append(Day_4N.iloc[c,0]) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_2N_1N, ax=ax_box, color='g') sns.distplot(Kit_per_day_2N_1N, ax=ax_hist, color='g') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 2N/1N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_3N_1N, ax=ax_box, color='r') sns.distplot(Kit_per_day_3N_1N, ax=ax_hist, color='r') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 3N/1N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_4N_1N, ax=ax_box, color='y') sns.distplot(Kit_per_day_4N_1N, ax=ax_hist, color='y') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 4N/1N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_3N_2N, ax=ax_box, color='m') sns.distplot(Kit_per_day_3N_2N, ax=ax_hist, color='m') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 3N/2N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_4N_3N, ax=ax_box, color='k') sns.distplot(Kit_per_day_4N_3N, ax=ax_hist, color='k') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 4N/3N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_4N_2N, ax=ax_box, color='tab:orange') sns.distplot(Kit_per_day_4N_2N, ax=ax_hist, color='tab:orange') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 4N/2N (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) fig, ax = plt.subplots() plt.title('No-Hood Kitchen PM per day') quant_1_1N = np.percentile(Kit_PM_per_day_1N, [25,50,75]) Top_lim_1_1N = quant_1_1N[2] + 1.5*(quant_1_1N[2] - quant_1_1N[0]) Low_lim_1_1N = quant_1_1N[0] - 1.5*(quant_1_1N[2] - quant_1_1N[0]) bp_1 = plt.boxplot(Kit_PM_per_day_1N, positions = [1], widths = 0.6) kitchen_pm_1N_outlier = [] for v,a in enumerate(Kit_PM_per_day_1N): if a > Top_lim_1_1N or a < Low_lim_1_1N: kitchen_pm_1N_outlier.append(K_PM_D_1N[v]) plt.text(1,a,K_PM_D_1N[v]) plt.text(1,0.1,'1N',color='b') quant_1_2N = np.percentile(Kit_PM_per_day_2N, [25,50,75]) Top_lim_1_2N = quant_1_2N[2] + 1.5*(quant_1_2N[2] - quant_1_2N[0]) Low_lim_1_2N = quant_1_2N[0] - 1.5*(quant_1_2N[2] - quant_1_2N[0]) bp_1 = plt.boxplot(Kit_PM_per_day_2N,positions = [2], widths = 0.6) kitchen_pm_2N_outlier = [] for v,a in enumerate(Kit_PM_per_day_2N): if a > Top_lim_1_2N or a < Low_lim_1_2N: kitchen_pm_2N_outlier.append(K_PM_D_2N[v]) plt.text(2,a,K_PM_D_2N[v]) plt.text(2,0.1,'2N', color= 'g') quant_1_3N = np.percentile(Kit_PM_per_day_3N, [25,50,75]) Top_lim_1_3N = quant_1_3N[2] + 1.5*(quant_1_3N[2] - quant_1_3N[0]) Low_lim_1_3N = quant_1_3N[0] - 1.5*(quant_1_3N[2] - quant_1_3N[0]) kitchen_pm_3N_outlier = [] bp_1 = plt.boxplot(Kit_PM_per_day_3N,positions = [3], widths = 0.6) count = 0 for v,a in enumerate(Kit_PM_per_day_3N): if a > Top_lim_1_3N or a < Low_lim_1_3N: kitchen_pm_3N_outlier.append(K_PM_D_3N[v]) count = count + 1 if count == (3): plt.text(3,a,K_PM_D_3N[v],ha='left', va='bottom') if count == (1): plt.text(3,a,K_PM_D_3N[v],ha='left', va='top') else: plt.text(3,a,K_PM_D_3N[v],ha='right', va='bottom') plt.text(3,0.1,'3N', color='r') quant_1_4N = np.percentile(Kit_PM_per_day_4N, [25,50,75]) Top_lim_1_4N = quant_1_4N[2] + 1.5*(quant_1_4N[2] - quant_1_4N[0]) Low_lim_1_4N = quant_1_4N[0] - 1.5*(quant_1_4N[2] - quant_1_4N[0]) bp_1 = plt.boxplot(Kit_PM_per_day_4N,positions = [4], widths = 0.6) kitchen_pm_4N_outlier = [] for v,a in enumerate(Kit_PM_per_day_4N): if a > Top_lim_1_4N or a < Low_lim_1_4N: kitchen_pm_4N_outlier.append(K_PM_D_4N[v]) plt.text(4,a,K_PM_D_4N[v]) plt.text(4,0.1,'4N', color='y') plt.xlim(0,5) plt.ylim(0,1200) print('Kitchen PM 1N had these values as outliers ', kitchen_pm_1N_outlier) print('Kitchen PM 2N had these values as outliers ', kitchen_pm_2N_outlier) print('Kitchen PM 3N had these values as outliers ', kitchen_pm_3N_outlier) print('Kitchen PM 4N had these values as outliers ', kitchen_pm_4N_outlier) plt.show() fig_2, ax2 = plt.subplots() plt.title('% No_hood PM per Day Change' ) quant_1_2N_1N = np.percentile(Kit_per_day_2N_1N, [25,50,75]) Top_lim_1_2N_1N = quant_1_2N_1N[2] + 1.5*(quant_1_2N_1N[2]-quant_1_2N_1N[0]) Low_lim_1_2N_1N = quant_1_2N_1N[0] - 1.5*(quant_1_2N_1N[2]-quant_1_2N_1N[0]) bp_1_1 = plt.boxplot(Kit_per_day_2N_1N, positions=[1], widths= 0.6) kitchen_pm_2N_1N_outlier = [] for v,a in enumerate(Kit_per_day_2N_1N): if a > Top_lim_1_2N_1N or a < Low_lim_1_2N_1N: kitchen_pm_2N_1N_outlier.append(K_PM_D_2N_1N[v]) plt.text(1, a, K_PM_D_2N_1N[v]) plt.text(0.5, -0.25, '2N / 1N', color= 'g') quant_1_3N_1N = np.percentile(Kit_per_day_3N_1N, [25,50,75]) Top_lim_1_3N_1N = quant_1_3N_1N[2] + 1.5*(quant_1_3N_1N[2]-quant_1_3N_1N[0]) Low_lim_1_3N_1N = quant_1_3N_1N[0] - 1.5*(quant_1_3N_1N[2]-quant_1_3N_1N[0]) bp_1_1 = plt.boxplot(Kit_per_day_3N_1N, positions=[2], widths= 0.6) kitchen_pm_3N_1N_outlier = [] for v,a in enumerate(Kit_per_day_3N_1N): if a > Top_lim_1_3N_1N or a < Low_lim_1_3N_1N: kitchen_pm_3N_1N_outlier.append(K_PM_D_3N_1N[v]) plt.text(2, a, K_PM_D_3N_1N[v]) plt.text(1.5, -0.25, '3N / 1N', color= 'r') quant_1_4N_1N = np.percentile(Kit_per_day_4N_1N, [25,50,75]) Top_lim_1_4N_1N = quant_1_4N_1N[2] + 1.5*(quant_1_4N_1N[2]-quant_1_4N_1N[0]) Low_lim_1_4N_1N = quant_1_4N_1N[0] - 1.5*(quant_1_4N_1N[2]-quant_1_4N_1N[0]) bp_1_1 = plt.boxplot(Kit_per_day_4N_1N, positions=[3], widths= 0.6) kitchen_pm_4N_1N_outlier = [] for v,a in enumerate(Kit_per_day_4N_1N): if a > Top_lim_1_4N_1N or a < Low_lim_1_4N_1N: kitchen_pm_4N_1N_outlier.append(K_PM_D_4N_1N[v]) plt.text(3, a, K_PM_D_4N_1N[v]) plt.text(2.5, -0.25, '4N / 1N', color= 'y') quant_1_3N_2N = np.percentile(Kit_per_day_3N_2N, [25,50,75]) Top_lim_1_3N_2N = quant_1_3N_2N[2] + 1.5*(quant_1_3N_2N[2]-quant_1_3N_2N[0]) Low_lim_1_3N_2N = quant_1_3N_2N[0] - 1.5*(quant_1_3N_2N[2]-quant_1_3N_2N[0]) bp_1_1 = plt.boxplot(Kit_per_day_3N_2N, positions=[4], widths= 0.6) kitchen_pm_3N_2N_outlier = [] for v,a in enumerate(Kit_per_day_3N_2N): if a > Top_lim_1_3N_2N or a < Low_lim_1_3N_2N: kitchen_pm_3N_2N_outlier.append(K_PM_D_3N_2N[v]) plt.text(4, a, K_PM_D_3N_2N[v]) plt.text(3.5, -0.25, '3N / 2N', color= 'm') quant_1_4N_3N = np.percentile(Kit_per_day_4N_3N, [25,50,75]) Top_lim_1_4N_3N = quant_1_4N_3N[2] + 1.5*(quant_1_4N_3N[2]-quant_1_4N_3N[0]) Low_lim_1_4N_3N = quant_1_4N_3N[0] - 1.5*(quant_1_4N_3N[2]-quant_1_4N_3N[0]) bp_1_1 = plt.boxplot(Kit_per_day_4N_3N, positions=[5], widths= 0.6) kitchen_pm_4N_3N_outlier = [] for v,a in enumerate(Kit_per_day_4N_3N): if a > Top_lim_1_4N_3N or a < Low_lim_1_4N_3N: kitchen_pm_4N_3N_outlier.append(K_PM_D_4N_3N[v]) plt.text(5, a, K_PM_D_4N_3N[v]) plt.text(4.5, -0.25, '4N / 3N', color= 'k') quant_1_4N_2N = np.percentile(Kit_per_day_4N_2N, [25,50,75]) Top_lim_1_4N_2N = quant_1_4N_2N[2] + 1.5*(quant_1_4N_2N[2]-quant_1_4N_2N[0]) Low_lim_1_4N_2N = quant_1_4N_2N[0] - 1.5*(quant_1_4N_2N[2]-quant_1_4N_2N[0]) bp_1_1 = plt.boxplot(Kit_per_day_4N_2N, positions=[6], widths= 0.6) kitchen_pm_4N_2N_outlier = [] for v,a in enumerate(Kit_per_day_4N_2N): if a > Top_lim_1_4N_2N or a < Low_lim_1_4N_2N: kitchen_pm_4N_2N_outlier.append(K_PM_D_4N_2N[v]) plt.text(6, a, K_PM_D_4N_2N[v]) plt.text(5.5, -0.25, '4N / 2N', color= 'tab:orange') plt.xlim(0,7) plt.ylim(-0.5,5) print('Kitchen PM 2N/1N had these values as outliers ', kitchen_pm_2N_1N_outlier) print('Kitchen PM 3N/1N had these values as outliers ', kitchen_pm_3N_1N_outlier) print('Kitchen PM 4N/1N had these values as outliers ', kitchen_pm_4N_1N_outlier) print('Kitchen PM 3N/2N had these values as outliers ', kitchen_pm_3N_2N_outlier) print('Kitchen PM 4N/3N had these values as outliers ', kitchen_pm_4N_3N_outlier) print('Kitchen PM 4N/2N had these values as outliers ', kitchen_pm_4N_2N_outlier) plt.show() quant_1_1N = np.append(quant_1_1N, np.average(Kit_PM_per_day_1N)) quant_1_2N = np.append(quant_1_2N, np.average(Kit_PM_per_day_2N)) quant_1_3N = np.append(quant_1_3N, np.average(Kit_PM_per_day_3N)) quant_1_4N = np.append(quant_1_4N, np.average(Kit_PM_per_day_4N)) D_50_quant_phase_PM_d = {'Percentile %': ['25','50','75', 'Avg'], '1N': quant_1_1N, '2N': quant_1_2N,'3N' : quant_1_3N,'4N': quant_1_4N} PM_D_50_phase_no_hood = pd.DataFrame(data=D_50_quant_phase_PM_d,columns=['Percentile %','1N', '2N', '3N','4N']) quant_1_2N_1N = np.append(quant_1_2N_1N , np.average(Kit_per_day_2N_1N)) quant_1_3N_1N = np.append(quant_1_3N_1N , np.average(Kit_per_day_3N_1N)) quant_1_4N_1N = np.append(quant_1_4N_1N , np.average(Kit_per_day_4N_1N)) quant_1_3N_2N = np.append(quant_1_3N_2N , np.average(Kit_per_day_3N_2N)) quant_1_4N_3N = np.append(quant_1_4N_3N , np.average(Kit_per_day_4N_3N)) quant_1_4N_2N = np.append(quant_1_4N_2N , np.average(Kit_per_day_4N_2N)) D_50_quant_percent_PM_d ={'Percentile %': ['25','50','75', 'Avg'],'2N / 1N': quant_1_2N_1N,'3N / 1N': quant_1_3N_1N,'4N / 1N': quant_1_4N_1N, '3N / 2N': quant_1_3N_2N,'4N / 3N': quant_1_4N_3N,'4N / 2N': quant_1_4N_2N} PM_D_50_percent_change_no_hood = pd.DataFrame(data=D_50_quant_percent_PM_d, columns=['Percentile %','2N / 1N','3N / 1N', '4N / 1N' ,'3N / 2N','4N / 3N','4N / 2N']) print(PM_D_50_phase_no_hood) print(PM_D_50_percent_change_no_hood) if Hood_or_no == 'hood': Kit_PM_per_day_1H = [] K_PM_D_1H = [] Kit_PM_per_day_2H = [] K_PM_D_2H = [] Kit_PM_per_day_3H = [] K_PM_D_3H = [] count_t = 0 count_pm = 0 for c in hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_Hood_PM[count_pm] - C_Place_holder): count_pm = count_pm + 1 if count_pm == len(Household_removal_Hood_PM): count_pm = 0 continue if Kit_PM_1H_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_1H.append(Kit_PM_1H_24hr.iloc[c,6]) K_PM_D_1H.append(Kit_PM_1H_24hr.iloc[c,0]) if Kit_PM_2H_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_2H.append(Kit_PM_2H_24hr.iloc[c,6]) K_PM_D_2H.append(Kit_PM_2H_24hr.iloc[c,0]) if Kit_PM_3H_24hr.iloc[c, 6] != -1.00: Kit_PM_per_day_3H.append(Kit_PM_3H_24hr.iloc[c,6]) K_PM_D_3H.append(Kit_PM_3H_24hr.iloc[c,0]) Kit_per_day_2H_1H = [] K_PM_D_2H_1H = [] Kit_per_day_3H_1H = [] K_PM_D_3H_1H = [] Kit_per_day_3H_2H = [] K_PM_D_3H_2H = [] count_t = 0 count_pm = 0 for c in NO_hood_counter: if c == (Household_removal[count_t] - C_Place_holder): count_t = count_t + 1 if count_t == len(Household_removal): count_t = 0 continue if c == (Household_removal_Hood_PM[count_pm] - C_Place_holder): count_pm = count_pm + 1 if count_pm == len(Household_removal_Hood_PM): count_pm = 0 continue if (len(Kit_PM_per_day_2H)-1) >= c and (len(Kit_PM_per_day_1H)-1) >= c: if Kit_PM_1H_24hr.iloc[c, 6] > 0 and Kit_PM_2H_24hr.iloc[c, 6] > 0 and Kit_PM_1H_24hr.iloc[c, 0] == Kit_PM_2H_24hr.iloc[c, 0]: Kit_per_day_2H_1H.append(Kit_PM_per_day_2H[c]/Kit_PM_per_day_1H[c]) K_PM_D_2H_1H.append(Day_1H.iloc[c,0]) if (len(Kit_PM_per_day_3H)-1) >= c and (len(Kit_PM_per_day_1H)-1) >= c: if Kit_PM_3H_24hr.iloc[c, 6] > 0 and Kit_PM_1H_24hr.iloc[c, 6] > 0 and Kit_PM_1H_24hr.iloc[c, 0] == \ Kit_PM_3H_24hr.iloc[c, 0]: Kit_per_day_3H_1H.append(Kit_PM_per_day_3H[c]/Kit_PM_per_day_1H[c]) K_PM_D_3H_1H.append(Day_1H.iloc[c,0]) if (len(Kit_PM_per_day_3H)-1) >= c and (len(Kit_PM_per_day_2H)-1) >= c: if Kit_PM_3H_24hr.iloc[c, 6] > 0 and Kit_PM_2H_24hr.iloc[c, 6] > 0 and Kit_PM_3H_24hr.iloc[c, 0] == \ Kit_PM_2H_24hr.iloc[c, 0]: Kit_per_day_3H_2H.append(Kit_PM_per_day_3H[c]/Kit_PM_per_day_2H[c]) K_PM_D_3H_2H.append(Day_2H.iloc[c,0]) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_2H_1H, ax=ax_box, color='g') sns.distplot(Kit_per_day_2H_1H, ax=ax_hist, color='g') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 2H/1H (Kitchen PM per Day)') plt.ylim(top=1.5) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_3H_1H, ax=ax_box, color='r') sns.distplot(Kit_per_day_3H_1H, ax=ax_hist, color='r') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 3H/1H (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) sns.set(style="ticks") f, (ax_box, ax_hist) = plt.subplots(2, sharex=True, gridspec_kw={"height_ratios": (0.15, 0.85)}) sns.boxplot(Kit_per_day_3H_2H, ax=ax_box, color='m') sns.distplot(Kit_per_day_3H_2H, ax=ax_hist, color='m') ax_box.set(yticks=[]) sns.despine(ax=ax_hist) sns.despine(ax=ax_box, left=True) plt.title('% 3H/2H (Kitchen PM per Day)') plt.ylim(top=2) plt.ylim(bottom = 0) fig, ax = plt.subplots() plt.title('Hood Kitchen PM per day') quant_1_1H = np.percentile(Kit_PM_per_day_1H, [25,50,75]) Top_lim_1_1H = quant_1_1H[2] + 1.5*(quant_1_1H[2] - quant_1_1H[0]) Low_lim_1_1H = quant_1_1H[0] - 1.5*(quant_1_1H[2] - quant_1_1H[0]) bp_1 = plt.boxplot(Kit_PM_per_day_1H, positions = [1], widths = 0.6) kitchen_pm_1H_outlier = [] for v,a in enumerate(Kit_PM_per_day_1H): if a > Top_lim_1_1H or a < Low_lim_1_1H: kitchen_pm_1H_outlier.append(K_PM_D_1H[v]) plt.text(1,a,K_PM_D_1H[v]) plt.text(0.5,0.1,'1H',color='b') quant_1_2H = np.percentile(Kit_PM_per_day_2H, [25,50,75]) Top_lim_1_2N = quant_1_2H[2] + 1.5*(quant_1_2H[2] - quant_1_2H[0]) Low_lim_1_2N = quant_1_2H[0] - 1.5*(quant_1_2H[2] - quant_1_2H[0]) bp_1 = plt.boxplot(Kit_PM_per_day_2H,positions = [2], widths = 0.6) kitchen_pm_2H_outlier = [] for v,a in enumerate(Kit_PM_per_day_2H): if a > Top_lim_1_2N or a < Low_lim_1_2N: kitchen_pm_2H_outlier.append(K_PM_D_2H[v]) plt.text(2,a,K_PM_D_2H[v]) plt.text(1.5,0.1,'2H', color= 'g') quant_1_3H = np.percentile(Kit_PM_per_day_3H, [25,50,75]) Top_lim_1_3N = quant_1_3H[2] + 1.5*(quant_1_3H[2] - quant_1_3H[0]) Low_lim_1_3N = quant_1_3H[0] - 1.5*(quant_1_3H[2] - quant_1_3H[0]) kitchen_3H_outlier = [] bp_1 = plt.boxplot(Kit_PM_per_day_3H,positions = [3], widths = 0.6) count = 0 kitchen_pm_3H_outlier = [] for v,a in enumerate(Kit_PM_per_day_3H): if a > Top_lim_1_3N or a < Low_lim_1_3N: kitchen_pm_3H_outlier.append(K_PM_D_3H[v]) plt.text(3,a,K_PM_D_3H[v]) plt.text(2.5,0.1,'3H', color='r') plt.xlim(0,4) plt.ylim(0,1200) print('Kitchen PM 1H had these values as outliers ', kitchen_pm_1H_outlier) print('Kitchen PM 2H had these values as outliers ', kitchen_pm_2H_outlier) print('Kitchen PM 3H had these values as outliers ', kitchen_pm_3H_outlier) plt.show() fig_2, ax2 = plt.subplots() plt.title('% hood PM per Day Change' ) quant_1_2H_1H = np.percentile(Kit_per_day_2H_1H, [25,50,75]) Top_lim_1_2N_1N = quant_1_2H_1H[2] + 1.5*(quant_1_2H_1H[2]-quant_1_2H_1H[0]) Low_lim_1_2N_1N = quant_1_2H_1H[0] - 1.5*(quant_1_2H_1H[2]-quant_1_2H_1H[0]) bp_1_1 = plt.boxplot(Kit_per_day_2H_1H, positions=[1], widths= 0.6) kitchen_pm_2H_1H_outlier = [] for v,a in enumerate(Kit_per_day_2H_1H): if a > Top_lim_1_2N_1N or a < Low_lim_1_2N_1N: kitchen_pm_2H_1H_outlier.append(K_PM_D_2H_1H[v]) plt.text(1, a, K_PM_D_2H_1H[v]) plt.text(0.75, -0.25, '2H / 1H', color= 'g') quant_1_3H_1H = np.percentile(Kit_per_day_3H_1H, [25,50,75]) Top_lim_1_3N_1N = quant_1_3H_1H[2] + 1.5*(quant_1_3H_1H[2]-quant_1_3H_1H[0]) Low_lim_1_3N_1N = quant_1_3H_1H[0] - 1.5*(quant_1_3H_1H[2]-quant_1_3H_1H[0]) bp_1_1 = plt.boxplot(Kit_per_day_3H_1H, positions=[2], widths= 0.6) kitchen_pm_3H_1H_outlier = [] for v,a in enumerate(Kit_per_day_3H_1H): if a > Top_lim_1_3N_1N or a < Low_lim_1_3N_1N: kitchen_pm_3H_1H_outlier.append(K_PM_D_3H_1H[v]) plt.text(2, a, K_PM_D_3H_1H[v]) plt.text(1.75, -0.25, '3H / 1H', color= 'r') quant_1_3H_2H = np.percentile(Kit_per_day_3H_2H, [25,50,75]) Top_lim_1_3N_2N = quant_1_3H_2H[2] + 1.5*(quant_1_3H_2H[2]-quant_1_3H_2H[0]) Low_lim_1_3N_2N = quant_1_3H_2H[0] - 1.5*(quant_1_3H_2H[2]-quant_1_3H_2H[0]) bp_1_1 = plt.boxplot(Kit_per_day_3H_2H, positions=[3], widths= 0.6) kitchen_pm_3H_2H_outlier = [] for v,a in enumerate(Kit_per_day_3H_2H): if a > Top_lim_1_3N_2N or a < Low_lim_1_3N_2N: kitchen_pm_3H_2H_outlier.append(K_PM_D_3H_2H[v]) plt.text(3, a, K_PM_D_3H_2H[v]) plt.text(2.75, -0.25, '3H / 2H', color= 'm') plt.xlim(0,4) plt.ylim(-0.5,5) print('Kitchen PM 2H/1H had these values as outliers ', kitchen_pm_2H_1H_outlier) print('Kitchen PM 3H/1H had these values as outliers ', kitchen_pm_3H_1H_outlier) print('Kitchen PM 3H/2H had these values as outliers ', kitchen_pm_3H_2H_outlier) plt.show() quant_1_1H = np.append(quant_1_1H, np.average(Kit_PM_per_day_1H)) quant_1_2H = np.append(quant_1_2H, np.average(Kit_PM_per_day_2H)) quant_1_3H = np.append(quant_1_3H, np.average(Kit_PM_per_day_3H)) D_50_quant_phase_PM_D_hood = {'Percentile %': ['25','50','75', 'Avg'], '1H': quant_1_1H, '2H': quant_1_2H,'3H' : quant_1_3H} PM_D_50_phase_hood = pd.DataFrame(data=D_50_quant_phase_PM_D_hood, columns= ['Percentile %','1H','2H','3H' ]) quant_1_2H_1H = np.append(quant_1_2H_1H , np.average(Kit_per_day_2H_1H)) quant_1_3H_1H = np.append(quant_1_3H_1H , np.average(Kit_per_day_3H_1H)) quant_1_3H_2H = np.append(quant_1_3H_2H , np.average(Kit_per_day_3H_2H)) D_50_quant_percent_PM_D_hood ={'Percentile %': ['25','50','75', 'Avg'],'2H / 1H': quant_1_2H_1H,'3H / 1H': quant_1_3H_1H,'3H / 2H': quant_1_3H_2H} PM_D_50_percent_change_hood = pd.DataFrame(data=D_50_quant_percent_PM_D_hood, columns=['Percentile %','2H / 1H','3H / 1H','3H / 2H']) print(PM_D_50_phase_hood) print(PM_D_50_percent_change_hood) if Hood_or_no == 'no_hood': plt.title('Histogram of Fuel per 24 Hours per Person - No Hood' ) plt.hist([Fuel_per_day_per_adult_1N], color=['b'], alpha=0.5, label='1N') plt.hist([Fuel_per_day_per_adult_2N], color=['g'], alpha=0.5, label='2N') plt.hist([Fuel_per_day_per_adult_3N], color=['r'], alpha=0.5, label='3N') plt.hist([Fuel_per_day_per_adult_4N], color=['y'], alpha=0.5, label='4N') plt.legend(loc='upper right') plt.show() plt.title('Histogram of Kitchen PM 24 Hours - No Hood' ) plt.hist([Kit_PM_per_day_1N], color=['b'], alpha=0.5, label='1N') plt.hist([Kit_PM_per_day_2N], color=['g'], alpha=0.5, label='2N') plt.hist([Kit_PM_per_day_3N], color=['r'], alpha=0.5, label='3N') plt.hist([Kit_PM_per_day_4N], color=['y'], alpha=0.5, label='4N') plt.legend(loc='upper right') plt.show() if Hood_or_no == 'hood': plt.title('Histogram of Fuel per 24 Hours per Person - Hood' ) plt.hist([Fuel_per_day_per_adult_1H], color=['b'], alpha=0.5, label='1H') plt.hist([Fuel_per_day_per_adult_2H], color=['g'], alpha=0.5, label='2H') plt.hist([Fuel_per_day_per_adult_3H], color=['r'], alpha=0.5, label='3H') plt.legend(loc='upper right') plt.show() plt.title('Histogram of Kitchen PM 24 Hours - Hood' ) plt.hist([Kit_PM_per_day_1H], color=['b'], alpha=0.5, label='1H') plt.hist([Kit_PM_per_day_2H], color=['g'], alpha=0.5, label='2H') plt.hist([Kit_PM_per_day_3H], color=['r'], alpha=0.5, label='3H') plt.legend(loc='upper right') plt.show()
true
true
f70ea318486da939ee5b7f2bf6eee6d6a675026f
7,269
py
Python
tests/metarl/torch/algos/test_torch_algo_utils.py
icml2020submission6857/metarl
9b66cefa2b6bcb6a38096d629ce8853b47c7171d
[ "MIT" ]
2
2020-03-15T14:35:15.000Z
2021-02-15T16:38:00.000Z
tests/metarl/torch/algos/test_torch_algo_utils.py
icml2020submission6857/metarl
9b66cefa2b6bcb6a38096d629ce8853b47c7171d
[ "MIT" ]
null
null
null
tests/metarl/torch/algos/test_torch_algo_utils.py
icml2020submission6857/metarl
9b66cefa2b6bcb6a38096d629ce8853b47c7171d
[ "MIT" ]
1
2020-02-24T03:04:23.000Z
2020-02-24T03:04:23.000Z
"""Test torch algo utility functions.""" import numpy as np import pytest import tensorflow as tf import torch import torch.nn.functional as F import metarl.tf.misc.tensor_utils as tf_utils import metarl.torch.algos._utils as torch_algo_utils from tests.fixtures import TfGraphTestCase def stack(d, arr): """Stack 'arr' 'd' times.""" return np.repeat(np.expand_dims(arr, axis=0), repeats=d, axis=0) ONES = np.ones((4, 6)) ZEROS = np.zeros((4, 6)) ARRANGE = stack(4, np.arange(6)) PI_DIGITS = stack(4, [3, 1, 4, 1, 5, 9]) E_DIGITS = stack(4, [2, 7, 1, 8, 2, 8]) FIBS = stack(4, [1, 1, 2, 3, 5, 8]) nums_1d = np.arange(0, 4).astype(float) nums_2d = np.arange(0, 4).astype(float).reshape(2, 2) nums_3d = np.arange(0, 8).astype(float).reshape(2, 2, 2) class TestTorchAlgoUtils(TfGraphTestCase): """Test class for torch algo utility functions.""" # yapf: disable @pytest.mark.parametrize('gae_lambda, rewards_val, baselines_val', [ (0.4, ONES, ZEROS), (0.8, PI_DIGITS, ARRANGE), (1.2, ONES, FIBS), (1.7, E_DIGITS, PI_DIGITS), ]) # yapf: enable def testcompute_advantages(self, gae_lambda, rewards_val, baselines_val): """Test compute_advantage function.""" discount = 0.99 max_len = rewards_val.shape[-1] torch_advs = torch_algo_utils.compute_advantages( discount, gae_lambda, max_len, torch.Tensor(baselines_val), torch.Tensor(rewards_val)) rewards = tf.compat.v1.placeholder(dtype=tf.float32, name='reward', shape=[None, None]) baselines = tf.compat.v1.placeholder(dtype=tf.float32, name='baseline', shape=[None, None]) adv = tf_utils.compute_advantages(discount, gae_lambda, max_len, baselines, rewards) tf_advs = self.sess.run(adv, feed_dict={ rewards: rewards_val, baselines: baselines_val, }) assert np.allclose(torch_advs.numpy(), tf_advs.reshape(torch_advs.shape), atol=1e-5) def test_add_padding_last_1d(self): """Test pad_to_last function for 1d.""" max_length = 10 expected = F.pad(torch.Tensor(nums_1d), (0, max_length - nums_1d.shape[-1])) tensor_padding = torch_algo_utils.pad_to_last(nums_1d, total_length=max_length) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_1d, total_length=10, axis=0) assert expected.eq(tensor_padding).all() def test_add_padding_last_2d(self): """Test pad_to_last function for 2d.""" max_length = 10 tensor_padding = torch_algo_utils.pad_to_last(nums_2d, total_length=10) expected = F.pad(torch.Tensor(nums_2d), (0, max_length - nums_2d.shape[-1])) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_2d, total_length=10, axis=0) expected = F.pad(torch.Tensor(nums_2d), (0, 0, 0, max_length - nums_2d.shape[0])) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_2d, total_length=10, axis=1) expected = F.pad(torch.Tensor(nums_2d), (0, max_length - nums_2d.shape[-1], 0, 0)) assert expected.eq(tensor_padding).all() def test_add_padding_last_3d(self): """Test pad_to_last function for 3d.""" max_length = 10 tensor_padding = torch_algo_utils.pad_to_last(nums_3d, total_length=10) expected = F.pad(torch.Tensor(nums_3d), (0, max_length - nums_3d.shape[-1], 0, 0, 0, 0)) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_3d, total_length=10, axis=0) expected = F.pad(torch.Tensor(nums_3d), (0, 0, 0, 0, 0, max_length - nums_3d.shape[0])) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_3d, total_length=10, axis=1) expected = F.pad(torch.Tensor(nums_3d), (0, 0, 0, max_length - nums_3d.shape[-1], 0, 0)) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_3d, total_length=10, axis=2) expected = F.pad(torch.Tensor(nums_3d), (0, max_length - nums_3d.shape[-1], 0, 0, 0, 0)) assert expected.eq(tensor_padding).all() @pytest.mark.parametrize('nums', [nums_1d, nums_2d, nums_3d]) def test_out_of_index_error(self, nums): """Test pad_to_last raises IndexError.""" with pytest.raises(IndexError): torch_algo_utils.pad_to_last(nums, total_length=10, axis=len(nums.shape)) def testmake_optimizer_with_type(self): """Test make_optimizer function with type as first argument.""" optimizer_type = torch.optim.Adam module = torch.nn.Linear(2, 1) lr = 0.123 optimizer = torch_algo_utils.make_optimizer(optimizer_type, module, lr=lr) assert isinstance(optimizer, optimizer_type) assert optimizer.defaults['lr'] == lr def testmake_optimizer_with_tuple(self): """Test make_optimizer function with tuple as first argument.""" optimizer_type = (torch.optim.Adam, {'lr': 0.1}) module = torch.nn.Linear(2, 1) optimizer = torch_algo_utils.make_optimizer(optimizer_type, module) assert isinstance(optimizer, optimizer_type) assert optimizer.defaults['lr'] == optimizer_type[1]['lr'] def testmake_optimizer_raise_value_error(self): """Test make_optimizer raises value error.""" optimizer_type = (torch.optim.Adam, {'lr': 0.1}) module = torch.nn.Linear(2, 1) with pytest.raises(ValueError): _ = torch_algo_utils.make_optimizer(optimizer_type, module, lr=0.123)
42.508772
79
0.526482
import numpy as np import pytest import tensorflow as tf import torch import torch.nn.functional as F import metarl.tf.misc.tensor_utils as tf_utils import metarl.torch.algos._utils as torch_algo_utils from tests.fixtures import TfGraphTestCase def stack(d, arr): return np.repeat(np.expand_dims(arr, axis=0), repeats=d, axis=0) ONES = np.ones((4, 6)) ZEROS = np.zeros((4, 6)) ARRANGE = stack(4, np.arange(6)) PI_DIGITS = stack(4, [3, 1, 4, 1, 5, 9]) E_DIGITS = stack(4, [2, 7, 1, 8, 2, 8]) FIBS = stack(4, [1, 1, 2, 3, 5, 8]) nums_1d = np.arange(0, 4).astype(float) nums_2d = np.arange(0, 4).astype(float).reshape(2, 2) nums_3d = np.arange(0, 8).astype(float).reshape(2, 2, 2) class TestTorchAlgoUtils(TfGraphTestCase): @pytest.mark.parametrize('gae_lambda, rewards_val, baselines_val', [ (0.4, ONES, ZEROS), (0.8, PI_DIGITS, ARRANGE), (1.2, ONES, FIBS), (1.7, E_DIGITS, PI_DIGITS), ]) def testcompute_advantages(self, gae_lambda, rewards_val, baselines_val): discount = 0.99 max_len = rewards_val.shape[-1] torch_advs = torch_algo_utils.compute_advantages( discount, gae_lambda, max_len, torch.Tensor(baselines_val), torch.Tensor(rewards_val)) rewards = tf.compat.v1.placeholder(dtype=tf.float32, name='reward', shape=[None, None]) baselines = tf.compat.v1.placeholder(dtype=tf.float32, name='baseline', shape=[None, None]) adv = tf_utils.compute_advantages(discount, gae_lambda, max_len, baselines, rewards) tf_advs = self.sess.run(adv, feed_dict={ rewards: rewards_val, baselines: baselines_val, }) assert np.allclose(torch_advs.numpy(), tf_advs.reshape(torch_advs.shape), atol=1e-5) def test_add_padding_last_1d(self): max_length = 10 expected = F.pad(torch.Tensor(nums_1d), (0, max_length - nums_1d.shape[-1])) tensor_padding = torch_algo_utils.pad_to_last(nums_1d, total_length=max_length) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_1d, total_length=10, axis=0) assert expected.eq(tensor_padding).all() def test_add_padding_last_2d(self): max_length = 10 tensor_padding = torch_algo_utils.pad_to_last(nums_2d, total_length=10) expected = F.pad(torch.Tensor(nums_2d), (0, max_length - nums_2d.shape[-1])) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_2d, total_length=10, axis=0) expected = F.pad(torch.Tensor(nums_2d), (0, 0, 0, max_length - nums_2d.shape[0])) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_2d, total_length=10, axis=1) expected = F.pad(torch.Tensor(nums_2d), (0, max_length - nums_2d.shape[-1], 0, 0)) assert expected.eq(tensor_padding).all() def test_add_padding_last_3d(self): max_length = 10 tensor_padding = torch_algo_utils.pad_to_last(nums_3d, total_length=10) expected = F.pad(torch.Tensor(nums_3d), (0, max_length - nums_3d.shape[-1], 0, 0, 0, 0)) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_3d, total_length=10, axis=0) expected = F.pad(torch.Tensor(nums_3d), (0, 0, 0, 0, 0, max_length - nums_3d.shape[0])) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_3d, total_length=10, axis=1) expected = F.pad(torch.Tensor(nums_3d), (0, 0, 0, max_length - nums_3d.shape[-1], 0, 0)) assert expected.eq(tensor_padding).all() tensor_padding = torch_algo_utils.pad_to_last(nums_3d, total_length=10, axis=2) expected = F.pad(torch.Tensor(nums_3d), (0, max_length - nums_3d.shape[-1], 0, 0, 0, 0)) assert expected.eq(tensor_padding).all() @pytest.mark.parametrize('nums', [nums_1d, nums_2d, nums_3d]) def test_out_of_index_error(self, nums): with pytest.raises(IndexError): torch_algo_utils.pad_to_last(nums, total_length=10, axis=len(nums.shape)) def testmake_optimizer_with_type(self): optimizer_type = torch.optim.Adam module = torch.nn.Linear(2, 1) lr = 0.123 optimizer = torch_algo_utils.make_optimizer(optimizer_type, module, lr=lr) assert isinstance(optimizer, optimizer_type) assert optimizer.defaults['lr'] == lr def testmake_optimizer_with_tuple(self): optimizer_type = (torch.optim.Adam, {'lr': 0.1}) module = torch.nn.Linear(2, 1) optimizer = torch_algo_utils.make_optimizer(optimizer_type, module) assert isinstance(optimizer, optimizer_type) assert optimizer.defaults['lr'] == optimizer_type[1]['lr'] def testmake_optimizer_raise_value_error(self): optimizer_type = (torch.optim.Adam, {'lr': 0.1}) module = torch.nn.Linear(2, 1) with pytest.raises(ValueError): _ = torch_algo_utils.make_optimizer(optimizer_type, module, lr=0.123)
true
true
f70ea35f73bd2981d6bbb6766833e038e16308d4
23,711
py
Python
src/sage/schemes/plane_curves/projective_curve.py
switzel/sage
7eb8510dacf61b691664cd8f1d2e75e5d473e5a0
[ "BSL-1.0" ]
null
null
null
src/sage/schemes/plane_curves/projective_curve.py
switzel/sage
7eb8510dacf61b691664cd8f1d2e75e5d473e5a0
[ "BSL-1.0" ]
null
null
null
src/sage/schemes/plane_curves/projective_curve.py
switzel/sage
7eb8510dacf61b691664cd8f1d2e75e5d473e5a0
[ "BSL-1.0" ]
1
2020-07-24T12:20:37.000Z
2020-07-24T12:20:37.000Z
""" Projective plane curves over a general ring AUTHORS: - William Stein (2005-11-13) - David Joyner (2005-11-13) - David Kohel (2006-01) - Moritz Minzlaff (2010-11) """ #***************************************************************************** # Copyright (C) 2005 William Stein <wstein@gmail.com> # # Distributed under the terms of the GNU General Public License (GPL) # # The full text of the GPL is available at: # # http://www.gnu.org/licenses/ #***************************************************************************** from sage.interfaces.all import singular from sage.misc.all import add, sage_eval from sage.rings.all import degree_lowest_rational_function from sage.schemes.projective.projective_space import is_ProjectiveSpace from curve import Curve_generic_projective class ProjectiveSpaceCurve_generic(Curve_generic_projective): def _repr_type(self): return "Projective Space" def __init__(self, A, X): if not is_ProjectiveSpace(A): raise TypeError("A (=%s) must be a projective space"%A) Curve_generic_projective.__init__(self, A, X) d = self.dimension() if d != 1: raise ValueError("defining equations (=%s) define a scheme of dimension %s != 1"%(X,d)) class ProjectiveCurve_generic(Curve_generic_projective): def __init__(self, A, f): if not (is_ProjectiveSpace(A) and A.dimension != 2): raise TypeError("Argument A (= %s) must be a projective plane."%A) Curve_generic_projective.__init__(self, A, [f]) def _repr_type(self): return "Projective" def arithmetic_genus(self): r""" Return the arithmetic genus of this curve. This is the arithmetic genus `g_a(C)` as defined in Hartshorne. If the curve has degree `d` then this is simply `(d-1)(d-2)/2`. It need *not* equal the geometric genus (the genus of the normalization of the curve). EXAMPLE:: sage: x,y,z = PolynomialRing(GF(5), 3, 'xyz').gens() sage: C = Curve(y^2*z^7 - x^9 - x*z^8); C Projective Curve over Finite Field of size 5 defined by -x^9 + y^2*z^7 - x*z^8 sage: C.arithmetic_genus() 28 sage: C.genus() 4 """ d = self.defining_polynomial().total_degree() return int((d-1)*(d-2)/2) def divisor_of_function(self, r): """ Return the divisor of a function on a curve. INPUT: r is a rational function on X OUTPUT: - ``list`` - The divisor of r represented as a list of coefficients and points. (TODO: This will change to a more structural output in the future.) EXAMPLES:: sage: FF = FiniteField(5) sage: P2 = ProjectiveSpace(2, FF, names = ['x','y','z']) sage: R = P2.coordinate_ring() sage: x, y, z = R.gens() sage: f = y^2*z^7 - x^9 - x*z^8 sage: C = Curve(f) sage: K = FractionField(R) sage: r = 1/x sage: C.divisor_of_function(r) # todo: not implemented !!!! [[-1, (0, 0, 1)]] sage: r = 1/x^3 sage: C.divisor_of_function(r) # todo: not implemented !!!! [[-3, (0, 0, 1)]] """ F = self.base_ring() f = self.defining_polynomial() x, y, z = f.parent().gens() pnts = self.rational_points() divf = [] for P in pnts: if P[2] != F(0): # What is the '5' in this line and the 'r()' in the next??? lcs = self.local_coordinates(P,5) ldg = degree_lowest_rational_function(r(lcs[0],lcs[1]),z) if ldg[0] != 0: divf.append([ldg[0],P]) return divf def local_coordinates(self, pt, n): r""" Return local coordinates to precision n at the given point. Behaviour is flaky - some choices of `n` are worst that others. INPUT: - ``pt`` - an F-rational point on X which is not a point of ramification for the projection (x,y) - x. - ``n`` - the number of terms desired OUTPUT: x = x0 + t y = y0 + power series in t EXAMPLES:: sage: FF = FiniteField(5) sage: P2 = ProjectiveSpace(2, FF, names = ['x','y','z']) sage: x, y, z = P2.coordinate_ring().gens() sage: C = Curve(y^2*z^7-x^9-x*z^8) sage: pt = C([2,3,1]) sage: C.local_coordinates(pt,9) # todo: not implemented !!!! [2 + t, 3 + 3*t^2 + t^3 + 3*t^4 + 3*t^6 + 3*t^7 + t^8 + 2*t^9 + 3*t^11 + 3*t^12] """ f = self.defining_polynomial() R = f.parent() F = self.base_ring() p = F.characteristic() x0 = F(pt[0]) y0 = F(pt[1]) astr = ["a"+str(i) for i in range(1,2*n)] x,y = R.gens() R0 = PolynomialRing(F,2*n+2,names = [str(x),str(y),"t"]+astr) vars0 = R0.gens() t = vars0[2] yt = y0*t**0 + add([vars0[i]*t**(i-2) for i in range(3,2*n+2)]) xt = x0+t ft = f(xt,yt) S = singular S.eval('ring s = '+str(p)+','+str(R0.gens())+',lp;') S.eval('poly f = '+str(ft)) cmd = 'matrix c = coeffs ('+str(ft)+',t)' S.eval(cmd) N = int(S.eval('size(c)')) b = ["c["+str(i)+",1]," for i in range(2,N/2-4)] b = ''.join(b) b = b[:len(b)-1] #to cut off the trailing comma cmd = 'ideal I = '+b S.eval(cmd) c = S.eval('slimgb(I)') d = c.split("=") d = d[1:] d[len(d)-1] += "\n" e = [x[:x.index("\n")] for x in d] vals = [] for x in e: for y in vars0: if str(y) in x: if len(x.replace(str(y),"")) != 0: i = x.find("-") if i>0: vals.append([eval(x[1:i]),x[:i],F(eval(x[i+1:]))]) i = x.find("+") if i>0: vals.append([eval(x[1:i]),x[:i],-F(eval(x[i+1:]))]) else: vals.append([eval(str(y)[1:]),str(y),F(0)]) vals.sort() k = len(vals) v = [x0+t,y0+add([vals[i][2]*t**(i+1) for i in range(k)])] return v def plot(self, *args, **kwds): """ Plot the real points of an affine patch of this projective plane curve. INPUT: - ``self`` - an affine plane curve - ``patch`` - (optional) the affine patch to be plotted; if not specified, the patch corresponding to the last projective coordinate being nonzero - ``*args`` - optional tuples (variable, minimum, maximum) for plotting dimensions - ``**kwds`` - optional keyword arguments passed on to ``implicit_plot`` EXAMPLES: A cuspidal curve:: sage: R.<x, y, z> = QQ[] sage: C = Curve(x^3 - y^2*z) sage: C.plot() Graphics object consisting of 1 graphics primitive The other affine patches of the same curve:: sage: C.plot(patch=0) Graphics object consisting of 1 graphics primitive sage: C.plot(patch=1) Graphics object consisting of 1 graphics primitive An elliptic curve:: sage: E = EllipticCurve('101a') sage: C = Curve(E) sage: C.plot() Graphics object consisting of 1 graphics primitive sage: C.plot(patch=0) Graphics object consisting of 1 graphics primitive sage: C.plot(patch=1) Graphics object consisting of 1 graphics primitive A hyperelliptic curve:: sage: P.<x> = QQ[] sage: f = 4*x^5 - 30*x^3 + 45*x - 22 sage: C = HyperellipticCurve(f) sage: C.plot() Graphics object consisting of 1 graphics primitive sage: C.plot(patch=0) Graphics object consisting of 1 graphics primitive sage: C.plot(patch=1) Graphics object consisting of 1 graphics primitive """ # if user hasn't specified a favourite affine patch, take the # one avoiding "infinity", i.e. the one corresponding to the # last projective coordinate being nonzero patch = kwds.pop('patch', self.ngens() - 1) from constructor import Curve C = Curve(self.affine_patch(patch)) return C.plot(*args, **kwds) def is_singular(C): r""" Returns whether the curve is singular or not. EXAMPLES: Over `\QQ`:: sage: F = QQ sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X^3-Y^2*Z) sage: C.is_singular() True Over a finite field:: sage: F = GF(19) sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X^3+Y^3+Z^3) sage: C.is_singular() False sage: D = Curve(X^4-X*Z^3) sage: D.is_singular() True sage: E = Curve(X^5+19*Y^5+Z^5) sage: E.is_singular() True sage: E = Curve(X^5+9*Y^5+Z^5) sage: E.is_singular() False Over `\CC`:: sage: F = CC sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X) sage: C.is_singular() False sage: D = Curve(Y^2*Z-X^3) sage: D.is_singular() True sage: E = Curve(Y^2*Z-X^3+Z^3) sage: E.is_singular() False Showing that ticket #12187 is fixed:: sage: F.<X,Y,Z> = GF(2)[] sage: G = Curve(X^2+Y*Z) sage: G.is_singular() False """ poly = C.defining_polynomial() return poly.parent().ideal(poly.gradient()+[poly]).dimension()> 0 class ProjectiveCurve_finite_field(ProjectiveCurve_generic): def rational_points_iterator(self): r""" Return a generator object for the rational points on this curve. INPUT: - ``self`` -- a projective curve OUTPUT: A generator of all the rational points on the curve defined over its base field. EXAMPLE:: sage: F = GF(37) sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X^7+Y*X*Z^5*55+Y^7*12) sage: len(list(C.rational_points_iterator())) 37 :: sage: F = GF(2) sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X*Y*Z) sage: a = C.rational_points_iterator() sage: next(a) (1 : 0 : 0) sage: next(a) (0 : 1 : 0) sage: next(a) (1 : 1 : 0) sage: next(a) (0 : 0 : 1) sage: next(a) (1 : 0 : 1) sage: next(a) (0 : 1 : 1) sage: next(a) Traceback (most recent call last): ... StopIteration :: sage: F = GF(3^2,'a') sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X^3+5*Y^2*Z-33*X*Y*X) sage: b = C.rational_points_iterator() sage: next(b) (0 : 1 : 0) sage: next(b) (0 : 0 : 1) sage: next(b) (2*a + 2 : a : 1) sage: next(b) (2 : a + 1 : 1) sage: next(b) (a + 1 : 2*a + 1 : 1) sage: next(b) (1 : 2 : 1) sage: next(b) (2*a + 2 : 2*a : 1) sage: next(b) (2 : 2*a + 2 : 1) sage: next(b) (a + 1 : a + 2 : 1) sage: next(b) (1 : 1 : 1) sage: next(b) Traceback (most recent call last): ... StopIteration """ g = self.defining_polynomial() K = g.parent().base_ring() from sage.rings.polynomial.all import PolynomialRing R = PolynomialRing(K,'X') X = R.gen() one = K.one() zero = K.zero() # the point with Z = 0 = Y try: t = self.point([one,zero,zero]) yield(t) except TypeError: pass # points with Z = 0, Y = 1 g10 = R(g(X,one,zero)) if g10.is_zero(): for x in K: yield(self.point([x,one,zero])) else: for x in g10.roots(multiplicities=False): yield(self.point([x,one,zero])) # points with Z = 1 for y in K: gy1 = R(g(X,y,one)) if gy1.is_zero(): for x in K: yield(self.point([x,y,one])) else: for x in gy1.roots(multiplicities=False): yield(self.point([x,y,one])) def rational_points(self, algorithm="enum", sort=True): r""" Return the rational points on this curve computed via enumeration. INPUT: - ``algorithm`` (string, default: 'enum') -- the algorithm to use. Currently this is ignored. - ``sort`` (boolean, default ``True``) -- whether the output points should be sorted. If False, the order of the output is non-deterministic. OUTPUT: A list of all the rational points on the curve defined over its base field, possibly sorted. .. note:: This is a slow Python-level implementation. EXAMPLES:: sage: F = GF(7) sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X^3+Y^3-Z^3) sage: C.rational_points() [(0 : 1 : 1), (0 : 2 : 1), (0 : 4 : 1), (1 : 0 : 1), (2 : 0 : 1), (3 : 1 : 0), (4 : 0 : 1), (5 : 1 : 0), (6 : 1 : 0)] :: sage: F = GF(1237) sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X^7+7*Y^6*Z+Z^4*X^2*Y*89) sage: len(C.rational_points()) 1237 :: sage: F = GF(2^6,'a') sage: P2.<X,Y,Z> = ProjectiveSpace(F,2) sage: C = Curve(X^5+11*X*Y*Z^3 + X^2*Y^3 - 13*Y^2*Z^3) sage: len(C.rational_points()) 104 :: sage: R.<x,y,z> = GF(2)[] sage: f = x^3*y + y^3*z + x*z^3 sage: C = Curve(f); pts = C.rational_points() sage: pts [(0 : 0 : 1), (0 : 1 : 0), (1 : 0 : 0)] """ points = list(self.rational_points_iterator()) if sort: points.sort() return points class ProjectiveCurve_prime_finite_field(ProjectiveCurve_finite_field): def _points_via_singular(self, sort=True): r""" Return all rational points on this curve, computed using Singular's Brill-Noether implementation. INPUT: - ``sort`` - bool (default: True), if True return the point list sorted. If False, returns the points in the order computed by Singular. EXAMPLE:: sage: x, y, z = PolynomialRing(GF(5), 3, 'xyz').gens() sage: f = y^2*z^7 - x^9 - x*z^8 sage: C = Curve(f); C Projective Curve over Finite Field of size 5 defined by -x^9 + y^2*z^7 - x*z^8 sage: C._points_via_singular() [(0 : 0 : 1), (0 : 1 : 0), (2 : 2 : 1), (2 : 3 : 1), (3 : 1 : 1), (3 : 4 : 1)] sage: C._points_via_singular(sort=False) #random [(0 : 1 : 0), (3 : 1 : 1), (3 : 4 : 1), (2 : 2 : 1), (0 : 0 : 1), (2 : 3 : 1)] .. note:: The Brill-Noether package does not always work (i.e., the 'bn' algorithm. When it fails a RuntimeError exception is raised. """ f = self.defining_polynomial()._singular_() singular = f.parent() singular.lib('brnoeth') try: X1 = f.Adj_div() except (TypeError, RuntimeError) as s: raise RuntimeError(str(s) + "\n\n ** Unable to use the\ Brill-Noether Singular package to\ compute all points (see above).") X2 = singular.NSplaces(1, X1) R = X2[5][1][1] singular.set_ring(R) # We use sage_flattened_str_list since iterating through # the entire list through the sage/singular interface directly # would involve hundreds of calls to singular, and timing issues with # the expect interface could crop up. Also, this is vastly # faster (and more robust). v = singular('POINTS').sage_flattened_str_list() pnts = [self(int(v[3*i]), int(v[3*i+1]), int(v[3*i+2])) for i in range(len(v)//3)] # singular always dehomogenizes with respect to the last variable # so if this variable divides the curve equation, we need to add # points at infinity F = self.defining_polynomial() z = F.parent().gens()[-1] if z.divides(F): pnts += [self(1,a,0) for a in self.base_ring()] pnts += [self(0,1,0)] # remove multiple points pnts = list(set(pnts)) if sort: pnts.sort() return pnts def riemann_roch_basis(self, D): r""" Return a basis for the Riemann-Roch space corresponding to `D`. This uses Singular's Brill-Noether implementation. INPUT: - ``D`` - a divisor OUTPUT: A list of function field elements that form a basis of the Riemann-Roch space EXAMPLE:: sage: R.<x,y,z> = GF(2)[] sage: f = x^3*y + y^3*z + x*z^3 sage: C = Curve(f); pts = C.rational_points() sage: D = C.divisor([ (4, pts[0]), (4, pts[2]) ]) sage: C.riemann_roch_basis(D) [x/y, 1, z/y, z^2/y^2, z/x, z^2/(x*y)] :: sage: R.<x,y,z> = GF(5)[] sage: f = x^7 + y^7 + z^7 sage: C = Curve(f); pts = C.rational_points() sage: D = C.divisor([ (3, pts[0]), (-1,pts[1]), (10, pts[5]) ]) sage: C.riemann_roch_basis(D) [(-2*x + y)/(x + y), (-x + z)/(x + y)] .. NOTE:: Currently this only works over prime field and divisors supported on rational points. """ f = self.defining_polynomial()._singular_() singular = f.parent() singular.lib('brnoeth') try: X1 = f.Adj_div() except (TypeError, RuntimeError) as s: raise RuntimeError(str(s) + "\n\n ** Unable to use the Brill-Noether Singular package to compute all points (see above).") X2 = singular.NSplaces(1, X1) # retrieve list of all computed closed points (possibly of degree >1) v = X2[3].sage_flattened_str_list() # We use sage_flattened_str_list since iterating through # the entire list through the sage/singular interface directly # would involve hundreds of calls to singular, and timing issues with # the expect interface could crop up. Also, this is vastly # faster (and more robust). v = [ v[i].partition(',') for i in range(len(v)) ] pnts = [ ( int(v[i][0]), int(v[i][2])-1 ) for i in range(len(v))] # retrieve coordinates of rational points R = X2[5][1][1] singular.set_ring(R) v = singular('POINTS').sage_flattened_str_list() coords = [self(int(v[3*i]), int(v[3*i+1]), int(v[3*i+2])) for i in range(len(v)//3)] # build correct representation of D for singular Dsupport = D.support() Dcoeffs = [] for x in pnts: if x[0] == 1: Dcoeffs.append(D.coefficient(coords[x[1]])) else: Dcoeffs.append(0) Dstr = str(tuple(Dcoeffs)) G = singular(','.join([str(x) for x in Dcoeffs]), type='intvec') # call singular's brill noether routine and return T = X2[1][2] T.set_ring() LG = G.BrillNoether(X2) LG = [X.split(',\n') for X in LG.sage_structured_str_list()] x,y,z = self.ambient_space().coordinate_ring().gens() vars = {'x':x, 'y':y, 'z':z} V = [(sage_eval(a, vars)/sage_eval(b, vars)) for a, b in LG] return V def rational_points(self, algorithm="enum", sort=True): r""" INPUT: - ``algorithm`` - string: - ``'enum'`` - straightforward enumeration - ``'bn'`` - via Singular's brnoeth package. EXAMPLE:: sage: x, y, z = PolynomialRing(GF(5), 3, 'xyz').gens() sage: f = y^2*z^7 - x^9 - x*z^8 sage: C = Curve(f); C Projective Curve over Finite Field of size 5 defined by -x^9 + y^2*z^7 - x*z^8 sage: C.rational_points() [(0 : 0 : 1), (0 : 1 : 0), (2 : 2 : 1), (2 : 3 : 1), (3 : 1 : 1), (3 : 4 : 1)] sage: C = Curve(x - y + z) sage: C.rational_points() [(0 : 1 : 1), (1 : 1 : 0), (1 : 2 : 1), (2 : 3 : 1), (3 : 4 : 1), (4 : 0 : 1)] sage: C = Curve(x*z+z^2) sage: C.rational_points('all') [(0 : 1 : 0), (1 : 0 : 0), (1 : 1 : 0), (2 : 1 : 0), (3 : 1 : 0), (4 : 0 : 1), (4 : 1 : 0), (4 : 1 : 1), (4 : 2 : 1), (4 : 3 : 1), (4 : 4 : 1)] .. note:: The Brill-Noether package does not always work (i.e., the 'bn' algorithm. When it fails a RuntimeError exception is raised. """ if algorithm == "enum": return ProjectiveCurve_finite_field.rational_points(self, algorithm="enum", sort=sort) elif algorithm == "bn": return self._points_via_singular(sort=sort) elif algorithm == "all": S_enum = self.rational_points(algorithm = "enum") S_bn = self.rational_points(algorithm = "bn") if S_enum != S_bn: raise RuntimeError("Bug in rational_points -- different\ algorithms give different answers for\ curve %s!"%self) return S_enum else: raise ValueError("No algorithm '%s' known"%algorithm) def Hasse_bounds(q, genus=1): r""" Return the Hasse-Weil bounds for the cardinality of a nonsingular curve defined over `\GF{q}` of given ``genus``. INPUT: - ``q`` (int) -- a prime power - ``genus`` (int, default 1) -- a non-negative integer, OUTPUT: (tuple) The Hasse bounds (lb,ub) for the cardinality of a curve of genus ``genus`` defined over `\GF{q}`. EXAMPLES:: sage: Hasse_bounds(2) (1, 5) sage: Hasse_bounds(next_prime(10^30)) (999999999999998000000000000058, 1000000000000002000000000000058) """ if genus==1: rq = (4*q).isqrt() else: rq = (4*(genus**2)*q).isqrt() return (q+1-rq,q+1+rq)
32.17232
134
0.484374
from sage.interfaces.all import singular from sage.misc.all import add, sage_eval from sage.rings.all import degree_lowest_rational_function from sage.schemes.projective.projective_space import is_ProjectiveSpace from curve import Curve_generic_projective class ProjectiveSpaceCurve_generic(Curve_generic_projective): def _repr_type(self): return "Projective Space" def __init__(self, A, X): if not is_ProjectiveSpace(A): raise TypeError("A (=%s) must be a projective space"%A) Curve_generic_projective.__init__(self, A, X) d = self.dimension() if d != 1: raise ValueError("defining equations (=%s) define a scheme of dimension %s != 1"%(X,d)) class ProjectiveCurve_generic(Curve_generic_projective): def __init__(self, A, f): if not (is_ProjectiveSpace(A) and A.dimension != 2): raise TypeError("Argument A (= %s) must be a projective plane."%A) Curve_generic_projective.__init__(self, A, [f]) def _repr_type(self): return "Projective" def arithmetic_genus(self): d = self.defining_polynomial().total_degree() return int((d-1)*(d-2)/2) def divisor_of_function(self, r): F = self.base_ring() f = self.defining_polynomial() x, y, z = f.parent().gens() pnts = self.rational_points() divf = [] for P in pnts: if P[2] != F(0): lcs = self.local_coordinates(P,5) ldg = degree_lowest_rational_function(r(lcs[0],lcs[1]),z) if ldg[0] != 0: divf.append([ldg[0],P]) return divf def local_coordinates(self, pt, n): f = self.defining_polynomial() R = f.parent() F = self.base_ring() p = F.characteristic() x0 = F(pt[0]) y0 = F(pt[1]) astr = ["a"+str(i) for i in range(1,2*n)] x,y = R.gens() R0 = PolynomialRing(F,2*n+2,names = [str(x),str(y),"t"]+astr) vars0 = R0.gens() t = vars0[2] yt = y0*t**0 + add([vars0[i]*t**(i-2) for i in range(3,2*n+2)]) xt = x0+t ft = f(xt,yt) S = singular S.eval('ring s = '+str(p)+','+str(R0.gens())+',lp;') S.eval('poly f = '+str(ft)) cmd = 'matrix c = coeffs ('+str(ft)+',t)' S.eval(cmd) N = int(S.eval('size(c)')) b = ["c["+str(i)+",1]," for i in range(2,N/2-4)] b = ''.join(b) b = b[:len(b)-1] cmd = 'ideal I = '+b S.eval(cmd) c = S.eval('slimgb(I)') d = c.split("=") d = d[1:] d[len(d)-1] += "\n" e = [x[:x.index("\n")] for x in d] vals = [] for x in e: for y in vars0: if str(y) in x: if len(x.replace(str(y),"")) != 0: i = x.find("-") if i>0: vals.append([eval(x[1:i]),x[:i],F(eval(x[i+1:]))]) i = x.find("+") if i>0: vals.append([eval(x[1:i]),x[:i],-F(eval(x[i+1:]))]) else: vals.append([eval(str(y)[1:]),str(y),F(0)]) vals.sort() k = len(vals) v = [x0+t,y0+add([vals[i][2]*t**(i+1) for i in range(k)])] return v def plot(self, *args, **kwds): # one avoiding "infinity", i.e. the one corresponding to the # last projective coordinate being nonzero patch = kwds.pop('patch', self.ngens() - 1) from constructor import Curve C = Curve(self.affine_patch(patch)) return C.plot(*args, **kwds) def is_singular(C): poly = C.defining_polynomial() return poly.parent().ideal(poly.gradient()+[poly]).dimension()> 0 class ProjectiveCurve_finite_field(ProjectiveCurve_generic): def rational_points_iterator(self): g = self.defining_polynomial() K = g.parent().base_ring() from sage.rings.polynomial.all import PolynomialRing R = PolynomialRing(K,'X') X = R.gen() one = K.one() zero = K.zero() # the point with Z = 0 = Y try: t = self.point([one,zero,zero]) yield(t) except TypeError: pass # points with Z = 0, Y = 1 g10 = R(g(X,one,zero)) if g10.is_zero(): for x in K: yield(self.point([x,one,zero])) else: for x in g10.roots(multiplicities=False): yield(self.point([x,one,zero])) # points with Z = 1 for y in K: gy1 = R(g(X,y,one)) if gy1.is_zero(): for x in K: yield(self.point([x,y,one])) else: for x in gy1.roots(multiplicities=False): yield(self.point([x,y,one])) def rational_points(self, algorithm="enum", sort=True): points = list(self.rational_points_iterator()) if sort: points.sort() return points class ProjectiveCurve_prime_finite_field(ProjectiveCurve_finite_field): def _points_via_singular(self, sort=True): f = self.defining_polynomial()._singular_() singular = f.parent() singular.lib('brnoeth') try: X1 = f.Adj_div() except (TypeError, RuntimeError) as s: raise RuntimeError(str(s) + "\n\n ** Unable to use the\ Brill-Noether Singular package to\ compute all points (see above).") X2 = singular.NSplaces(1, X1) R = X2[5][1][1] singular.set_ring(R) # We use sage_flattened_str_list since iterating through # the entire list through the sage/singular interface directly # would involve hundreds of calls to singular, and timing issues with # the expect interface could crop up. Also, this is vastly # faster (and more robust). v = singular('POINTS').sage_flattened_str_list() pnts = [self(int(v[3*i]), int(v[3*i+1]), int(v[3*i+2])) for i in range(len(v)//3)] # singular always dehomogenizes with respect to the last variable # so if this variable divides the curve equation, we need to add # points at infinity F = self.defining_polynomial() z = F.parent().gens()[-1] if z.divides(F): pnts += [self(1,a,0) for a in self.base_ring()] pnts += [self(0,1,0)] # remove multiple points pnts = list(set(pnts)) if sort: pnts.sort() return pnts def riemann_roch_basis(self, D): f = self.defining_polynomial()._singular_() singular = f.parent() singular.lib('brnoeth') try: X1 = f.Adj_div() except (TypeError, RuntimeError) as s: raise RuntimeError(str(s) + "\n\n ** Unable to use the Brill-Noether Singular package to compute all points (see above).") X2 = singular.NSplaces(1, X1) # retrieve list of all computed closed points (possibly of degree >1) v = X2[3].sage_flattened_str_list() # We use sage_flattened_str_list since iterating through # the entire list through the sage/singular interface directly # would involve hundreds of calls to singular, and timing issues with # the expect interface could crop up. Also, this is vastly # faster (and more robust). v = [ v[i].partition(',') for i in range(len(v)) ] pnts = [ ( int(v[i][0]), int(v[i][2])-1 ) for i in range(len(v))] # retrieve coordinates of rational points R = X2[5][1][1] singular.set_ring(R) v = singular('POINTS').sage_flattened_str_list() coords = [self(int(v[3*i]), int(v[3*i+1]), int(v[3*i+2])) for i in range(len(v)//3)] # build correct representation of D for singular Dsupport = D.support() Dcoeffs = [] for x in pnts: if x[0] == 1: Dcoeffs.append(D.coefficient(coords[x[1]])) else: Dcoeffs.append(0) Dstr = str(tuple(Dcoeffs)) G = singular(','.join([str(x) for x in Dcoeffs]), type='intvec') # call singular's brill noether routine and return T = X2[1][2] T.set_ring() LG = G.BrillNoether(X2) LG = [X.split(',\n') for X in LG.sage_structured_str_list()] x,y,z = self.ambient_space().coordinate_ring().gens() vars = {'x':x, 'y':y, 'z':z} V = [(sage_eval(a, vars)/sage_eval(b, vars)) for a, b in LG] return V def rational_points(self, algorithm="enum", sort=True): if algorithm == "enum": return ProjectiveCurve_finite_field.rational_points(self, algorithm="enum", sort=sort) elif algorithm == "bn": return self._points_via_singular(sort=sort) elif algorithm == "all": S_enum = self.rational_points(algorithm = "enum") S_bn = self.rational_points(algorithm = "bn") if S_enum != S_bn: raise RuntimeError("Bug in rational_points -- different\ algorithms give different answers for\ curve %s!"%self) return S_enum else: raise ValueError("No algorithm '%s' known"%algorithm) def Hasse_bounds(q, genus=1): if genus==1: rq = (4*q).isqrt() else: rq = (4*(genus**2)*q).isqrt() return (q+1-rq,q+1+rq)
true
true
f70ea37895116b00771e48f3d3576aa318e30354
5,209
py
Python
HelloWorldOpenCV.py
jhbrito/HelloWorlds
7e2247ca7f312a516ce6a5054913d59e2f1de0f9
[ "MIT" ]
6
2020-02-14T15:18:14.000Z
2022-02-22T15:40:33.000Z
HelloWorldOpenCV.py
jhbrito/HelloWorlds
7e2247ca7f312a516ce6a5054913d59e2f1de0f9
[ "MIT" ]
null
null
null
HelloWorldOpenCV.py
jhbrito/HelloWorlds
7e2247ca7f312a516ce6a5054913d59e2f1de0f9
[ "MIT" ]
null
null
null
# Demo with a few examples of using OpenCV functions and UI # packages: opencv-python # uses lena: https://upload.wikimedia.org/wikipedia/en/7/7d/Lenna_%28test_image%29.png import numpy as np import cv2 print("Hello World OpenCV") print("OpenCV Version:", cv2.__version__) image = np.ones((256, 256), dtype="uint8") image = image * 127 image[0:128, 0:128] = 0 image[128:, 128:] = 255 cv2.imshow("Image", image) cv2.waitKey(0) # Opening and Viewing an Image import os.path if os.path.isfile('lena.png'): print("Test Image File exist") else: print("Test Image File does not exist; downloading...") import urllib.request as urllib_request urllib_request.urlretrieve("https://upload.wikimedia.org/wikipedia/en/7/7d/Lenna_%28test_image%29.png", "lena.png") image = cv2.imread("./lena.png") cv2.imshow("Image", image) cv2.waitKey(0) cv2.destroyAllWindows() rgb_image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB) cv2.imshow("Image RGB", rgb_image) cv2.waitKey(0) cv2.destroyAllWindows() def viewImage(image, name_of_window): cv2.namedWindow(name_of_window, cv2.WINDOW_AUTOSIZE) cv2.imshow(name_of_window, image) cv2.waitKey(0) cv2.destroyAllWindows() viewImage(image, "Lena") # Edit pixels edited = image.copy() edited[200:390, 200:360, 0] = 255 viewImage(edited, "Lena edited") # Cropping cropped = image[200:390, 200:360] viewImage(cropped, "Lena cropped") # Resizing scale_percent = 10 # percent of original size width = int(image.shape[1] * scale_percent / 100) height = int(image.shape[0] * scale_percent / 100) dim = (width, height) resized = cv2.resize(image, dim, interpolation=cv2.INTER_AREA) viewImage(resized, "Lena resized to {}%".format(scale_percent)) # Drawing a Rectangle output = image.copy() cv2.rectangle(output, (200, 200), (360, 390), (255, 0, 0), 10) viewImage(output, "Lena with a rectangle") # Drawing a line cv2.line(output, (256, 390), (256, 512), (0, 0, 255), 5) viewImage(output, "Lena with a line") # Writing on an image cv2.putText(output, "Lena", (360, 390), cv2.FONT_HERSHEY_SIMPLEX, 2, (0, 255, 0), 2) viewImage(output, "Lena with text") # Saving an image cv2.imwrite("./output.jpg", output) # Blurring/Smoothing blurred = cv2.GaussianBlur(image, (15, 15), 0) viewImage(blurred, "Lena blurred") # Rotating (h, w, d) = image.shape center = (w // 2, h // 2) rot = 45 M = cv2.getRotationMatrix2D(center, rot, 1.0) rotated = cv2.warpAffine(image, M, (w, h)) viewImage(rotated, "Lena rotated by {} degrees".format(rot)) # Blend alpha_slider_max = 100 def on_trackbar_weight(val): alpha = val / alpha_slider_max beta = (1.0 - alpha) blend = cv2.addWeighted(image, alpha, rotated, beta, 0.0) cv2.imshow('Lena blended', blend) cv2.namedWindow('Lena blended') trackbar_name = 'Alpha 0 - {}'.format(alpha_slider_max) cv2.createTrackbar(trackbar_name, 'Lena blended', 50, alpha_slider_max, on_trackbar_weight) on_trackbar_weight(50) cv2.waitKey() cv2.destroyWindow('Lena blended') # Grayscaling gray_image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) viewImage(gray_image, "Lena gray-scale") # Thresholding threshold_slider_max = 255 threshold = 200 ret, threshold_image = cv2.threshold(gray_image, threshold, 255, 0) def on_trackbar_threshold(val): threshold = val ret, threshold_image = cv2.threshold(gray_image, threshold, 255, 0) cv2.imshow("Lena thresholded", threshold_image) cv2.namedWindow("Lena thresholded") trackbar_name = "Threshold 0 - {}".format(threshold_slider_max) cv2.createTrackbar(trackbar_name, "Lena thresholded", threshold, threshold_slider_max, on_trackbar_threshold) on_trackbar_threshold(threshold) cv2.waitKey() cv2.destroyWindow("Lena thresholded") # Contours contours, hierarchy = cv2.findContours(threshold_image, cv2.RETR_TREE, cv2.CHAIN_APPROX_NONE) image_with_contours = image.copy() cv2.drawContours(image_with_contours, contours, -1, (255, 0, 0), 1) viewImage(image_with_contours, "Lena contours") # Face Detection face_cascade = cv2.CascadeClassifier('venv\Lib\site-packages\cv2\data\haarcascade_frontalface_default.xml') faces = face_cascade.detectMultiScale(gray_image) print("Lena with {} faces detected".format(len(faces))) image_faces = image.copy() for (x, y, w, h) in faces: cv2.rectangle(image_faces, (x, y), (x + w, y + h), (0, 255, 0), 2) viewImage(image_faces, "Lena with {} faces detected".format(len(faces))) def display_box(im, bbox): n_boxes = len(bbox) for j_box in range(n_boxes): for j in range(4): cv2.line(im, (int(bbox[j_box][j][0]), int(bbox[j_box][j][1])), (int(bbox[j_box][(j + 1) % 4][0]), int(bbox[j_box][(j + 1) % 4][1])), (255, 0, 0), 3) # Display results cv2.imshow("Results", im) inputImage = cv2.imread("qrcode.jpg") qrDecoder = cv2.QRCodeDetector() data, bbox, rectifiedImage = qrDecoder.detectAndDecode(inputImage) if len(data) > 0: print("Decoded Data : {}".format(data)) display_box(inputImage, bbox) rectifiedImage = np.uint8(rectifiedImage) cv2.imshow("Rectified QRCode", rectifiedImage) else: print("QR Code not detected") cv2.imshow("Results", inputImage) cv2.waitKey(0) cv2.destroyAllWindows()
29.429379
119
0.710501
import numpy as np import cv2 print("Hello World OpenCV") print("OpenCV Version:", cv2.__version__) image = np.ones((256, 256), dtype="uint8") image = image * 127 image[0:128, 0:128] = 0 image[128:, 128:] = 255 cv2.imshow("Image", image) cv2.waitKey(0) import os.path if os.path.isfile('lena.png'): print("Test Image File exist") else: print("Test Image File does not exist; downloading...") import urllib.request as urllib_request urllib_request.urlretrieve("https://upload.wikimedia.org/wikipedia/en/7/7d/Lenna_%28test_image%29.png", "lena.png") image = cv2.imread("./lena.png") cv2.imshow("Image", image) cv2.waitKey(0) cv2.destroyAllWindows() rgb_image = cv2.cvtColor(image, cv2.COLOR_BGR2RGB) cv2.imshow("Image RGB", rgb_image) cv2.waitKey(0) cv2.destroyAllWindows() def viewImage(image, name_of_window): cv2.namedWindow(name_of_window, cv2.WINDOW_AUTOSIZE) cv2.imshow(name_of_window, image) cv2.waitKey(0) cv2.destroyAllWindows() viewImage(image, "Lena") edited = image.copy() edited[200:390, 200:360, 0] = 255 viewImage(edited, "Lena edited") cropped = image[200:390, 200:360] viewImage(cropped, "Lena cropped") scale_percent = 10 width = int(image.shape[1] * scale_percent / 100) height = int(image.shape[0] * scale_percent / 100) dim = (width, height) resized = cv2.resize(image, dim, interpolation=cv2.INTER_AREA) viewImage(resized, "Lena resized to {}%".format(scale_percent)) output = image.copy() cv2.rectangle(output, (200, 200), (360, 390), (255, 0, 0), 10) viewImage(output, "Lena with a rectangle") cv2.line(output, (256, 390), (256, 512), (0, 0, 255), 5) viewImage(output, "Lena with a line") cv2.putText(output, "Lena", (360, 390), cv2.FONT_HERSHEY_SIMPLEX, 2, (0, 255, 0), 2) viewImage(output, "Lena with text") cv2.imwrite("./output.jpg", output) blurred = cv2.GaussianBlur(image, (15, 15), 0) viewImage(blurred, "Lena blurred") (h, w, d) = image.shape center = (w // 2, h // 2) rot = 45 M = cv2.getRotationMatrix2D(center, rot, 1.0) rotated = cv2.warpAffine(image, M, (w, h)) viewImage(rotated, "Lena rotated by {} degrees".format(rot)) alpha_slider_max = 100 def on_trackbar_weight(val): alpha = val / alpha_slider_max beta = (1.0 - alpha) blend = cv2.addWeighted(image, alpha, rotated, beta, 0.0) cv2.imshow('Lena blended', blend) cv2.namedWindow('Lena blended') trackbar_name = 'Alpha 0 - {}'.format(alpha_slider_max) cv2.createTrackbar(trackbar_name, 'Lena blended', 50, alpha_slider_max, on_trackbar_weight) on_trackbar_weight(50) cv2.waitKey() cv2.destroyWindow('Lena blended') gray_image = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) viewImage(gray_image, "Lena gray-scale") threshold_slider_max = 255 threshold = 200 ret, threshold_image = cv2.threshold(gray_image, threshold, 255, 0) def on_trackbar_threshold(val): threshold = val ret, threshold_image = cv2.threshold(gray_image, threshold, 255, 0) cv2.imshow("Lena thresholded", threshold_image) cv2.namedWindow("Lena thresholded") trackbar_name = "Threshold 0 - {}".format(threshold_slider_max) cv2.createTrackbar(trackbar_name, "Lena thresholded", threshold, threshold_slider_max, on_trackbar_threshold) on_trackbar_threshold(threshold) cv2.waitKey() cv2.destroyWindow("Lena thresholded") contours, hierarchy = cv2.findContours(threshold_image, cv2.RETR_TREE, cv2.CHAIN_APPROX_NONE) image_with_contours = image.copy() cv2.drawContours(image_with_contours, contours, -1, (255, 0, 0), 1) viewImage(image_with_contours, "Lena contours") face_cascade = cv2.CascadeClassifier('venv\Lib\site-packages\cv2\data\haarcascade_frontalface_default.xml') faces = face_cascade.detectMultiScale(gray_image) print("Lena with {} faces detected".format(len(faces))) image_faces = image.copy() for (x, y, w, h) in faces: cv2.rectangle(image_faces, (x, y), (x + w, y + h), (0, 255, 0), 2) viewImage(image_faces, "Lena with {} faces detected".format(len(faces))) def display_box(im, bbox): n_boxes = len(bbox) for j_box in range(n_boxes): for j in range(4): cv2.line(im, (int(bbox[j_box][j][0]), int(bbox[j_box][j][1])), (int(bbox[j_box][(j + 1) % 4][0]), int(bbox[j_box][(j + 1) % 4][1])), (255, 0, 0), 3) cv2.imshow("Results", im) inputImage = cv2.imread("qrcode.jpg") qrDecoder = cv2.QRCodeDetector() data, bbox, rectifiedImage = qrDecoder.detectAndDecode(inputImage) if len(data) > 0: print("Decoded Data : {}".format(data)) display_box(inputImage, bbox) rectifiedImage = np.uint8(rectifiedImage) cv2.imshow("Rectified QRCode", rectifiedImage) else: print("QR Code not detected") cv2.imshow("Results", inputImage) cv2.waitKey(0) cv2.destroyAllWindows()
true
true
f70ea378aaf48671277dab43c5ea3b4387cbb5a8
7,007
py
Python
vitrage/datasources/zabbix/driver.py
openstack/vitrage
95b33dbf39b040e23915882a2879c87aec239ca9
[ "Apache-2.0" ]
89
2015-09-30T21:42:17.000Z
2022-03-28T16:31:19.000Z
vitrage/datasources/zabbix/driver.py
openstack/vitrage
95b33dbf39b040e23915882a2879c87aec239ca9
[ "Apache-2.0" ]
4
2015-12-13T13:06:53.000Z
2016-01-03T19:51:28.000Z
vitrage/datasources/zabbix/driver.py
openstack/vitrage
95b33dbf39b040e23915882a2879c87aec239ca9
[ "Apache-2.0" ]
43
2015-11-04T15:54:27.000Z
2021-12-10T14:24:03.000Z
# Copyright 2016 - Nokia # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from collections import namedtuple from oslo_config import cfg from oslo_log import log from oslo_utils import importutils as utils from vitrage.common.constants import DatasourceAction from vitrage.common.constants import DatasourceOpts as DSOpts from vitrage.common.constants import DatasourceProperties as DSProps from vitrage.datasources.alarm_driver_base import AlarmDriverBase from vitrage.datasources.zabbix.properties import ZabbixProperties as ZProps from vitrage.datasources.zabbix.properties import ZabbixTriggerStatus \ as TriggerStatus from vitrage.datasources.zabbix.properties import ZabbixTriggerValue \ as TriggerValue from vitrage.datasources.zabbix import ZABBIX_DATASOURCE from vitrage.utils import file as file_utils CONF = cfg.CONF LOG = log.getLogger(__name__) class ZabbixDriver(AlarmDriverBase): ServiceKey = namedtuple('ServiceKey', ['hostname', 'triggerid']) conf_map = None def __init__(self): super(ZabbixDriver, self).__init__() if not ZabbixDriver.conf_map: ZabbixDriver.conf_map =\ ZabbixDriver._configuration_mapping() self._client = None def zabbix_client_login(self): if not CONF.zabbix.user: LOG.warning('Zabbix user is not defined') if not CONF.zabbix.password: LOG.warning('Zabbix password is not defined') if not CONF.zabbix.url: LOG.warning('Zabbix url is not defined') try: if not self._client: self._client = utils.import_object( 'pyzabbix.ZabbixAPI', CONF.zabbix.url) self._client.login( CONF.zabbix.user, CONF.zabbix.password) except Exception: LOG.exception('pyzabbix.ZabbixAPI error occurred.') self._client = None def _vitrage_type(self): return ZABBIX_DATASOURCE def _alarm_key(self, alarm): return self.ServiceKey(hostname=alarm[ZProps.RESOURCE_NAME], triggerid=alarm[ZProps.TRIGGER_ID]) def _get_alarms(self): self.zabbix_client_login() if not self._client: return [] alarms = [] valid_hosts = (host for host in self._client.host.get(output=[ZProps.HOST]) if host[ZProps.HOST] in ZabbixDriver.conf_map) for host in valid_hosts: self._get_triggers_per_host(host, alarms) return alarms def _get_triggers_per_host(self, host, alarms): host_id = host[ZProps.HOST_ID] triggers = self._client.trigger.get(hostids=host_id, expandDescription=True) triggers_rawtexts = self._get_triggers_rawtexts(host_id) for trigger in triggers: trigger[ZProps.ZABBIX_RESOURCE_NAME] = host[ZProps.HOST] trigger_id = trigger[ZProps.TRIGGER_ID] trigger[ZProps.RAWTEXT] = triggers_rawtexts[trigger_id] alarms.append(trigger) def _get_triggers_rawtexts(self, host_id): output = [ZProps.TRIGGER_ID, ZProps.DESCRIPTION] triggers = self._client.trigger.get(hostids=host_id, output=output) return {trigger[ZProps.TRIGGER_ID]: trigger[ZProps.DESCRIPTION] for trigger in triggers} def _enrich_alarms(self, alarms): """Enrich zabbix alarm using zabbix configuration file converting Zabbix host name to Vitrage resource type and name :param alarms: Zabbix alarm :return: enriched alarm """ for alarm in alarms: alarm[ZProps.VALUE] = self._get_value(alarm) zabbix_host = alarm[ZProps.ZABBIX_RESOURCE_NAME] vitrage_host = ZabbixDriver.conf_map[zabbix_host] alarm[ZProps.RESOURCE_TYPE] = vitrage_host[ZProps.RESOURCE_TYPE] alarm[ZProps.RESOURCE_NAME] = vitrage_host[ZProps.RESOURCE_NAME] def _is_erroneous(self, alarm): return alarm and \ alarm[ZProps.VALUE] == TriggerValue.PROBLEM def _status_changed(self, new_alarm, old_alarm): if not (new_alarm and old_alarm): return False if new_alarm[ZProps.VALUE] != old_alarm[ZProps.VALUE]: return True if new_alarm[ZProps.VALUE] == TriggerValue.PROBLEM: priority_changed = \ new_alarm[ZProps.PRIORITY] != old_alarm[ZProps.PRIORITY] description_changed = \ new_alarm[ZProps.DESCRIPTION] != old_alarm[ZProps.DESCRIPTION] return priority_changed or description_changed def _is_valid(self, alarm): return alarm[ZProps.RESOURCE_TYPE] is not None and \ alarm[ZProps.RESOURCE_NAME] is not None @staticmethod def _get_value(alarm): if alarm[ZProps.STATUS] == TriggerStatus.DISABLED: return TriggerValue.OK return alarm[ZProps.VALUE] @staticmethod def _configuration_mapping(): try: zabbix_config_file = CONF.zabbix[DSOpts.CONFIG_FILE] zabbix_config = file_utils.load_yaml_file(zabbix_config_file) zabbix_config_elements = zabbix_config[ZABBIX_DATASOURCE] mappings = {} for element_config in zabbix_config_elements: mappings[element_config['zabbix_host']] = { ZProps.RESOURCE_TYPE: element_config['type'], ZProps.RESOURCE_NAME: element_config['name'] } return mappings except Exception: LOG.exception('Failed in init.') return {} def enrich_event(self, event, event_type): event[DSProps.EVENT_TYPE] = event_type if ZabbixDriver.conf_map: zabbix_host = event[ZProps.HOST] event[ZProps.ZABBIX_RESOURCE_NAME] = zabbix_host v_resource = ZabbixDriver.conf_map[zabbix_host] event[ZProps.RESOURCE_NAME] = v_resource[ZProps.RESOURCE_NAME] event[ZProps.RESOURCE_TYPE] = v_resource[ZProps.RESOURCE_TYPE] return ZabbixDriver.make_pickleable([event], ZABBIX_DATASOURCE, DatasourceAction.UPDATE)[0] @staticmethod def get_event_types(): return ['zabbix.alarm.ok', 'zabbix.alarm.problem'] @staticmethod def should_delete_outdated_entities(): return True
35.75
78
0.657057
from collections import namedtuple from oslo_config import cfg from oslo_log import log from oslo_utils import importutils as utils from vitrage.common.constants import DatasourceAction from vitrage.common.constants import DatasourceOpts as DSOpts from vitrage.common.constants import DatasourceProperties as DSProps from vitrage.datasources.alarm_driver_base import AlarmDriverBase from vitrage.datasources.zabbix.properties import ZabbixProperties as ZProps from vitrage.datasources.zabbix.properties import ZabbixTriggerStatus \ as TriggerStatus from vitrage.datasources.zabbix.properties import ZabbixTriggerValue \ as TriggerValue from vitrage.datasources.zabbix import ZABBIX_DATASOURCE from vitrage.utils import file as file_utils CONF = cfg.CONF LOG = log.getLogger(__name__) class ZabbixDriver(AlarmDriverBase): ServiceKey = namedtuple('ServiceKey', ['hostname', 'triggerid']) conf_map = None def __init__(self): super(ZabbixDriver, self).__init__() if not ZabbixDriver.conf_map: ZabbixDriver.conf_map =\ ZabbixDriver._configuration_mapping() self._client = None def zabbix_client_login(self): if not CONF.zabbix.user: LOG.warning('Zabbix user is not defined') if not CONF.zabbix.password: LOG.warning('Zabbix password is not defined') if not CONF.zabbix.url: LOG.warning('Zabbix url is not defined') try: if not self._client: self._client = utils.import_object( 'pyzabbix.ZabbixAPI', CONF.zabbix.url) self._client.login( CONF.zabbix.user, CONF.zabbix.password) except Exception: LOG.exception('pyzabbix.ZabbixAPI error occurred.') self._client = None def _vitrage_type(self): return ZABBIX_DATASOURCE def _alarm_key(self, alarm): return self.ServiceKey(hostname=alarm[ZProps.RESOURCE_NAME], triggerid=alarm[ZProps.TRIGGER_ID]) def _get_alarms(self): self.zabbix_client_login() if not self._client: return [] alarms = [] valid_hosts = (host for host in self._client.host.get(output=[ZProps.HOST]) if host[ZProps.HOST] in ZabbixDriver.conf_map) for host in valid_hosts: self._get_triggers_per_host(host, alarms) return alarms def _get_triggers_per_host(self, host, alarms): host_id = host[ZProps.HOST_ID] triggers = self._client.trigger.get(hostids=host_id, expandDescription=True) triggers_rawtexts = self._get_triggers_rawtexts(host_id) for trigger in triggers: trigger[ZProps.ZABBIX_RESOURCE_NAME] = host[ZProps.HOST] trigger_id = trigger[ZProps.TRIGGER_ID] trigger[ZProps.RAWTEXT] = triggers_rawtexts[trigger_id] alarms.append(trigger) def _get_triggers_rawtexts(self, host_id): output = [ZProps.TRIGGER_ID, ZProps.DESCRIPTION] triggers = self._client.trigger.get(hostids=host_id, output=output) return {trigger[ZProps.TRIGGER_ID]: trigger[ZProps.DESCRIPTION] for trigger in triggers} def _enrich_alarms(self, alarms): for alarm in alarms: alarm[ZProps.VALUE] = self._get_value(alarm) zabbix_host = alarm[ZProps.ZABBIX_RESOURCE_NAME] vitrage_host = ZabbixDriver.conf_map[zabbix_host] alarm[ZProps.RESOURCE_TYPE] = vitrage_host[ZProps.RESOURCE_TYPE] alarm[ZProps.RESOURCE_NAME] = vitrage_host[ZProps.RESOURCE_NAME] def _is_erroneous(self, alarm): return alarm and \ alarm[ZProps.VALUE] == TriggerValue.PROBLEM def _status_changed(self, new_alarm, old_alarm): if not (new_alarm and old_alarm): return False if new_alarm[ZProps.VALUE] != old_alarm[ZProps.VALUE]: return True if new_alarm[ZProps.VALUE] == TriggerValue.PROBLEM: priority_changed = \ new_alarm[ZProps.PRIORITY] != old_alarm[ZProps.PRIORITY] description_changed = \ new_alarm[ZProps.DESCRIPTION] != old_alarm[ZProps.DESCRIPTION] return priority_changed or description_changed def _is_valid(self, alarm): return alarm[ZProps.RESOURCE_TYPE] is not None and \ alarm[ZProps.RESOURCE_NAME] is not None @staticmethod def _get_value(alarm): if alarm[ZProps.STATUS] == TriggerStatus.DISABLED: return TriggerValue.OK return alarm[ZProps.VALUE] @staticmethod def _configuration_mapping(): try: zabbix_config_file = CONF.zabbix[DSOpts.CONFIG_FILE] zabbix_config = file_utils.load_yaml_file(zabbix_config_file) zabbix_config_elements = zabbix_config[ZABBIX_DATASOURCE] mappings = {} for element_config in zabbix_config_elements: mappings[element_config['zabbix_host']] = { ZProps.RESOURCE_TYPE: element_config['type'], ZProps.RESOURCE_NAME: element_config['name'] } return mappings except Exception: LOG.exception('Failed in init.') return {} def enrich_event(self, event, event_type): event[DSProps.EVENT_TYPE] = event_type if ZabbixDriver.conf_map: zabbix_host = event[ZProps.HOST] event[ZProps.ZABBIX_RESOURCE_NAME] = zabbix_host v_resource = ZabbixDriver.conf_map[zabbix_host] event[ZProps.RESOURCE_NAME] = v_resource[ZProps.RESOURCE_NAME] event[ZProps.RESOURCE_TYPE] = v_resource[ZProps.RESOURCE_TYPE] return ZabbixDriver.make_pickleable([event], ZABBIX_DATASOURCE, DatasourceAction.UPDATE)[0] @staticmethod def get_event_types(): return ['zabbix.alarm.ok', 'zabbix.alarm.problem'] @staticmethod def should_delete_outdated_entities(): return True
true
true
f70ea3d37cea2ad796cf374b9a4ab9e73fc5c35c
1,057
py
Python
migrations/versions/5c80010c853a_posts_table.py
ChanForPres/Social-Blogging-App
525c4588dfa50780a8b7c067111c101ade4e20f2
[ "MIT" ]
null
null
null
migrations/versions/5c80010c853a_posts_table.py
ChanForPres/Social-Blogging-App
525c4588dfa50780a8b7c067111c101ade4e20f2
[ "MIT" ]
null
null
null
migrations/versions/5c80010c853a_posts_table.py
ChanForPres/Social-Blogging-App
525c4588dfa50780a8b7c067111c101ade4e20f2
[ "MIT" ]
null
null
null
"""posts table Revision ID: 5c80010c853a Revises: 6ca7139bbbf2 Create Date: 2018-06-25 17:18:29.165993 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = '5c80010c853a' down_revision = '6ca7139bbbf2' branch_labels = None depends_on = None def upgrade(): # ### commands auto generated by Alembic - please adjust! ### op.create_table('post', sa.Column('id', sa.Integer(), nullable=False), sa.Column('body', sa.String(length=140), nullable=True), sa.Column('timestamp', sa.DateTime(), nullable=True), sa.Column('user_id', sa.Integer(), nullable=True), sa.ForeignKeyConstraint(['user_id'], ['user.id'], ), sa.PrimaryKeyConstraint('id') ) op.create_index(op.f('ix_post_timestamp'), 'post', ['timestamp'], unique=False) # ### end Alembic commands ### def downgrade(): # ### commands auto generated by Alembic - please adjust! ### op.drop_index(op.f('ix_post_timestamp'), table_name='post') op.drop_table('post') # ### end Alembic commands ###
27.815789
83
0.676443
from alembic import op import sqlalchemy as sa revision = '5c80010c853a' down_revision = '6ca7139bbbf2' branch_labels = None depends_on = None def upgrade(): ), nullable=True), sa.Column('user_id', sa.Integer(), nullable=True), sa.ForeignKeyConstraint(['user_id'], ['user.id'], ), sa.PrimaryKeyConstraint('id') ) op.create_index(op.f('ix_post_timestamp'), 'post', ['timestamp'], unique=False)
true
true
f70ea4478d7916ef0eca7ea387f9da960ada87f9
27
py
Python
src/euler_python_package/euler_python/medium/p255.py
wilsonify/euler
5214b776175e6d76a7c6d8915d0e062d189d9b79
[ "MIT" ]
null
null
null
src/euler_python_package/euler_python/medium/p255.py
wilsonify/euler
5214b776175e6d76a7c6d8915d0e062d189d9b79
[ "MIT" ]
null
null
null
src/euler_python_package/euler_python/medium/p255.py
wilsonify/euler
5214b776175e6d76a7c6d8915d0e062d189d9b79
[ "MIT" ]
null
null
null
def problem255(): pass
9
17
0.62963
def problem255(): pass
true
true
f70ea613433f02ed837e0b8f877dedbe1826238c
3,471
py
Python
factory-ai-vision/EdgeSolution/modules/WebModule/backend/configs/api_router.py
piyushka17/azure-intelligent-edge-patterns
0d088899afb0022daa2ac434226824dba2c997c1
[ "MIT" ]
null
null
null
factory-ai-vision/EdgeSolution/modules/WebModule/backend/configs/api_router.py
piyushka17/azure-intelligent-edge-patterns
0d088899afb0022daa2ac434226824dba2c997c1
[ "MIT" ]
null
null
null
factory-ai-vision/EdgeSolution/modules/WebModule/backend/configs/api_router.py
piyushka17/azure-intelligent-edge-patterns
0d088899afb0022daa2ac434226824dba2c997c1
[ "MIT" ]
null
null
null
"""API router """ from django.conf.urls import url from django.urls import path from rest_framework.routers import DefaultRouter from vision_on_edge.azure_app_insight.api import views as app_insight_views from vision_on_edge.azure_parts.api import views as azure_part_views from vision_on_edge.azure_settings.api import views as azure_setting_views from vision_on_edge.azure_training.api import views as azure_training_views from vision_on_edge.azure_training_status.api import \ views as azure_training_status_views from vision_on_edge.cameras.api import util_views as camera_util_views from vision_on_edge.cameras.api import views from vision_on_edge.feedback.api import views as feedback_views from vision_on_edge.image_predictions.api import \ views as image_prediction_views from vision_on_edge.images.api import views as image_views from vision_on_edge.locations.api import views as location_views from vision_on_edge.notifications.api import views as notifications_views from vision_on_edge.relabeling.api import views as relabel_views from vision_on_edge.streams.api import views as stream_views router = DefaultRouter() router.trailing_slash = '/?' router.register('settings', azure_setting_views.SettingViewSet) router.register('cameras', views.CameraViewSet) router.register('parts', azure_part_views.PartViewSet) router.register('locations', location_views.LocationViewSet) router.register('image_predictions', image_prediction_views.ImagePredictionViewSet) router.register('projects', azure_training_views.ProjectViewSet) router.register('training_status', azure_training_status_views.TrainingStatusViewSet) router.register('tasks', azure_training_views.TaskViewSet) router.register('images', image_views.ImageViewSet) router.register('feedback', feedback_views.FeedbackViewSet) router.register('notifications', notifications_views.NotificationViewSet) router.register('images', image_views.ImageViewSet) urlpatterns = router.urls urlpatterns += [ url('streams/connect', stream_views.connect_stream), path('streams/<int:stream_id>/disconnect', stream_views.disconnect_stream), path('streams/<int:stream_id>/video_feed', stream_views.video_feed), path('streams/<int:stream_id>/capture', stream_views.capture), path('streams/<int:stream_id>/keep_alive', stream_views.keep_alive), path('projects/<int:project_id>/train', azure_training_views.train), path('projects/<int:project_id>/export', azure_training_views.export), path('projects/<int:project_id>/train_performance', azure_training_views.train_performance), path('projects/<int:project_id>/inference_video_feed', stream_views.inference_video_feed), path('projects/<int:project_id>/pull_cv_project', azure_training_views.pull_cv_project), path('projects/<int:project_id>/update_prob_threshold', azure_training_views.update_prob_threshold), path('projects/<int:project_id>/reset_project', azure_training_views.reset_project), path('projects/<int:project_id>/reset_camera', azure_training_views.project_reset_camera), path('projects/null/export', azure_training_views.export_null), path('relabel', relabel_views.upload_relabel_image), path('relabel/update', relabel_views.relabel_update), path('appinsight/key', app_insight_views.instrumentation_key), path('camera_utils/verify_rtsp', camera_util_views.verify_rtsp) ] app_name = "api"
47.547945
79
0.803803
from django.conf.urls import url from django.urls import path from rest_framework.routers import DefaultRouter from vision_on_edge.azure_app_insight.api import views as app_insight_views from vision_on_edge.azure_parts.api import views as azure_part_views from vision_on_edge.azure_settings.api import views as azure_setting_views from vision_on_edge.azure_training.api import views as azure_training_views from vision_on_edge.azure_training_status.api import \ views as azure_training_status_views from vision_on_edge.cameras.api import util_views as camera_util_views from vision_on_edge.cameras.api import views from vision_on_edge.feedback.api import views as feedback_views from vision_on_edge.image_predictions.api import \ views as image_prediction_views from vision_on_edge.images.api import views as image_views from vision_on_edge.locations.api import views as location_views from vision_on_edge.notifications.api import views as notifications_views from vision_on_edge.relabeling.api import views as relabel_views from vision_on_edge.streams.api import views as stream_views router = DefaultRouter() router.trailing_slash = '/?' router.register('settings', azure_setting_views.SettingViewSet) router.register('cameras', views.CameraViewSet) router.register('parts', azure_part_views.PartViewSet) router.register('locations', location_views.LocationViewSet) router.register('image_predictions', image_prediction_views.ImagePredictionViewSet) router.register('projects', azure_training_views.ProjectViewSet) router.register('training_status', azure_training_status_views.TrainingStatusViewSet) router.register('tasks', azure_training_views.TaskViewSet) router.register('images', image_views.ImageViewSet) router.register('feedback', feedback_views.FeedbackViewSet) router.register('notifications', notifications_views.NotificationViewSet) router.register('images', image_views.ImageViewSet) urlpatterns = router.urls urlpatterns += [ url('streams/connect', stream_views.connect_stream), path('streams/<int:stream_id>/disconnect', stream_views.disconnect_stream), path('streams/<int:stream_id>/video_feed', stream_views.video_feed), path('streams/<int:stream_id>/capture', stream_views.capture), path('streams/<int:stream_id>/keep_alive', stream_views.keep_alive), path('projects/<int:project_id>/train', azure_training_views.train), path('projects/<int:project_id>/export', azure_training_views.export), path('projects/<int:project_id>/train_performance', azure_training_views.train_performance), path('projects/<int:project_id>/inference_video_feed', stream_views.inference_video_feed), path('projects/<int:project_id>/pull_cv_project', azure_training_views.pull_cv_project), path('projects/<int:project_id>/update_prob_threshold', azure_training_views.update_prob_threshold), path('projects/<int:project_id>/reset_project', azure_training_views.reset_project), path('projects/<int:project_id>/reset_camera', azure_training_views.project_reset_camera), path('projects/null/export', azure_training_views.export_null), path('relabel', relabel_views.upload_relabel_image), path('relabel/update', relabel_views.relabel_update), path('appinsight/key', app_insight_views.instrumentation_key), path('camera_utils/verify_rtsp', camera_util_views.verify_rtsp) ] app_name = "api"
true
true
f70ea783739509cd8b2366a0314a6db0627abca1
6,569
py
Python
lib/sqlalchemy/orm/exc.py
petit87/sqlalchemy
67d674bd63ca36ac32b23f96e2b19e9dac6b0863
[ "MIT" ]
null
null
null
lib/sqlalchemy/orm/exc.py
petit87/sqlalchemy
67d674bd63ca36ac32b23f96e2b19e9dac6b0863
[ "MIT" ]
null
null
null
lib/sqlalchemy/orm/exc.py
petit87/sqlalchemy
67d674bd63ca36ac32b23f96e2b19e9dac6b0863
[ "MIT" ]
null
null
null
# orm/exc.py # Copyright (C) 2005-2022 the SQLAlchemy authors and contributors # <see AUTHORS file> # # This module is part of SQLAlchemy and is released under # the MIT License: https://www.opensource.org/licenses/mit-license.php """SQLAlchemy ORM exceptions.""" from __future__ import annotations from .. import exc as sa_exc from .. import util from ..exc import MultipleResultsFound # noqa from ..exc import NoResultFound # noqa NO_STATE = (AttributeError, KeyError) """Exception types that may be raised by instrumentation implementations.""" class StaleDataError(sa_exc.SQLAlchemyError): """An operation encountered database state that is unaccounted for. Conditions which cause this to happen include: * A flush may have attempted to update or delete rows and an unexpected number of rows were matched during the UPDATE or DELETE statement. Note that when version_id_col is used, rows in UPDATE or DELETE statements are also matched against the current known version identifier. * A mapped object with version_id_col was refreshed, and the version number coming back from the database does not match that of the object itself. * A object is detached from its parent object, however the object was previously attached to a different parent identity which was garbage collected, and a decision cannot be made if the new parent was really the most recent "parent". """ ConcurrentModificationError = StaleDataError class FlushError(sa_exc.SQLAlchemyError): """A invalid condition was detected during flush().""" class UnmappedError(sa_exc.InvalidRequestError): """Base for exceptions that involve expected mappings not present.""" class ObjectDereferencedError(sa_exc.SQLAlchemyError): """An operation cannot complete due to an object being garbage collected. """ class DetachedInstanceError(sa_exc.SQLAlchemyError): """An attempt to access unloaded attributes on a mapped instance that is detached.""" code = "bhk3" class UnmappedInstanceError(UnmappedError): """An mapping operation was requested for an unknown instance.""" @util.preload_module("sqlalchemy.orm.base") def __init__(self, obj, msg=None): base = util.preloaded.orm_base if not msg: try: base.class_mapper(type(obj)) name = _safe_cls_name(type(obj)) msg = ( "Class %r is mapped, but this instance lacks " "instrumentation. This occurs when the instance " "is created before sqlalchemy.orm.mapper(%s) " "was called." % (name, name) ) except UnmappedClassError: msg = _default_unmapped(type(obj)) if isinstance(obj, type): msg += ( "; was a class (%s) supplied where an instance was " "required?" % _safe_cls_name(obj) ) UnmappedError.__init__(self, msg) def __reduce__(self): return self.__class__, (None, self.args[0]) class UnmappedClassError(UnmappedError): """An mapping operation was requested for an unknown class.""" def __init__(self, cls, msg=None): if not msg: msg = _default_unmapped(cls) UnmappedError.__init__(self, msg) def __reduce__(self): return self.__class__, (None, self.args[0]) class ObjectDeletedError(sa_exc.InvalidRequestError): """A refresh operation failed to retrieve the database row corresponding to an object's known primary key identity. A refresh operation proceeds when an expired attribute is accessed on an object, or when :meth:`_query.Query.get` is used to retrieve an object which is, upon retrieval, detected as expired. A SELECT is emitted for the target row based on primary key; if no row is returned, this exception is raised. The true meaning of this exception is simply that no row exists for the primary key identifier associated with a persistent object. The row may have been deleted, or in some cases the primary key updated to a new value, outside of the ORM's management of the target object. """ @util.preload_module("sqlalchemy.orm.base") def __init__(self, state, msg=None): base = util.preloaded.orm_base if not msg: msg = ( "Instance '%s' has been deleted, or its " "row is otherwise not present." % base.state_str(state) ) sa_exc.InvalidRequestError.__init__(self, msg) def __reduce__(self): return self.__class__, (None, self.args[0]) class UnmappedColumnError(sa_exc.InvalidRequestError): """Mapping operation was requested on an unknown column.""" class LoaderStrategyException(sa_exc.InvalidRequestError): """A loader strategy for an attribute does not exist.""" def __init__( self, applied_to_property_type, requesting_property, applies_to, actual_strategy_type, strategy_key, ): if actual_strategy_type is None: sa_exc.InvalidRequestError.__init__( self, "Can't find strategy %s for %s" % (strategy_key, requesting_property), ) else: sa_exc.InvalidRequestError.__init__( self, 'Can\'t apply "%s" strategy to property "%s", ' 'which is a "%s"; this loader strategy is intended ' 'to be used with a "%s".' % ( util.clsname_as_plain_name(actual_strategy_type), requesting_property, util.clsname_as_plain_name(applied_to_property_type), util.clsname_as_plain_name(applies_to), ), ) def _safe_cls_name(cls): try: cls_name = ".".join((cls.__module__, cls.__name__)) except AttributeError: cls_name = getattr(cls, "__name__", None) if cls_name is None: cls_name = repr(cls) return cls_name @util.preload_module("sqlalchemy.orm.base") def _default_unmapped(cls): base = util.preloaded.orm_base try: mappers = base.manager_of_class(cls).mappers except (TypeError,) + NO_STATE: mappers = {} name = _safe_cls_name(cls) if not mappers: return "Class '%s' is not mapped" % name
31.581731
76
0.642716
from __future__ import annotations from .. import exc as sa_exc from .. import util from ..exc import MultipleResultsFound from ..exc import NoResultFound NO_STATE = (AttributeError, KeyError) class StaleDataError(sa_exc.SQLAlchemyError): ConcurrentModificationError = StaleDataError class FlushError(sa_exc.SQLAlchemyError): class UnmappedError(sa_exc.InvalidRequestError): class ObjectDereferencedError(sa_exc.SQLAlchemyError): class DetachedInstanceError(sa_exc.SQLAlchemyError): code = "bhk3" class UnmappedInstanceError(UnmappedError): @util.preload_module("sqlalchemy.orm.base") def __init__(self, obj, msg=None): base = util.preloaded.orm_base if not msg: try: base.class_mapper(type(obj)) name = _safe_cls_name(type(obj)) msg = ( "Class %r is mapped, but this instance lacks " "instrumentation. This occurs when the instance " "is created before sqlalchemy.orm.mapper(%s) " "was called." % (name, name) ) except UnmappedClassError: msg = _default_unmapped(type(obj)) if isinstance(obj, type): msg += ( "; was a class (%s) supplied where an instance was " "required?" % _safe_cls_name(obj) ) UnmappedError.__init__(self, msg) def __reduce__(self): return self.__class__, (None, self.args[0]) class UnmappedClassError(UnmappedError): def __init__(self, cls, msg=None): if not msg: msg = _default_unmapped(cls) UnmappedError.__init__(self, msg) def __reduce__(self): return self.__class__, (None, self.args[0]) class ObjectDeletedError(sa_exc.InvalidRequestError): @util.preload_module("sqlalchemy.orm.base") def __init__(self, state, msg=None): base = util.preloaded.orm_base if not msg: msg = ( "Instance '%s' has been deleted, or its " "row is otherwise not present." % base.state_str(state) ) sa_exc.InvalidRequestError.__init__(self, msg) def __reduce__(self): return self.__class__, (None, self.args[0]) class UnmappedColumnError(sa_exc.InvalidRequestError): class LoaderStrategyException(sa_exc.InvalidRequestError): def __init__( self, applied_to_property_type, requesting_property, applies_to, actual_strategy_type, strategy_key, ): if actual_strategy_type is None: sa_exc.InvalidRequestError.__init__( self, "Can't find strategy %s for %s" % (strategy_key, requesting_property), ) else: sa_exc.InvalidRequestError.__init__( self, 'Can\'t apply "%s" strategy to property "%s", ' 'which is a "%s"; this loader strategy is intended ' 'to be used with a "%s".' % ( util.clsname_as_plain_name(actual_strategy_type), requesting_property, util.clsname_as_plain_name(applied_to_property_type), util.clsname_as_plain_name(applies_to), ), ) def _safe_cls_name(cls): try: cls_name = ".".join((cls.__module__, cls.__name__)) except AttributeError: cls_name = getattr(cls, "__name__", None) if cls_name is None: cls_name = repr(cls) return cls_name @util.preload_module("sqlalchemy.orm.base") def _default_unmapped(cls): base = util.preloaded.orm_base try: mappers = base.manager_of_class(cls).mappers except (TypeError,) + NO_STATE: mappers = {} name = _safe_cls_name(cls) if not mappers: return "Class '%s' is not mapped" % name
true
true
f70ea910deb64c851f94887e577c45916aab7cf2
12,652
py
Python
doc/conf.py
gitter-badger/SoCo
65977466057748ea522a6d8b7f2a649091485a07
[ "MIT" ]
1
2019-03-09T14:23:48.000Z
2019-03-09T14:23:48.000Z
doc/conf.py
gitter-badger/SoCo
65977466057748ea522a6d8b7f2a649091485a07
[ "MIT" ]
null
null
null
doc/conf.py
gitter-badger/SoCo
65977466057748ea522a6d8b7f2a649091485a07
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # # soco documentation build configuration file, created by # sphinx-quickstart on Mon Sep 14 08:03:37 2015. # # This file is execfile()d with the current directory set to its # containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. import sys import os import shlex sys.path.insert(0, os.path.abspath('..')) import soco # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. # sys.path.insert(0, os.path.abspath('.')) # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. needs_sphinx = '1.3' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ 'sphinx.ext.autodoc', 'sphinx.ext.autosummary', 'sphinx.ext.extlinks', 'sphinx.ext.inheritance_diagram', 'sphinx.ext.intersphinx', 'sphinx.ext.todo', 'sphinx.ext.viewcode', 'sphinx.ext.napoleon', ] # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix(es) of source filenames. # You can specify multiple suffix as a list of string: # source_suffix = ['.rst', '.md'] source_suffix = '.rst' # The encoding of source files. # source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' # General information about the project. project = 'SoCo' copyright = '2015, The SoCo Team' author = "`The SoCo Team" # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = soco.__version__ # The full version, including alpha/beta/rc tags. release = soco.__version__ # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. language = 'en' # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: # today = '' # Else, today_fmt is used as the format for a strftime call. # today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = ['_build'] # The reST default role (used for this markup: `text`) to use for all # documents. default_role = 'any' # If true, '()' will be appended to :func: etc. cross-reference text. # add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). # add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. # show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. modindex_common_prefix = ['soco.', 'soco.music_services.'] # If true, keep warnings as "system message" paragraphs in the built documents. keep_warnings = True # If true, `todo` and `todoList` produce output, else they produce nothing. todo_include_todos = True # Allow auto links into the Python and Requests docs intersphinx_mapping = { 'python': ('https://docs.python.org/3', None), 'requests': ('http://www.python-requests.org/en/latest/', None) } # Shortcuts to Github Issues etc. Use them like this: # :issue:`123` (which will generate a link to issue 123) extlinks = { 'issue': ('https://github.com/SoCo/SoCo/issues/%s', '#'), 'PR': ('https://github.com/SoCo/SoCo/pull/%s', '#') } # Document members by default, and in source order. This allows the stub files # in the api directory to be much shorter. autodoc_default_flags = ['members'] autodoc_member_order = 'bysource' # Concatenate the class and __init__ docstrings autoclass_content = 'both' # Nicer inheritance graphs for RTD theme. NB the image map does not rescale # properly, so we have had to add some javascript to handle it. See # _templates and _static inheritance_node_attrs = dict( fontsize=14, height=0.75, color='dodgerblue', style='rounded', ) inheritance_graph_attrs = dict( rankdir="LR", size='""', ) # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'sphinx_rtd_theme' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. # html_theme_options = {} # Add any paths that contain custom themes here, relative to this directory. # html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". # html_title = None # A shorter title for the navigation bar. Default is the same as html_title. # html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. # html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. # html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. # html_extra_path = [] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. # html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. # html_use_smartypants = True # Custom sidebar templates, maps document names to template names. # html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. # html_additional_pages = {} # If false, no module index is generated. # html_domain_indices = True # If false, no index is generated. # html_use_index = True # If true, the index is split into individual pages for each letter. # html_split_index = False # If true, links to the reST sources are added to the pages. # html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. # html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. # html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. # html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). # html_file_suffix = None # Language to be used for generating the HTML full-text search index. # Sphinx supports the following languages: # 'da', 'de', 'en', 'es', 'fi', 'fr', 'h', 'it', 'ja' # 'nl', 'no', 'pt', 'ro', 'r', 'sv', 'tr' # html_search_language = 'en' # A dictionary with options for the search language support, empty by default. # Now only 'ja' uses this config value # html_search_options = {'type': 'default'} # The name of a javascript file (relative to the configuration directory) that # implements a search results scorer. If empty, the default will be used. # html_search_scorer = 'scorer.js' # Output file base name for HTML help builder. htmlhelp_basename = 'socodoc' # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). # 'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). # 'pointsize': '10pt', # Additional stuff for the LaTeX preamble. # 'preamble': '', # Latex figure (float) alignment # 'figure_align': 'htbp', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ (master_doc, 'soco.tex', 'soco Documentation', 'Author', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. # latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. # latex_use_parts = False # If true, show page references after internal links. # latex_show_pagerefs = False # If true, show URL addresses after external links. # latex_show_urls = False # Documents to append as an appendix to all manuals. # latex_appendices = [] # If false, no module index is generated. # latex_domain_indices = True # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ (master_doc, 'soco', 'soco Documentation', [author], 1) ] # If true, show URL addresses after external links. # man_show_urls = False # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ (master_doc, 'soco', 'soco Documentation', author, 'soco', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. # texinfo_appendices = [] # If false, no module index is generated. # texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. # texinfo_show_urls = 'footnote' # If true, do not generate a @detailmenu in the "Top" node's menu. # texinfo_no_detailmenu = False # -- Options for Epub output ---------------------------------------------- # Bibliographic Dublin Core info. epub_title = project epub_author = author epub_publisher = author epub_copyright = copyright # The basename for the epub file. It defaults to the project name. # epub_basename = project # The HTML theme for the epub output. Since the default themes are not # optimized for small screen space, using the same theme for HTML and epub # output is usually not wise. This defaults to 'epub', a theme designed to # save visual space. # epub_theme = 'epub' # The language of the text. It defaults to the language option # or 'en' if the language is not set. # epub_language = '' # The scheme of the identifier. Typical schemes are ISBN or URL. # epub_scheme = '' # The unique identifier of the text. This can be a ISBN number # or the project homepage. # epub_identifier = '' # A unique identification for the text. # epub_uid = '' # A tuple containing the cover image and cover page html template filenames. # epub_cover = () # A sequence of (type, uri, title) tuples for the guide element of content.opf. # epub_guide = () # HTML files that should be inserted before the pages created by sphinx. # The format is a list of tuples containing the path and title. # epub_pre_files = [] # HTML files shat should be inserted after the pages created by sphinx. # The format is a list of tuples containing the path and title. # epub_post_files = [] # A list of files that should not be packed into the epub file. epub_exclude_files = ['search.html'] # The depth of the table of contents in toc.ncx. # epub_tocdepth = 3 # Allow duplicate toc entries. # epub_tocdup = True # Choose between 'default' and 'includehidden'. # epub_tocscope = 'default' # Fix unsupported image types using the Pillow. # epub_fix_images = False # Scale large images. # epub_max_image_width = 0 # How to display URL addresses: 'footnote', 'no', or 'inline'. # epub_show_urls = 'inline' # If false, no index is generated. # epub_use_index = True
31.788945
79
0.710639
import sys import os import shlex sys.path.insert(0, os.path.abspath('..')) import soco needs_sphinx = '1.3' extensions = [ 'sphinx.ext.autodoc', 'sphinx.ext.autosummary', 'sphinx.ext.extlinks', 'sphinx.ext.inheritance_diagram', 'sphinx.ext.intersphinx', 'sphinx.ext.todo', 'sphinx.ext.viewcode', 'sphinx.ext.napoleon', ] templates_path = ['_templates'] source_suffix = '.rst' master_doc = 'index' project = 'SoCo' copyright = '2015, The SoCo Team' author = "`The SoCo Team" # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = soco.__version__ # The full version, including alpha/beta/rc tags. release = soco.__version__ # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. # # This is also used if you do content translation via gettext catalogs. # Usually you set "language" from the command line for these cases. language = 'en' # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: # today = '' # Else, today_fmt is used as the format for a strftime call. # today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = ['_build'] # The reST default role (used for this markup: `text`) to use for all # documents. default_role = 'any' # If true, '()' will be appended to :func: etc. cross-reference text. # add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). # add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. # show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. modindex_common_prefix = ['soco.', 'soco.music_services.'] # If true, keep warnings as "system message" paragraphs in the built documents. keep_warnings = True # If true, `todo` and `todoList` produce output, else they produce nothing. todo_include_todos = True # Allow auto links into the Python and Requests docs intersphinx_mapping = { 'python': ('https://docs.python.org/3', None), 'requests': ('http://www.python-requests.org/en/latest/', None) } # Shortcuts to Github Issues etc. Use them like this: # :issue:`123` (which will generate a link to issue 123) extlinks = { 'issue': ('https://github.com/SoCo/SoCo/issues/%s', ' 'PR': ('https://github.com/SoCo/SoCo/pull/%s', ' } # Document members by default, and in source order. This allows the stub files # in the api directory to be much shorter. autodoc_default_flags = ['members'] autodoc_member_order = 'bysource' # Concatenate the class and __init__ docstrings autoclass_content = 'both' # Nicer inheritance graphs for RTD theme. NB the image map does not rescale # properly, so we have had to add some javascript to handle it. See # _templates and _static inheritance_node_attrs = dict( fontsize=14, height=0.75, color='dodgerblue', style='rounded', ) inheritance_graph_attrs = dict( rankdir="LR", size='""', ) # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'sphinx_rtd_theme' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. # html_theme_options = {} # Add any paths that contain custom themes here, relative to this directory. # html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". # html_title = None # A shorter title for the navigation bar. Default is the same as html_title. # html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. # html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. # html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. # html_extra_path = [] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. # html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. # html_use_smartypants = True # Custom sidebar templates, maps document names to template names. # html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. # html_additional_pages = {} # If false, no module index is generated. # html_domain_indices = True # If false, no index is generated. # html_use_index = True # If true, the index is split into individual pages for each letter. # html_split_index = False # If true, links to the reST sources are added to the pages. # html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. # html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. # html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. # html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). # html_file_suffix = None # Language to be used for generating the HTML full-text search index. # Sphinx supports the following languages: # 'da', 'de', 'en', 'es', 'fi', 'fr', 'h', 'it', 'ja' # 'nl', 'no', 'pt', 'ro', 'r', 'sv', 'tr' # html_search_language = 'en' # A dictionary with options for the search language support, empty by default. # Now only 'ja' uses this config value # html_search_options = {'type': 'default'} # The name of a javascript file (relative to the configuration directory) that # implements a search results scorer. If empty, the default will be used. # html_search_scorer = 'scorer.js' # Output file base name for HTML help builder. htmlhelp_basename = 'socodoc' # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). # 'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). # 'pointsize': '10pt', # Additional stuff for the LaTeX preamble. # 'preamble': '', # Latex figure (float) alignment # 'figure_align': 'htbp', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ (master_doc, 'soco.tex', 'soco Documentation', 'Author', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. # latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. # latex_use_parts = False # If true, show page references after internal links. # latex_show_pagerefs = False # If true, show URL addresses after external links. # latex_show_urls = False # Documents to append as an appendix to all manuals. # latex_appendices = [] # If false, no module index is generated. # latex_domain_indices = True # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ (master_doc, 'soco', 'soco Documentation', [author], 1) ] # If true, show URL addresses after external links. # man_show_urls = False # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ (master_doc, 'soco', 'soco Documentation', author, 'soco', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. # texinfo_appendices = [] # If false, no module index is generated. # texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. # texinfo_show_urls = 'footnote' # If true, do not generate a @detailmenu in the "Top" node's menu. epub_title = project epub_author = author epub_publisher = author epub_copyright = copyright epub_exclude_files = ['search.html']
true
true
f70eaaf12a9fe00936523a8b54e0345d5ae2821e
768
py
Python
src/deployable/__main__.py
cpuabuse/py-deployment-automation
aea0c48ac4c5a81f2e027c984ab65f911ad29d0d
[ "0BSD" ]
1
2020-02-23T22:35:28.000Z
2020-02-23T22:35:28.000Z
src/deployable/__main__.py
cpuabuse/py-deployment-automation
aea0c48ac4c5a81f2e027c984ab65f911ad29d0d
[ "0BSD" ]
null
null
null
src/deployable/__main__.py
cpuabuse/py-deployment-automation
aea0c48ac4c5a81f2e027c984ab65f911ad29d0d
[ "0BSD" ]
null
null
null
#!/usr/bin/env python3 """ Main module for the deployable project. """ # Bootstrap to be able to perform absolute imports as standalone code if __name__ == "__main__": from absolute_import import absolute_import absolute_import(file=__file__, name=__name__, path=__path__) # Normal imports from argparse import ArgumentParser, RawDescriptionHelpFormatter from deployable.defaults.args import description, epilog from typing import Any, Tuple def get_args() -> Tuple[Any]: """ Retrieves arguments from command line. """ # Create parser and groups parser = ArgumentParser(description=description, epilog=epilog, formatter_class=RawDescriptionHelpFormatter) def main() -> None: """ Entrypoint. """ # Call main method if __name__ == "__main__": main()
21.942857
109
0.763021
if __name__ == "__main__": from absolute_import import absolute_import absolute_import(file=__file__, name=__name__, path=__path__) from argparse import ArgumentParser, RawDescriptionHelpFormatter from deployable.defaults.args import description, epilog from typing import Any, Tuple def get_args() -> Tuple[Any]: parser = ArgumentParser(description=description, epilog=epilog, formatter_class=RawDescriptionHelpFormatter) def main() -> None: if __name__ == "__main__": main()
true
true
f70eab155736800811a7f28862b37c609211be8a
4,889
py
Python
src/cartpole.py
rish-16/gym-navmaze
cc21d730ec6ab1e96a4a1a8f602a5bbb951d2929
[ "MIT" ]
1
2021-06-10T13:40:09.000Z
2021-06-10T13:40:09.000Z
src/cartpole.py
rish-16/gym-navmaze
cc21d730ec6ab1e96a4a1a8f602a5bbb951d2929
[ "MIT" ]
null
null
null
src/cartpole.py
rish-16/gym-navmaze
cc21d730ec6ab1e96a4a1a8f602a5bbb951d2929
[ "MIT" ]
1
2021-06-10T13:40:11.000Z
2021-06-10T13:40:11.000Z
import numpy as np from collections import deque import pickle import torch from utils import collect_trajectories, random_sample from PPO import PPO import matplotlib.pyplot as plt from parallelEnv import * import gym env = gym.make("CartPole-v0") env.reset() env.seed(2) obs_dim = env.observation_space.shape[0] n_actions = env.action_space.n act_dist = [0 for i in range(n_actions)] def train(episode, env_name): gamma = .99 gae_lambda = 0.95 use_gae = True beta = .01 cliprange = 0.1 best_score = -np.inf goal_score = 195.0 ep_length = [] nenvs = 1 rollout_length = 200 minibatches = 10*8 nbatch = nenvs * rollout_length optimization_epochs = 4 device=torch.device("cuda:0" if torch.cuda.is_available() else "cpu") envs = parallelEnv(env_name, nenvs, seed=1234) agent = PPO(state_size=obs_dim, action_size=n_actions, seed=0, hidden_layers=[64,64], lr_policy=1e-4, use_reset=True, device=device) print(agent.policy) # keep track of progress mean_rewards = [] scores_window = deque(maxlen=100) loss_storage = [] for i_episode in range(episode+1): log_probs_old, states, actions, rewards, values, dones, vals_last, infos, ep_length = collect_trajectories(envs, act_dist, ep_length, agent.policy, rollout_length) returns = np.zeros_like(rewards) advantages = np.zeros_like(rewards) if not use_gae: for t in reversed(range(rollout_length)): if t == rollout_length - 1: returns[t] = rewards[t] + gamma * (1-dones[t]) * vals_last else: returns[t] = rewards[t] + gamma * (1-dones[t]) * returns[t+1] advantages[t] = returns[t] - values[t] else: for t in reversed(range(rollout_length)): if t == rollout_length - 1: returns[t] = rewards[t] + gamma * (1-dones[t]) * vals_last td_error = returns[t] - values[t] else: returns[t] = rewards[t] + gamma * (1-dones[t]) * returns[t+1] td_error = rewards[t] + gamma * (1-dones[t]) * values[t+1] - values[t] advantages[t] = advantages[t] * gae_lambda * gamma * (1-dones[t]) + td_error # convert to pytorch tensors and move to gpu if available returns = torch.from_numpy(returns).float().to(device).view(-1,) advantages = torch.from_numpy(advantages).float().to(device).view(-1,) advantages = (advantages - advantages.mean()) / (advantages.std() + 1e-10) for _ in range(optimization_epochs): sampler = random_sample(nbatch, minibatches) for inds in sampler: mb_log_probs_old = log_probs_old[inds] mb_states = states[inds] mb_actions = actions[inds] mb_returns = returns[inds] mb_advantages = advantages[inds] loss_p, loss_v, loss_ent = agent.update(mb_log_probs_old, mb_states, mb_actions, mb_returns, mb_advantages, cliprange=cliprange, beta=beta) loss_storage.append([loss_p, loss_v, loss_ent]) total_rewards = np.sum(rewards, axis=0) scores_window.append(np.mean(total_rewards)) # last 100 scores mean_rewards.append(np.mean(total_rewards)) # get the average reward of the parallel environments cliprange *= 0.999 # the clipping parameter reduces as time goes on beta *= 0.999 # the regulation term reduces if i_episode % 100 == 0: print('\rEpisode {}\tAverage Score: {:.2f}'.format(i_episode, np.mean(scores_window))) print(total_rewards) if np.mean(scores_window)>=goal_score and np.mean(scores_window)>=best_score: torch.save(agent.policy.state_dict(), "policy_cartpole.pth") best_score = np.mean(scores_window) return mean_rewards, loss_storage, act_dist, ep_length mean_rewards, loss, new_act_dist, ep_length = train(10000, 'CartPole-v0') print (new_act_dist[-1]) print (ep_length) plt.rcParams['xtick.direction'] = 'in' plt.rcParams['ytick.direction'] = 'in' plt.rcParams['font.size'] = 10 plt.title("PPO + MLP + GAE for 10000 episodes") plt.subplot(131) plt.plot(mean_rewards) plt.ylabel('Average score') plt.xlabel('Episode') plt.subplot(132) plt.plot(list(range(len(ep_length))), ep_length, color="red") plt.ylabel('Episode Length') plt.xlabel('Episode') plt.subplot(133) plt.ylabel('Frequency') plt.xlabel('Actions') plt.bar(['Action {}'.format(i) for i in range(len(new_act_dist))], new_act_dist[-1]) plt.show()
36.759398
171
0.606668
import numpy as np from collections import deque import pickle import torch from utils import collect_trajectories, random_sample from PPO import PPO import matplotlib.pyplot as plt from parallelEnv import * import gym env = gym.make("CartPole-v0") env.reset() env.seed(2) obs_dim = env.observation_space.shape[0] n_actions = env.action_space.n act_dist = [0 for i in range(n_actions)] def train(episode, env_name): gamma = .99 gae_lambda = 0.95 use_gae = True beta = .01 cliprange = 0.1 best_score = -np.inf goal_score = 195.0 ep_length = [] nenvs = 1 rollout_length = 200 minibatches = 10*8 nbatch = nenvs * rollout_length optimization_epochs = 4 device=torch.device("cuda:0" if torch.cuda.is_available() else "cpu") envs = parallelEnv(env_name, nenvs, seed=1234) agent = PPO(state_size=obs_dim, action_size=n_actions, seed=0, hidden_layers=[64,64], lr_policy=1e-4, use_reset=True, device=device) print(agent.policy) mean_rewards = [] scores_window = deque(maxlen=100) loss_storage = [] for i_episode in range(episode+1): log_probs_old, states, actions, rewards, values, dones, vals_last, infos, ep_length = collect_trajectories(envs, act_dist, ep_length, agent.policy, rollout_length) returns = np.zeros_like(rewards) advantages = np.zeros_like(rewards) if not use_gae: for t in reversed(range(rollout_length)): if t == rollout_length - 1: returns[t] = rewards[t] + gamma * (1-dones[t]) * vals_last else: returns[t] = rewards[t] + gamma * (1-dones[t]) * returns[t+1] advantages[t] = returns[t] - values[t] else: for t in reversed(range(rollout_length)): if t == rollout_length - 1: returns[t] = rewards[t] + gamma * (1-dones[t]) * vals_last td_error = returns[t] - values[t] else: returns[t] = rewards[t] + gamma * (1-dones[t]) * returns[t+1] td_error = rewards[t] + gamma * (1-dones[t]) * values[t+1] - values[t] advantages[t] = advantages[t] * gae_lambda * gamma * (1-dones[t]) + td_error returns = torch.from_numpy(returns).float().to(device).view(-1,) advantages = torch.from_numpy(advantages).float().to(device).view(-1,) advantages = (advantages - advantages.mean()) / (advantages.std() + 1e-10) for _ in range(optimization_epochs): sampler = random_sample(nbatch, minibatches) for inds in sampler: mb_log_probs_old = log_probs_old[inds] mb_states = states[inds] mb_actions = actions[inds] mb_returns = returns[inds] mb_advantages = advantages[inds] loss_p, loss_v, loss_ent = agent.update(mb_log_probs_old, mb_states, mb_actions, mb_returns, mb_advantages, cliprange=cliprange, beta=beta) loss_storage.append([loss_p, loss_v, loss_ent]) total_rewards = np.sum(rewards, axis=0) scores_window.append(np.mean(total_rewards)) mean_rewards.append(np.mean(total_rewards)) cliprange *= 0.999 beta *= 0.999 if i_episode % 100 == 0: print('\rEpisode {}\tAverage Score: {:.2f}'.format(i_episode, np.mean(scores_window))) print(total_rewards) if np.mean(scores_window)>=goal_score and np.mean(scores_window)>=best_score: torch.save(agent.policy.state_dict(), "policy_cartpole.pth") best_score = np.mean(scores_window) return mean_rewards, loss_storage, act_dist, ep_length mean_rewards, loss, new_act_dist, ep_length = train(10000, 'CartPole-v0') print (new_act_dist[-1]) print (ep_length) plt.rcParams['xtick.direction'] = 'in' plt.rcParams['ytick.direction'] = 'in' plt.rcParams['font.size'] = 10 plt.title("PPO + MLP + GAE for 10000 episodes") plt.subplot(131) plt.plot(mean_rewards) plt.ylabel('Average score') plt.xlabel('Episode') plt.subplot(132) plt.plot(list(range(len(ep_length))), ep_length, color="red") plt.ylabel('Episode Length') plt.xlabel('Episode') plt.subplot(133) plt.ylabel('Frequency') plt.xlabel('Actions') plt.bar(['Action {}'.format(i) for i in range(len(new_act_dist))], new_act_dist[-1]) plt.show()
true
true
f70eab2e8fd598cbdc5417a99d3c40fb7b2434e9
643
py
Python
lofo/infer_defaults.py
williamberrios/lofo-importance
34967cf47dc1c2797d3a77f8926918ae91e4197a
[ "MIT" ]
1
2020-10-31T10:05:45.000Z
2020-10-31T10:05:45.000Z
lofo/infer_defaults.py
williamberrios/lofo-importance
34967cf47dc1c2797d3a77f8926918ae91e4197a
[ "MIT" ]
null
null
null
lofo/infer_defaults.py
williamberrios/lofo-importance
34967cf47dc1c2797d3a77f8926918ae91e4197a
[ "MIT" ]
null
null
null
import numpy as np from sklearn.preprocessing import LabelEncoder from lightgbm import LGBMClassifier, LGBMRegressor def infer_model(df, features, y, n_jobs): model_class = LGBMRegressor if len(np.unique(y)) == 2: y = LabelEncoder().fit_transform(y) model_class = LGBMClassifier categoricals = df[features].select_dtypes(exclude=[np.number]).columns.tolist() for f in categoricals: df[f] = LabelEncoder().fit_transform(df[f].apply(str)) min_child_samples = int(0.01*df.shape[0]) model = model_class(min_child_samples=min_child_samples, n_jobs=n_jobs) return model, df, categoricals, y
30.619048
83
0.720062
import numpy as np from sklearn.preprocessing import LabelEncoder from lightgbm import LGBMClassifier, LGBMRegressor def infer_model(df, features, y, n_jobs): model_class = LGBMRegressor if len(np.unique(y)) == 2: y = LabelEncoder().fit_transform(y) model_class = LGBMClassifier categoricals = df[features].select_dtypes(exclude=[np.number]).columns.tolist() for f in categoricals: df[f] = LabelEncoder().fit_transform(df[f].apply(str)) min_child_samples = int(0.01*df.shape[0]) model = model_class(min_child_samples=min_child_samples, n_jobs=n_jobs) return model, df, categoricals, y
true
true
f70eab61632b9f861e510d3c8482608efe5ebe97
449,278
py
Python
rnn_cell_impl.py
narutowang/indrnn
434e1200b5e742a0eac92bed661c69e97b8b8711
[ "Apache-2.0" ]
null
null
null
rnn_cell_impl.py
narutowang/indrnn
434e1200b5e742a0eac92bed661c69e97b8b8711
[ "Apache-2.0" ]
null
null
null
rnn_cell_impl.py
narutowang/indrnn
434e1200b5e742a0eac92bed661c69e97b8b8711
[ "Apache-2.0" ]
null
null
null
<!DOCTYPE html> <html lang="en"> <head> <meta charset="utf-8"> <link rel="dns-prefetch" href="https://assets-cdn.github.com"> <link rel="dns-prefetch" href="https://avatars0.githubusercontent.com"> <link rel="dns-prefetch" href="https://avatars1.githubusercontent.com"> <link rel="dns-prefetch" href="https://avatars2.githubusercontent.com"> <link rel="dns-prefetch" href="https://avatars3.githubusercontent.com"> <link rel="dns-prefetch" href="https://github-cloud.s3.amazonaws.com"> <link rel="dns-prefetch" href="https://user-images.githubusercontent.com/"> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://assets-cdn.github.com/assets/frameworks-0b6eee1f89d4460d83bdbee0a4cb0020.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://assets-cdn.github.com/assets/github-c4d5af25254cfc8bb30ae5cb6f074c97.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://assets-cdn.github.com/assets/site-83dc1f7ebc9c7461fe1eab799b56c4c4.css" /> <meta name="viewport" content="width=device-width"> <title>tensorflow/rnn_cell_impl.py at r1.6 · tensorflow/tensorflow · GitHub</title> <meta name="description" content="GitHub is where people build software. More than 27 million people use GitHub to discover, fork, and contribute to over 80 million projects."> <link rel="search" type="application/opensearchdescription+xml" href="/opensearch.xml" title="GitHub"> <link rel="fluid-icon" href="https://github.com/fluidicon.png" title="GitHub"> <meta property="fb:app_id" content="1401488693436528"> <meta property="og:image" content="https://avatars2.githubusercontent.com/u/15658638?s=400&amp;v=4" /><meta property="og:site_name" content="GitHub" /><meta property="og:type" content="object" /><meta property="og:title" content="tensorflow/tensorflow" /><meta property="og:url" content="https://github.com/tensorflow/tensorflow" /><meta property="og:description" content="tensorflow - Computation using data flow graphs for scalable machine learning" /> <link rel="assets" href="https://assets-cdn.github.com/"> <meta name="pjax-timeout" content="1000"> <meta name="request-id" content="D9CE:22519:FDC93F:17A0312:5AB1D278" data-pjax-transient> <meta name="selected-link" value="repo_source" data-pjax-transient> <meta name="google-site-verification" content="KT5gs8h0wvaagLKAVWq8bbeNwnZZK1r1XQysX3xurLU"> <meta name="google-site-verification" content="ZzhVyEFwb7w3e0-uOTltm8Jsck2F5StVihD0exw2fsA"> <meta name="google-site-verification" content="GXs5KoUUkNCoaAZn7wPN-t01Pywp9M3sEjnt_3_ZWPc"> <meta name="google-analytics" content="UA-3769691-2"> <meta name="octolytics-host" content="collector.githubapp.com" /><meta name="octolytics-app-id" content="github" /><meta name="octolytics-event-url" content="https://collector.githubapp.com/github-external/browser_event" /><meta name="octolytics-dimension-request_id" content="D9CE:22519:FDC93F:17A0312:5AB1D278" /><meta name="octolytics-dimension-region_edge" content="sea" /><meta name="octolytics-dimension-region_render" content="iad" /> <meta name="hydro-events-url" content="https://github.com/hydro_browser_events" /> <meta name="analytics-location" content="/&lt;user-name&gt;/&lt;repo-name&gt;/blob/show" data-pjax-transient="true" /> <meta class="js-ga-set" name="dimension1" content="Logged Out"> <meta name="hostname" content="github.com"> <meta name="user-login" content=""> <meta name="expected-hostname" content="github.com"> <meta name="js-proxy-site-detection-payload" content="YTI3MjdkNDE4ODBjNzIwYjgyNTVlMzFjZTEwODkzZDRmYTJkNTZhMTdhZGZmNGM0ODc3MTZmNTg3MzBjNmM2Nnx7InJlbW90ZV9hZGRyZXNzIjoiMTY3LjIyMC4yMzIuMTI0IiwicmVxdWVzdF9pZCI6IkQ5Q0U6MjI1MTk6RkRDOTNGOjE3QTAzMTI6NUFCMUQyNzgiLCJ0aW1lc3RhbXAiOjE1MjE2MDMxOTMsImhvc3QiOiJnaXRodWIuY29tIn0="> <meta name="enabled-features" content="UNIVERSE_BANNER,FREE_TRIALS,MARKETPLACE_INSIGHTS,MARKETPLACE_INSIGHTS_CONVERSION_PERCENTAGES"> <meta name="html-safe-nonce" content="54cbe43f921a89e1942c69748da62cb0fab66263"> <meta http-equiv="x-pjax-version" content="f5d037ad7a1b600b21fb2f3474bd0c48"> <link href="https://github.com/tensorflow/tensorflow/commits/r1.6.atom" rel="alternate" title="Recent Commits to tensorflow:r1.6" type="application/atom+xml"> <meta name="description" content="tensorflow - Computation using data flow graphs for scalable machine learning"> <meta name="go-import" content="github.com/tensorflow/tensorflow git https://github.com/tensorflow/tensorflow.git"> <meta name="octolytics-dimension-user_id" content="15658638" /><meta name="octolytics-dimension-user_login" content="tensorflow" /><meta name="octolytics-dimension-repository_id" content="45717250" /><meta name="octolytics-dimension-repository_nwo" content="tensorflow/tensorflow" /><meta name="octolytics-dimension-repository_public" content="true" /><meta name="octolytics-dimension-repository_is_fork" content="false" /><meta name="octolytics-dimension-repository_network_root_id" content="45717250" /><meta name="octolytics-dimension-repository_network_root_nwo" content="tensorflow/tensorflow" /><meta name="octolytics-dimension-repository_explore_github_marketplace_ci_cta_shown" content="false" /> <link rel="canonical" href="https://github.com/tensorflow/tensorflow/blob/r1.6/tensorflow/python/ops/rnn_cell_impl.py" data-pjax-transient> <meta name="browser-stats-url" content="https://api.github.com/_private/browser/stats"> <meta name="browser-errors-url" content="https://api.github.com/_private/browser/errors"> <link rel="mask-icon" href="https://assets-cdn.github.com/pinned-octocat.svg" color="#000000"> <link rel="icon" type="image/x-icon" class="js-site-favicon" href="https://assets-cdn.github.com/favicon.ico"> <meta name="theme-color" content="#1e2327"> <link rel="manifest" href="/manifest.json" crossOrigin="use-credentials"> </head> <body class="logged-out env-production page-blob"> <div class="position-relative js-header-wrapper "> <a href="#start-of-content" tabindex="1" class="px-2 py-4 show-on-focus js-skip-to-content">Skip to content</a> <div id="js-pjax-loader-bar" class="pjax-loader-bar"><div class="progress"></div></div> <header class="Header header-logged-out position-relative f4 py-3" role="banner"> <div class="container-lg d-flex px-3"> <div class="d-flex flex-justify-between flex-items-center"> <a class="header-logo-invertocat my-0" href="https://github.com/" aria-label="Homepage" data-ga-click="(Logged out) Header, go to homepage, icon:logo-wordmark"> <svg height="32" class="octicon octicon-mark-github" viewBox="0 0 16 16" version="1.1" width="32" aria-hidden="true"><path fill-rule="evenodd" d="M8 0C3.58 0 0 3.58 0 8c0 3.54 2.29 6.53 5.47 7.59.4.07.55-.17.55-.38 0-.19-.01-.82-.01-1.49-2.01.37-2.53-.49-2.69-.94-.09-.23-.48-.94-.82-1.13-.28-.15-.68-.52-.01-.53.63-.01 1.08.58 1.23.82.72 1.21 1.87.87 2.33.66.07-.52.28-.87.51-1.07-1.78-.2-3.64-.89-3.64-3.95 0-.87.31-1.59.82-2.15-.08-.2-.36-1.02.08-2.12 0 0 .67-.21 2.2.82.64-.18 1.32-.27 2-.27.68 0 1.36.09 2 .27 1.53-1.04 2.2-.82 2.2-.82.44 1.1.16 1.92.08 2.12.51.56.82 1.27.82 2.15 0 3.07-1.87 3.75-3.65 3.95.29.25.54.73.54 1.48 0 1.07-.01 1.93-.01 2.2 0 .21.15.46.55.38A8.013 8.013 0 0 0 16 8c0-4.42-3.58-8-8-8z"/></svg> </a> </div> <div class="HeaderMenu HeaderMenu--bright d-flex flex-justify-between flex-auto"> <nav class="mt-0"> <ul class="d-flex list-style-none"> <li class="ml-2"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:features" data-selected-links="/features /features/project-management /features/code-review /features/project-management /features/integrations /features" href="/features"> Features </a> </li> <li class="ml-4"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:business" data-selected-links="/business /business/security /business/customers /business" href="/business"> Business </a> </li> <li class="ml-4"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:explore" data-selected-links="/explore /trending /trending/developers /integrations /integrations/feature/code /integrations/feature/collaborate /integrations/feature/ship showcases showcases_search showcases_landing /explore" href="/explore"> Explore </a> </li> <li class="ml-4"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:marketplace" data-selected-links=" /marketplace" href="/marketplace"> Marketplace </a> </li> <li class="ml-4"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:pricing" data-selected-links="/pricing /pricing/developer /pricing/team /pricing/business-hosted /pricing/business-enterprise /pricing" href="/pricing"> Pricing </a> </li> </ul> </nav> <div class="d-flex"> <div class="d-lg-flex flex-items-center mr-3"> <div class="header-search scoped-search site-scoped-search js-site-search" role="search"> <!-- '"` --><!-- </textarea></xmp> --></option></form><form class="js-site-search-form" data-scoped-search-url="/tensorflow/tensorflow/search" data-unscoped-search-url="/search" action="/tensorflow/tensorflow/search" accept-charset="UTF-8" method="get"><input name="utf8" type="hidden" value="&#x2713;" /> <label class="form-control header-search-wrapper js-chromeless-input-container"> <a class="header-search-scope no-underline" href="/tensorflow/tensorflow/blob/r1.6/tensorflow/python/ops/rnn_cell_impl.py">This repository</a> <input type="text" class="form-control header-search-input js-site-search-focus js-site-search-field is-clearable" data-hotkey="s,/" name="q" value="" placeholder="Search" aria-label="Search this repository" data-unscoped-placeholder="Search GitHub" data-scoped-placeholder="Search" autocapitalize="off"> <input type="hidden" class="js-site-search-type-field" name="type" > </label> </form></div> </div> <span class="d-inline-block"> <div class="HeaderNavlink px-0 py-2 m-0"> <a class="text-bold text-white no-underline" href="/login?return_to=%2Ftensorflow%2Ftensorflow%2Fblob%2Fr1.6%2Ftensorflow%2Fpython%2Fops%2Frnn_cell_impl.py" data-ga-click="(Logged out) Header, clicked Sign in, text:sign-in">Sign in</a> <span class="text-gray">or</span> <a class="text-bold text-white no-underline" href="/join?source=header-repo" data-ga-click="(Logged out) Header, clicked Sign up, text:sign-up">Sign up</a> </div> </span> </div> </div> </div> </header> </div> <div id="start-of-content" class="show-on-focus"></div> <div id="js-flash-container"> </div> <div role="main" class="application-main " > <div itemscope itemtype="http://schema.org/SoftwareSourceCode" class=""> <div id="js-repo-pjax-container" data-pjax-container > <div class="pagehead repohead instapaper_ignore readability-menu experiment-repo-nav "> <div class="repohead-details-container clearfix container"> <ul class="pagehead-actions"> <li> <a href="/login?return_to=%2Ftensorflow%2Ftensorflow" class="btn btn-sm btn-with-count tooltipped tooltipped-n" aria-label="You must be signed in to watch a repository" rel="nofollow"> <svg class="octicon octicon-eye" viewBox="0 0 16 16" version="1.1" width="16" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M8.06 2C3 2 0 8 0 8s3 6 8.06 6C13 14 16 8 16 8s-3-6-7.94-6zM8 12c-2.2 0-4-1.78-4-4 0-2.2 1.8-4 4-4 2.22 0 4 1.8 4 4 0 2.22-1.78 4-4 4zm2-4c0 1.11-.89 2-2 2-1.11 0-2-.89-2-2 0-1.11.89-2 2-2 1.11 0 2 .89 2 2z"/></svg> Watch </a> <a class="social-count" href="/tensorflow/tensorflow/watchers" aria-label="7632 users are watching this repository"> 7,632 </a> </li> <li> <a href="/login?return_to=%2Ftensorflow%2Ftensorflow" class="btn btn-sm btn-with-count tooltipped tooltipped-n" aria-label="You must be signed in to star a repository" rel="nofollow"> <svg class="octicon octicon-star" viewBox="0 0 14 16" version="1.1" width="14" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M14 6l-4.9-.64L7 1 4.9 5.36 0 6l3.6 3.26L2.67 14 7 11.67 11.33 14l-.93-4.74z"/></svg> Star </a> <a class="social-count js-social-count" href="/tensorflow/tensorflow/stargazers" aria-label="93301 users starred this repository"> 93,301 </a> </li> <li> <a href="/login?return_to=%2Ftensorflow%2Ftensorflow" class="btn btn-sm btn-with-count tooltipped tooltipped-n" aria-label="You must be signed in to fork a repository" rel="nofollow"> <svg class="octicon octicon-repo-forked" viewBox="0 0 10 16" version="1.1" width="10" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M8 1a1.993 1.993 0 0 0-1 3.72V6L5 8 3 6V4.72A1.993 1.993 0 0 0 2 1a1.993 1.993 0 0 0-1 3.72V6.5l3 3v1.78A1.993 1.993 0 0 0 5 15a1.993 1.993 0 0 0 1-3.72V9.5l3-3V4.72A1.993 1.993 0 0 0 8 1zM2 4.2C1.34 4.2.8 3.65.8 3c0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2zm3 10c-.66 0-1.2-.55-1.2-1.2 0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2zm3-10c-.66 0-1.2-.55-1.2-1.2 0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2z"/></svg> Fork </a> <a href="/tensorflow/tensorflow/network" class="social-count" aria-label="59890 users forked this repository"> 59,890 </a> </li> </ul> <h1 class="public "> <svg class="octicon octicon-repo" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M4 9H3V8h1v1zm0-3H3v1h1V6zm0-2H3v1h1V4zm0-2H3v1h1V2zm8-1v12c0 .55-.45 1-1 1H6v2l-1.5-1.5L3 16v-2H1c-.55 0-1-.45-1-1V1c0-.55.45-1 1-1h10c.55 0 1 .45 1 1zm-1 10H1v2h2v-1h3v1h5v-2zm0-10H2v9h9V1z"/></svg> <span class="author" itemprop="author"><a class="url fn" rel="author" href="/tensorflow">tensorflow</a></span><!-- --><span class="path-divider">/</span><!-- --><strong itemprop="name"><a data-pjax="#js-repo-pjax-container" href="/tensorflow/tensorflow">tensorflow</a></strong> </h1> </div> <nav class="reponav js-repo-nav js-sidenav-container-pjax container" itemscope itemtype="http://schema.org/BreadcrumbList" role="navigation" data-pjax="#js-repo-pjax-container"> <span itemscope itemtype="http://schema.org/ListItem" itemprop="itemListElement"> <a class="js-selected-navigation-item selected reponav-item" itemprop="url" data-hotkey="g c" data-selected-links="repo_source repo_downloads repo_commits repo_releases repo_tags repo_branches repo_packages /tensorflow/tensorflow/tree/r1.6" href="/tensorflow/tensorflow/tree/r1.6"> <svg class="octicon octicon-code" viewBox="0 0 14 16" version="1.1" width="14" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M9.5 3L8 4.5 11.5 8 8 11.5 9.5 13 14 8 9.5 3zm-5 0L0 8l4.5 5L6 11.5 2.5 8 6 4.5 4.5 3z"/></svg> <span itemprop="name">Code</span> <meta itemprop="position" content="1"> </a> </span> <span itemscope itemtype="http://schema.org/ListItem" itemprop="itemListElement"> <a itemprop="url" data-hotkey="g i" class="js-selected-navigation-item reponav-item" data-selected-links="repo_issues repo_labels repo_milestones /tensorflow/tensorflow/issues" href="/tensorflow/tensorflow/issues"> <svg class="octicon octicon-issue-opened" viewBox="0 0 14 16" version="1.1" width="14" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M7 2.3c3.14 0 5.7 2.56 5.7 5.7s-2.56 5.7-5.7 5.7A5.71 5.71 0 0 1 1.3 8c0-3.14 2.56-5.7 5.7-5.7zM7 1C3.14 1 0 4.14 0 8s3.14 7 7 7 7-3.14 7-7-3.14-7-7-7zm1 3H6v5h2V4zm0 6H6v2h2v-2z"/></svg> <span itemprop="name">Issues</span> <span class="Counter">1,287</span> <meta itemprop="position" content="2"> </a> </span> <span itemscope itemtype="http://schema.org/ListItem" itemprop="itemListElement"> <a data-hotkey="g p" itemprop="url" class="js-selected-navigation-item reponav-item" data-selected-links="repo_pulls checks /tensorflow/tensorflow/pulls" href="/tensorflow/tensorflow/pulls"> <svg class="octicon octicon-git-pull-request" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M11 11.28V5c-.03-.78-.34-1.47-.94-2.06C9.46 2.35 8.78 2.03 8 2H7V0L4 3l3 3V4h1c.27.02.48.11.69.31.21.2.3.42.31.69v6.28A1.993 1.993 0 0 0 10 15a1.993 1.993 0 0 0 1-3.72zm-1 2.92c-.66 0-1.2-.55-1.2-1.2 0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2zM4 3c0-1.11-.89-2-2-2a1.993 1.993 0 0 0-1 3.72v6.56A1.993 1.993 0 0 0 2 15a1.993 1.993 0 0 0 1-3.72V4.72c.59-.34 1-.98 1-1.72zm-.8 10c0 .66-.55 1.2-1.2 1.2-.65 0-1.2-.55-1.2-1.2 0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2zM2 4.2C1.34 4.2.8 3.65.8 3c0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2z"/></svg> <span itemprop="name">Pull requests</span> <span class="Counter">197</span> <meta itemprop="position" content="3"> </a> </span> <a data-hotkey="g b" class="js-selected-navigation-item reponav-item" data-selected-links="repo_projects new_repo_project repo_project /tensorflow/tensorflow/projects" href="/tensorflow/tensorflow/projects"> <svg class="octicon octicon-project" viewBox="0 0 15 16" version="1.1" width="15" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M10 12h3V2h-3v10zm-4-2h3V2H6v8zm-4 4h3V2H2v12zm-1 1h13V1H1v14zM14 0H1a1 1 0 0 0-1 1v14a1 1 0 0 0 1 1h13a1 1 0 0 0 1-1V1a1 1 0 0 0-1-1z"/></svg> Projects <span class="Counter" >0</span> </a> <a class="js-selected-navigation-item reponav-item" data-selected-links="repo_graphs repo_contributors dependency_graph pulse /tensorflow/tensorflow/pulse" href="/tensorflow/tensorflow/pulse"> <svg class="octicon octicon-graph" viewBox="0 0 16 16" version="1.1" width="16" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M16 14v1H0V0h1v14h15zM5 13H3V8h2v5zm4 0H7V3h2v10zm4 0h-2V6h2v7z"/></svg> Insights </a> </nav> </div> <div class="container new-discussion-timeline experiment-repo-nav "> <div class="repository-content "> <a class="d-none js-permalink-shortcut" data-hotkey="y" href="/tensorflow/tensorflow/blob/cbc658095ae228f2f557af47e4901d552573aa15/tensorflow/python/ops/rnn_cell_impl.py">Permalink</a> <!-- blob contrib key: blob_contributors:v21:5895d9f5b73fd7bcfb782c5d3257ea21 --> <div class="file-navigation"> <div class="select-menu branch-select-menu js-menu-container js-select-menu float-left"> <button class=" btn btn-sm select-menu-button js-menu-target css-truncate" data-hotkey="w" type="button" aria-label="Switch branches or tags" aria-expanded="false" aria-haspopup="true"> <i>Branch:</i> <span class="js-select-button css-truncate-target">r1.6</span> </button> <div class="select-menu-modal-holder js-menu-content js-navigation-container" data-pjax> <div class="select-menu-modal"> <div class="select-menu-header"> <svg class="octicon octicon-x js-menu-close" role="img" aria-label="Close" viewBox="0 0 12 16" version="1.1" width="12" height="16"><path fill-rule="evenodd" d="M7.48 8l3.75 3.75-1.48 1.48L6 9.48l-3.75 3.75-1.48-1.48L4.52 8 .77 4.25l1.48-1.48L6 6.52l3.75-3.75 1.48 1.48z"/></svg> <span class="select-menu-title">Switch branches/tags</span> </div> <div class="select-menu-filters"> <div class="select-menu-text-filter"> <input type="text" aria-label="Filter branches/tags" id="context-commitish-filter-field" class="form-control js-filterable-field js-navigation-enable" placeholder="Filter branches/tags"> </div> <div class="select-menu-tabs"> <ul> <li class="select-menu-tab"> <a href="#" data-tab-filter="branches" data-filter-placeholder="Filter branches/tags" class="js-select-menu-tab" role="tab">Branches</a> </li> <li class="select-menu-tab"> <a href="#" data-tab-filter="tags" data-filter-placeholder="Find a tag…" class="js-select-menu-tab" role="tab">Tags</a> </li> </ul> </div> </div> <div class="select-menu-list select-menu-tab-bucket js-select-menu-tab-bucket" data-tab-filter="branches" role="menu"> <div data-filterable-for="context-commitish-filter-field" data-filterable-type="substring"> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/0.6.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.6.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> 0.6.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/branch_189041421/tensorflow/python/ops/rnn_cell_impl.py" data-name="branch_189041421" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> branch_189041421 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/martinwicke-patch-1/tensorflow/python/ops/rnn_cell_impl.py" data-name="martinwicke-patch-1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> martinwicke-patch-1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/martinwicke-patch-2/tensorflow/python/ops/rnn_cell_impl.py" data-name="martinwicke-patch-2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> martinwicke-patch-2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/master/tensorflow/python/ops/rnn_cell_impl.py" data-name="master" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> master </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/mrry-patch-1/tensorflow/python/ops/rnn_cell_impl.py" data-name="mrry-patch-1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> mrry-patch-1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.7/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.7" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.7 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.8/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.8" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.8 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.9/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.9" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.9 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.10/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.10" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.10 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.11/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.11" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.11 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.12/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.12" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.12 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.2/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.3/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.3" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.3 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.4/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.4" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.4 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.5/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.5" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.5 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open selected" href="/tensorflow/tensorflow/blob/r1.6/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.6" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.6 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.7/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.7" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.7 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/yifeif-patch-1/tensorflow/python/ops/rnn_cell_impl.py" data-name="yifeif-patch-1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> yifeif-patch-1 </span> </a> </div> <div class="select-menu-no-results">Nothing to show</div> </div> <div class="select-menu-list select-menu-tab-bucket js-select-menu-tab-bucket" data-tab-filter="tags"> <div data-filterable-for="context-commitish-filter-field" data-filterable-type="substring"> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.7.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.7.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.7.0-rc0"> v1.7.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.6.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.6.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.6.0"> v1.6.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.6.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.6.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.6.0-rc1"> v1.6.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.6.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.6.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.6.0-rc0"> v1.6.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.5.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.5.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.5.1"> v1.5.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.5.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.5.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.5.0"> v1.5.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.5.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.5.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.5.0-rc1"> v1.5.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.5.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.5.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.5.0-rc0"> v1.5.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.4.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.4.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.4.1"> v1.4.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.4.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.4.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.4.0"> v1.4.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.4.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.4.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.4.0-rc1"> v1.4.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.4.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.4.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.4.0-rc0"> v1.4.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.1"> v1.3.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.0"> v1.3.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.0-rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.0-rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.0-rc2"> v1.3.0-rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.0-rc1"> v1.3.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.0-rc0"> v1.3.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.1"> v1.2.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.0"> v1.2.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.0-rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.0-rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.0-rc2"> v1.2.0-rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.0-rc1"> v1.2.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.0-rc0"> v1.2.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.1.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.1.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.1.0"> v1.1.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.1.0-rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.1.0-rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.1.0-rc2"> v1.1.0-rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.1.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.1.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.1.0-rc1"> v1.1.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.1.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.1.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.1.0-rc0"> v1.1.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.1"> v1.0.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0"> v1.0.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0-rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0-rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0-rc2"> v1.0.0-rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0-rc1"> v1.0.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0-rc0"> v1.0.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0-alpha/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0-alpha" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0-alpha"> v1.0.0-alpha </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.12.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.12.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.12.0"> v0.12.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.11.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.11.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.11.0"> v0.11.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.11.0rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.11.0rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.11.0rc2"> v0.11.0rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.11.0rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.11.0rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.11.0rc1"> v0.11.0rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.11.0rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.11.0rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.11.0rc0"> v0.11.0rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.10.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.10.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.10.0"> v0.10.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.10.0rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.10.0rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.10.0rc0"> v0.10.0rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.9.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.9.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.9.0"> v0.9.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.9.0rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.9.0rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.9.0rc0"> v0.9.0rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.8.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.8.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.8.0"> v0.8.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.8.0rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.8.0rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.8.0rc0"> v0.8.0rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.7.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.7.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.7.1"> v0.7.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.7.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.7.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.7.0"> v0.7.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.6.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.6.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.6.0"> v0.6.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.12.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.12.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.12.1"> 0.12.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.12.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.12.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.12.0-rc1"> 0.12.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.12.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.12.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.12.0-rc0"> 0.12.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.6.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.6.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.6.0"> 0.6.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.5.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.5.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.5.0"> 0.5.0 </span> </a> </div> <div class="select-menu-no-results">Nothing to show</div> </div> </div> </div> </div> <div class="BtnGroup float-right"> <a href="/tensorflow/tensorflow/find/r1.6" class="js-pjax-capture-input btn btn-sm BtnGroup-item" data-pjax data-hotkey="t"> Find file </a> <clipboard-copy for="blob-path" aria-label="Copy file path to clipboard" class="btn btn-sm BtnGroup-item tooltipped tooltipped-s" data-copied-hint="Copied!"> Copy path </clipboard-copy> </div> <div id="blob-path" class="breadcrumb"> <span class="repo-root js-repo-root"><span class="js-path-segment"><a data-pjax="true" href="/tensorflow/tensorflow/tree/r1.6"><span>tensorflow</span></a></span></span><span class="separator">/</span><span class="js-path-segment"><a data-pjax="true" href="/tensorflow/tensorflow/tree/r1.6/tensorflow"><span>tensorflow</span></a></span><span class="separator">/</span><span class="js-path-segment"><a data-pjax="true" href="/tensorflow/tensorflow/tree/r1.6/tensorflow/python"><span>python</span></a></span><span class="separator">/</span><span class="js-path-segment"><a data-pjax="true" href="/tensorflow/tensorflow/tree/r1.6/tensorflow/python/ops"><span>ops</span></a></span><span class="separator">/</span><strong class="final-path">rnn_cell_impl.py</strong> </div> </div> <div class="commit-tease"> <span class="float-right"> <a class="commit-tease-sha" href="/tensorflow/tensorflow/commit/dff64beac8570d910f83774087642bb6a3fda96a" data-pjax> dff64be </a> <relative-time datetime="2018-01-23T21:46:25Z">Jan 23, 2018</relative-time> </span> <div> <img alt="@michaelkhan3" class="avatar" height="20" src="https://avatars3.githubusercontent.com/u/4420471?s=40&amp;v=4" width="20" /> <a href="/michaelkhan3" class="user-mention" rel="contributor">michaelkhan3</a> <a href="/tensorflow/tensorflow/commit/dff64beac8570d910f83774087642bb6a3fda96a" class="message" data-pjax="true" title="Add property to get cell wrapped by DropoutWrapper (#16006) * add property to get cell wrapped by dropout * updating golden files">Add property to get cell wrapped by DropoutWrapper (</a><a href="https://github.com/tensorflow/tensorflow/pull/16006" class="issue-link js-issue-link" data-error-text="Failed to load issue title" data-id="287414630" data-permission-text="Issue title is private" data-url="https://github.com/tensorflow/tensorflow/issues/16006">#16006</a><a href="/tensorflow/tensorflow/commit/dff64beac8570d910f83774087642bb6a3fda96a" class="message" data-pjax="true" title="Add property to get cell wrapped by DropoutWrapper (#16006) * add property to get cell wrapped by dropout * updating golden files">)</a> </div> <div class="commit-tease-contributors"> <button type="button" class="btn-link muted-link contributors-toggle" data-facebox="#blob_contributors_box"> <strong>13</strong> contributors </button> <a class="avatar-link tooltipped tooltipped-s" aria-label="tensorflower-gardener" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=tensorflower-gardener"><img alt="@tensorflower-gardener" class="avatar" height="20" src="https://avatars1.githubusercontent.com/u/17151892?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="ebrevdo" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=ebrevdo"><img alt="@ebrevdo" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/1794715?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="jhseu" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=jhseu"><img alt="@jhseu" class="avatar" height="20" src="https://avatars1.githubusercontent.com/u/170179?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="protoget" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=protoget"><img alt="@protoget" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/5117188?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="xiejw" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=xiejw"><img alt="@xiejw" class="avatar" height="20" src="https://avatars0.githubusercontent.com/u/1184671?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="martinwicke" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=martinwicke"><img alt="@martinwicke" class="avatar" height="20" src="https://avatars1.githubusercontent.com/u/577277?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="michaelkhan3" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=michaelkhan3"><img alt="@michaelkhan3" class="avatar" height="20" src="https://avatars3.githubusercontent.com/u/4420471?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="fchollet" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=fchollet"><img alt="@fchollet" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/710255?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="drpngx" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=drpngx"><img alt="@drpngx" class="avatar" height="20" src="https://avatars3.githubusercontent.com/u/20959853?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="caisq" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=caisq"><img alt="@caisq" class="avatar" height="20" src="https://avatars0.githubusercontent.com/u/16824702?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="alextp" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=alextp"><img alt="@alextp" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/5061?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="annarev" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=annarev"><img alt="@annarev" class="avatar" height="20" src="https://avatars0.githubusercontent.com/u/22060313?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="kosklain" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=kosklain"><img alt="@kosklain" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/1104579?s=40&amp;v=4" width="20" /> </a> </div> <div id="blob_contributors_box" style="display:none"> <h2 class="facebox-header" data-facebox-id="facebox-header">Users who have contributed to this file</h2> <ul class="facebox-user-list" data-facebox-id="facebox-description"> <li class="facebox-user-list-item"> <img alt="@tensorflower-gardener" height="24" src="https://avatars0.githubusercontent.com/u/17151892?s=48&amp;v=4" width="24" /> <a href="/tensorflower-gardener">tensorflower-gardener</a> </li> <li class="facebox-user-list-item"> <img alt="@ebrevdo" height="24" src="https://avatars3.githubusercontent.com/u/1794715?s=48&amp;v=4" width="24" /> <a href="/ebrevdo">ebrevdo</a> </li> <li class="facebox-user-list-item"> <img alt="@jhseu" height="24" src="https://avatars0.githubusercontent.com/u/170179?s=48&amp;v=4" width="24" /> <a href="/jhseu">jhseu</a> </li> <li class="facebox-user-list-item"> <img alt="@protoget" height="24" src="https://avatars3.githubusercontent.com/u/5117188?s=48&amp;v=4" width="24" /> <a href="/protoget">protoget</a> </li> <li class="facebox-user-list-item"> <img alt="@xiejw" height="24" src="https://avatars1.githubusercontent.com/u/1184671?s=48&amp;v=4" width="24" /> <a href="/xiejw">xiejw</a> </li> <li class="facebox-user-list-item"> <img alt="@martinwicke" height="24" src="https://avatars0.githubusercontent.com/u/577277?s=48&amp;v=4" width="24" /> <a href="/martinwicke">martinwicke</a> </li> <li class="facebox-user-list-item"> <img alt="@michaelkhan3" height="24" src="https://avatars2.githubusercontent.com/u/4420471?s=48&amp;v=4" width="24" /> <a href="/michaelkhan3">michaelkhan3</a> </li> <li class="facebox-user-list-item"> <img alt="@fchollet" height="24" src="https://avatars3.githubusercontent.com/u/710255?s=48&amp;v=4" width="24" /> <a href="/fchollet">fchollet</a> </li> <li class="facebox-user-list-item"> <img alt="@drpngx" height="24" src="https://avatars2.githubusercontent.com/u/20959853?s=48&amp;v=4" width="24" /> <a href="/drpngx">drpngx</a> </li> <li class="facebox-user-list-item"> <img alt="@caisq" height="24" src="https://avatars1.githubusercontent.com/u/16824702?s=48&amp;v=4" width="24" /> <a href="/caisq">caisq</a> </li> <li class="facebox-user-list-item"> <img alt="@alextp" height="24" src="https://avatars3.githubusercontent.com/u/5061?s=48&amp;v=4" width="24" /> <a href="/alextp">alextp</a> </li> <li class="facebox-user-list-item"> <img alt="@annarev" height="24" src="https://avatars1.githubusercontent.com/u/22060313?s=48&amp;v=4" width="24" /> <a href="/annarev">annarev</a> </li> <li class="facebox-user-list-item"> <img alt="@kosklain" height="24" src="https://avatars3.githubusercontent.com/u/1104579?s=48&amp;v=4" width="24" /> <a href="/kosklain">kosklain</a> </li> </ul> </div> </div> <div class="file"> <div class="file-header"> <div class="file-actions"> <div class="BtnGroup"> <a id="raw-url" class="btn btn-sm BtnGroup-item" href="/tensorflow/tensorflow/raw/r1.6/tensorflow/python/ops/rnn_cell_impl.py">Raw</a> <a class="btn btn-sm js-update-url-with-hash BtnGroup-item" data-hotkey="b" href="/tensorflow/tensorflow/blame/r1.6/tensorflow/python/ops/rnn_cell_impl.py">Blame</a> <a rel="nofollow" class="btn btn-sm BtnGroup-item" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py">History</a> </div> <button type="button" class="btn-octicon disabled tooltipped tooltipped-nw" aria-label="You must be signed in to make or propose changes"> <svg class="octicon octicon-pencil" viewBox="0 0 14 16" version="1.1" width="14" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M0 12v3h3l8-8-3-3-8 8zm3 2H1v-2h1v1h1v1zm10.3-9.3L12 6 9 3l1.3-1.3a.996.996 0 0 1 1.41 0l1.59 1.59c.39.39.39 1.02 0 1.41z"/></svg> </button> <button type="button" class="btn-octicon btn-octicon-danger disabled tooltipped tooltipped-nw" aria-label="You must be signed in to make or propose changes"> <svg class="octicon octicon-trashcan" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M11 2H9c0-.55-.45-1-1-1H5c-.55 0-1 .45-1 1H2c-.55 0-1 .45-1 1v1c0 .55.45 1 1 1v9c0 .55.45 1 1 1h7c.55 0 1-.45 1-1V5c.55 0 1-.45 1-1V3c0-.55-.45-1-1-1zm-1 12H3V5h1v8h1V5h1v8h1V5h1v8h1V5h1v9zm1-10H2V3h9v1z"/></svg> </button> </div> <div class="file-info"> 1286 lines (1078 sloc) <span class="file-info-divider"></span> 49.7 KB </div> </div> <div itemprop="text" class="blob-wrapper data type-python"> <table class="highlight tab-size js-file-line-container" data-tab-size="8"> <tr> <td id="L1" class="blob-num js-line-number" data-line-number="1"></td> <td id="LC1" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> Copyright 2015 The TensorFlow Authors. All Rights Reserved.</span></td> </tr> <tr> <td id="L2" class="blob-num js-line-number" data-line-number="2"></td> <td id="LC2" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span></span></td> </tr> <tr> <td id="L3" class="blob-num js-line-number" data-line-number="3"></td> <td id="LC3" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> Licensed under the Apache License, Version 2.0 (the &quot;License&quot;);</span></td> </tr> <tr> <td id="L4" class="blob-num js-line-number" data-line-number="4"></td> <td id="LC4" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> you may not use this file except in compliance with the License.</span></td> </tr> <tr> <td id="L5" class="blob-num js-line-number" data-line-number="5"></td> <td id="LC5" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> You may obtain a copy of the License at</span></td> </tr> <tr> <td id="L6" class="blob-num js-line-number" data-line-number="6"></td> <td id="LC6" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span></span></td> </tr> <tr> <td id="L7" class="blob-num js-line-number" data-line-number="7"></td> <td id="LC7" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> http://www.apache.org/licenses/LICENSE-2.0</span></td> </tr> <tr> <td id="L8" class="blob-num js-line-number" data-line-number="8"></td> <td id="LC8" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span></span></td> </tr> <tr> <td id="L9" class="blob-num js-line-number" data-line-number="9"></td> <td id="LC9" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> Unless required by applicable law or agreed to in writing, software</span></td> </tr> <tr> <td id="L10" class="blob-num js-line-number" data-line-number="10"></td> <td id="LC10" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> distributed under the License is distributed on an &quot;AS IS&quot; BASIS,</span></td> </tr> <tr> <td id="L11" class="blob-num js-line-number" data-line-number="11"></td> <td id="LC11" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.</span></td> </tr> <tr> <td id="L12" class="blob-num js-line-number" data-line-number="12"></td> <td id="LC12" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> See the License for the specific language governing permissions and</span></td> </tr> <tr> <td id="L13" class="blob-num js-line-number" data-line-number="13"></td> <td id="LC13" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> limitations under the License.</span></td> </tr> <tr> <td id="L14" class="blob-num js-line-number" data-line-number="14"></td> <td id="LC14" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> ==============================================================================</span></td> </tr> <tr> <td id="L15" class="blob-num js-line-number" data-line-number="15"></td> <td id="LC15" class="blob-code blob-code-inner js-file-line"><span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Module implementing RNN Cells.</span></td> </tr> <tr> <td id="L16" class="blob-num js-line-number" data-line-number="16"></td> <td id="LC16" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L17" class="blob-num js-line-number" data-line-number="17"></td> <td id="LC17" class="blob-code blob-code-inner js-file-line"><span class="pl-s">This module provides a number of basic commonly used RNN cells, such as LSTM</span></td> </tr> <tr> <td id="L18" class="blob-num js-line-number" data-line-number="18"></td> <td id="LC18" class="blob-code blob-code-inner js-file-line"><span class="pl-s">(Long Short Term Memory) or GRU (Gated Recurrent Unit), and a number of</span></td> </tr> <tr> <td id="L19" class="blob-num js-line-number" data-line-number="19"></td> <td id="LC19" class="blob-code blob-code-inner js-file-line"><span class="pl-s">operators that allow adding dropouts, projections, or embeddings for inputs.</span></td> </tr> <tr> <td id="L20" class="blob-num js-line-number" data-line-number="20"></td> <td id="LC20" class="blob-code blob-code-inner js-file-line"><span class="pl-s">Constructing multi-layer cells is supported by the class `MultiRNNCell`, or by</span></td> </tr> <tr> <td id="L21" class="blob-num js-line-number" data-line-number="21"></td> <td id="LC21" class="blob-code blob-code-inner js-file-line"><span class="pl-s">calling the `rnn` ops several times.</span></td> </tr> <tr> <td id="L22" class="blob-num js-line-number" data-line-number="22"></td> <td id="LC22" class="blob-code blob-code-inner js-file-line"><span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L23" class="blob-num js-line-number" data-line-number="23"></td> <td id="LC23" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-c1">__future__</span> <span class="pl-k">import</span> absolute_import</td> </tr> <tr> <td id="L24" class="blob-num js-line-number" data-line-number="24"></td> <td id="LC24" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-c1">__future__</span> <span class="pl-k">import</span> division</td> </tr> <tr> <td id="L25" class="blob-num js-line-number" data-line-number="25"></td> <td id="LC25" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-c1">__future__</span> <span class="pl-k">import</span> print_function</td> </tr> <tr> <td id="L26" class="blob-num js-line-number" data-line-number="26"></td> <td id="LC26" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L27" class="blob-num js-line-number" data-line-number="27"></td> <td id="LC27" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> collections</td> </tr> <tr> <td id="L28" class="blob-num js-line-number" data-line-number="28"></td> <td id="LC28" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> hashlib</td> </tr> <tr> <td id="L29" class="blob-num js-line-number" data-line-number="29"></td> <td id="LC29" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> numbers</td> </tr> <tr> <td id="L30" class="blob-num js-line-number" data-line-number="30"></td> <td id="LC30" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L31" class="blob-num js-line-number" data-line-number="31"></td> <td id="LC31" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.eager <span class="pl-k">import</span> context</td> </tr> <tr> <td id="L32" class="blob-num js-line-number" data-line-number="32"></td> <td id="LC32" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> constant_op</td> </tr> <tr> <td id="L33" class="blob-num js-line-number" data-line-number="33"></td> <td id="LC33" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> dtypes</td> </tr> <tr> <td id="L34" class="blob-num js-line-number" data-line-number="34"></td> <td id="LC34" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> ops</td> </tr> <tr> <td id="L35" class="blob-num js-line-number" data-line-number="35"></td> <td id="LC35" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> tensor_shape</td> </tr> <tr> <td id="L36" class="blob-num js-line-number" data-line-number="36"></td> <td id="LC36" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> tensor_util</td> </tr> <tr> <td id="L37" class="blob-num js-line-number" data-line-number="37"></td> <td id="LC37" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.layers <span class="pl-k">import</span> base <span class="pl-k">as</span> base_layer</td> </tr> <tr> <td id="L38" class="blob-num js-line-number" data-line-number="38"></td> <td id="LC38" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> array_ops</td> </tr> <tr> <td id="L39" class="blob-num js-line-number" data-line-number="39"></td> <td id="LC39" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> clip_ops</td> </tr> <tr> <td id="L40" class="blob-num js-line-number" data-line-number="40"></td> <td id="LC40" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> init_ops</td> </tr> <tr> <td id="L41" class="blob-num js-line-number" data-line-number="41"></td> <td id="LC41" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> math_ops</td> </tr> <tr> <td id="L42" class="blob-num js-line-number" data-line-number="42"></td> <td id="LC42" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> nn_ops</td> </tr> <tr> <td id="L43" class="blob-num js-line-number" data-line-number="43"></td> <td id="LC43" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> partitioned_variables</td> </tr> <tr> <td id="L44" class="blob-num js-line-number" data-line-number="44"></td> <td id="LC44" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> random_ops</td> </tr> <tr> <td id="L45" class="blob-num js-line-number" data-line-number="45"></td> <td id="LC45" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> tensor_array_ops</td> </tr> <tr> <td id="L46" class="blob-num js-line-number" data-line-number="46"></td> <td id="LC46" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> variable_scope <span class="pl-k">as</span> vs</td> </tr> <tr> <td id="L47" class="blob-num js-line-number" data-line-number="47"></td> <td id="LC47" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> variables <span class="pl-k">as</span> tf_variables</td> </tr> <tr> <td id="L48" class="blob-num js-line-number" data-line-number="48"></td> <td id="LC48" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.platform <span class="pl-k">import</span> tf_logging <span class="pl-k">as</span> logging</td> </tr> <tr> <td id="L49" class="blob-num js-line-number" data-line-number="49"></td> <td id="LC49" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.util <span class="pl-k">import</span> nest</td> </tr> <tr> <td id="L50" class="blob-num js-line-number" data-line-number="50"></td> <td id="LC50" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.util.tf_export <span class="pl-k">import</span> tf_export</td> </tr> <tr> <td id="L51" class="blob-num js-line-number" data-line-number="51"></td> <td id="LC51" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L52" class="blob-num js-line-number" data-line-number="52"></td> <td id="LC52" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L53" class="blob-num js-line-number" data-line-number="53"></td> <td id="LC53" class="blob-code blob-code-inner js-file-line"><span class="pl-c1">_BIAS_VARIABLE_NAME</span> <span class="pl-k">=</span> <span class="pl-s"><span class="pl-pds">&quot;</span>bias<span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L54" class="blob-num js-line-number" data-line-number="54"></td> <td id="LC54" class="blob-code blob-code-inner js-file-line"><span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span> <span class="pl-k">=</span> <span class="pl-s"><span class="pl-pds">&quot;</span>kernel<span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L55" class="blob-num js-line-number" data-line-number="55"></td> <td id="LC55" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L56" class="blob-num js-line-number" data-line-number="56"></td> <td id="LC56" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L57" class="blob-num js-line-number" data-line-number="57"></td> <td id="LC57" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_like_rnncell</span>(<span class="pl-smi">cell</span>):</td> </tr> <tr> <td id="L58" class="blob-num js-line-number" data-line-number="58"></td> <td id="LC58" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Checks that a given object is an RNNCell by using duck typing.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L59" class="blob-num js-line-number" data-line-number="59"></td> <td id="LC59" class="blob-code blob-code-inner js-file-line"> conditions <span class="pl-k">=</span> [<span class="pl-c1">hasattr</span>(cell, <span class="pl-s"><span class="pl-pds">&quot;</span>output_size<span class="pl-pds">&quot;</span></span>), <span class="pl-c1">hasattr</span>(cell, <span class="pl-s"><span class="pl-pds">&quot;</span>state_size<span class="pl-pds">&quot;</span></span>),</td> </tr> <tr> <td id="L60" class="blob-num js-line-number" data-line-number="60"></td> <td id="LC60" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">hasattr</span>(cell, <span class="pl-s"><span class="pl-pds">&quot;</span>zero_state<span class="pl-pds">&quot;</span></span>), <span class="pl-c1">callable</span>(cell)]</td> </tr> <tr> <td id="L61" class="blob-num js-line-number" data-line-number="61"></td> <td id="LC61" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">all</span>(conditions)</td> </tr> <tr> <td id="L62" class="blob-num js-line-number" data-line-number="62"></td> <td id="LC62" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L63" class="blob-num js-line-number" data-line-number="63"></td> <td id="LC63" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L64" class="blob-num js-line-number" data-line-number="64"></td> <td id="LC64" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_concat</span>(<span class="pl-smi">prefix</span>, <span class="pl-smi">suffix</span>, <span class="pl-smi">static</span><span class="pl-k">=</span><span class="pl-c1">False</span>):</td> </tr> <tr> <td id="L65" class="blob-num js-line-number" data-line-number="65"></td> <td id="LC65" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Concat that enables int, Tensor, or TensorShape values.</span></td> </tr> <tr> <td id="L66" class="blob-num js-line-number" data-line-number="66"></td> <td id="LC66" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L67" class="blob-num js-line-number" data-line-number="67"></td> <td id="LC67" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> This function takes a size specification, which can be an integer, a</span></td> </tr> <tr> <td id="L68" class="blob-num js-line-number" data-line-number="68"></td> <td id="LC68" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> TensorShape, or a Tensor, and converts it into a concatenated Tensor</span></td> </tr> <tr> <td id="L69" class="blob-num js-line-number" data-line-number="69"></td> <td id="LC69" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> (if static = False) or a list of integers (if static = True).</span></td> </tr> <tr> <td id="L70" class="blob-num js-line-number" data-line-number="70"></td> <td id="LC70" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L71" class="blob-num js-line-number" data-line-number="71"></td> <td id="LC71" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L72" class="blob-num js-line-number" data-line-number="72"></td> <td id="LC72" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> prefix: The prefix; usually the batch size (and/or time step size).</span></td> </tr> <tr> <td id="L73" class="blob-num js-line-number" data-line-number="73"></td> <td id="LC73" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> (TensorShape, int, or Tensor.)</span></td> </tr> <tr> <td id="L74" class="blob-num js-line-number" data-line-number="74"></td> <td id="LC74" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> suffix: TensorShape, int, or Tensor.</span></td> </tr> <tr> <td id="L75" class="blob-num js-line-number" data-line-number="75"></td> <td id="LC75" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> static: If `True`, return a python list with possibly unknown dimensions.</span></td> </tr> <tr> <td id="L76" class="blob-num js-line-number" data-line-number="76"></td> <td id="LC76" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Otherwise return a `Tensor`.</span></td> </tr> <tr> <td id="L77" class="blob-num js-line-number" data-line-number="77"></td> <td id="LC77" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L78" class="blob-num js-line-number" data-line-number="78"></td> <td id="LC78" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L79" class="blob-num js-line-number" data-line-number="79"></td> <td id="LC79" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> shape: the concatenation of prefix and suffix.</span></td> </tr> <tr> <td id="L80" class="blob-num js-line-number" data-line-number="80"></td> <td id="LC80" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L81" class="blob-num js-line-number" data-line-number="81"></td> <td id="LC81" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L82" class="blob-num js-line-number" data-line-number="82"></td> <td id="LC82" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if `suffix` is not a scalar or vector (or TensorShape).</span></td> </tr> <tr> <td id="L83" class="blob-num js-line-number" data-line-number="83"></td> <td id="LC83" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if prefix or suffix was `None` and asked for dynamic</span></td> </tr> <tr> <td id="L84" class="blob-num js-line-number" data-line-number="84"></td> <td id="LC84" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Tensors out.</span></td> </tr> <tr> <td id="L85" class="blob-num js-line-number" data-line-number="85"></td> <td id="LC85" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L86" class="blob-num js-line-number" data-line-number="86"></td> <td id="LC86" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">isinstance</span>(prefix, ops.Tensor):</td> </tr> <tr> <td id="L87" class="blob-num js-line-number" data-line-number="87"></td> <td id="LC87" class="blob-code blob-code-inner js-file-line"> p <span class="pl-k">=</span> prefix</td> </tr> <tr> <td id="L88" class="blob-num js-line-number" data-line-number="88"></td> <td id="LC88" class="blob-code blob-code-inner js-file-line"> p_static <span class="pl-k">=</span> tensor_util.constant_value(prefix)</td> </tr> <tr> <td id="L89" class="blob-num js-line-number" data-line-number="89"></td> <td id="LC89" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> p.shape.ndims <span class="pl-k">==</span> <span class="pl-c1">0</span>:</td> </tr> <tr> <td id="L90" class="blob-num js-line-number" data-line-number="90"></td> <td id="LC90" class="blob-code blob-code-inner js-file-line"> p <span class="pl-k">=</span> array_ops.expand_dims(p, <span class="pl-c1">0</span>)</td> </tr> <tr> <td id="L91" class="blob-num js-line-number" data-line-number="91"></td> <td id="LC91" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">elif</span> p.shape.ndims <span class="pl-k">!=</span> <span class="pl-c1">1</span>:</td> </tr> <tr> <td id="L92" class="blob-num js-line-number" data-line-number="92"></td> <td id="LC92" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>prefix tensor must be either a scalar or vector, <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L93" class="blob-num js-line-number" data-line-number="93"></td> <td id="LC93" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>but saw tensor: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> p)</td> </tr> <tr> <td id="L94" class="blob-num js-line-number" data-line-number="94"></td> <td id="LC94" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L95" class="blob-num js-line-number" data-line-number="95"></td> <td id="LC95" class="blob-code blob-code-inner js-file-line"> p <span class="pl-k">=</span> tensor_shape.as_shape(prefix)</td> </tr> <tr> <td id="L96" class="blob-num js-line-number" data-line-number="96"></td> <td id="LC96" class="blob-code blob-code-inner js-file-line"> p_static <span class="pl-k">=</span> p.as_list() <span class="pl-k">if</span> p.ndims <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L97" class="blob-num js-line-number" data-line-number="97"></td> <td id="LC97" class="blob-code blob-code-inner js-file-line"> p <span class="pl-k">=</span> (constant_op.constant(p.as_list(), <span class="pl-v">dtype</span><span class="pl-k">=</span>dtypes.int32)</td> </tr> <tr> <td id="L98" class="blob-num js-line-number" data-line-number="98"></td> <td id="LC98" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> p.is_fully_defined() <span class="pl-k">else</span> <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L99" class="blob-num js-line-number" data-line-number="99"></td> <td id="LC99" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">isinstance</span>(suffix, ops.Tensor):</td> </tr> <tr> <td id="L100" class="blob-num js-line-number" data-line-number="100"></td> <td id="LC100" class="blob-code blob-code-inner js-file-line"> s <span class="pl-k">=</span> suffix</td> </tr> <tr> <td id="L101" class="blob-num js-line-number" data-line-number="101"></td> <td id="LC101" class="blob-code blob-code-inner js-file-line"> s_static <span class="pl-k">=</span> tensor_util.constant_value(suffix)</td> </tr> <tr> <td id="L102" class="blob-num js-line-number" data-line-number="102"></td> <td id="LC102" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> s.shape.ndims <span class="pl-k">==</span> <span class="pl-c1">0</span>:</td> </tr> <tr> <td id="L103" class="blob-num js-line-number" data-line-number="103"></td> <td id="LC103" class="blob-code blob-code-inner js-file-line"> s <span class="pl-k">=</span> array_ops.expand_dims(s, <span class="pl-c1">0</span>)</td> </tr> <tr> <td id="L104" class="blob-num js-line-number" data-line-number="104"></td> <td id="LC104" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">elif</span> s.shape.ndims <span class="pl-k">!=</span> <span class="pl-c1">1</span>:</td> </tr> <tr> <td id="L105" class="blob-num js-line-number" data-line-number="105"></td> <td id="LC105" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>suffix tensor must be either a scalar or vector, <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L106" class="blob-num js-line-number" data-line-number="106"></td> <td id="LC106" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>but saw tensor: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> s)</td> </tr> <tr> <td id="L107" class="blob-num js-line-number" data-line-number="107"></td> <td id="LC107" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L108" class="blob-num js-line-number" data-line-number="108"></td> <td id="LC108" class="blob-code blob-code-inner js-file-line"> s <span class="pl-k">=</span> tensor_shape.as_shape(suffix)</td> </tr> <tr> <td id="L109" class="blob-num js-line-number" data-line-number="109"></td> <td id="LC109" class="blob-code blob-code-inner js-file-line"> s_static <span class="pl-k">=</span> s.as_list() <span class="pl-k">if</span> s.ndims <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L110" class="blob-num js-line-number" data-line-number="110"></td> <td id="LC110" class="blob-code blob-code-inner js-file-line"> s <span class="pl-k">=</span> (constant_op.constant(s.as_list(), <span class="pl-v">dtype</span><span class="pl-k">=</span>dtypes.int32)</td> </tr> <tr> <td id="L111" class="blob-num js-line-number" data-line-number="111"></td> <td id="LC111" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> s.is_fully_defined() <span class="pl-k">else</span> <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L112" class="blob-num js-line-number" data-line-number="112"></td> <td id="LC112" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L113" class="blob-num js-line-number" data-line-number="113"></td> <td id="LC113" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> static:</td> </tr> <tr> <td id="L114" class="blob-num js-line-number" data-line-number="114"></td> <td id="LC114" class="blob-code blob-code-inner js-file-line"> shape <span class="pl-k">=</span> tensor_shape.as_shape(p_static).concatenate(s_static)</td> </tr> <tr> <td id="L115" class="blob-num js-line-number" data-line-number="115"></td> <td id="LC115" class="blob-code blob-code-inner js-file-line"> shape <span class="pl-k">=</span> shape.as_list() <span class="pl-k">if</span> shape.ndims <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L116" class="blob-num js-line-number" data-line-number="116"></td> <td id="LC116" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L117" class="blob-num js-line-number" data-line-number="117"></td> <td id="LC117" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> p <span class="pl-k">is</span> <span class="pl-c1">None</span> <span class="pl-k">or</span> s <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L118" class="blob-num js-line-number" data-line-number="118"></td> <td id="LC118" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Provided a prefix or suffix of None: <span class="pl-c1">%s</span> and <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L119" class="blob-num js-line-number" data-line-number="119"></td> <td id="LC119" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> (prefix, suffix))</td> </tr> <tr> <td id="L120" class="blob-num js-line-number" data-line-number="120"></td> <td id="LC120" class="blob-code blob-code-inner js-file-line"> shape <span class="pl-k">=</span> array_ops.concat((p, s), <span class="pl-c1">0</span>)</td> </tr> <tr> <td id="L121" class="blob-num js-line-number" data-line-number="121"></td> <td id="LC121" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> shape</td> </tr> <tr> <td id="L122" class="blob-num js-line-number" data-line-number="122"></td> <td id="LC122" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L123" class="blob-num js-line-number" data-line-number="123"></td> <td id="LC123" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L124" class="blob-num js-line-number" data-line-number="124"></td> <td id="LC124" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_zero_state_tensors</span>(<span class="pl-smi">state_size</span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L125" class="blob-num js-line-number" data-line-number="125"></td> <td id="LC125" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Create tensors of zeros based on state_size, batch_size, and dtype.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L126" class="blob-num js-line-number" data-line-number="126"></td> <td id="LC126" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">get_state_shape</span>(<span class="pl-smi">s</span>):</td> </tr> <tr> <td id="L127" class="blob-num js-line-number" data-line-number="127"></td> <td id="LC127" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Combine s with batch_size to get a proper tensor shape.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L128" class="blob-num js-line-number" data-line-number="128"></td> <td id="LC128" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> _concat(batch_size, s)</td> </tr> <tr> <td id="L129" class="blob-num js-line-number" data-line-number="129"></td> <td id="LC129" class="blob-code blob-code-inner js-file-line"> size <span class="pl-k">=</span> array_ops.zeros(c, <span class="pl-v">dtype</span><span class="pl-k">=</span>dtype)</td> </tr> <tr> <td id="L130" class="blob-num js-line-number" data-line-number="130"></td> <td id="LC130" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> context.in_graph_mode():</td> </tr> <tr> <td id="L131" class="blob-num js-line-number" data-line-number="131"></td> <td id="LC131" class="blob-code blob-code-inner js-file-line"> c_static <span class="pl-k">=</span> _concat(batch_size, s, <span class="pl-v">static</span><span class="pl-k">=</span><span class="pl-c1">True</span>)</td> </tr> <tr> <td id="L132" class="blob-num js-line-number" data-line-number="132"></td> <td id="LC132" class="blob-code blob-code-inner js-file-line"> size.set_shape(c_static)</td> </tr> <tr> <td id="L133" class="blob-num js-line-number" data-line-number="133"></td> <td id="LC133" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> size</td> </tr> <tr> <td id="L134" class="blob-num js-line-number" data-line-number="134"></td> <td id="LC134" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> nest.map_structure(get_state_shape, state_size)</td> </tr> <tr> <td id="L135" class="blob-num js-line-number" data-line-number="135"></td> <td id="LC135" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L136" class="blob-num js-line-number" data-line-number="136"></td> <td id="LC136" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L137" class="blob-num js-line-number" data-line-number="137"></td> <td id="LC137" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.RNNCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L138" class="blob-num js-line-number" data-line-number="138"></td> <td id="LC138" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">RNNCell</span>(<span class="pl-e">base_layer</span>.<span class="pl-e">Layer</span>):</td> </tr> <tr> <td id="L139" class="blob-num js-line-number" data-line-number="139"></td> <td id="LC139" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Abstract object representing an RNN cell.</span></td> </tr> <tr> <td id="L140" class="blob-num js-line-number" data-line-number="140"></td> <td id="LC140" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L141" class="blob-num js-line-number" data-line-number="141"></td> <td id="LC141" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Every `RNNCell` must have the properties below and implement `call` with</span></td> </tr> <tr> <td id="L142" class="blob-num js-line-number" data-line-number="142"></td> <td id="LC142" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the signature `(output, next_state) = call(input, state)`. The optional</span></td> </tr> <tr> <td id="L143" class="blob-num js-line-number" data-line-number="143"></td> <td id="LC143" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> third input argument, `scope`, is allowed for backwards compatibility</span></td> </tr> <tr> <td id="L144" class="blob-num js-line-number" data-line-number="144"></td> <td id="LC144" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> purposes; but should be left off for new subclasses.</span></td> </tr> <tr> <td id="L145" class="blob-num js-line-number" data-line-number="145"></td> <td id="LC145" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L146" class="blob-num js-line-number" data-line-number="146"></td> <td id="LC146" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> This definition of cell differs from the definition used in the literature.</span></td> </tr> <tr> <td id="L147" class="blob-num js-line-number" data-line-number="147"></td> <td id="LC147" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> In the literature, &#39;cell&#39; refers to an object with a single scalar output.</span></td> </tr> <tr> <td id="L148" class="blob-num js-line-number" data-line-number="148"></td> <td id="LC148" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> This definition refers to a horizontal array of such units.</span></td> </tr> <tr> <td id="L149" class="blob-num js-line-number" data-line-number="149"></td> <td id="LC149" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L150" class="blob-num js-line-number" data-line-number="150"></td> <td id="LC150" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> An RNN cell, in the most abstract setting, is anything that has</span></td> </tr> <tr> <td id="L151" class="blob-num js-line-number" data-line-number="151"></td> <td id="LC151" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> a state and performs some operation that takes a matrix of inputs.</span></td> </tr> <tr> <td id="L152" class="blob-num js-line-number" data-line-number="152"></td> <td id="LC152" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> This operation results in an output matrix with `self.output_size` columns.</span></td> </tr> <tr> <td id="L153" class="blob-num js-line-number" data-line-number="153"></td> <td id="LC153" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If `self.state_size` is an integer, this operation also results in a new</span></td> </tr> <tr> <td id="L154" class="blob-num js-line-number" data-line-number="154"></td> <td id="LC154" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state matrix with `self.state_size` columns. If `self.state_size` is a</span></td> </tr> <tr> <td id="L155" class="blob-num js-line-number" data-line-number="155"></td> <td id="LC155" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> (possibly nested tuple of) TensorShape object(s), then it should return a</span></td> </tr> <tr> <td id="L156" class="blob-num js-line-number" data-line-number="156"></td> <td id="LC156" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> matching structure of Tensors having shape `[batch_size].concatenate(s)`</span></td> </tr> <tr> <td id="L157" class="blob-num js-line-number" data-line-number="157"></td> <td id="LC157" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> for each `s` in `self.batch_size`.</span></td> </tr> <tr> <td id="L158" class="blob-num js-line-number" data-line-number="158"></td> <td id="LC158" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L159" class="blob-num js-line-number" data-line-number="159"></td> <td id="LC159" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L160" class="blob-num js-line-number" data-line-number="160"></td> <td id="LC160" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L161" class="blob-num js-line-number" data-line-number="161"></td> <td id="LC161" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run this RNN cell on inputs, starting from the given state.</span></td> </tr> <tr> <td id="L162" class="blob-num js-line-number" data-line-number="162"></td> <td id="LC162" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L163" class="blob-num js-line-number" data-line-number="163"></td> <td id="LC163" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L164" class="blob-num js-line-number" data-line-number="164"></td> <td id="LC164" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: `2-D` tensor with shape `[batch_size, input_size]`.</span></td> </tr> <tr> <td id="L165" class="blob-num js-line-number" data-line-number="165"></td> <td id="LC165" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: if `self.state_size` is an integer, this should be a `2-D Tensor`</span></td> </tr> <tr> <td id="L166" class="blob-num js-line-number" data-line-number="166"></td> <td id="LC166" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> with shape `[batch_size, self.state_size]`. Otherwise, if</span></td> </tr> <tr> <td id="L167" class="blob-num js-line-number" data-line-number="167"></td> <td id="LC167" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `self.state_size` is a tuple of integers, this should be a tuple</span></td> </tr> <tr> <td id="L168" class="blob-num js-line-number" data-line-number="168"></td> <td id="LC168" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> with shapes `[batch_size, s] for s in self.state_size`.</span></td> </tr> <tr> <td id="L169" class="blob-num js-line-number" data-line-number="169"></td> <td id="LC169" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> scope: VariableScope for the created subgraph; defaults to class name.</span></td> </tr> <tr> <td id="L170" class="blob-num js-line-number" data-line-number="170"></td> <td id="LC170" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L171" class="blob-num js-line-number" data-line-number="171"></td> <td id="LC171" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L172" class="blob-num js-line-number" data-line-number="172"></td> <td id="LC172" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> A pair containing:</span></td> </tr> <tr> <td id="L173" class="blob-num js-line-number" data-line-number="173"></td> <td id="LC173" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L174" class="blob-num js-line-number" data-line-number="174"></td> <td id="LC174" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - Output: A `2-D` tensor with shape `[batch_size, self.output_size]`.</span></td> </tr> <tr> <td id="L175" class="blob-num js-line-number" data-line-number="175"></td> <td id="LC175" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - New state: Either a single `2-D` tensor, or a tuple of tensors matching</span></td> </tr> <tr> <td id="L176" class="blob-num js-line-number" data-line-number="176"></td> <td id="LC176" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the arity and shapes of `state`.</span></td> </tr> <tr> <td id="L177" class="blob-num js-line-number" data-line-number="177"></td> <td id="LC177" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L178" class="blob-num js-line-number" data-line-number="178"></td> <td id="LC178" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> scope <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L179" class="blob-num js-line-number" data-line-number="179"></td> <td id="LC179" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> vs.variable_scope(scope,</td> </tr> <tr> <td id="L180" class="blob-num js-line-number" data-line-number="180"></td> <td id="LC180" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">custom_getter</span><span class="pl-k">=</span><span class="pl-c1">self</span>._rnn_get_variable) <span class="pl-k">as</span> scope:</td> </tr> <tr> <td id="L181" class="blob-num js-line-number" data-line-number="181"></td> <td id="LC181" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">super</span>(RNNCell, <span class="pl-c1">self</span>).<span class="pl-c1">__call__</span>(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L182" class="blob-num js-line-number" data-line-number="182"></td> <td id="LC182" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L183" class="blob-num js-line-number" data-line-number="183"></td> <td id="LC183" class="blob-code blob-code-inner js-file-line"> scope_attrname <span class="pl-k">=</span> <span class="pl-s"><span class="pl-pds">&quot;</span>rnncell_scope<span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L184" class="blob-num js-line-number" data-line-number="184"></td> <td id="LC184" class="blob-code blob-code-inner js-file-line"> scope <span class="pl-k">=</span> <span class="pl-c1">getattr</span>(<span class="pl-c1">self</span>, scope_attrname, <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L185" class="blob-num js-line-number" data-line-number="185"></td> <td id="LC185" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> scope <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L186" class="blob-num js-line-number" data-line-number="186"></td> <td id="LC186" class="blob-code blob-code-inner js-file-line"> scope <span class="pl-k">=</span> vs.variable_scope(vs.get_variable_scope(),</td> </tr> <tr> <td id="L187" class="blob-num js-line-number" data-line-number="187"></td> <td id="LC187" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">custom_getter</span><span class="pl-k">=</span><span class="pl-c1">self</span>._rnn_get_variable)</td> </tr> <tr> <td id="L188" class="blob-num js-line-number" data-line-number="188"></td> <td id="LC188" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">setattr</span>(<span class="pl-c1">self</span>, scope_attrname, scope)</td> </tr> <tr> <td id="L189" class="blob-num js-line-number" data-line-number="189"></td> <td id="LC189" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> scope:</td> </tr> <tr> <td id="L190" class="blob-num js-line-number" data-line-number="190"></td> <td id="LC190" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">super</span>(RNNCell, <span class="pl-c1">self</span>).<span class="pl-c1">__call__</span>(inputs, state)</td> </tr> <tr> <td id="L191" class="blob-num js-line-number" data-line-number="191"></td> <td id="LC191" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L192" class="blob-num js-line-number" data-line-number="192"></td> <td id="LC192" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_rnn_get_variable</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">getter</span>, <span class="pl-k">*</span><span class="pl-smi">args</span>, <span class="pl-k">**</span><span class="pl-smi">kwargs</span>):</td> </tr> <tr> <td id="L193" class="blob-num js-line-number" data-line-number="193"></td> <td id="LC193" class="blob-code blob-code-inner js-file-line"> variable <span class="pl-k">=</span> getter(<span class="pl-k">*</span>args, <span class="pl-k">**</span>kwargs)</td> </tr> <tr> <td id="L194" class="blob-num js-line-number" data-line-number="194"></td> <td id="LC194" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> context.in_graph_mode():</td> </tr> <tr> <td id="L195" class="blob-num js-line-number" data-line-number="195"></td> <td id="LC195" class="blob-code blob-code-inner js-file-line"> trainable <span class="pl-k">=</span> (variable <span class="pl-k">in</span> tf_variables.trainable_variables() <span class="pl-k">or</span></td> </tr> <tr> <td id="L196" class="blob-num js-line-number" data-line-number="196"></td> <td id="LC196" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">isinstance</span>(variable, tf_variables.PartitionedVariable) <span class="pl-k">and</span></td> </tr> <tr> <td id="L197" class="blob-num js-line-number" data-line-number="197"></td> <td id="LC197" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">list</span>(variable)[<span class="pl-c1">0</span>] <span class="pl-k">in</span> tf_variables.trainable_variables()))</td> </tr> <tr> <td id="L198" class="blob-num js-line-number" data-line-number="198"></td> <td id="LC198" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L199" class="blob-num js-line-number" data-line-number="199"></td> <td id="LC199" class="blob-code blob-code-inner js-file-line"> trainable <span class="pl-k">=</span> variable._trainable <span class="pl-c"><span class="pl-c">#</span> pylint: disable=protected-access</span></td> </tr> <tr> <td id="L200" class="blob-num js-line-number" data-line-number="200"></td> <td id="LC200" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> trainable <span class="pl-k">and</span> variable <span class="pl-k">not</span> <span class="pl-k">in</span> <span class="pl-c1">self</span>._trainable_weights:</td> </tr> <tr> <td id="L201" class="blob-num js-line-number" data-line-number="201"></td> <td id="LC201" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._trainable_weights.append(variable)</td> </tr> <tr> <td id="L202" class="blob-num js-line-number" data-line-number="202"></td> <td id="LC202" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">elif</span> <span class="pl-k">not</span> trainable <span class="pl-k">and</span> variable <span class="pl-k">not</span> <span class="pl-k">in</span> <span class="pl-c1">self</span>._non_trainable_weights:</td> </tr> <tr> <td id="L203" class="blob-num js-line-number" data-line-number="203"></td> <td id="LC203" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._non_trainable_weights.append(variable)</td> </tr> <tr> <td id="L204" class="blob-num js-line-number" data-line-number="204"></td> <td id="LC204" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> variable</td> </tr> <tr> <td id="L205" class="blob-num js-line-number" data-line-number="205"></td> <td id="LC205" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L206" class="blob-num js-line-number" data-line-number="206"></td> <td id="LC206" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L207" class="blob-num js-line-number" data-line-number="207"></td> <td id="LC207" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L208" class="blob-num js-line-number" data-line-number="208"></td> <td id="LC208" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>size(s) of state(s) used by this cell.</span></td> </tr> <tr> <td id="L209" class="blob-num js-line-number" data-line-number="209"></td> <td id="LC209" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L210" class="blob-num js-line-number" data-line-number="210"></td> <td id="LC210" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> It can be represented by an Integer, a TensorShape or a tuple of Integers</span></td> </tr> <tr> <td id="L211" class="blob-num js-line-number" data-line-number="211"></td> <td id="LC211" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> or TensorShapes.</span></td> </tr> <tr> <td id="L212" class="blob-num js-line-number" data-line-number="212"></td> <td id="LC212" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L213" class="blob-num js-line-number" data-line-number="213"></td> <td id="LC213" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">NotImplementedError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Abstract method<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L214" class="blob-num js-line-number" data-line-number="214"></td> <td id="LC214" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L215" class="blob-num js-line-number" data-line-number="215"></td> <td id="LC215" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L216" class="blob-num js-line-number" data-line-number="216"></td> <td id="LC216" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L217" class="blob-num js-line-number" data-line-number="217"></td> <td id="LC217" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Integer or TensorShape: size of outputs produced by this cell.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L218" class="blob-num js-line-number" data-line-number="218"></td> <td id="LC218" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">NotImplementedError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Abstract method<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L219" class="blob-num js-line-number" data-line-number="219"></td> <td id="LC219" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L220" class="blob-num js-line-number" data-line-number="220"></td> <td id="LC220" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">_</span>):</td> </tr> <tr> <td id="L221" class="blob-num js-line-number" data-line-number="221"></td> <td id="LC221" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> This tells the parent Layer object that it&#39;s OK to call</span></td> </tr> <tr> <td id="L222" class="blob-num js-line-number" data-line-number="222"></td> <td id="LC222" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> self.add_variable() inside the call() method.</span></td> </tr> <tr> <td id="L223" class="blob-num js-line-number" data-line-number="223"></td> <td id="LC223" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">pass</span></td> </tr> <tr> <td id="L224" class="blob-num js-line-number" data-line-number="224"></td> <td id="LC224" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L225" class="blob-num js-line-number" data-line-number="225"></td> <td id="LC225" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L226" class="blob-num js-line-number" data-line-number="226"></td> <td id="LC226" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Return zero-filled state tensor(s).</span></td> </tr> <tr> <td id="L227" class="blob-num js-line-number" data-line-number="227"></td> <td id="LC227" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L228" class="blob-num js-line-number" data-line-number="228"></td> <td id="LC228" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L229" class="blob-num js-line-number" data-line-number="229"></td> <td id="LC229" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> batch_size: int, float, or unit Tensor representing the batch size.</span></td> </tr> <tr> <td id="L230" class="blob-num js-line-number" data-line-number="230"></td> <td id="LC230" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> dtype: the data type to use for the state.</span></td> </tr> <tr> <td id="L231" class="blob-num js-line-number" data-line-number="231"></td> <td id="LC231" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L232" class="blob-num js-line-number" data-line-number="232"></td> <td id="LC232" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L233" class="blob-num js-line-number" data-line-number="233"></td> <td id="LC233" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If `state_size` is an int or TensorShape, then the return value is a</span></td> </tr> <tr> <td id="L234" class="blob-num js-line-number" data-line-number="234"></td> <td id="LC234" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `N-D` tensor of shape `[batch_size, state_size]` filled with zeros.</span></td> </tr> <tr> <td id="L235" class="blob-num js-line-number" data-line-number="235"></td> <td id="LC235" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L236" class="blob-num js-line-number" data-line-number="236"></td> <td id="LC236" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If `state_size` is a nested list or tuple, then the return value is</span></td> </tr> <tr> <td id="L237" class="blob-num js-line-number" data-line-number="237"></td> <td id="LC237" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> a nested list or tuple (of the same structure) of `2-D` tensors with</span></td> </tr> <tr> <td id="L238" class="blob-num js-line-number" data-line-number="238"></td> <td id="LC238" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the shapes `[batch_size, s]` for each s in `state_size`.</span></td> </tr> <tr> <td id="L239" class="blob-num js-line-number" data-line-number="239"></td> <td id="LC239" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L240" class="blob-num js-line-number" data-line-number="240"></td> <td id="LC240" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Try to use the last cached zero_state. This is done to avoid recreating</span></td> </tr> <tr> <td id="L241" class="blob-num js-line-number" data-line-number="241"></td> <td id="LC241" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> zeros, especially when eager execution is enabled.</span></td> </tr> <tr> <td id="L242" class="blob-num js-line-number" data-line-number="242"></td> <td id="LC242" class="blob-code blob-code-inner js-file-line"> state_size <span class="pl-k">=</span> <span class="pl-c1">self</span>.state_size</td> </tr> <tr> <td id="L243" class="blob-num js-line-number" data-line-number="243"></td> <td id="LC243" class="blob-code blob-code-inner js-file-line"> is_eager <span class="pl-k">=</span> context.in_eager_mode()</td> </tr> <tr> <td id="L244" class="blob-num js-line-number" data-line-number="244"></td> <td id="LC244" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> is_eager <span class="pl-k">and</span> <span class="pl-c1">hasattr</span>(<span class="pl-c1">self</span>, <span class="pl-s"><span class="pl-pds">&quot;</span>_last_zero_state<span class="pl-pds">&quot;</span></span>):</td> </tr> <tr> <td id="L245" class="blob-num js-line-number" data-line-number="245"></td> <td id="LC245" class="blob-code blob-code-inner js-file-line"> (last_state_size, last_batch_size, last_dtype,</td> </tr> <tr> <td id="L246" class="blob-num js-line-number" data-line-number="246"></td> <td id="LC246" class="blob-code blob-code-inner js-file-line"> last_output) <span class="pl-k">=</span> <span class="pl-c1">getattr</span>(<span class="pl-c1">self</span>, <span class="pl-s"><span class="pl-pds">&quot;</span>_last_zero_state<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L247" class="blob-num js-line-number" data-line-number="247"></td> <td id="LC247" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> (last_batch_size <span class="pl-k">==</span> batch_size <span class="pl-k">and</span></td> </tr> <tr> <td id="L248" class="blob-num js-line-number" data-line-number="248"></td> <td id="LC248" class="blob-code blob-code-inner js-file-line"> last_dtype <span class="pl-k">==</span> dtype <span class="pl-k">and</span></td> </tr> <tr> <td id="L249" class="blob-num js-line-number" data-line-number="249"></td> <td id="LC249" class="blob-code blob-code-inner js-file-line"> last_state_size <span class="pl-k">==</span> state_size):</td> </tr> <tr> <td id="L250" class="blob-num js-line-number" data-line-number="250"></td> <td id="LC250" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> last_output</td> </tr> <tr> <td id="L251" class="blob-num js-line-number" data-line-number="251"></td> <td id="LC251" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L252" class="blob-num js-line-number" data-line-number="252"></td> <td id="LC252" class="blob-code blob-code-inner js-file-line"> output <span class="pl-k">=</span> _zero_state_tensors(state_size, batch_size, dtype)</td> </tr> <tr> <td id="L253" class="blob-num js-line-number" data-line-number="253"></td> <td id="LC253" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> is_eager:</td> </tr> <tr> <td id="L254" class="blob-num js-line-number" data-line-number="254"></td> <td id="LC254" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._last_zero_state <span class="pl-k">=</span> (state_size, batch_size, dtype, output)</td> </tr> <tr> <td id="L255" class="blob-num js-line-number" data-line-number="255"></td> <td id="LC255" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> output</td> </tr> <tr> <td id="L256" class="blob-num js-line-number" data-line-number="256"></td> <td id="LC256" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L257" class="blob-num js-line-number" data-line-number="257"></td> <td id="LC257" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L258" class="blob-num js-line-number" data-line-number="258"></td> <td id="LC258" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">_LayerRNNCell</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L259" class="blob-num js-line-number" data-line-number="259"></td> <td id="LC259" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Subclass of RNNCells that act like proper `tf.Layer` objects.</span></td> </tr> <tr> <td id="L260" class="blob-num js-line-number" data-line-number="260"></td> <td id="LC260" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L261" class="blob-num js-line-number" data-line-number="261"></td> <td id="LC261" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> For backwards compatibility purposes, most `RNNCell` instances allow their</span></td> </tr> <tr> <td id="L262" class="blob-num js-line-number" data-line-number="262"></td> <td id="LC262" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `call` methods to instantiate variables via `tf.get_variable`. The underlying</span></td> </tr> <tr> <td id="L263" class="blob-num js-line-number" data-line-number="263"></td> <td id="LC263" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> variable scope thus keeps track of any variables, and returning cached</span></td> </tr> <tr> <td id="L264" class="blob-num js-line-number" data-line-number="264"></td> <td id="LC264" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> versions. This is atypical of `tf.layer` objects, which separate this</span></td> </tr> <tr> <td id="L265" class="blob-num js-line-number" data-line-number="265"></td> <td id="LC265" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> part of layer building into a `build` method that is only called once.</span></td> </tr> <tr> <td id="L266" class="blob-num js-line-number" data-line-number="266"></td> <td id="LC266" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L267" class="blob-num js-line-number" data-line-number="267"></td> <td id="LC267" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Here we provide a subclass for `RNNCell` objects that act exactly as</span></td> </tr> <tr> <td id="L268" class="blob-num js-line-number" data-line-number="268"></td> <td id="LC268" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `Layer` objects do. They must provide a `build` method and their</span></td> </tr> <tr> <td id="L269" class="blob-num js-line-number" data-line-number="269"></td> <td id="LC269" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `call` methods do not access Variables `tf.get_variable`.</span></td> </tr> <tr> <td id="L270" class="blob-num js-line-number" data-line-number="270"></td> <td id="LC270" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L271" class="blob-num js-line-number" data-line-number="271"></td> <td id="LC271" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L272" class="blob-num js-line-number" data-line-number="272"></td> <td id="LC272" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-k">*</span><span class="pl-smi">args</span>, <span class="pl-k">**</span><span class="pl-smi">kwargs</span>):</td> </tr> <tr> <td id="L273" class="blob-num js-line-number" data-line-number="273"></td> <td id="LC273" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run this RNN cell on inputs, starting from the given state.</span></td> </tr> <tr> <td id="L274" class="blob-num js-line-number" data-line-number="274"></td> <td id="LC274" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L275" class="blob-num js-line-number" data-line-number="275"></td> <td id="LC275" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L276" class="blob-num js-line-number" data-line-number="276"></td> <td id="LC276" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: `2-D` tensor with shape `[batch_size, input_size]`.</span></td> </tr> <tr> <td id="L277" class="blob-num js-line-number" data-line-number="277"></td> <td id="LC277" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: if `self.state_size` is an integer, this should be a `2-D Tensor`</span></td> </tr> <tr> <td id="L278" class="blob-num js-line-number" data-line-number="278"></td> <td id="LC278" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> with shape `[batch_size, self.state_size]`. Otherwise, if</span></td> </tr> <tr> <td id="L279" class="blob-num js-line-number" data-line-number="279"></td> <td id="LC279" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `self.state_size` is a tuple of integers, this should be a tuple</span></td> </tr> <tr> <td id="L280" class="blob-num js-line-number" data-line-number="280"></td> <td id="LC280" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> with shapes `[batch_size, s] for s in self.state_size`.</span></td> </tr> <tr> <td id="L281" class="blob-num js-line-number" data-line-number="281"></td> <td id="LC281" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> scope: optional cell scope.</span></td> </tr> <tr> <td id="L282" class="blob-num js-line-number" data-line-number="282"></td> <td id="LC282" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> *args: Additional positional arguments.</span></td> </tr> <tr> <td id="L283" class="blob-num js-line-number" data-line-number="283"></td> <td id="LC283" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> **kwargs: Additional keyword arguments.</span></td> </tr> <tr> <td id="L284" class="blob-num js-line-number" data-line-number="284"></td> <td id="LC284" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L285" class="blob-num js-line-number" data-line-number="285"></td> <td id="LC285" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L286" class="blob-num js-line-number" data-line-number="286"></td> <td id="LC286" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> A pair containing:</span></td> </tr> <tr> <td id="L287" class="blob-num js-line-number" data-line-number="287"></td> <td id="LC287" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L288" class="blob-num js-line-number" data-line-number="288"></td> <td id="LC288" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - Output: A `2-D` tensor with shape `[batch_size, self.output_size]`.</span></td> </tr> <tr> <td id="L289" class="blob-num js-line-number" data-line-number="289"></td> <td id="LC289" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - New state: Either a single `2-D` tensor, or a tuple of tensors matching</span></td> </tr> <tr> <td id="L290" class="blob-num js-line-number" data-line-number="290"></td> <td id="LC290" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the arity and shapes of `state`.</span></td> </tr> <tr> <td id="L291" class="blob-num js-line-number" data-line-number="291"></td> <td id="LC291" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L292" class="blob-num js-line-number" data-line-number="292"></td> <td id="LC292" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Bypass RNNCell&#39;s variable capturing semantics for LayerRNNCell.</span></td> </tr> <tr> <td id="L293" class="blob-num js-line-number" data-line-number="293"></td> <td id="LC293" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Instead, it is up to subclasses to provide a proper build</span></td> </tr> <tr> <td id="L294" class="blob-num js-line-number" data-line-number="294"></td> <td id="LC294" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> method. See the class docstring for more details.</span></td> </tr> <tr> <td id="L295" class="blob-num js-line-number" data-line-number="295"></td> <td id="LC295" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> base_layer.Layer.<span class="pl-c1">__call__</span>(<span class="pl-c1">self</span>, inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope,</td> </tr> <tr> <td id="L296" class="blob-num js-line-number" data-line-number="296"></td> <td id="LC296" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">*</span>args, <span class="pl-k">**</span>kwargs)</td> </tr> <tr> <td id="L297" class="blob-num js-line-number" data-line-number="297"></td> <td id="LC297" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L298" class="blob-num js-line-number" data-line-number="298"></td> <td id="LC298" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L299" class="blob-num js-line-number" data-line-number="299"></td> <td id="LC299" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.BasicRNNCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L300" class="blob-num js-line-number" data-line-number="300"></td> <td id="LC300" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">BasicRNNCell</span>(<span class="pl-e">_LayerRNNCell</span>):</td> </tr> <tr> <td id="L301" class="blob-num js-line-number" data-line-number="301"></td> <td id="LC301" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>The most basic RNN cell.</span></td> </tr> <tr> <td id="L302" class="blob-num js-line-number" data-line-number="302"></td> <td id="LC302" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L303" class="blob-num js-line-number" data-line-number="303"></td> <td id="LC303" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L304" class="blob-num js-line-number" data-line-number="304"></td> <td id="LC304" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units: int, The number of units in the RNN cell.</span></td> </tr> <tr> <td id="L305" class="blob-num js-line-number" data-line-number="305"></td> <td id="LC305" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> activation: Nonlinearity to use. Default: `tanh`.</span></td> </tr> <tr> <td id="L306" class="blob-num js-line-number" data-line-number="306"></td> <td id="LC306" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reuse: (optional) Python boolean describing whether to reuse variables</span></td> </tr> <tr> <td id="L307" class="blob-num js-line-number" data-line-number="307"></td> <td id="LC307" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in an existing scope. If not `True`, and the existing scope already has</span></td> </tr> <tr> <td id="L308" class="blob-num js-line-number" data-line-number="308"></td> <td id="LC308" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the given variables, an error is raised.</span></td> </tr> <tr> <td id="L309" class="blob-num js-line-number" data-line-number="309"></td> <td id="LC309" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> name: String, the name of the layer. Layers with the same name will</span></td> </tr> <tr> <td id="L310" class="blob-num js-line-number" data-line-number="310"></td> <td id="LC310" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> share weights, but to avoid mistakes we require reuse=True in such</span></td> </tr> <tr> <td id="L311" class="blob-num js-line-number" data-line-number="311"></td> <td id="LC311" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cases.</span></td> </tr> <tr> <td id="L312" class="blob-num js-line-number" data-line-number="312"></td> <td id="LC312" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L313" class="blob-num js-line-number" data-line-number="313"></td> <td id="LC313" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L314" class="blob-num js-line-number" data-line-number="314"></td> <td id="LC314" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">num_units</span>, <span class="pl-smi">activation</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">reuse</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">name</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L315" class="blob-num js-line-number" data-line-number="315"></td> <td id="LC315" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(BasicRNNCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>(<span class="pl-v">_reuse</span><span class="pl-k">=</span>reuse, <span class="pl-v">name</span><span class="pl-k">=</span>name)</td> </tr> <tr> <td id="L316" class="blob-num js-line-number" data-line-number="316"></td> <td id="LC316" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L317" class="blob-num js-line-number" data-line-number="317"></td> <td id="LC317" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Inputs must be 2-dimensional.</span></td> </tr> <tr> <td id="L318" class="blob-num js-line-number" data-line-number="318"></td> <td id="LC318" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.input_spec <span class="pl-k">=</span> base_layer.InputSpec(<span class="pl-v">ndim</span><span class="pl-k">=</span><span class="pl-c1">2</span>)</td> </tr> <tr> <td id="L319" class="blob-num js-line-number" data-line-number="319"></td> <td id="LC319" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L320" class="blob-num js-line-number" data-line-number="320"></td> <td id="LC320" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_units <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L321" class="blob-num js-line-number" data-line-number="321"></td> <td id="LC321" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation <span class="pl-k">=</span> activation <span class="pl-k">or</span> math_ops.tanh</td> </tr> <tr> <td id="L322" class="blob-num js-line-number" data-line-number="322"></td> <td id="LC322" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L323" class="blob-num js-line-number" data-line-number="323"></td> <td id="LC323" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L324" class="blob-num js-line-number" data-line-number="324"></td> <td id="LC324" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L325" class="blob-num js-line-number" data-line-number="325"></td> <td id="LC325" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L326" class="blob-num js-line-number" data-line-number="326"></td> <td id="LC326" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L327" class="blob-num js-line-number" data-line-number="327"></td> <td id="LC327" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L328" class="blob-num js-line-number" data-line-number="328"></td> <td id="LC328" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L329" class="blob-num js-line-number" data-line-number="329"></td> <td id="LC329" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L330" class="blob-num js-line-number" data-line-number="330"></td> <td id="LC330" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L331" class="blob-num js-line-number" data-line-number="331"></td> <td id="LC331" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs_shape</span>):</td> </tr> <tr> <td id="L332" class="blob-num js-line-number" data-line-number="332"></td> <td id="LC332" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> inputs_shape[<span class="pl-c1">1</span>].value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L333" class="blob-num js-line-number" data-line-number="333"></td> <td id="LC333" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Expected inputs.shape[-1] to be known, saw shape: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L334" class="blob-num js-line-number" data-line-number="334"></td> <td id="LC334" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> inputs_shape)</td> </tr> <tr> <td id="L335" class="blob-num js-line-number" data-line-number="335"></td> <td id="LC335" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L336" class="blob-num js-line-number" data-line-number="336"></td> <td id="LC336" class="blob-code blob-code-inner js-file-line"> input_depth <span class="pl-k">=</span> inputs_shape[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L337" class="blob-num js-line-number" data-line-number="337"></td> <td id="LC337" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L338" class="blob-num js-line-number" data-line-number="338"></td> <td id="LC338" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L339" class="blob-num js-line-number" data-line-number="339"></td> <td id="LC339" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> <span class="pl-c1">self</span>._num_units, <span class="pl-c1">self</span>._num_units])</td> </tr> <tr> <td id="L340" class="blob-num js-line-number" data-line-number="340"></td> <td id="LC340" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L341" class="blob-num js-line-number" data-line-number="341"></td> <td id="LC341" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L342" class="blob-num js-line-number" data-line-number="342"></td> <td id="LC342" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L343" class="blob-num js-line-number" data-line-number="343"></td> <td id="LC343" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>init_ops.zeros_initializer(<span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype))</td> </tr> <tr> <td id="L344" class="blob-num js-line-number" data-line-number="344"></td> <td id="LC344" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L345" class="blob-num js-line-number" data-line-number="345"></td> <td id="LC345" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.built <span class="pl-k">=</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L346" class="blob-num js-line-number" data-line-number="346"></td> <td id="LC346" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L347" class="blob-num js-line-number" data-line-number="347"></td> <td id="LC347" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L348" class="blob-num js-line-number" data-line-number="348"></td> <td id="LC348" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Most basic RNN: output = new_state = act(W * input + U * state + B).<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L349" class="blob-num js-line-number" data-line-number="349"></td> <td id="LC349" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L350" class="blob-num js-line-number" data-line-number="350"></td> <td id="LC350" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L351" class="blob-num js-line-number" data-line-number="351"></td> <td id="LC351" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, state], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._kernel)</td> </tr> <tr> <td id="L352" class="blob-num js-line-number" data-line-number="352"></td> <td id="LC352" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> nn_ops.bias_add(gate_inputs, <span class="pl-c1">self</span>._bias)</td> </tr> <tr> <td id="L353" class="blob-num js-line-number" data-line-number="353"></td> <td id="LC353" class="blob-code blob-code-inner js-file-line"> output <span class="pl-k">=</span> <span class="pl-c1">self</span>._activation(gate_inputs)</td> </tr> <tr> <td id="L354" class="blob-num js-line-number" data-line-number="354"></td> <td id="LC354" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> output, output</td> </tr> <tr> <td id="L355" class="blob-num js-line-number" data-line-number="355"></td> <td id="LC355" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L356" class="blob-num js-line-number" data-line-number="356"></td> <td id="LC356" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L357" class="blob-num js-line-number" data-line-number="357"></td> <td id="LC357" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.GRUCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L358" class="blob-num js-line-number" data-line-number="358"></td> <td id="LC358" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">GRUCell</span>(<span class="pl-e">_LayerRNNCell</span>):</td> </tr> <tr> <td id="L359" class="blob-num js-line-number" data-line-number="359"></td> <td id="LC359" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Gated Recurrent Unit cell (cf. http://arxiv.org/abs/1406.1078).</span></td> </tr> <tr> <td id="L360" class="blob-num js-line-number" data-line-number="360"></td> <td id="LC360" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L361" class="blob-num js-line-number" data-line-number="361"></td> <td id="LC361" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L362" class="blob-num js-line-number" data-line-number="362"></td> <td id="LC362" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units: int, The number of units in the GRU cell.</span></td> </tr> <tr> <td id="L363" class="blob-num js-line-number" data-line-number="363"></td> <td id="LC363" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> activation: Nonlinearity to use. Default: `tanh`.</span></td> </tr> <tr> <td id="L364" class="blob-num js-line-number" data-line-number="364"></td> <td id="LC364" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reuse: (optional) Python boolean describing whether to reuse variables</span></td> </tr> <tr> <td id="L365" class="blob-num js-line-number" data-line-number="365"></td> <td id="LC365" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in an existing scope. If not `True`, and the existing scope already has</span></td> </tr> <tr> <td id="L366" class="blob-num js-line-number" data-line-number="366"></td> <td id="LC366" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the given variables, an error is raised.</span></td> </tr> <tr> <td id="L367" class="blob-num js-line-number" data-line-number="367"></td> <td id="LC367" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> kernel_initializer: (optional) The initializer to use for the weight and</span></td> </tr> <tr> <td id="L368" class="blob-num js-line-number" data-line-number="368"></td> <td id="LC368" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> projection matrices.</span></td> </tr> <tr> <td id="L369" class="blob-num js-line-number" data-line-number="369"></td> <td id="LC369" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> bias_initializer: (optional) The initializer to use for the bias.</span></td> </tr> <tr> <td id="L370" class="blob-num js-line-number" data-line-number="370"></td> <td id="LC370" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> name: String, the name of the layer. Layers with the same name will</span></td> </tr> <tr> <td id="L371" class="blob-num js-line-number" data-line-number="371"></td> <td id="LC371" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> share weights, but to avoid mistakes we require reuse=True in such</span></td> </tr> <tr> <td id="L372" class="blob-num js-line-number" data-line-number="372"></td> <td id="LC372" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cases.</span></td> </tr> <tr> <td id="L373" class="blob-num js-line-number" data-line-number="373"></td> <td id="LC373" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L374" class="blob-num js-line-number" data-line-number="374"></td> <td id="LC374" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L375" class="blob-num js-line-number" data-line-number="375"></td> <td id="LC375" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>,</td> </tr> <tr> <td id="L376" class="blob-num js-line-number" data-line-number="376"></td> <td id="LC376" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">num_units</span>,</td> </tr> <tr> <td id="L377" class="blob-num js-line-number" data-line-number="377"></td> <td id="LC377" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">activation</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L378" class="blob-num js-line-number" data-line-number="378"></td> <td id="LC378" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">reuse</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L379" class="blob-num js-line-number" data-line-number="379"></td> <td id="LC379" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">kernel_initializer</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L380" class="blob-num js-line-number" data-line-number="380"></td> <td id="LC380" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">bias_initializer</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L381" class="blob-num js-line-number" data-line-number="381"></td> <td id="LC381" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">name</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L382" class="blob-num js-line-number" data-line-number="382"></td> <td id="LC382" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(GRUCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>(<span class="pl-v">_reuse</span><span class="pl-k">=</span>reuse, <span class="pl-v">name</span><span class="pl-k">=</span>name)</td> </tr> <tr> <td id="L383" class="blob-num js-line-number" data-line-number="383"></td> <td id="LC383" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L384" class="blob-num js-line-number" data-line-number="384"></td> <td id="LC384" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Inputs must be 2-dimensional.</span></td> </tr> <tr> <td id="L385" class="blob-num js-line-number" data-line-number="385"></td> <td id="LC385" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.input_spec <span class="pl-k">=</span> base_layer.InputSpec(<span class="pl-v">ndim</span><span class="pl-k">=</span><span class="pl-c1">2</span>)</td> </tr> <tr> <td id="L386" class="blob-num js-line-number" data-line-number="386"></td> <td id="LC386" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L387" class="blob-num js-line-number" data-line-number="387"></td> <td id="LC387" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_units <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L388" class="blob-num js-line-number" data-line-number="388"></td> <td id="LC388" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation <span class="pl-k">=</span> activation <span class="pl-k">or</span> math_ops.tanh</td> </tr> <tr> <td id="L389" class="blob-num js-line-number" data-line-number="389"></td> <td id="LC389" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._kernel_initializer <span class="pl-k">=</span> kernel_initializer</td> </tr> <tr> <td id="L390" class="blob-num js-line-number" data-line-number="390"></td> <td id="LC390" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias_initializer <span class="pl-k">=</span> bias_initializer</td> </tr> <tr> <td id="L391" class="blob-num js-line-number" data-line-number="391"></td> <td id="LC391" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L392" class="blob-num js-line-number" data-line-number="392"></td> <td id="LC392" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L393" class="blob-num js-line-number" data-line-number="393"></td> <td id="LC393" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L394" class="blob-num js-line-number" data-line-number="394"></td> <td id="LC394" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L395" class="blob-num js-line-number" data-line-number="395"></td> <td id="LC395" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L396" class="blob-num js-line-number" data-line-number="396"></td> <td id="LC396" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L397" class="blob-num js-line-number" data-line-number="397"></td> <td id="LC397" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L398" class="blob-num js-line-number" data-line-number="398"></td> <td id="LC398" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L399" class="blob-num js-line-number" data-line-number="399"></td> <td id="LC399" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L400" class="blob-num js-line-number" data-line-number="400"></td> <td id="LC400" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs_shape</span>):</td> </tr> <tr> <td id="L401" class="blob-num js-line-number" data-line-number="401"></td> <td id="LC401" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> inputs_shape[<span class="pl-c1">1</span>].value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L402" class="blob-num js-line-number" data-line-number="402"></td> <td id="LC402" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Expected inputs.shape[-1] to be known, saw shape: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L403" class="blob-num js-line-number" data-line-number="403"></td> <td id="LC403" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> inputs_shape)</td> </tr> <tr> <td id="L404" class="blob-num js-line-number" data-line-number="404"></td> <td id="LC404" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L405" class="blob-num js-line-number" data-line-number="405"></td> <td id="LC405" class="blob-code blob-code-inner js-file-line"> input_depth <span class="pl-k">=</span> inputs_shape[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L406" class="blob-num js-line-number" data-line-number="406"></td> <td id="LC406" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._gate_kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L407" class="blob-num js-line-number" data-line-number="407"></td> <td id="LC407" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>gates/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L408" class="blob-num js-line-number" data-line-number="408"></td> <td id="LC408" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> <span class="pl-c1">self</span>._num_units, <span class="pl-c1">2</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L409" class="blob-num js-line-number" data-line-number="409"></td> <td id="LC409" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._kernel_initializer)</td> </tr> <tr> <td id="L410" class="blob-num js-line-number" data-line-number="410"></td> <td id="LC410" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._gate_bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L411" class="blob-num js-line-number" data-line-number="411"></td> <td id="LC411" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>gates/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L412" class="blob-num js-line-number" data-line-number="412"></td> <td id="LC412" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">2</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L413" class="blob-num js-line-number" data-line-number="413"></td> <td id="LC413" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>(</td> </tr> <tr> <td id="L414" class="blob-num js-line-number" data-line-number="414"></td> <td id="LC414" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias_initializer</td> </tr> <tr> <td id="L415" class="blob-num js-line-number" data-line-number="415"></td> <td id="LC415" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._bias_initializer <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L416" class="blob-num js-line-number" data-line-number="416"></td> <td id="LC416" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span> init_ops.constant_initializer(<span class="pl-c1">1.0</span>, <span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype)))</td> </tr> <tr> <td id="L417" class="blob-num js-line-number" data-line-number="417"></td> <td id="LC417" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._candidate_kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L418" class="blob-num js-line-number" data-line-number="418"></td> <td id="LC418" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>candidate/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L419" class="blob-num js-line-number" data-line-number="419"></td> <td id="LC419" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> <span class="pl-c1">self</span>._num_units, <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L420" class="blob-num js-line-number" data-line-number="420"></td> <td id="LC420" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._kernel_initializer)</td> </tr> <tr> <td id="L421" class="blob-num js-line-number" data-line-number="421"></td> <td id="LC421" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._candidate_bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L422" class="blob-num js-line-number" data-line-number="422"></td> <td id="LC422" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>candidate/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L423" class="blob-num js-line-number" data-line-number="423"></td> <td id="LC423" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L424" class="blob-num js-line-number" data-line-number="424"></td> <td id="LC424" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>(</td> </tr> <tr> <td id="L425" class="blob-num js-line-number" data-line-number="425"></td> <td id="LC425" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias_initializer</td> </tr> <tr> <td id="L426" class="blob-num js-line-number" data-line-number="426"></td> <td id="LC426" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._bias_initializer <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L427" class="blob-num js-line-number" data-line-number="427"></td> <td id="LC427" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span> init_ops.zeros_initializer(<span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype)))</td> </tr> <tr> <td id="L428" class="blob-num js-line-number" data-line-number="428"></td> <td id="LC428" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L429" class="blob-num js-line-number" data-line-number="429"></td> <td id="LC429" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.built <span class="pl-k">=</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L430" class="blob-num js-line-number" data-line-number="430"></td> <td id="LC430" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L431" class="blob-num js-line-number" data-line-number="431"></td> <td id="LC431" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L432" class="blob-num js-line-number" data-line-number="432"></td> <td id="LC432" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Gated recurrent unit (GRU) with nunits cells.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L433" class="blob-num js-line-number" data-line-number="433"></td> <td id="LC433" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L434" class="blob-num js-line-number" data-line-number="434"></td> <td id="LC434" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L435" class="blob-num js-line-number" data-line-number="435"></td> <td id="LC435" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, state], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._gate_kernel)</td> </tr> <tr> <td id="L436" class="blob-num js-line-number" data-line-number="436"></td> <td id="LC436" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> nn_ops.bias_add(gate_inputs, <span class="pl-c1">self</span>._gate_bias)</td> </tr> <tr> <td id="L437" class="blob-num js-line-number" data-line-number="437"></td> <td id="LC437" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L438" class="blob-num js-line-number" data-line-number="438"></td> <td id="LC438" class="blob-code blob-code-inner js-file-line"> value <span class="pl-k">=</span> math_ops.sigmoid(gate_inputs)</td> </tr> <tr> <td id="L439" class="blob-num js-line-number" data-line-number="439"></td> <td id="LC439" class="blob-code blob-code-inner js-file-line"> r, u <span class="pl-k">=</span> array_ops.split(<span class="pl-v">value</span><span class="pl-k">=</span>value, <span class="pl-v">num_or_size_splits</span><span class="pl-k">=</span><span class="pl-c1">2</span>, <span class="pl-v">axis</span><span class="pl-k">=</span><span class="pl-c1">1</span>)</td> </tr> <tr> <td id="L440" class="blob-num js-line-number" data-line-number="440"></td> <td id="LC440" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L441" class="blob-num js-line-number" data-line-number="441"></td> <td id="LC441" class="blob-code blob-code-inner js-file-line"> r_state <span class="pl-k">=</span> r <span class="pl-k">*</span> state</td> </tr> <tr> <td id="L442" class="blob-num js-line-number" data-line-number="442"></td> <td id="LC442" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L443" class="blob-num js-line-number" data-line-number="443"></td> <td id="LC443" class="blob-code blob-code-inner js-file-line"> candidate <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L444" class="blob-num js-line-number" data-line-number="444"></td> <td id="LC444" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, r_state], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._candidate_kernel)</td> </tr> <tr> <td id="L445" class="blob-num js-line-number" data-line-number="445"></td> <td id="LC445" class="blob-code blob-code-inner js-file-line"> candidate <span class="pl-k">=</span> nn_ops.bias_add(candidate, <span class="pl-c1">self</span>._candidate_bias)</td> </tr> <tr> <td id="L446" class="blob-num js-line-number" data-line-number="446"></td> <td id="LC446" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L447" class="blob-num js-line-number" data-line-number="447"></td> <td id="LC447" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> <span class="pl-c1">self</span>._activation(candidate)</td> </tr> <tr> <td id="L448" class="blob-num js-line-number" data-line-number="448"></td> <td id="LC448" class="blob-code blob-code-inner js-file-line"> new_h <span class="pl-k">=</span> u <span class="pl-k">*</span> state <span class="pl-k">+</span> (<span class="pl-c1">1</span> <span class="pl-k">-</span> u) <span class="pl-k">*</span> c</td> </tr> <tr> <td id="L449" class="blob-num js-line-number" data-line-number="449"></td> <td id="LC449" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> new_h, new_h</td> </tr> <tr> <td id="L450" class="blob-num js-line-number" data-line-number="450"></td> <td id="LC450" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L451" class="blob-num js-line-number" data-line-number="451"></td> <td id="LC451" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L452" class="blob-num js-line-number" data-line-number="452"></td> <td id="LC452" class="blob-code blob-code-inner js-file-line">_LSTMStateTuple <span class="pl-k">=</span> collections.namedtuple(<span class="pl-s"><span class="pl-pds">&quot;</span>LSTMStateTuple<span class="pl-pds">&quot;</span></span>, (<span class="pl-s"><span class="pl-pds">&quot;</span>c<span class="pl-pds">&quot;</span></span>, <span class="pl-s"><span class="pl-pds">&quot;</span>h<span class="pl-pds">&quot;</span></span>))</td> </tr> <tr> <td id="L453" class="blob-num js-line-number" data-line-number="453"></td> <td id="LC453" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L454" class="blob-num js-line-number" data-line-number="454"></td> <td id="LC454" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L455" class="blob-num js-line-number" data-line-number="455"></td> <td id="LC455" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.LSTMStateTuple<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L456" class="blob-num js-line-number" data-line-number="456"></td> <td id="LC456" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">LSTMStateTuple</span>(<span class="pl-e">_LSTMStateTuple</span>):</td> </tr> <tr> <td id="L457" class="blob-num js-line-number" data-line-number="457"></td> <td id="LC457" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Tuple used by LSTM Cells for `state_size`, `zero_state`, and output state.</span></td> </tr> <tr> <td id="L458" class="blob-num js-line-number" data-line-number="458"></td> <td id="LC458" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L459" class="blob-num js-line-number" data-line-number="459"></td> <td id="LC459" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Stores two elements: `(c, h)`, in that order. Where `c` is the hidden state</span></td> </tr> <tr> <td id="L460" class="blob-num js-line-number" data-line-number="460"></td> <td id="LC460" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> and `h` is the output.</span></td> </tr> <tr> <td id="L461" class="blob-num js-line-number" data-line-number="461"></td> <td id="LC461" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L462" class="blob-num js-line-number" data-line-number="462"></td> <td id="LC462" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Only used when `state_is_tuple=True`.</span></td> </tr> <tr> <td id="L463" class="blob-num js-line-number" data-line-number="463"></td> <td id="LC463" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L464" class="blob-num js-line-number" data-line-number="464"></td> <td id="LC464" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">__slots__</span> <span class="pl-k">=</span> ()</td> </tr> <tr> <td id="L465" class="blob-num js-line-number" data-line-number="465"></td> <td id="LC465" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L466" class="blob-num js-line-number" data-line-number="466"></td> <td id="LC466" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L467" class="blob-num js-line-number" data-line-number="467"></td> <td id="LC467" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">dtype</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L468" class="blob-num js-line-number" data-line-number="468"></td> <td id="LC468" class="blob-code blob-code-inner js-file-line"> (c, h) <span class="pl-k">=</span> <span class="pl-c1">self</span></td> </tr> <tr> <td id="L469" class="blob-num js-line-number" data-line-number="469"></td> <td id="LC469" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> c.dtype <span class="pl-k">!=</span> h.dtype:</td> </tr> <tr> <td id="L470" class="blob-num js-line-number" data-line-number="470"></td> <td id="LC470" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Inconsistent internal state: <span class="pl-c1">%s</span> vs <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span></td> </tr> <tr> <td id="L471" class="blob-num js-line-number" data-line-number="471"></td> <td id="LC471" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">str</span>(c.dtype), <span class="pl-c1">str</span>(h.dtype)))</td> </tr> <tr> <td id="L472" class="blob-num js-line-number" data-line-number="472"></td> <td id="LC472" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> c.dtype</td> </tr> <tr> <td id="L473" class="blob-num js-line-number" data-line-number="473"></td> <td id="LC473" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L474" class="blob-num js-line-number" data-line-number="474"></td> <td id="LC474" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L475" class="blob-num js-line-number" data-line-number="475"></td> <td id="LC475" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.BasicLSTMCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L476" class="blob-num js-line-number" data-line-number="476"></td> <td id="LC476" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">BasicLSTMCell</span>(<span class="pl-e">_LayerRNNCell</span>):</td> </tr> <tr> <td id="L477" class="blob-num js-line-number" data-line-number="477"></td> <td id="LC477" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Basic LSTM recurrent network cell.</span></td> </tr> <tr> <td id="L478" class="blob-num js-line-number" data-line-number="478"></td> <td id="LC478" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L479" class="blob-num js-line-number" data-line-number="479"></td> <td id="LC479" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> The implementation is based on: http://arxiv.org/abs/1409.2329.</span></td> </tr> <tr> <td id="L480" class="blob-num js-line-number" data-line-number="480"></td> <td id="LC480" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L481" class="blob-num js-line-number" data-line-number="481"></td> <td id="LC481" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> We add forget_bias (default: 1) to the biases of the forget gate in order to</span></td> </tr> <tr> <td id="L482" class="blob-num js-line-number" data-line-number="482"></td> <td id="LC482" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reduce the scale of forgetting in the beginning of the training.</span></td> </tr> <tr> <td id="L483" class="blob-num js-line-number" data-line-number="483"></td> <td id="LC483" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L484" class="blob-num js-line-number" data-line-number="484"></td> <td id="LC484" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> It does not allow cell clipping, a projection layer, and does not</span></td> </tr> <tr> <td id="L485" class="blob-num js-line-number" data-line-number="485"></td> <td id="LC485" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> use peep-hole connections: it is the basic baseline.</span></td> </tr> <tr> <td id="L486" class="blob-num js-line-number" data-line-number="486"></td> <td id="LC486" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L487" class="blob-num js-line-number" data-line-number="487"></td> <td id="LC487" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> For advanced models, please use the full @{tf.nn.rnn_cell.LSTMCell}</span></td> </tr> <tr> <td id="L488" class="blob-num js-line-number" data-line-number="488"></td> <td id="LC488" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> that follows.</span></td> </tr> <tr> <td id="L489" class="blob-num js-line-number" data-line-number="489"></td> <td id="LC489" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L490" class="blob-num js-line-number" data-line-number="490"></td> <td id="LC490" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L491" class="blob-num js-line-number" data-line-number="491"></td> <td id="LC491" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">num_units</span>, <span class="pl-smi">forget_bias</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>,</td> </tr> <tr> <td id="L492" class="blob-num js-line-number" data-line-number="492"></td> <td id="LC492" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">state_is_tuple</span><span class="pl-k">=</span><span class="pl-c1">True</span>, <span class="pl-smi">activation</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">reuse</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">name</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L493" class="blob-num js-line-number" data-line-number="493"></td> <td id="LC493" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Initialize the basic LSTM cell.</span></td> </tr> <tr> <td id="L494" class="blob-num js-line-number" data-line-number="494"></td> <td id="LC494" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L495" class="blob-num js-line-number" data-line-number="495"></td> <td id="LC495" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L496" class="blob-num js-line-number" data-line-number="496"></td> <td id="LC496" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units: int, The number of units in the LSTM cell.</span></td> </tr> <tr> <td id="L497" class="blob-num js-line-number" data-line-number="497"></td> <td id="LC497" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> forget_bias: float, The bias added to forget gates (see above).</span></td> </tr> <tr> <td id="L498" class="blob-num js-line-number" data-line-number="498"></td> <td id="LC498" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Must set to `0.0` manually when restoring from CudnnLSTM-trained</span></td> </tr> <tr> <td id="L499" class="blob-num js-line-number" data-line-number="499"></td> <td id="LC499" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> checkpoints.</span></td> </tr> <tr> <td id="L500" class="blob-num js-line-number" data-line-number="500"></td> <td id="LC500" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state_is_tuple: If True, accepted and returned states are 2-tuples of</span></td> </tr> <tr> <td id="L501" class="blob-num js-line-number" data-line-number="501"></td> <td id="LC501" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the `c_state` and `m_state`. If False, they are concatenated</span></td> </tr> <tr> <td id="L502" class="blob-num js-line-number" data-line-number="502"></td> <td id="LC502" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> along the column axis. The latter behavior will soon be deprecated.</span></td> </tr> <tr> <td id="L503" class="blob-num js-line-number" data-line-number="503"></td> <td id="LC503" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> activation: Activation function of the inner states. Default: `tanh`.</span></td> </tr> <tr> <td id="L504" class="blob-num js-line-number" data-line-number="504"></td> <td id="LC504" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reuse: (optional) Python boolean describing whether to reuse variables</span></td> </tr> <tr> <td id="L505" class="blob-num js-line-number" data-line-number="505"></td> <td id="LC505" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in an existing scope. If not `True`, and the existing scope already has</span></td> </tr> <tr> <td id="L506" class="blob-num js-line-number" data-line-number="506"></td> <td id="LC506" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the given variables, an error is raised.</span></td> </tr> <tr> <td id="L507" class="blob-num js-line-number" data-line-number="507"></td> <td id="LC507" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> name: String, the name of the layer. Layers with the same name will</span></td> </tr> <tr> <td id="L508" class="blob-num js-line-number" data-line-number="508"></td> <td id="LC508" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> share weights, but to avoid mistakes we require reuse=True in such</span></td> </tr> <tr> <td id="L509" class="blob-num js-line-number" data-line-number="509"></td> <td id="LC509" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cases.</span></td> </tr> <tr> <td id="L510" class="blob-num js-line-number" data-line-number="510"></td> <td id="LC510" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L511" class="blob-num js-line-number" data-line-number="511"></td> <td id="LC511" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> When restoring from CudnnLSTM-trained checkpoints, must use</span></td> </tr> <tr> <td id="L512" class="blob-num js-line-number" data-line-number="512"></td> <td id="LC512" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `CudnnCompatibleLSTMCell` instead.</span></td> </tr> <tr> <td id="L513" class="blob-num js-line-number" data-line-number="513"></td> <td id="LC513" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L514" class="blob-num js-line-number" data-line-number="514"></td> <td id="LC514" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(BasicLSTMCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>(<span class="pl-v">_reuse</span><span class="pl-k">=</span>reuse, <span class="pl-v">name</span><span class="pl-k">=</span>name)</td> </tr> <tr> <td id="L515" class="blob-num js-line-number" data-line-number="515"></td> <td id="LC515" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> state_is_tuple:</td> </tr> <tr> <td id="L516" class="blob-num js-line-number" data-line-number="516"></td> <td id="LC516" class="blob-code blob-code-inner js-file-line"> logging.warn(<span class="pl-s"><span class="pl-pds">&quot;</span><span class="pl-c1">%s</span>: Using a concatenated state is slower and will soon be <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L517" class="blob-num js-line-number" data-line-number="517"></td> <td id="LC517" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>deprecated. Use state_is_tuple=True.<span class="pl-pds">&quot;</span></span>, <span class="pl-c1">self</span>)</td> </tr> <tr> <td id="L518" class="blob-num js-line-number" data-line-number="518"></td> <td id="LC518" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L519" class="blob-num js-line-number" data-line-number="519"></td> <td id="LC519" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Inputs must be 2-dimensional.</span></td> </tr> <tr> <td id="L520" class="blob-num js-line-number" data-line-number="520"></td> <td id="LC520" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.input_spec <span class="pl-k">=</span> base_layer.InputSpec(<span class="pl-v">ndim</span><span class="pl-k">=</span><span class="pl-c1">2</span>)</td> </tr> <tr> <td id="L521" class="blob-num js-line-number" data-line-number="521"></td> <td id="LC521" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L522" class="blob-num js-line-number" data-line-number="522"></td> <td id="LC522" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_units <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L523" class="blob-num js-line-number" data-line-number="523"></td> <td id="LC523" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._forget_bias <span class="pl-k">=</span> forget_bias</td> </tr> <tr> <td id="L524" class="blob-num js-line-number" data-line-number="524"></td> <td id="LC524" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">=</span> state_is_tuple</td> </tr> <tr> <td id="L525" class="blob-num js-line-number" data-line-number="525"></td> <td id="LC525" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation <span class="pl-k">=</span> activation <span class="pl-k">or</span> math_ops.tanh</td> </tr> <tr> <td id="L526" class="blob-num js-line-number" data-line-number="526"></td> <td id="LC526" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L527" class="blob-num js-line-number" data-line-number="527"></td> <td id="LC527" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L528" class="blob-num js-line-number" data-line-number="528"></td> <td id="LC528" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L529" class="blob-num js-line-number" data-line-number="529"></td> <td id="LC529" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> (LSTMStateTuple(<span class="pl-c1">self</span>._num_units, <span class="pl-c1">self</span>._num_units)</td> </tr> <tr> <td id="L530" class="blob-num js-line-number" data-line-number="530"></td> <td id="LC530" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">else</span> <span class="pl-c1">2</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units)</td> </tr> <tr> <td id="L531" class="blob-num js-line-number" data-line-number="531"></td> <td id="LC531" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L532" class="blob-num js-line-number" data-line-number="532"></td> <td id="LC532" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L533" class="blob-num js-line-number" data-line-number="533"></td> <td id="LC533" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L534" class="blob-num js-line-number" data-line-number="534"></td> <td id="LC534" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L535" class="blob-num js-line-number" data-line-number="535"></td> <td id="LC535" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L536" class="blob-num js-line-number" data-line-number="536"></td> <td id="LC536" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs_shape</span>):</td> </tr> <tr> <td id="L537" class="blob-num js-line-number" data-line-number="537"></td> <td id="LC537" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> inputs_shape[<span class="pl-c1">1</span>].value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L538" class="blob-num js-line-number" data-line-number="538"></td> <td id="LC538" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Expected inputs.shape[-1] to be known, saw shape: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L539" class="blob-num js-line-number" data-line-number="539"></td> <td id="LC539" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> inputs_shape)</td> </tr> <tr> <td id="L540" class="blob-num js-line-number" data-line-number="540"></td> <td id="LC540" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L541" class="blob-num js-line-number" data-line-number="541"></td> <td id="LC541" class="blob-code blob-code-inner js-file-line"> input_depth <span class="pl-k">=</span> inputs_shape[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L542" class="blob-num js-line-number" data-line-number="542"></td> <td id="LC542" class="blob-code blob-code-inner js-file-line"> h_depth <span class="pl-k">=</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L543" class="blob-num js-line-number" data-line-number="543"></td> <td id="LC543" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L544" class="blob-num js-line-number" data-line-number="544"></td> <td id="LC544" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L545" class="blob-num js-line-number" data-line-number="545"></td> <td id="LC545" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> h_depth, <span class="pl-c1">4</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units])</td> </tr> <tr> <td id="L546" class="blob-num js-line-number" data-line-number="546"></td> <td id="LC546" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L547" class="blob-num js-line-number" data-line-number="547"></td> <td id="LC547" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L548" class="blob-num js-line-number" data-line-number="548"></td> <td id="LC548" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">4</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L549" class="blob-num js-line-number" data-line-number="549"></td> <td id="LC549" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>init_ops.zeros_initializer(<span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype))</td> </tr> <tr> <td id="L550" class="blob-num js-line-number" data-line-number="550"></td> <td id="LC550" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L551" class="blob-num js-line-number" data-line-number="551"></td> <td id="LC551" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.built <span class="pl-k">=</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L552" class="blob-num js-line-number" data-line-number="552"></td> <td id="LC552" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L553" class="blob-num js-line-number" data-line-number="553"></td> <td id="LC553" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L554" class="blob-num js-line-number" data-line-number="554"></td> <td id="LC554" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Long short-term memory cell (LSTM).</span></td> </tr> <tr> <td id="L555" class="blob-num js-line-number" data-line-number="555"></td> <td id="LC555" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L556" class="blob-num js-line-number" data-line-number="556"></td> <td id="LC556" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L557" class="blob-num js-line-number" data-line-number="557"></td> <td id="LC557" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: `2-D` tensor with shape `[batch_size, input_size]`.</span></td> </tr> <tr> <td id="L558" class="blob-num js-line-number" data-line-number="558"></td> <td id="LC558" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: An `LSTMStateTuple` of state tensors, each shaped</span></td> </tr> <tr> <td id="L559" class="blob-num js-line-number" data-line-number="559"></td> <td id="LC559" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `[batch_size, self.state_size]`, if `state_is_tuple` has been set to</span></td> </tr> <tr> <td id="L560" class="blob-num js-line-number" data-line-number="560"></td> <td id="LC560" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `True`. Otherwise, a `Tensor` shaped</span></td> </tr> <tr> <td id="L561" class="blob-num js-line-number" data-line-number="561"></td> <td id="LC561" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `[batch_size, 2 * self.state_size]`.</span></td> </tr> <tr> <td id="L562" class="blob-num js-line-number" data-line-number="562"></td> <td id="LC562" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L563" class="blob-num js-line-number" data-line-number="563"></td> <td id="LC563" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L564" class="blob-num js-line-number" data-line-number="564"></td> <td id="LC564" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> A pair containing the new hidden state, and the new state (either a</span></td> </tr> <tr> <td id="L565" class="blob-num js-line-number" data-line-number="565"></td> <td id="LC565" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `LSTMStateTuple` or a concatenated state, depending on</span></td> </tr> <tr> <td id="L566" class="blob-num js-line-number" data-line-number="566"></td> <td id="LC566" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `state_is_tuple`).</span></td> </tr> <tr> <td id="L567" class="blob-num js-line-number" data-line-number="567"></td> <td id="LC567" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L568" class="blob-num js-line-number" data-line-number="568"></td> <td id="LC568" class="blob-code blob-code-inner js-file-line"> sigmoid <span class="pl-k">=</span> math_ops.sigmoid</td> </tr> <tr> <td id="L569" class="blob-num js-line-number" data-line-number="569"></td> <td id="LC569" class="blob-code blob-code-inner js-file-line"> one <span class="pl-k">=</span> constant_op.constant(<span class="pl-c1">1</span>, <span class="pl-v">dtype</span><span class="pl-k">=</span>dtypes.int32)</td> </tr> <tr> <td id="L570" class="blob-num js-line-number" data-line-number="570"></td> <td id="LC570" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Parameters of gates are concatenated into one multiply for efficiency.</span></td> </tr> <tr> <td id="L571" class="blob-num js-line-number" data-line-number="571"></td> <td id="LC571" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L572" class="blob-num js-line-number" data-line-number="572"></td> <td id="LC572" class="blob-code blob-code-inner js-file-line"> c, h <span class="pl-k">=</span> state</td> </tr> <tr> <td id="L573" class="blob-num js-line-number" data-line-number="573"></td> <td id="LC573" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L574" class="blob-num js-line-number" data-line-number="574"></td> <td id="LC574" class="blob-code blob-code-inner js-file-line"> c, h <span class="pl-k">=</span> array_ops.split(<span class="pl-v">value</span><span class="pl-k">=</span>state, <span class="pl-v">num_or_size_splits</span><span class="pl-k">=</span><span class="pl-c1">2</span>, <span class="pl-v">axis</span><span class="pl-k">=</span>one)</td> </tr> <tr> <td id="L575" class="blob-num js-line-number" data-line-number="575"></td> <td id="LC575" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L576" class="blob-num js-line-number" data-line-number="576"></td> <td id="LC576" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L577" class="blob-num js-line-number" data-line-number="577"></td> <td id="LC577" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, h], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._kernel)</td> </tr> <tr> <td id="L578" class="blob-num js-line-number" data-line-number="578"></td> <td id="LC578" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> nn_ops.bias_add(gate_inputs, <span class="pl-c1">self</span>._bias)</td> </tr> <tr> <td id="L579" class="blob-num js-line-number" data-line-number="579"></td> <td id="LC579" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L580" class="blob-num js-line-number" data-line-number="580"></td> <td id="LC580" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> i = input_gate, j = new_input, f = forget_gate, o = output_gate</span></td> </tr> <tr> <td id="L581" class="blob-num js-line-number" data-line-number="581"></td> <td id="LC581" class="blob-code blob-code-inner js-file-line"> i, j, f, o <span class="pl-k">=</span> array_ops.split(</td> </tr> <tr> <td id="L582" class="blob-num js-line-number" data-line-number="582"></td> <td id="LC582" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">value</span><span class="pl-k">=</span>gate_inputs, <span class="pl-v">num_or_size_splits</span><span class="pl-k">=</span><span class="pl-c1">4</span>, <span class="pl-v">axis</span><span class="pl-k">=</span>one)</td> </tr> <tr> <td id="L583" class="blob-num js-line-number" data-line-number="583"></td> <td id="LC583" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L584" class="blob-num js-line-number" data-line-number="584"></td> <td id="LC584" class="blob-code blob-code-inner js-file-line"> forget_bias_tensor <span class="pl-k">=</span> constant_op.constant(<span class="pl-c1">self</span>._forget_bias, <span class="pl-v">dtype</span><span class="pl-k">=</span>f.dtype)</td> </tr> <tr> <td id="L585" class="blob-num js-line-number" data-line-number="585"></td> <td id="LC585" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Note that using `add` and `multiply` instead of `+` and `*` gives a</span></td> </tr> <tr> <td id="L586" class="blob-num js-line-number" data-line-number="586"></td> <td id="LC586" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> performance improvement. So using those at the cost of readability.</span></td> </tr> <tr> <td id="L587" class="blob-num js-line-number" data-line-number="587"></td> <td id="LC587" class="blob-code blob-code-inner js-file-line"> add <span class="pl-k">=</span> math_ops.add</td> </tr> <tr> <td id="L588" class="blob-num js-line-number" data-line-number="588"></td> <td id="LC588" class="blob-code blob-code-inner js-file-line"> multiply <span class="pl-k">=</span> math_ops.multiply</td> </tr> <tr> <td id="L589" class="blob-num js-line-number" data-line-number="589"></td> <td id="LC589" class="blob-code blob-code-inner js-file-line"> new_c <span class="pl-k">=</span> add(multiply(c, sigmoid(add(f, forget_bias_tensor))),</td> </tr> <tr> <td id="L590" class="blob-num js-line-number" data-line-number="590"></td> <td id="LC590" class="blob-code blob-code-inner js-file-line"> multiply(sigmoid(i), <span class="pl-c1">self</span>._activation(j)))</td> </tr> <tr> <td id="L591" class="blob-num js-line-number" data-line-number="591"></td> <td id="LC591" class="blob-code blob-code-inner js-file-line"> new_h <span class="pl-k">=</span> multiply(<span class="pl-c1">self</span>._activation(new_c), sigmoid(o))</td> </tr> <tr> <td id="L592" class="blob-num js-line-number" data-line-number="592"></td> <td id="LC592" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L593" class="blob-num js-line-number" data-line-number="593"></td> <td id="LC593" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L594" class="blob-num js-line-number" data-line-number="594"></td> <td id="LC594" class="blob-code blob-code-inner js-file-line"> new_state <span class="pl-k">=</span> LSTMStateTuple(new_c, new_h)</td> </tr> <tr> <td id="L595" class="blob-num js-line-number" data-line-number="595"></td> <td id="LC595" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L596" class="blob-num js-line-number" data-line-number="596"></td> <td id="LC596" class="blob-code blob-code-inner js-file-line"> new_state <span class="pl-k">=</span> array_ops.concat([new_c, new_h], <span class="pl-c1">1</span>)</td> </tr> <tr> <td id="L597" class="blob-num js-line-number" data-line-number="597"></td> <td id="LC597" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> new_h, new_state</td> </tr> <tr> <td id="L598" class="blob-num js-line-number" data-line-number="598"></td> <td id="LC598" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L599" class="blob-num js-line-number" data-line-number="599"></td> <td id="LC599" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L600" class="blob-num js-line-number" data-line-number="600"></td> <td id="LC600" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.LSTMCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L601" class="blob-num js-line-number" data-line-number="601"></td> <td id="LC601" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">LSTMCell</span>(<span class="pl-e">_LayerRNNCell</span>):</td> </tr> <tr> <td id="L602" class="blob-num js-line-number" data-line-number="602"></td> <td id="LC602" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Long short-term memory unit (LSTM) recurrent network cell.</span></td> </tr> <tr> <td id="L603" class="blob-num js-line-number" data-line-number="603"></td> <td id="LC603" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L604" class="blob-num js-line-number" data-line-number="604"></td> <td id="LC604" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> The default non-peephole implementation is based on:</span></td> </tr> <tr> <td id="L605" class="blob-num js-line-number" data-line-number="605"></td> <td id="LC605" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L606" class="blob-num js-line-number" data-line-number="606"></td> <td id="LC606" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> http://www.bioinf.jku.at/publications/older/2604.pdf</span></td> </tr> <tr> <td id="L607" class="blob-num js-line-number" data-line-number="607"></td> <td id="LC607" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L608" class="blob-num js-line-number" data-line-number="608"></td> <td id="LC608" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> S. Hochreiter and J. Schmidhuber.</span></td> </tr> <tr> <td id="L609" class="blob-num js-line-number" data-line-number="609"></td> <td id="LC609" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> &quot;Long Short-Term Memory&quot;. Neural Computation, 9(8):1735-1780, 1997.</span></td> </tr> <tr> <td id="L610" class="blob-num js-line-number" data-line-number="610"></td> <td id="LC610" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L611" class="blob-num js-line-number" data-line-number="611"></td> <td id="LC611" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> The peephole implementation is based on:</span></td> </tr> <tr> <td id="L612" class="blob-num js-line-number" data-line-number="612"></td> <td id="LC612" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L613" class="blob-num js-line-number" data-line-number="613"></td> <td id="LC613" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> https://research.google.com/pubs/archive/43905.pdf</span></td> </tr> <tr> <td id="L614" class="blob-num js-line-number" data-line-number="614"></td> <td id="LC614" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L615" class="blob-num js-line-number" data-line-number="615"></td> <td id="LC615" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Hasim Sak, Andrew Senior, and Francoise Beaufays.</span></td> </tr> <tr> <td id="L616" class="blob-num js-line-number" data-line-number="616"></td> <td id="LC616" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> &quot;Long short-term memory recurrent neural network architectures for</span></td> </tr> <tr> <td id="L617" class="blob-num js-line-number" data-line-number="617"></td> <td id="LC617" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> large scale acoustic modeling.&quot; INTERSPEECH, 2014.</span></td> </tr> <tr> <td id="L618" class="blob-num js-line-number" data-line-number="618"></td> <td id="LC618" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L619" class="blob-num js-line-number" data-line-number="619"></td> <td id="LC619" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> The class uses optional peep-hole connections, optional cell clipping, and</span></td> </tr> <tr> <td id="L620" class="blob-num js-line-number" data-line-number="620"></td> <td id="LC620" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> an optional projection layer.</span></td> </tr> <tr> <td id="L621" class="blob-num js-line-number" data-line-number="621"></td> <td id="LC621" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L622" class="blob-num js-line-number" data-line-number="622"></td> <td id="LC622" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L623" class="blob-num js-line-number" data-line-number="623"></td> <td id="LC623" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">num_units</span>,</td> </tr> <tr> <td id="L624" class="blob-num js-line-number" data-line-number="624"></td> <td id="LC624" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">use_peepholes</span><span class="pl-k">=</span><span class="pl-c1">False</span>, <span class="pl-smi">cell_clip</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L625" class="blob-num js-line-number" data-line-number="625"></td> <td id="LC625" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">initializer</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">num_proj</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">proj_clip</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L626" class="blob-num js-line-number" data-line-number="626"></td> <td id="LC626" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">num_unit_shards</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">num_proj_shards</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L627" class="blob-num js-line-number" data-line-number="627"></td> <td id="LC627" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">forget_bias</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>, <span class="pl-smi">state_is_tuple</span><span class="pl-k">=</span><span class="pl-c1">True</span>,</td> </tr> <tr> <td id="L628" class="blob-num js-line-number" data-line-number="628"></td> <td id="LC628" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">activation</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">reuse</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">name</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L629" class="blob-num js-line-number" data-line-number="629"></td> <td id="LC629" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Initialize the parameters for an LSTM cell.</span></td> </tr> <tr> <td id="L630" class="blob-num js-line-number" data-line-number="630"></td> <td id="LC630" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L631" class="blob-num js-line-number" data-line-number="631"></td> <td id="LC631" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L632" class="blob-num js-line-number" data-line-number="632"></td> <td id="LC632" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units: int, The number of units in the LSTM cell.</span></td> </tr> <tr> <td id="L633" class="blob-num js-line-number" data-line-number="633"></td> <td id="LC633" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> use_peepholes: bool, set True to enable diagonal/peephole connections.</span></td> </tr> <tr> <td id="L634" class="blob-num js-line-number" data-line-number="634"></td> <td id="LC634" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell_clip: (optional) A float value, if provided the cell state is clipped</span></td> </tr> <tr> <td id="L635" class="blob-num js-line-number" data-line-number="635"></td> <td id="LC635" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> by this value prior to the cell output activation.</span></td> </tr> <tr> <td id="L636" class="blob-num js-line-number" data-line-number="636"></td> <td id="LC636" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> initializer: (optional) The initializer to use for the weight and</span></td> </tr> <tr> <td id="L637" class="blob-num js-line-number" data-line-number="637"></td> <td id="LC637" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> projection matrices.</span></td> </tr> <tr> <td id="L638" class="blob-num js-line-number" data-line-number="638"></td> <td id="LC638" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_proj: (optional) int, The output dimensionality for the projection</span></td> </tr> <tr> <td id="L639" class="blob-num js-line-number" data-line-number="639"></td> <td id="LC639" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> matrices. If None, no projection is performed.</span></td> </tr> <tr> <td id="L640" class="blob-num js-line-number" data-line-number="640"></td> <td id="LC640" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> proj_clip: (optional) A float value. If `num_proj &gt; 0` and `proj_clip` is</span></td> </tr> <tr> <td id="L641" class="blob-num js-line-number" data-line-number="641"></td> <td id="LC641" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> provided, then the projected values are clipped elementwise to within</span></td> </tr> <tr> <td id="L642" class="blob-num js-line-number" data-line-number="642"></td> <td id="LC642" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `[-proj_clip, proj_clip]`.</span></td> </tr> <tr> <td id="L643" class="blob-num js-line-number" data-line-number="643"></td> <td id="LC643" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_unit_shards: Deprecated, will be removed by Jan. 2017.</span></td> </tr> <tr> <td id="L644" class="blob-num js-line-number" data-line-number="644"></td> <td id="LC644" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Use a variable_scope partitioner instead.</span></td> </tr> <tr> <td id="L645" class="blob-num js-line-number" data-line-number="645"></td> <td id="LC645" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_proj_shards: Deprecated, will be removed by Jan. 2017.</span></td> </tr> <tr> <td id="L646" class="blob-num js-line-number" data-line-number="646"></td> <td id="LC646" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Use a variable_scope partitioner instead.</span></td> </tr> <tr> <td id="L647" class="blob-num js-line-number" data-line-number="647"></td> <td id="LC647" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> forget_bias: Biases of the forget gate are initialized by default to 1</span></td> </tr> <tr> <td id="L648" class="blob-num js-line-number" data-line-number="648"></td> <td id="LC648" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in order to reduce the scale of forgetting at the beginning of</span></td> </tr> <tr> <td id="L649" class="blob-num js-line-number" data-line-number="649"></td> <td id="LC649" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the training. Must set it manually to `0.0` when restoring from</span></td> </tr> <tr> <td id="L650" class="blob-num js-line-number" data-line-number="650"></td> <td id="LC650" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> CudnnLSTM trained checkpoints.</span></td> </tr> <tr> <td id="L651" class="blob-num js-line-number" data-line-number="651"></td> <td id="LC651" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state_is_tuple: If True, accepted and returned states are 2-tuples of</span></td> </tr> <tr> <td id="L652" class="blob-num js-line-number" data-line-number="652"></td> <td id="LC652" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the `c_state` and `m_state`. If False, they are concatenated</span></td> </tr> <tr> <td id="L653" class="blob-num js-line-number" data-line-number="653"></td> <td id="LC653" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> along the column axis. This latter behavior will soon be deprecated.</span></td> </tr> <tr> <td id="L654" class="blob-num js-line-number" data-line-number="654"></td> <td id="LC654" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> activation: Activation function of the inner states. Default: `tanh`.</span></td> </tr> <tr> <td id="L655" class="blob-num js-line-number" data-line-number="655"></td> <td id="LC655" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reuse: (optional) Python boolean describing whether to reuse variables</span></td> </tr> <tr> <td id="L656" class="blob-num js-line-number" data-line-number="656"></td> <td id="LC656" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in an existing scope. If not `True`, and the existing scope already has</span></td> </tr> <tr> <td id="L657" class="blob-num js-line-number" data-line-number="657"></td> <td id="LC657" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the given variables, an error is raised.</span></td> </tr> <tr> <td id="L658" class="blob-num js-line-number" data-line-number="658"></td> <td id="LC658" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> name: String, the name of the layer. Layers with the same name will</span></td> </tr> <tr> <td id="L659" class="blob-num js-line-number" data-line-number="659"></td> <td id="LC659" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> share weights, but to avoid mistakes we require reuse=True in such</span></td> </tr> <tr> <td id="L660" class="blob-num js-line-number" data-line-number="660"></td> <td id="LC660" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cases.</span></td> </tr> <tr> <td id="L661" class="blob-num js-line-number" data-line-number="661"></td> <td id="LC661" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L662" class="blob-num js-line-number" data-line-number="662"></td> <td id="LC662" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> When restoring from CudnnLSTM-trained checkpoints, use</span></td> </tr> <tr> <td id="L663" class="blob-num js-line-number" data-line-number="663"></td> <td id="LC663" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `CudnnCompatibleLSTMCell` instead.</span></td> </tr> <tr> <td id="L664" class="blob-num js-line-number" data-line-number="664"></td> <td id="LC664" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L665" class="blob-num js-line-number" data-line-number="665"></td> <td id="LC665" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(LSTMCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>(<span class="pl-v">_reuse</span><span class="pl-k">=</span>reuse, <span class="pl-v">name</span><span class="pl-k">=</span>name)</td> </tr> <tr> <td id="L666" class="blob-num js-line-number" data-line-number="666"></td> <td id="LC666" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> state_is_tuple:</td> </tr> <tr> <td id="L667" class="blob-num js-line-number" data-line-number="667"></td> <td id="LC667" class="blob-code blob-code-inner js-file-line"> logging.warn(<span class="pl-s"><span class="pl-pds">&quot;</span><span class="pl-c1">%s</span>: Using a concatenated state is slower and will soon be <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L668" class="blob-num js-line-number" data-line-number="668"></td> <td id="LC668" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>deprecated. Use state_is_tuple=True.<span class="pl-pds">&quot;</span></span>, <span class="pl-c1">self</span>)</td> </tr> <tr> <td id="L669" class="blob-num js-line-number" data-line-number="669"></td> <td id="LC669" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> num_unit_shards <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span> <span class="pl-k">or</span> num_proj_shards <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L670" class="blob-num js-line-number" data-line-number="670"></td> <td id="LC670" class="blob-code blob-code-inner js-file-line"> logging.warn(</td> </tr> <tr> <td id="L671" class="blob-num js-line-number" data-line-number="671"></td> <td id="LC671" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span><span class="pl-c1">%s</span>: The num_unit_shards and proj_unit_shards parameters are <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L672" class="blob-num js-line-number" data-line-number="672"></td> <td id="LC672" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>deprecated and will be removed in Jan 2017. <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L673" class="blob-num js-line-number" data-line-number="673"></td> <td id="LC673" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>Use a variable scope with a partitioner instead.<span class="pl-pds">&quot;</span></span>, <span class="pl-c1">self</span>)</td> </tr> <tr> <td id="L674" class="blob-num js-line-number" data-line-number="674"></td> <td id="LC674" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L675" class="blob-num js-line-number" data-line-number="675"></td> <td id="LC675" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Inputs must be 2-dimensional.</span></td> </tr> <tr> <td id="L676" class="blob-num js-line-number" data-line-number="676"></td> <td id="LC676" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.input_spec <span class="pl-k">=</span> base_layer.InputSpec(<span class="pl-v">ndim</span><span class="pl-k">=</span><span class="pl-c1">2</span>)</td> </tr> <tr> <td id="L677" class="blob-num js-line-number" data-line-number="677"></td> <td id="LC677" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L678" class="blob-num js-line-number" data-line-number="678"></td> <td id="LC678" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_units <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L679" class="blob-num js-line-number" data-line-number="679"></td> <td id="LC679" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._use_peepholes <span class="pl-k">=</span> use_peepholes</td> </tr> <tr> <td id="L680" class="blob-num js-line-number" data-line-number="680"></td> <td id="LC680" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell_clip <span class="pl-k">=</span> cell_clip</td> </tr> <tr> <td id="L681" class="blob-num js-line-number" data-line-number="681"></td> <td id="LC681" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._initializer <span class="pl-k">=</span> initializer</td> </tr> <tr> <td id="L682" class="blob-num js-line-number" data-line-number="682"></td> <td id="LC682" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_proj <span class="pl-k">=</span> num_proj</td> </tr> <tr> <td id="L683" class="blob-num js-line-number" data-line-number="683"></td> <td id="LC683" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._proj_clip <span class="pl-k">=</span> proj_clip</td> </tr> <tr> <td id="L684" class="blob-num js-line-number" data-line-number="684"></td> <td id="LC684" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_unit_shards <span class="pl-k">=</span> num_unit_shards</td> </tr> <tr> <td id="L685" class="blob-num js-line-number" data-line-number="685"></td> <td id="LC685" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_proj_shards <span class="pl-k">=</span> num_proj_shards</td> </tr> <tr> <td id="L686" class="blob-num js-line-number" data-line-number="686"></td> <td id="LC686" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._forget_bias <span class="pl-k">=</span> forget_bias</td> </tr> <tr> <td id="L687" class="blob-num js-line-number" data-line-number="687"></td> <td id="LC687" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">=</span> state_is_tuple</td> </tr> <tr> <td id="L688" class="blob-num js-line-number" data-line-number="688"></td> <td id="LC688" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation <span class="pl-k">=</span> activation <span class="pl-k">or</span> math_ops.tanh</td> </tr> <tr> <td id="L689" class="blob-num js-line-number" data-line-number="689"></td> <td id="LC689" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L690" class="blob-num js-line-number" data-line-number="690"></td> <td id="LC690" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> num_proj:</td> </tr> <tr> <td id="L691" class="blob-num js-line-number" data-line-number="691"></td> <td id="LC691" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_size <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L692" class="blob-num js-line-number" data-line-number="692"></td> <td id="LC692" class="blob-code blob-code-inner js-file-line"> LSTMStateTuple(num_units, num_proj)</td> </tr> <tr> <td id="L693" class="blob-num js-line-number" data-line-number="693"></td> <td id="LC693" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> state_is_tuple <span class="pl-k">else</span> num_units <span class="pl-k">+</span> num_proj)</td> </tr> <tr> <td id="L694" class="blob-num js-line-number" data-line-number="694"></td> <td id="LC694" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._output_size <span class="pl-k">=</span> num_proj</td> </tr> <tr> <td id="L695" class="blob-num js-line-number" data-line-number="695"></td> <td id="LC695" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L696" class="blob-num js-line-number" data-line-number="696"></td> <td id="LC696" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_size <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L697" class="blob-num js-line-number" data-line-number="697"></td> <td id="LC697" class="blob-code blob-code-inner js-file-line"> LSTMStateTuple(num_units, num_units)</td> </tr> <tr> <td id="L698" class="blob-num js-line-number" data-line-number="698"></td> <td id="LC698" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> state_is_tuple <span class="pl-k">else</span> <span class="pl-c1">2</span> <span class="pl-k">*</span> num_units)</td> </tr> <tr> <td id="L699" class="blob-num js-line-number" data-line-number="699"></td> <td id="LC699" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._output_size <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L700" class="blob-num js-line-number" data-line-number="700"></td> <td id="LC700" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L701" class="blob-num js-line-number" data-line-number="701"></td> <td id="LC701" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L702" class="blob-num js-line-number" data-line-number="702"></td> <td id="LC702" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L703" class="blob-num js-line-number" data-line-number="703"></td> <td id="LC703" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._state_size</td> </tr> <tr> <td id="L704" class="blob-num js-line-number" data-line-number="704"></td> <td id="LC704" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L705" class="blob-num js-line-number" data-line-number="705"></td> <td id="LC705" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L706" class="blob-num js-line-number" data-line-number="706"></td> <td id="LC706" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L707" class="blob-num js-line-number" data-line-number="707"></td> <td id="LC707" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._output_size</td> </tr> <tr> <td id="L708" class="blob-num js-line-number" data-line-number="708"></td> <td id="LC708" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L709" class="blob-num js-line-number" data-line-number="709"></td> <td id="LC709" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs_shape</span>):</td> </tr> <tr> <td id="L710" class="blob-num js-line-number" data-line-number="710"></td> <td id="LC710" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> inputs_shape[<span class="pl-c1">1</span>].value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L711" class="blob-num js-line-number" data-line-number="711"></td> <td id="LC711" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Expected inputs.shape[-1] to be known, saw shape: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L712" class="blob-num js-line-number" data-line-number="712"></td> <td id="LC712" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> inputs_shape)</td> </tr> <tr> <td id="L713" class="blob-num js-line-number" data-line-number="713"></td> <td id="LC713" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L714" class="blob-num js-line-number" data-line-number="714"></td> <td id="LC714" class="blob-code blob-code-inner js-file-line"> input_depth <span class="pl-k">=</span> inputs_shape[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L715" class="blob-num js-line-number" data-line-number="715"></td> <td id="LC715" class="blob-code blob-code-inner js-file-line"> h_depth <span class="pl-k">=</span> <span class="pl-c1">self</span>._num_units <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj <span class="pl-k">is</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">self</span>._num_proj</td> </tr> <tr> <td id="L716" class="blob-num js-line-number" data-line-number="716"></td> <td id="LC716" class="blob-code blob-code-inner js-file-line"> maybe_partitioner <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L717" class="blob-num js-line-number" data-line-number="717"></td> <td id="LC717" class="blob-code blob-code-inner js-file-line"> partitioned_variables.fixed_size_partitioner(<span class="pl-c1">self</span>._num_unit_shards)</td> </tr> <tr> <td id="L718" class="blob-num js-line-number" data-line-number="718"></td> <td id="LC718" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_unit_shards <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L719" class="blob-num js-line-number" data-line-number="719"></td> <td id="LC719" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span> <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L720" class="blob-num js-line-number" data-line-number="720"></td> <td id="LC720" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L721" class="blob-num js-line-number" data-line-number="721"></td> <td id="LC721" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L722" class="blob-num js-line-number" data-line-number="722"></td> <td id="LC722" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> h_depth, <span class="pl-c1">4</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L723" class="blob-num js-line-number" data-line-number="723"></td> <td id="LC723" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer,</td> </tr> <tr> <td id="L724" class="blob-num js-line-number" data-line-number="724"></td> <td id="LC724" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">partitioner</span><span class="pl-k">=</span>maybe_partitioner)</td> </tr> <tr> <td id="L725" class="blob-num js-line-number" data-line-number="725"></td> <td id="LC725" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L726" class="blob-num js-line-number" data-line-number="726"></td> <td id="LC726" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L727" class="blob-num js-line-number" data-line-number="727"></td> <td id="LC727" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">4</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L728" class="blob-num js-line-number" data-line-number="728"></td> <td id="LC728" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>init_ops.zeros_initializer(<span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype))</td> </tr> <tr> <td id="L729" class="blob-num js-line-number" data-line-number="729"></td> <td id="LC729" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._use_peepholes:</td> </tr> <tr> <td id="L730" class="blob-num js-line-number" data-line-number="730"></td> <td id="LC730" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._w_f_diag <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(<span class="pl-s"><span class="pl-pds">&quot;</span>w_f_diag<span class="pl-pds">&quot;</span></span>, <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L731" class="blob-num js-line-number" data-line-number="731"></td> <td id="LC731" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer)</td> </tr> <tr> <td id="L732" class="blob-num js-line-number" data-line-number="732"></td> <td id="LC732" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._w_i_diag <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(<span class="pl-s"><span class="pl-pds">&quot;</span>w_i_diag<span class="pl-pds">&quot;</span></span>, <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L733" class="blob-num js-line-number" data-line-number="733"></td> <td id="LC733" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer)</td> </tr> <tr> <td id="L734" class="blob-num js-line-number" data-line-number="734"></td> <td id="LC734" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._w_o_diag <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(<span class="pl-s"><span class="pl-pds">&quot;</span>w_o_diag<span class="pl-pds">&quot;</span></span>, <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L735" class="blob-num js-line-number" data-line-number="735"></td> <td id="LC735" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer)</td> </tr> <tr> <td id="L736" class="blob-num js-line-number" data-line-number="736"></td> <td id="LC736" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L737" class="blob-num js-line-number" data-line-number="737"></td> <td id="LC737" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L738" class="blob-num js-line-number" data-line-number="738"></td> <td id="LC738" class="blob-code blob-code-inner js-file-line"> maybe_proj_partitioner <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L739" class="blob-num js-line-number" data-line-number="739"></td> <td id="LC739" class="blob-code blob-code-inner js-file-line"> partitioned_variables.fixed_size_partitioner(<span class="pl-c1">self</span>._num_proj_shards)</td> </tr> <tr> <td id="L740" class="blob-num js-line-number" data-line-number="740"></td> <td id="LC740" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj_shards <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L741" class="blob-num js-line-number" data-line-number="741"></td> <td id="LC741" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span> <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L742" class="blob-num js-line-number" data-line-number="742"></td> <td id="LC742" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._proj_kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L743" class="blob-num js-line-number" data-line-number="743"></td> <td id="LC743" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>projection/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L744" class="blob-num js-line-number" data-line-number="744"></td> <td id="LC744" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units, <span class="pl-c1">self</span>._num_proj],</td> </tr> <tr> <td id="L745" class="blob-num js-line-number" data-line-number="745"></td> <td id="LC745" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer,</td> </tr> <tr> <td id="L746" class="blob-num js-line-number" data-line-number="746"></td> <td id="LC746" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">partitioner</span><span class="pl-k">=</span>maybe_proj_partitioner)</td> </tr> <tr> <td id="L747" class="blob-num js-line-number" data-line-number="747"></td> <td id="LC747" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L748" class="blob-num js-line-number" data-line-number="748"></td> <td id="LC748" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.built <span class="pl-k">=</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L749" class="blob-num js-line-number" data-line-number="749"></td> <td id="LC749" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L750" class="blob-num js-line-number" data-line-number="750"></td> <td id="LC750" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L751" class="blob-num js-line-number" data-line-number="751"></td> <td id="LC751" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run one step of LSTM.</span></td> </tr> <tr> <td id="L752" class="blob-num js-line-number" data-line-number="752"></td> <td id="LC752" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L753" class="blob-num js-line-number" data-line-number="753"></td> <td id="LC753" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L754" class="blob-num js-line-number" data-line-number="754"></td> <td id="LC754" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: input Tensor, 2D, `[batch, num_units].</span></td> </tr> <tr> <td id="L755" class="blob-num js-line-number" data-line-number="755"></td> <td id="LC755" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: if `state_is_tuple` is False, this must be a state Tensor,</span></td> </tr> <tr> <td id="L756" class="blob-num js-line-number" data-line-number="756"></td> <td id="LC756" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `2-D, [batch, state_size]`. If `state_is_tuple` is True, this must be a</span></td> </tr> <tr> <td id="L757" class="blob-num js-line-number" data-line-number="757"></td> <td id="LC757" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> tuple of state Tensors, both `2-D`, with column sizes `c_state` and</span></td> </tr> <tr> <td id="L758" class="blob-num js-line-number" data-line-number="758"></td> <td id="LC758" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `m_state`.</span></td> </tr> <tr> <td id="L759" class="blob-num js-line-number" data-line-number="759"></td> <td id="LC759" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L760" class="blob-num js-line-number" data-line-number="760"></td> <td id="LC760" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L761" class="blob-num js-line-number" data-line-number="761"></td> <td id="LC761" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> A tuple containing:</span></td> </tr> <tr> <td id="L762" class="blob-num js-line-number" data-line-number="762"></td> <td id="LC762" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L763" class="blob-num js-line-number" data-line-number="763"></td> <td id="LC763" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - A `2-D, [batch, output_dim]`, Tensor representing the output of the</span></td> </tr> <tr> <td id="L764" class="blob-num js-line-number" data-line-number="764"></td> <td id="LC764" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> LSTM after reading `inputs` when previous state was `state`.</span></td> </tr> <tr> <td id="L765" class="blob-num js-line-number" data-line-number="765"></td> <td id="LC765" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Here output_dim is:</span></td> </tr> <tr> <td id="L766" class="blob-num js-line-number" data-line-number="766"></td> <td id="LC766" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_proj if num_proj was set,</span></td> </tr> <tr> <td id="L767" class="blob-num js-line-number" data-line-number="767"></td> <td id="LC767" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units otherwise.</span></td> </tr> <tr> <td id="L768" class="blob-num js-line-number" data-line-number="768"></td> <td id="LC768" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - Tensor(s) representing the new state of LSTM after reading `inputs` when</span></td> </tr> <tr> <td id="L769" class="blob-num js-line-number" data-line-number="769"></td> <td id="LC769" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the previous state was `state`. Same type and shape(s) as `state`.</span></td> </tr> <tr> <td id="L770" class="blob-num js-line-number" data-line-number="770"></td> <td id="LC770" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L771" class="blob-num js-line-number" data-line-number="771"></td> <td id="LC771" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L772" class="blob-num js-line-number" data-line-number="772"></td> <td id="LC772" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: If input size cannot be inferred from inputs via</span></td> </tr> <tr> <td id="L773" class="blob-num js-line-number" data-line-number="773"></td> <td id="LC773" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> static shape inference.</span></td> </tr> <tr> <td id="L774" class="blob-num js-line-number" data-line-number="774"></td> <td id="LC774" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L775" class="blob-num js-line-number" data-line-number="775"></td> <td id="LC775" class="blob-code blob-code-inner js-file-line"> num_proj <span class="pl-k">=</span> <span class="pl-c1">self</span>._num_units <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj <span class="pl-k">is</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">self</span>._num_proj</td> </tr> <tr> <td id="L776" class="blob-num js-line-number" data-line-number="776"></td> <td id="LC776" class="blob-code blob-code-inner js-file-line"> sigmoid <span class="pl-k">=</span> math_ops.sigmoid</td> </tr> <tr> <td id="L777" class="blob-num js-line-number" data-line-number="777"></td> <td id="LC777" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L778" class="blob-num js-line-number" data-line-number="778"></td> <td id="LC778" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L779" class="blob-num js-line-number" data-line-number="779"></td> <td id="LC779" class="blob-code blob-code-inner js-file-line"> (c_prev, m_prev) <span class="pl-k">=</span> state</td> </tr> <tr> <td id="L780" class="blob-num js-line-number" data-line-number="780"></td> <td id="LC780" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L781" class="blob-num js-line-number" data-line-number="781"></td> <td id="LC781" class="blob-code blob-code-inner js-file-line"> c_prev <span class="pl-k">=</span> array_ops.slice(state, [<span class="pl-c1">0</span>, <span class="pl-c1">0</span>], [<span class="pl-k">-</span><span class="pl-c1">1</span>, <span class="pl-c1">self</span>._num_units])</td> </tr> <tr> <td id="L782" class="blob-num js-line-number" data-line-number="782"></td> <td id="LC782" class="blob-code blob-code-inner js-file-line"> m_prev <span class="pl-k">=</span> array_ops.slice(state, [<span class="pl-c1">0</span>, <span class="pl-c1">self</span>._num_units], [<span class="pl-k">-</span><span class="pl-c1">1</span>, num_proj])</td> </tr> <tr> <td id="L783" class="blob-num js-line-number" data-line-number="783"></td> <td id="LC783" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L784" class="blob-num js-line-number" data-line-number="784"></td> <td id="LC784" class="blob-code blob-code-inner js-file-line"> input_size <span class="pl-k">=</span> inputs.get_shape().with_rank(<span class="pl-c1">2</span>)[<span class="pl-c1">1</span>]</td> </tr> <tr> <td id="L785" class="blob-num js-line-number" data-line-number="785"></td> <td id="LC785" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> input_size.value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L786" class="blob-num js-line-number" data-line-number="786"></td> <td id="LC786" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Could not infer input size from inputs.get_shape()[-1]<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L787" class="blob-num js-line-number" data-line-number="787"></td> <td id="LC787" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L788" class="blob-num js-line-number" data-line-number="788"></td> <td id="LC788" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> i = input_gate, j = new_input, f = forget_gate, o = output_gate</span></td> </tr> <tr> <td id="L789" class="blob-num js-line-number" data-line-number="789"></td> <td id="LC789" class="blob-code blob-code-inner js-file-line"> lstm_matrix <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L790" class="blob-num js-line-number" data-line-number="790"></td> <td id="LC790" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, m_prev], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._kernel)</td> </tr> <tr> <td id="L791" class="blob-num js-line-number" data-line-number="791"></td> <td id="LC791" class="blob-code blob-code-inner js-file-line"> lstm_matrix <span class="pl-k">=</span> nn_ops.bias_add(lstm_matrix, <span class="pl-c1">self</span>._bias)</td> </tr> <tr> <td id="L792" class="blob-num js-line-number" data-line-number="792"></td> <td id="LC792" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L793" class="blob-num js-line-number" data-line-number="793"></td> <td id="LC793" class="blob-code blob-code-inner js-file-line"> i, j, f, o <span class="pl-k">=</span> array_ops.split(</td> </tr> <tr> <td id="L794" class="blob-num js-line-number" data-line-number="794"></td> <td id="LC794" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">value</span><span class="pl-k">=</span>lstm_matrix, <span class="pl-v">num_or_size_splits</span><span class="pl-k">=</span><span class="pl-c1">4</span>, <span class="pl-v">axis</span><span class="pl-k">=</span><span class="pl-c1">1</span>)</td> </tr> <tr> <td id="L795" class="blob-num js-line-number" data-line-number="795"></td> <td id="LC795" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Diagonal connections</span></td> </tr> <tr> <td id="L796" class="blob-num js-line-number" data-line-number="796"></td> <td id="LC796" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._use_peepholes:</td> </tr> <tr> <td id="L797" class="blob-num js-line-number" data-line-number="797"></td> <td id="LC797" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> (sigmoid(f <span class="pl-k">+</span> <span class="pl-c1">self</span>._forget_bias <span class="pl-k">+</span> <span class="pl-c1">self</span>._w_f_diag <span class="pl-k">*</span> c_prev) <span class="pl-k">*</span> c_prev <span class="pl-k">+</span></td> </tr> <tr> <td id="L798" class="blob-num js-line-number" data-line-number="798"></td> <td id="LC798" class="blob-code blob-code-inner js-file-line"> sigmoid(i <span class="pl-k">+</span> <span class="pl-c1">self</span>._w_i_diag <span class="pl-k">*</span> c_prev) <span class="pl-k">*</span> <span class="pl-c1">self</span>._activation(j))</td> </tr> <tr> <td id="L799" class="blob-num js-line-number" data-line-number="799"></td> <td id="LC799" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L800" class="blob-num js-line-number" data-line-number="800"></td> <td id="LC800" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> (sigmoid(f <span class="pl-k">+</span> <span class="pl-c1">self</span>._forget_bias) <span class="pl-k">*</span> c_prev <span class="pl-k">+</span> sigmoid(i) <span class="pl-k">*</span></td> </tr> <tr> <td id="L801" class="blob-num js-line-number" data-line-number="801"></td> <td id="LC801" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation(j))</td> </tr> <tr> <td id="L802" class="blob-num js-line-number" data-line-number="802"></td> <td id="LC802" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L803" class="blob-num js-line-number" data-line-number="803"></td> <td id="LC803" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._cell_clip <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L804" class="blob-num js-line-number" data-line-number="804"></td> <td id="LC804" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> pylint: disable=invalid-unary-operand-type</span></td> </tr> <tr> <td id="L805" class="blob-num js-line-number" data-line-number="805"></td> <td id="LC805" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> clip_ops.clip_by_value(c, <span class="pl-k">-</span><span class="pl-c1">self</span>._cell_clip, <span class="pl-c1">self</span>._cell_clip)</td> </tr> <tr> <td id="L806" class="blob-num js-line-number" data-line-number="806"></td> <td id="LC806" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> pylint: enable=invalid-unary-operand-type</span></td> </tr> <tr> <td id="L807" class="blob-num js-line-number" data-line-number="807"></td> <td id="LC807" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._use_peepholes:</td> </tr> <tr> <td id="L808" class="blob-num js-line-number" data-line-number="808"></td> <td id="LC808" class="blob-code blob-code-inner js-file-line"> m <span class="pl-k">=</span> sigmoid(o <span class="pl-k">+</span> <span class="pl-c1">self</span>._w_o_diag <span class="pl-k">*</span> c) <span class="pl-k">*</span> <span class="pl-c1">self</span>._activation(c)</td> </tr> <tr> <td id="L809" class="blob-num js-line-number" data-line-number="809"></td> <td id="LC809" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L810" class="blob-num js-line-number" data-line-number="810"></td> <td id="LC810" class="blob-code blob-code-inner js-file-line"> m <span class="pl-k">=</span> sigmoid(o) <span class="pl-k">*</span> <span class="pl-c1">self</span>._activation(c)</td> </tr> <tr> <td id="L811" class="blob-num js-line-number" data-line-number="811"></td> <td id="LC811" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L812" class="blob-num js-line-number" data-line-number="812"></td> <td id="LC812" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L813" class="blob-num js-line-number" data-line-number="813"></td> <td id="LC813" class="blob-code blob-code-inner js-file-line"> m <span class="pl-k">=</span> math_ops.matmul(m, <span class="pl-c1">self</span>._proj_kernel)</td> </tr> <tr> <td id="L814" class="blob-num js-line-number" data-line-number="814"></td> <td id="LC814" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L815" class="blob-num js-line-number" data-line-number="815"></td> <td id="LC815" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._proj_clip <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L816" class="blob-num js-line-number" data-line-number="816"></td> <td id="LC816" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> pylint: disable=invalid-unary-operand-type</span></td> </tr> <tr> <td id="L817" class="blob-num js-line-number" data-line-number="817"></td> <td id="LC817" class="blob-code blob-code-inner js-file-line"> m <span class="pl-k">=</span> clip_ops.clip_by_value(m, <span class="pl-k">-</span><span class="pl-c1">self</span>._proj_clip, <span class="pl-c1">self</span>._proj_clip)</td> </tr> <tr> <td id="L818" class="blob-num js-line-number" data-line-number="818"></td> <td id="LC818" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> pylint: enable=invalid-unary-operand-type</span></td> </tr> <tr> <td id="L819" class="blob-num js-line-number" data-line-number="819"></td> <td id="LC819" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L820" class="blob-num js-line-number" data-line-number="820"></td> <td id="LC820" class="blob-code blob-code-inner js-file-line"> new_state <span class="pl-k">=</span> (LSTMStateTuple(c, m) <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">else</span></td> </tr> <tr> <td id="L821" class="blob-num js-line-number" data-line-number="821"></td> <td id="LC821" class="blob-code blob-code-inner js-file-line"> array_ops.concat([c, m], <span class="pl-c1">1</span>))</td> </tr> <tr> <td id="L822" class="blob-num js-line-number" data-line-number="822"></td> <td id="LC822" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> m, new_state</td> </tr> <tr> <td id="L823" class="blob-num js-line-number" data-line-number="823"></td> <td id="LC823" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L824" class="blob-num js-line-number" data-line-number="824"></td> <td id="LC824" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L825" class="blob-num js-line-number" data-line-number="825"></td> <td id="LC825" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_enumerated_map_structure_up_to</span>(<span class="pl-smi">shallow_structure</span>, <span class="pl-smi">map_fn</span>, <span class="pl-k">*</span><span class="pl-smi">args</span>, <span class="pl-k">**</span><span class="pl-smi">kwargs</span>):</td> </tr> <tr> <td id="L826" class="blob-num js-line-number" data-line-number="826"></td> <td id="LC826" class="blob-code blob-code-inner js-file-line"> ix <span class="pl-k">=</span> [<span class="pl-c1">0</span>]</td> </tr> <tr> <td id="L827" class="blob-num js-line-number" data-line-number="827"></td> <td id="LC827" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">enumerated_fn</span>(<span class="pl-k">*</span><span class="pl-smi">inner_args</span>, <span class="pl-k">**</span><span class="pl-smi">inner_kwargs</span>):</td> </tr> <tr> <td id="L828" class="blob-num js-line-number" data-line-number="828"></td> <td id="LC828" class="blob-code blob-code-inner js-file-line"> r <span class="pl-k">=</span> map_fn(ix[<span class="pl-c1">0</span>], <span class="pl-k">*</span>inner_args, <span class="pl-k">**</span>inner_kwargs)</td> </tr> <tr> <td id="L829" class="blob-num js-line-number" data-line-number="829"></td> <td id="LC829" class="blob-code blob-code-inner js-file-line"> ix[<span class="pl-c1">0</span>] <span class="pl-k">+=</span> <span class="pl-c1">1</span></td> </tr> <tr> <td id="L830" class="blob-num js-line-number" data-line-number="830"></td> <td id="LC830" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> r</td> </tr> <tr> <td id="L831" class="blob-num js-line-number" data-line-number="831"></td> <td id="LC831" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> nest.map_structure_up_to(shallow_structure,</td> </tr> <tr> <td id="L832" class="blob-num js-line-number" data-line-number="832"></td> <td id="LC832" class="blob-code blob-code-inner js-file-line"> enumerated_fn, <span class="pl-k">*</span>args, <span class="pl-k">**</span>kwargs)</td> </tr> <tr> <td id="L833" class="blob-num js-line-number" data-line-number="833"></td> <td id="LC833" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L834" class="blob-num js-line-number" data-line-number="834"></td> <td id="LC834" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L835" class="blob-num js-line-number" data-line-number="835"></td> <td id="LC835" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_default_dropout_state_filter_visitor</span>(<span class="pl-smi">substate</span>):</td> </tr> <tr> <td id="L836" class="blob-num js-line-number" data-line-number="836"></td> <td id="LC836" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">isinstance</span>(substate, LSTMStateTuple):</td> </tr> <tr> <td id="L837" class="blob-num js-line-number" data-line-number="837"></td> <td id="LC837" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Do not perform dropout on the memory state.</span></td> </tr> <tr> <td id="L838" class="blob-num js-line-number" data-line-number="838"></td> <td id="LC838" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> LSTMStateTuple(<span class="pl-v">c</span><span class="pl-k">=</span><span class="pl-c1">False</span>, <span class="pl-v">h</span><span class="pl-k">=</span><span class="pl-c1">True</span>)</td> </tr> <tr> <td id="L839" class="blob-num js-line-number" data-line-number="839"></td> <td id="LC839" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">elif</span> <span class="pl-c1">isinstance</span>(substate, tensor_array_ops.TensorArray):</td> </tr> <tr> <td id="L840" class="blob-num js-line-number" data-line-number="840"></td> <td id="LC840" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">False</span></td> </tr> <tr> <td id="L841" class="blob-num js-line-number" data-line-number="841"></td> <td id="LC841" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L842" class="blob-num js-line-number" data-line-number="842"></td> <td id="LC842" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L843" class="blob-num js-line-number" data-line-number="843"></td> <td id="LC843" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L844" class="blob-num js-line-number" data-line-number="844"></td> <td id="LC844" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.DropoutWrapper<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L845" class="blob-num js-line-number" data-line-number="845"></td> <td id="LC845" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">DropoutWrapper</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L846" class="blob-num js-line-number" data-line-number="846"></td> <td id="LC846" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Operator adding dropout to inputs and outputs of the given cell.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L847" class="blob-num js-line-number" data-line-number="847"></td> <td id="LC847" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L848" class="blob-num js-line-number" data-line-number="848"></td> <td id="LC848" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cell</span>, <span class="pl-smi">input_keep_prob</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>, <span class="pl-smi">output_keep_prob</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>,</td> </tr> <tr> <td id="L849" class="blob-num js-line-number" data-line-number="849"></td> <td id="LC849" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">state_keep_prob</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>, <span class="pl-smi">variational_recurrent</span><span class="pl-k">=</span><span class="pl-c1">False</span>,</td> </tr> <tr> <td id="L850" class="blob-num js-line-number" data-line-number="850"></td> <td id="LC850" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">input_size</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">dtype</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">seed</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L851" class="blob-num js-line-number" data-line-number="851"></td> <td id="LC851" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">dropout_state_filter_visitor</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L852" class="blob-num js-line-number" data-line-number="852"></td> <td id="LC852" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Create a cell with added input, state, and/or output dropout.</span></td> </tr> <tr> <td id="L853" class="blob-num js-line-number" data-line-number="853"></td> <td id="LC853" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L854" class="blob-num js-line-number" data-line-number="854"></td> <td id="LC854" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If `variational_recurrent` is set to `True` (**NOT** the default behavior),</span></td> </tr> <tr> <td id="L855" class="blob-num js-line-number" data-line-number="855"></td> <td id="LC855" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> then the same dropout mask is applied at every step, as described in:</span></td> </tr> <tr> <td id="L856" class="blob-num js-line-number" data-line-number="856"></td> <td id="LC856" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L857" class="blob-num js-line-number" data-line-number="857"></td> <td id="LC857" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Y. Gal, Z Ghahramani. &quot;A Theoretically Grounded Application of Dropout in</span></td> </tr> <tr> <td id="L858" class="blob-num js-line-number" data-line-number="858"></td> <td id="LC858" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Recurrent Neural Networks&quot;. https://arxiv.org/abs/1512.05287</span></td> </tr> <tr> <td id="L859" class="blob-num js-line-number" data-line-number="859"></td> <td id="LC859" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L860" class="blob-num js-line-number" data-line-number="860"></td> <td id="LC860" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Otherwise a different dropout mask is applied at every time step.</span></td> </tr> <tr> <td id="L861" class="blob-num js-line-number" data-line-number="861"></td> <td id="LC861" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L862" class="blob-num js-line-number" data-line-number="862"></td> <td id="LC862" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Note, by default (unless a custom `dropout_state_filter` is provided),</span></td> </tr> <tr> <td id="L863" class="blob-num js-line-number" data-line-number="863"></td> <td id="LC863" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the memory state (`c` component of any `LSTMStateTuple`) passing through</span></td> </tr> <tr> <td id="L864" class="blob-num js-line-number" data-line-number="864"></td> <td id="LC864" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> a `DropoutWrapper` is never modified. This behavior is described in the</span></td> </tr> <tr> <td id="L865" class="blob-num js-line-number" data-line-number="865"></td> <td id="LC865" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> above article.</span></td> </tr> <tr> <td id="L866" class="blob-num js-line-number" data-line-number="866"></td> <td id="LC866" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L867" class="blob-num js-line-number" data-line-number="867"></td> <td id="LC867" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L868" class="blob-num js-line-number" data-line-number="868"></td> <td id="LC868" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell: an RNNCell, a projection to output_size is added to it.</span></td> </tr> <tr> <td id="L869" class="blob-num js-line-number" data-line-number="869"></td> <td id="LC869" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> input_keep_prob: unit Tensor or float between 0 and 1, input keep</span></td> </tr> <tr> <td id="L870" class="blob-num js-line-number" data-line-number="870"></td> <td id="LC870" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> probability; if it is constant and 1, no input dropout will be added.</span></td> </tr> <tr> <td id="L871" class="blob-num js-line-number" data-line-number="871"></td> <td id="LC871" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> output_keep_prob: unit Tensor or float between 0 and 1, output keep</span></td> </tr> <tr> <td id="L872" class="blob-num js-line-number" data-line-number="872"></td> <td id="LC872" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> probability; if it is constant and 1, no output dropout will be added.</span></td> </tr> <tr> <td id="L873" class="blob-num js-line-number" data-line-number="873"></td> <td id="LC873" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state_keep_prob: unit Tensor or float between 0 and 1, output keep</span></td> </tr> <tr> <td id="L874" class="blob-num js-line-number" data-line-number="874"></td> <td id="LC874" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> probability; if it is constant and 1, no output dropout will be added.</span></td> </tr> <tr> <td id="L875" class="blob-num js-line-number" data-line-number="875"></td> <td id="LC875" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> State dropout is performed on the outgoing states of the cell.</span></td> </tr> <tr> <td id="L876" class="blob-num js-line-number" data-line-number="876"></td> <td id="LC876" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> **Note** the state components to which dropout is applied when</span></td> </tr> <tr> <td id="L877" class="blob-num js-line-number" data-line-number="877"></td> <td id="LC877" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `state_keep_prob` is in `(0, 1)` are also determined by</span></td> </tr> <tr> <td id="L878" class="blob-num js-line-number" data-line-number="878"></td> <td id="LC878" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the argument `dropout_state_filter_visitor` (e.g. by default dropout</span></td> </tr> <tr> <td id="L879" class="blob-num js-line-number" data-line-number="879"></td> <td id="LC879" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> is never applied to the `c` component of an `LSTMStateTuple`).</span></td> </tr> <tr> <td id="L880" class="blob-num js-line-number" data-line-number="880"></td> <td id="LC880" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> variational_recurrent: Python bool. If `True`, then the same</span></td> </tr> <tr> <td id="L881" class="blob-num js-line-number" data-line-number="881"></td> <td id="LC881" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> dropout pattern is applied across all time steps per run call.</span></td> </tr> <tr> <td id="L882" class="blob-num js-line-number" data-line-number="882"></td> <td id="LC882" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If this parameter is set, `input_size` **must** be provided.</span></td> </tr> <tr> <td id="L883" class="blob-num js-line-number" data-line-number="883"></td> <td id="LC883" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> input_size: (optional) (possibly nested tuple of) `TensorShape` objects</span></td> </tr> <tr> <td id="L884" class="blob-num js-line-number" data-line-number="884"></td> <td id="LC884" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> containing the depth(s) of the input tensors expected to be passed in to</span></td> </tr> <tr> <td id="L885" class="blob-num js-line-number" data-line-number="885"></td> <td id="LC885" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the `DropoutWrapper`. Required and used **iff**</span></td> </tr> <tr> <td id="L886" class="blob-num js-line-number" data-line-number="886"></td> <td id="LC886" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `variational_recurrent = True` and `input_keep_prob &lt; 1`.</span></td> </tr> <tr> <td id="L887" class="blob-num js-line-number" data-line-number="887"></td> <td id="LC887" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> dtype: (optional) The `dtype` of the input, state, and output tensors.</span></td> </tr> <tr> <td id="L888" class="blob-num js-line-number" data-line-number="888"></td> <td id="LC888" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Required and used **iff** `variational_recurrent = True`.</span></td> </tr> <tr> <td id="L889" class="blob-num js-line-number" data-line-number="889"></td> <td id="LC889" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> seed: (optional) integer, the randomness seed.</span></td> </tr> <tr> <td id="L890" class="blob-num js-line-number" data-line-number="890"></td> <td id="LC890" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> dropout_state_filter_visitor: (optional), default: (see below). Function</span></td> </tr> <tr> <td id="L891" class="blob-num js-line-number" data-line-number="891"></td> <td id="LC891" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> that takes any hierarchical level of the state and returns</span></td> </tr> <tr> <td id="L892" class="blob-num js-line-number" data-line-number="892"></td> <td id="LC892" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> a scalar or depth=1 structure of Python booleans describing</span></td> </tr> <tr> <td id="L893" class="blob-num js-line-number" data-line-number="893"></td> <td id="LC893" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> which terms in the state should be dropped out. In addition, if the</span></td> </tr> <tr> <td id="L894" class="blob-num js-line-number" data-line-number="894"></td> <td id="LC894" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> function returns `True`, dropout is applied across this sublevel. If</span></td> </tr> <tr> <td id="L895" class="blob-num js-line-number" data-line-number="895"></td> <td id="LC895" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the function returns `False`, dropout is not applied across this entire</span></td> </tr> <tr> <td id="L896" class="blob-num js-line-number" data-line-number="896"></td> <td id="LC896" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> sublevel.</span></td> </tr> <tr> <td id="L897" class="blob-num js-line-number" data-line-number="897"></td> <td id="LC897" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Default behavior: perform dropout on all terms except the memory (`c`)</span></td> </tr> <tr> <td id="L898" class="blob-num js-line-number" data-line-number="898"></td> <td id="LC898" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state of `LSTMCellState` objects, and don&#39;t try to apply dropout to</span></td> </tr> <tr> <td id="L899" class="blob-num js-line-number" data-line-number="899"></td> <td id="LC899" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `TensorArray` objects:</span></td> </tr> <tr> <td id="L900" class="blob-num js-line-number" data-line-number="900"></td> <td id="LC900" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ```</span></td> </tr> <tr> <td id="L901" class="blob-num js-line-number" data-line-number="901"></td> <td id="LC901" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> def dropout_state_filter_visitor(s):</span></td> </tr> <tr> <td id="L902" class="blob-num js-line-number" data-line-number="902"></td> <td id="LC902" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> if isinstance(s, LSTMCellState):</span></td> </tr> <tr> <td id="L903" class="blob-num js-line-number" data-line-number="903"></td> <td id="LC903" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> # Never perform dropout on the c state.</span></td> </tr> <tr> <td id="L904" class="blob-num js-line-number" data-line-number="904"></td> <td id="LC904" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> return LSTMCellState(c=False, h=True)</span></td> </tr> <tr> <td id="L905" class="blob-num js-line-number" data-line-number="905"></td> <td id="LC905" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> elif isinstance(s, TensorArray):</span></td> </tr> <tr> <td id="L906" class="blob-num js-line-number" data-line-number="906"></td> <td id="LC906" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> return False</span></td> </tr> <tr> <td id="L907" class="blob-num js-line-number" data-line-number="907"></td> <td id="LC907" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> return True</span></td> </tr> <tr> <td id="L908" class="blob-num js-line-number" data-line-number="908"></td> <td id="LC908" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ```</span></td> </tr> <tr> <td id="L909" class="blob-num js-line-number" data-line-number="909"></td> <td id="LC909" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L910" class="blob-num js-line-number" data-line-number="910"></td> <td id="LC910" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L911" class="blob-num js-line-number" data-line-number="911"></td> <td id="LC911" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> TypeError: if `cell` is not an `RNNCell`, or `keep_state_fn` is provided</span></td> </tr> <tr> <td id="L912" class="blob-num js-line-number" data-line-number="912"></td> <td id="LC912" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> but not `callable`.</span></td> </tr> <tr> <td id="L913" class="blob-num js-line-number" data-line-number="913"></td> <td id="LC913" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if any of the keep_probs are not between 0 and 1.</span></td> </tr> <tr> <td id="L914" class="blob-num js-line-number" data-line-number="914"></td> <td id="LC914" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L915" class="blob-num js-line-number" data-line-number="915"></td> <td id="LC915" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> _like_rnncell(cell):</td> </tr> <tr> <td id="L916" class="blob-num js-line-number" data-line-number="916"></td> <td id="LC916" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>The parameter cell is not a RNNCell.<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L917" class="blob-num js-line-number" data-line-number="917"></td> <td id="LC917" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> (dropout_state_filter_visitor <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L918" class="blob-num js-line-number" data-line-number="918"></td> <td id="LC918" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">and</span> <span class="pl-k">not</span> <span class="pl-c1">callable</span>(dropout_state_filter_visitor)):</td> </tr> <tr> <td id="L919" class="blob-num js-line-number" data-line-number="919"></td> <td id="LC919" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>dropout_state_filter_visitor must be callable<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L920" class="blob-num js-line-number" data-line-number="920"></td> <td id="LC920" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._dropout_state_filter <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L921" class="blob-num js-line-number" data-line-number="921"></td> <td id="LC921" class="blob-code blob-code-inner js-file-line"> dropout_state_filter_visitor <span class="pl-k">or</span> _default_dropout_state_filter_visitor)</td> </tr> <tr> <td id="L922" class="blob-num js-line-number" data-line-number="922"></td> <td id="LC922" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-s"><span class="pl-pds">&quot;</span>DropoutWrapperInit<span class="pl-pds">&quot;</span></span>):</td> </tr> <tr> <td id="L923" class="blob-num js-line-number" data-line-number="923"></td> <td id="LC923" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">tensor_and_const_value</span>(<span class="pl-smi">v</span>):</td> </tr> <tr> <td id="L924" class="blob-num js-line-number" data-line-number="924"></td> <td id="LC924" class="blob-code blob-code-inner js-file-line"> tensor_value <span class="pl-k">=</span> ops.convert_to_tensor(v)</td> </tr> <tr> <td id="L925" class="blob-num js-line-number" data-line-number="925"></td> <td id="LC925" class="blob-code blob-code-inner js-file-line"> const_value <span class="pl-k">=</span> tensor_util.constant_value(tensor_value)</td> </tr> <tr> <td id="L926" class="blob-num js-line-number" data-line-number="926"></td> <td id="LC926" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> (tensor_value, const_value)</td> </tr> <tr> <td id="L927" class="blob-num js-line-number" data-line-number="927"></td> <td id="LC927" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">for</span> prob, attr <span class="pl-k">in</span> [(input_keep_prob, <span class="pl-s"><span class="pl-pds">&quot;</span>input_keep_prob<span class="pl-pds">&quot;</span></span>),</td> </tr> <tr> <td id="L928" class="blob-num js-line-number" data-line-number="928"></td> <td id="LC928" class="blob-code blob-code-inner js-file-line"> (state_keep_prob, <span class="pl-s"><span class="pl-pds">&quot;</span>state_keep_prob<span class="pl-pds">&quot;</span></span>),</td> </tr> <tr> <td id="L929" class="blob-num js-line-number" data-line-number="929"></td> <td id="LC929" class="blob-code blob-code-inner js-file-line"> (output_keep_prob, <span class="pl-s"><span class="pl-pds">&quot;</span>output_keep_prob<span class="pl-pds">&quot;</span></span>)]:</td> </tr> <tr> <td id="L930" class="blob-num js-line-number" data-line-number="930"></td> <td id="LC930" class="blob-code blob-code-inner js-file-line"> tensor_prob, const_prob <span class="pl-k">=</span> tensor_and_const_value(prob)</td> </tr> <tr> <td id="L931" class="blob-num js-line-number" data-line-number="931"></td> <td id="LC931" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> const_prob <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L932" class="blob-num js-line-number" data-line-number="932"></td> <td id="LC932" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> const_prob <span class="pl-k">&lt;</span> <span class="pl-c1">0</span> <span class="pl-k">or</span> const_prob <span class="pl-k">&gt;</span> <span class="pl-c1">1</span>:</td> </tr> <tr> <td id="L933" class="blob-num js-line-number" data-line-number="933"></td> <td id="LC933" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Parameter <span class="pl-c1">%s</span> must be between 0 and 1: <span class="pl-c1">%d</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L934" class="blob-num js-line-number" data-line-number="934"></td> <td id="LC934" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> (attr, const_prob))</td> </tr> <tr> <td id="L935" class="blob-num js-line-number" data-line-number="935"></td> <td id="LC935" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">setattr</span>(<span class="pl-c1">self</span>, <span class="pl-s"><span class="pl-pds">&quot;</span>_<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> attr, <span class="pl-c1">float</span>(const_prob))</td> </tr> <tr> <td id="L936" class="blob-num js-line-number" data-line-number="936"></td> <td id="LC936" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L937" class="blob-num js-line-number" data-line-number="937"></td> <td id="LC937" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">setattr</span>(<span class="pl-c1">self</span>, <span class="pl-s"><span class="pl-pds">&quot;</span>_<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> attr, tensor_prob)</td> </tr> <tr> <td id="L938" class="blob-num js-line-number" data-line-number="938"></td> <td id="LC938" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L939" class="blob-num js-line-number" data-line-number="939"></td> <td id="LC939" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Set cell, variational_recurrent, seed before running the code below</span></td> </tr> <tr> <td id="L940" class="blob-num js-line-number" data-line-number="940"></td> <td id="LC940" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell <span class="pl-k">=</span> cell</td> </tr> <tr> <td id="L941" class="blob-num js-line-number" data-line-number="941"></td> <td id="LC941" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._variational_recurrent <span class="pl-k">=</span> variational_recurrent</td> </tr> <tr> <td id="L942" class="blob-num js-line-number" data-line-number="942"></td> <td id="LC942" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._seed <span class="pl-k">=</span> seed</td> </tr> <tr> <td id="L943" class="blob-num js-line-number" data-line-number="943"></td> <td id="LC943" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L944" class="blob-num js-line-number" data-line-number="944"></td> <td id="LC944" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_input_noise <span class="pl-k">=</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L945" class="blob-num js-line-number" data-line-number="945"></td> <td id="LC945" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_state_noise <span class="pl-k">=</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L946" class="blob-num js-line-number" data-line-number="946"></td> <td id="LC946" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_output_noise <span class="pl-k">=</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L947" class="blob-num js-line-number" data-line-number="947"></td> <td id="LC947" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L948" class="blob-num js-line-number" data-line-number="948"></td> <td id="LC948" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> variational_recurrent:</td> </tr> <tr> <td id="L949" class="blob-num js-line-number" data-line-number="949"></td> <td id="LC949" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> dtype <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L950" class="blob-num js-line-number" data-line-number="950"></td> <td id="LC950" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(</td> </tr> <tr> <td id="L951" class="blob-num js-line-number" data-line-number="951"></td> <td id="LC951" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>When variational_recurrent=True, dtype must be provided<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L952" class="blob-num js-line-number" data-line-number="952"></td> <td id="LC952" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L953" class="blob-num js-line-number" data-line-number="953"></td> <td id="LC953" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">convert_to_batch_shape</span>(<span class="pl-smi">s</span>):</td> </tr> <tr> <td id="L954" class="blob-num js-line-number" data-line-number="954"></td> <td id="LC954" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Prepend a 1 for the batch dimension; for recurrent</span></td> </tr> <tr> <td id="L955" class="blob-num js-line-number" data-line-number="955"></td> <td id="LC955" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> variational dropout we use the same dropout mask for all</span></td> </tr> <tr> <td id="L956" class="blob-num js-line-number" data-line-number="956"></td> <td id="LC956" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> batch elements.</span></td> </tr> <tr> <td id="L957" class="blob-num js-line-number" data-line-number="957"></td> <td id="LC957" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> array_ops.concat(</td> </tr> <tr> <td id="L958" class="blob-num js-line-number" data-line-number="958"></td> <td id="LC958" class="blob-code blob-code-inner js-file-line"> ([<span class="pl-c1">1</span>], tensor_shape.TensorShape(s).as_list()), <span class="pl-c1">0</span>)</td> </tr> <tr> <td id="L959" class="blob-num js-line-number" data-line-number="959"></td> <td id="LC959" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L960" class="blob-num js-line-number" data-line-number="960"></td> <td id="LC960" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">batch_noise</span>(<span class="pl-smi">s</span>, <span class="pl-smi">inner_seed</span>):</td> </tr> <tr> <td id="L961" class="blob-num js-line-number" data-line-number="961"></td> <td id="LC961" class="blob-code blob-code-inner js-file-line"> shape <span class="pl-k">=</span> convert_to_batch_shape(s)</td> </tr> <tr> <td id="L962" class="blob-num js-line-number" data-line-number="962"></td> <td id="LC962" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> random_ops.random_uniform(shape, <span class="pl-v">seed</span><span class="pl-k">=</span>inner_seed, <span class="pl-v">dtype</span><span class="pl-k">=</span>dtype)</td> </tr> <tr> <td id="L963" class="blob-num js-line-number" data-line-number="963"></td> <td id="LC963" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L964" class="blob-num js-line-number" data-line-number="964"></td> <td id="LC964" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> (<span class="pl-k">not</span> <span class="pl-c1">isinstance</span>(<span class="pl-c1">self</span>._input_keep_prob, numbers.Real) <span class="pl-k">or</span></td> </tr> <tr> <td id="L965" class="blob-num js-line-number" data-line-number="965"></td> <td id="LC965" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._input_keep_prob <span class="pl-k">&lt;</span> <span class="pl-c1">1.0</span>):</td> </tr> <tr> <td id="L966" class="blob-num js-line-number" data-line-number="966"></td> <td id="LC966" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> input_size <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L967" class="blob-num js-line-number" data-line-number="967"></td> <td id="LC967" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(</td> </tr> <tr> <td id="L968" class="blob-num js-line-number" data-line-number="968"></td> <td id="LC968" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>When variational_recurrent=True and input_keep_prob &lt; 1.0 or <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L969" class="blob-num js-line-number" data-line-number="969"></td> <td id="LC969" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>is unknown, input_size must be provided<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L970" class="blob-num js-line-number" data-line-number="970"></td> <td id="LC970" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_input_noise <span class="pl-k">=</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L971" class="blob-num js-line-number" data-line-number="971"></td> <td id="LC971" class="blob-code blob-code-inner js-file-line"> input_size,</td> </tr> <tr> <td id="L972" class="blob-num js-line-number" data-line-number="972"></td> <td id="LC972" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">lambda</span> <span class="pl-smi">i</span>, <span class="pl-smi">s</span>: batch_noise(s, <span class="pl-v">inner_seed</span><span class="pl-k">=</span><span class="pl-c1">self</span>._gen_seed(<span class="pl-s"><span class="pl-pds">&quot;</span>input<span class="pl-pds">&quot;</span></span>, i)),</td> </tr> <tr> <td id="L973" class="blob-num js-line-number" data-line-number="973"></td> <td id="LC973" class="blob-code blob-code-inner js-file-line"> input_size)</td> </tr> <tr> <td id="L974" class="blob-num js-line-number" data-line-number="974"></td> <td id="LC974" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_state_noise <span class="pl-k">=</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L975" class="blob-num js-line-number" data-line-number="975"></td> <td id="LC975" class="blob-code blob-code-inner js-file-line"> cell.state_size,</td> </tr> <tr> <td id="L976" class="blob-num js-line-number" data-line-number="976"></td> <td id="LC976" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">lambda</span> <span class="pl-smi">i</span>, <span class="pl-smi">s</span>: batch_noise(s, <span class="pl-v">inner_seed</span><span class="pl-k">=</span><span class="pl-c1">self</span>._gen_seed(<span class="pl-s"><span class="pl-pds">&quot;</span>state<span class="pl-pds">&quot;</span></span>, i)),</td> </tr> <tr> <td id="L977" class="blob-num js-line-number" data-line-number="977"></td> <td id="LC977" class="blob-code blob-code-inner js-file-line"> cell.state_size)</td> </tr> <tr> <td id="L978" class="blob-num js-line-number" data-line-number="978"></td> <td id="LC978" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_output_noise <span class="pl-k">=</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L979" class="blob-num js-line-number" data-line-number="979"></td> <td id="LC979" class="blob-code blob-code-inner js-file-line"> cell.output_size,</td> </tr> <tr> <td id="L980" class="blob-num js-line-number" data-line-number="980"></td> <td id="LC980" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">lambda</span> <span class="pl-smi">i</span>, <span class="pl-smi">s</span>: batch_noise(s, <span class="pl-v">inner_seed</span><span class="pl-k">=</span><span class="pl-c1">self</span>._gen_seed(<span class="pl-s"><span class="pl-pds">&quot;</span>output<span class="pl-pds">&quot;</span></span>, i)),</td> </tr> <tr> <td id="L981" class="blob-num js-line-number" data-line-number="981"></td> <td id="LC981" class="blob-code blob-code-inner js-file-line"> cell.output_size)</td> </tr> <tr> <td id="L982" class="blob-num js-line-number" data-line-number="982"></td> <td id="LC982" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L983" class="blob-num js-line-number" data-line-number="983"></td> <td id="LC983" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_gen_seed</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">salt_prefix</span>, <span class="pl-smi">index</span>):</td> </tr> <tr> <td id="L984" class="blob-num js-line-number" data-line-number="984"></td> <td id="LC984" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._seed <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L985" class="blob-num js-line-number" data-line-number="985"></td> <td id="LC985" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L986" class="blob-num js-line-number" data-line-number="986"></td> <td id="LC986" class="blob-code blob-code-inner js-file-line"> salt <span class="pl-k">=</span> <span class="pl-s"><span class="pl-pds">&quot;</span><span class="pl-c1">%s</span>_<span class="pl-c1">%d</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> (salt_prefix, index)</td> </tr> <tr> <td id="L987" class="blob-num js-line-number" data-line-number="987"></td> <td id="LC987" class="blob-code blob-code-inner js-file-line"> string <span class="pl-k">=</span> (<span class="pl-c1">str</span>(<span class="pl-c1">self</span>._seed) <span class="pl-k">+</span> salt).encode(<span class="pl-s"><span class="pl-pds">&quot;</span>utf-8<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L988" class="blob-num js-line-number" data-line-number="988"></td> <td id="LC988" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">int</span>(hashlib.md5(string).hexdigest()[:<span class="pl-c1">8</span>], <span class="pl-c1">16</span>) <span class="pl-k">&amp;</span> <span class="pl-c1"><span class="pl-k">0x</span>7FFFFFFF</span></td> </tr> <tr> <td id="L989" class="blob-num js-line-number" data-line-number="989"></td> <td id="LC989" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L990" class="blob-num js-line-number" data-line-number="990"></td> <td id="LC990" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L991" class="blob-num js-line-number" data-line-number="991"></td> <td id="LC991" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">wrapped_cell</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L992" class="blob-num js-line-number" data-line-number="992"></td> <td id="LC992" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell</td> </tr> <tr> <td id="L993" class="blob-num js-line-number" data-line-number="993"></td> <td id="LC993" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L994" class="blob-num js-line-number" data-line-number="994"></td> <td id="LC994" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L995" class="blob-num js-line-number" data-line-number="995"></td> <td id="LC995" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L996" class="blob-num js-line-number" data-line-number="996"></td> <td id="LC996" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.state_size</td> </tr> <tr> <td id="L997" class="blob-num js-line-number" data-line-number="997"></td> <td id="LC997" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L998" class="blob-num js-line-number" data-line-number="998"></td> <td id="LC998" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L999" class="blob-num js-line-number" data-line-number="999"></td> <td id="LC999" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1000" class="blob-num js-line-number" data-line-number="1000"></td> <td id="LC1000" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.output_size</td> </tr> <tr> <td id="L1001" class="blob-num js-line-number" data-line-number="1001"></td> <td id="LC1001" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1002" class="blob-num js-line-number" data-line-number="1002"></td> <td id="LC1002" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L1003" class="blob-num js-line-number" data-line-number="1003"></td> <td id="LC1003" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L1004" class="blob-num js-line-number" data-line-number="1004"></td> <td id="LC1004" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.zero_state(batch_size, dtype)</td> </tr> <tr> <td id="L1005" class="blob-num js-line-number" data-line-number="1005"></td> <td id="LC1005" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1006" class="blob-num js-line-number" data-line-number="1006"></td> <td id="LC1006" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_variational_recurrent_dropout_value</span>(</td> </tr> <tr> <td id="L1007" class="blob-num js-line-number" data-line-number="1007"></td> <td id="LC1007" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">index</span>, <span class="pl-smi">value</span>, <span class="pl-smi">noise</span>, <span class="pl-smi">keep_prob</span>):</td> </tr> <tr> <td id="L1008" class="blob-num js-line-number" data-line-number="1008"></td> <td id="LC1008" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Performs dropout given the pre-calculated noise tensor.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1009" class="blob-num js-line-number" data-line-number="1009"></td> <td id="LC1009" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> uniform [keep_prob, 1.0 + keep_prob)</span></td> </tr> <tr> <td id="L1010" class="blob-num js-line-number" data-line-number="1010"></td> <td id="LC1010" class="blob-code blob-code-inner js-file-line"> random_tensor <span class="pl-k">=</span> keep_prob <span class="pl-k">+</span> noise</td> </tr> <tr> <td id="L1011" class="blob-num js-line-number" data-line-number="1011"></td> <td id="LC1011" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1012" class="blob-num js-line-number" data-line-number="1012"></td> <td id="LC1012" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> 0. if [keep_prob, 1.0) and 1. if [1.0, 1.0 + keep_prob)</span></td> </tr> <tr> <td id="L1013" class="blob-num js-line-number" data-line-number="1013"></td> <td id="LC1013" class="blob-code blob-code-inner js-file-line"> binary_tensor <span class="pl-k">=</span> math_ops.floor(random_tensor)</td> </tr> <tr> <td id="L1014" class="blob-num js-line-number" data-line-number="1014"></td> <td id="LC1014" class="blob-code blob-code-inner js-file-line"> ret <span class="pl-k">=</span> math_ops.div(value, keep_prob) <span class="pl-k">*</span> binary_tensor</td> </tr> <tr> <td id="L1015" class="blob-num js-line-number" data-line-number="1015"></td> <td id="LC1015" class="blob-code blob-code-inner js-file-line"> ret.set_shape(value.get_shape())</td> </tr> <tr> <td id="L1016" class="blob-num js-line-number" data-line-number="1016"></td> <td id="LC1016" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> ret</td> </tr> <tr> <td id="L1017" class="blob-num js-line-number" data-line-number="1017"></td> <td id="LC1017" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1018" class="blob-num js-line-number" data-line-number="1018"></td> <td id="LC1018" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_dropout</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">values</span>, <span class="pl-smi">salt_prefix</span>, <span class="pl-smi">recurrent_noise</span>, <span class="pl-smi">keep_prob</span>,</td> </tr> <tr> <td id="L1019" class="blob-num js-line-number" data-line-number="1019"></td> <td id="LC1019" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">shallow_filtered_substructure</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1020" class="blob-num js-line-number" data-line-number="1020"></td> <td id="LC1020" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Decides whether to perform standard dropout or recurrent dropout.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1021" class="blob-num js-line-number" data-line-number="1021"></td> <td id="LC1021" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1022" class="blob-num js-line-number" data-line-number="1022"></td> <td id="LC1022" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> shallow_filtered_substructure <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L1023" class="blob-num js-line-number" data-line-number="1023"></td> <td id="LC1023" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Put something so we traverse the entire structure; inside the</span></td> </tr> <tr> <td id="L1024" class="blob-num js-line-number" data-line-number="1024"></td> <td id="LC1024" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> dropout function we check to see if leafs of this are bool or not.</span></td> </tr> <tr> <td id="L1025" class="blob-num js-line-number" data-line-number="1025"></td> <td id="LC1025" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure <span class="pl-k">=</span> values</td> </tr> <tr> <td id="L1026" class="blob-num js-line-number" data-line-number="1026"></td> <td id="LC1026" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1027" class="blob-num js-line-number" data-line-number="1027"></td> <td id="LC1027" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">self</span>._variational_recurrent:</td> </tr> <tr> <td id="L1028" class="blob-num js-line-number" data-line-number="1028"></td> <td id="LC1028" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">dropout</span>(<span class="pl-smi">i</span>, <span class="pl-smi">do_dropout</span>, <span class="pl-smi">v</span>):</td> </tr> <tr> <td id="L1029" class="blob-num js-line-number" data-line-number="1029"></td> <td id="LC1029" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">isinstance</span>(do_dropout, <span class="pl-c1">bool</span>) <span class="pl-k">or</span> do_dropout:</td> </tr> <tr> <td id="L1030" class="blob-num js-line-number" data-line-number="1030"></td> <td id="LC1030" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> nn_ops.dropout(</td> </tr> <tr> <td id="L1031" class="blob-num js-line-number" data-line-number="1031"></td> <td id="LC1031" class="blob-code blob-code-inner js-file-line"> v, <span class="pl-v">keep_prob</span><span class="pl-k">=</span>keep_prob, <span class="pl-v">seed</span><span class="pl-k">=</span><span class="pl-c1">self</span>._gen_seed(salt_prefix, i))</td> </tr> <tr> <td id="L1032" class="blob-num js-line-number" data-line-number="1032"></td> <td id="LC1032" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1033" class="blob-num js-line-number" data-line-number="1033"></td> <td id="LC1033" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> v</td> </tr> <tr> <td id="L1034" class="blob-num js-line-number" data-line-number="1034"></td> <td id="LC1034" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L1035" class="blob-num js-line-number" data-line-number="1035"></td> <td id="LC1035" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure, dropout,</td> </tr> <tr> <td id="L1036" class="blob-num js-line-number" data-line-number="1036"></td> <td id="LC1036" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">*</span>[shallow_filtered_substructure, values])</td> </tr> <tr> <td id="L1037" class="blob-num js-line-number" data-line-number="1037"></td> <td id="LC1037" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1038" class="blob-num js-line-number" data-line-number="1038"></td> <td id="LC1038" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">dropout</span>(<span class="pl-smi">i</span>, <span class="pl-smi">do_dropout</span>, <span class="pl-smi">v</span>, <span class="pl-smi">n</span>):</td> </tr> <tr> <td id="L1039" class="blob-num js-line-number" data-line-number="1039"></td> <td id="LC1039" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">isinstance</span>(do_dropout, <span class="pl-c1">bool</span>) <span class="pl-k">or</span> do_dropout:</td> </tr> <tr> <td id="L1040" class="blob-num js-line-number" data-line-number="1040"></td> <td id="LC1040" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._variational_recurrent_dropout_value(i, v, n, keep_prob)</td> </tr> <tr> <td id="L1041" class="blob-num js-line-number" data-line-number="1041"></td> <td id="LC1041" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1042" class="blob-num js-line-number" data-line-number="1042"></td> <td id="LC1042" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> v</td> </tr> <tr> <td id="L1043" class="blob-num js-line-number" data-line-number="1043"></td> <td id="LC1043" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L1044" class="blob-num js-line-number" data-line-number="1044"></td> <td id="LC1044" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure, dropout,</td> </tr> <tr> <td id="L1045" class="blob-num js-line-number" data-line-number="1045"></td> <td id="LC1045" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">*</span>[shallow_filtered_substructure, values, recurrent_noise])</td> </tr> <tr> <td id="L1046" class="blob-num js-line-number" data-line-number="1046"></td> <td id="LC1046" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1047" class="blob-num js-line-number" data-line-number="1047"></td> <td id="LC1047" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1048" class="blob-num js-line-number" data-line-number="1048"></td> <td id="LC1048" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run the cell with the declared dropouts.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1049" class="blob-num js-line-number" data-line-number="1049"></td> <td id="LC1049" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_should_dropout</span>(<span class="pl-smi">p</span>):</td> </tr> <tr> <td id="L1050" class="blob-num js-line-number" data-line-number="1050"></td> <td id="LC1050" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> (<span class="pl-k">not</span> <span class="pl-c1">isinstance</span>(p, <span class="pl-c1">float</span>)) <span class="pl-k">or</span> p <span class="pl-k">&lt;</span> <span class="pl-c1">1</span></td> </tr> <tr> <td id="L1051" class="blob-num js-line-number" data-line-number="1051"></td> <td id="LC1051" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1052" class="blob-num js-line-number" data-line-number="1052"></td> <td id="LC1052" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> _should_dropout(<span class="pl-c1">self</span>._input_keep_prob):</td> </tr> <tr> <td id="L1053" class="blob-num js-line-number" data-line-number="1053"></td> <td id="LC1053" class="blob-code blob-code-inner js-file-line"> inputs <span class="pl-k">=</span> <span class="pl-c1">self</span>._dropout(inputs, <span class="pl-s"><span class="pl-pds">&quot;</span>input<span class="pl-pds">&quot;</span></span>,</td> </tr> <tr> <td id="L1054" class="blob-num js-line-number" data-line-number="1054"></td> <td id="LC1054" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_input_noise,</td> </tr> <tr> <td id="L1055" class="blob-num js-line-number" data-line-number="1055"></td> <td id="LC1055" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._input_keep_prob)</td> </tr> <tr> <td id="L1056" class="blob-num js-line-number" data-line-number="1056"></td> <td id="LC1056" class="blob-code blob-code-inner js-file-line"> output, new_state <span class="pl-k">=</span> <span class="pl-c1">self</span>._cell(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L1057" class="blob-num js-line-number" data-line-number="1057"></td> <td id="LC1057" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> _should_dropout(<span class="pl-c1">self</span>._state_keep_prob):</td> </tr> <tr> <td id="L1058" class="blob-num js-line-number" data-line-number="1058"></td> <td id="LC1058" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Identify which subsets of the state to perform dropout on and</span></td> </tr> <tr> <td id="L1059" class="blob-num js-line-number" data-line-number="1059"></td> <td id="LC1059" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> which ones to keep.</span></td> </tr> <tr> <td id="L1060" class="blob-num js-line-number" data-line-number="1060"></td> <td id="LC1060" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure <span class="pl-k">=</span> nest.get_traverse_shallow_structure(</td> </tr> <tr> <td id="L1061" class="blob-num js-line-number" data-line-number="1061"></td> <td id="LC1061" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._dropout_state_filter, new_state)</td> </tr> <tr> <td id="L1062" class="blob-num js-line-number" data-line-number="1062"></td> <td id="LC1062" class="blob-code blob-code-inner js-file-line"> new_state <span class="pl-k">=</span> <span class="pl-c1">self</span>._dropout(new_state, <span class="pl-s"><span class="pl-pds">&quot;</span>state<span class="pl-pds">&quot;</span></span>,</td> </tr> <tr> <td id="L1063" class="blob-num js-line-number" data-line-number="1063"></td> <td id="LC1063" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_state_noise,</td> </tr> <tr> <td id="L1064" class="blob-num js-line-number" data-line-number="1064"></td> <td id="LC1064" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_keep_prob,</td> </tr> <tr> <td id="L1065" class="blob-num js-line-number" data-line-number="1065"></td> <td id="LC1065" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure)</td> </tr> <tr> <td id="L1066" class="blob-num js-line-number" data-line-number="1066"></td> <td id="LC1066" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> _should_dropout(<span class="pl-c1">self</span>._output_keep_prob):</td> </tr> <tr> <td id="L1067" class="blob-num js-line-number" data-line-number="1067"></td> <td id="LC1067" class="blob-code blob-code-inner js-file-line"> output <span class="pl-k">=</span> <span class="pl-c1">self</span>._dropout(output, <span class="pl-s"><span class="pl-pds">&quot;</span>output<span class="pl-pds">&quot;</span></span>,</td> </tr> <tr> <td id="L1068" class="blob-num js-line-number" data-line-number="1068"></td> <td id="LC1068" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_output_noise,</td> </tr> <tr> <td id="L1069" class="blob-num js-line-number" data-line-number="1069"></td> <td id="LC1069" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._output_keep_prob)</td> </tr> <tr> <td id="L1070" class="blob-num js-line-number" data-line-number="1070"></td> <td id="LC1070" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> output, new_state</td> </tr> <tr> <td id="L1071" class="blob-num js-line-number" data-line-number="1071"></td> <td id="LC1071" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1072" class="blob-num js-line-number" data-line-number="1072"></td> <td id="LC1072" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1073" class="blob-num js-line-number" data-line-number="1073"></td> <td id="LC1073" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.ResidualWrapper<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L1074" class="blob-num js-line-number" data-line-number="1074"></td> <td id="LC1074" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">ResidualWrapper</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L1075" class="blob-num js-line-number" data-line-number="1075"></td> <td id="LC1075" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>RNNCell wrapper that ensures cell inputs are added to the outputs.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1076" class="blob-num js-line-number" data-line-number="1076"></td> <td id="LC1076" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1077" class="blob-num js-line-number" data-line-number="1077"></td> <td id="LC1077" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cell</span>, <span class="pl-smi">residual_fn</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1078" class="blob-num js-line-number" data-line-number="1078"></td> <td id="LC1078" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Constructs a `ResidualWrapper` for `cell`.</span></td> </tr> <tr> <td id="L1079" class="blob-num js-line-number" data-line-number="1079"></td> <td id="LC1079" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1080" class="blob-num js-line-number" data-line-number="1080"></td> <td id="LC1080" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1081" class="blob-num js-line-number" data-line-number="1081"></td> <td id="LC1081" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell: An instance of `RNNCell`.</span></td> </tr> <tr> <td id="L1082" class="blob-num js-line-number" data-line-number="1082"></td> <td id="LC1082" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> residual_fn: (Optional) The function to map raw cell inputs and raw cell</span></td> </tr> <tr> <td id="L1083" class="blob-num js-line-number" data-line-number="1083"></td> <td id="LC1083" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> outputs to the actual cell outputs of the residual network.</span></td> </tr> <tr> <td id="L1084" class="blob-num js-line-number" data-line-number="1084"></td> <td id="LC1084" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Defaults to calling nest.map_structure on (lambda i, o: i + o), inputs</span></td> </tr> <tr> <td id="L1085" class="blob-num js-line-number" data-line-number="1085"></td> <td id="LC1085" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> and outputs.</span></td> </tr> <tr> <td id="L1086" class="blob-num js-line-number" data-line-number="1086"></td> <td id="LC1086" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1087" class="blob-num js-line-number" data-line-number="1087"></td> <td id="LC1087" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell <span class="pl-k">=</span> cell</td> </tr> <tr> <td id="L1088" class="blob-num js-line-number" data-line-number="1088"></td> <td id="LC1088" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._residual_fn <span class="pl-k">=</span> residual_fn</td> </tr> <tr> <td id="L1089" class="blob-num js-line-number" data-line-number="1089"></td> <td id="LC1089" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1090" class="blob-num js-line-number" data-line-number="1090"></td> <td id="LC1090" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1091" class="blob-num js-line-number" data-line-number="1091"></td> <td id="LC1091" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1092" class="blob-num js-line-number" data-line-number="1092"></td> <td id="LC1092" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.state_size</td> </tr> <tr> <td id="L1093" class="blob-num js-line-number" data-line-number="1093"></td> <td id="LC1093" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1094" class="blob-num js-line-number" data-line-number="1094"></td> <td id="LC1094" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1095" class="blob-num js-line-number" data-line-number="1095"></td> <td id="LC1095" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1096" class="blob-num js-line-number" data-line-number="1096"></td> <td id="LC1096" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.output_size</td> </tr> <tr> <td id="L1097" class="blob-num js-line-number" data-line-number="1097"></td> <td id="LC1097" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1098" class="blob-num js-line-number" data-line-number="1098"></td> <td id="LC1098" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L1099" class="blob-num js-line-number" data-line-number="1099"></td> <td id="LC1099" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L1100" class="blob-num js-line-number" data-line-number="1100"></td> <td id="LC1100" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.zero_state(batch_size, dtype)</td> </tr> <tr> <td id="L1101" class="blob-num js-line-number" data-line-number="1101"></td> <td id="LC1101" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1102" class="blob-num js-line-number" data-line-number="1102"></td> <td id="LC1102" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1103" class="blob-num js-line-number" data-line-number="1103"></td> <td id="LC1103" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run the cell and then apply the residual_fn on its inputs to its outputs.</span></td> </tr> <tr> <td id="L1104" class="blob-num js-line-number" data-line-number="1104"></td> <td id="LC1104" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1105" class="blob-num js-line-number" data-line-number="1105"></td> <td id="LC1105" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1106" class="blob-num js-line-number" data-line-number="1106"></td> <td id="LC1106" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: cell inputs.</span></td> </tr> <tr> <td id="L1107" class="blob-num js-line-number" data-line-number="1107"></td> <td id="LC1107" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: cell state.</span></td> </tr> <tr> <td id="L1108" class="blob-num js-line-number" data-line-number="1108"></td> <td id="LC1108" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> scope: optional cell scope.</span></td> </tr> <tr> <td id="L1109" class="blob-num js-line-number" data-line-number="1109"></td> <td id="LC1109" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1110" class="blob-num js-line-number" data-line-number="1110"></td> <td id="LC1110" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L1111" class="blob-num js-line-number" data-line-number="1111"></td> <td id="LC1111" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Tuple of cell outputs and new state.</span></td> </tr> <tr> <td id="L1112" class="blob-num js-line-number" data-line-number="1112"></td> <td id="LC1112" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1113" class="blob-num js-line-number" data-line-number="1113"></td> <td id="LC1113" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L1114" class="blob-num js-line-number" data-line-number="1114"></td> <td id="LC1114" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> TypeError: If cell inputs and outputs have different structure (type).</span></td> </tr> <tr> <td id="L1115" class="blob-num js-line-number" data-line-number="1115"></td> <td id="LC1115" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: If cell inputs and outputs have different structure (value).</span></td> </tr> <tr> <td id="L1116" class="blob-num js-line-number" data-line-number="1116"></td> <td id="LC1116" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1117" class="blob-num js-line-number" data-line-number="1117"></td> <td id="LC1117" class="blob-code blob-code-inner js-file-line"> outputs, new_state <span class="pl-k">=</span> <span class="pl-c1">self</span>._cell(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L1118" class="blob-num js-line-number" data-line-number="1118"></td> <td id="LC1118" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Ensure shapes match</span></td> </tr> <tr> <td id="L1119" class="blob-num js-line-number" data-line-number="1119"></td> <td id="LC1119" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">assert_shape_match</span>(<span class="pl-smi">inp</span>, <span class="pl-smi">out</span>):</td> </tr> <tr> <td id="L1120" class="blob-num js-line-number" data-line-number="1120"></td> <td id="LC1120" class="blob-code blob-code-inner js-file-line"> inp.get_shape().assert_is_compatible_with(out.get_shape())</td> </tr> <tr> <td id="L1121" class="blob-num js-line-number" data-line-number="1121"></td> <td id="LC1121" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">default_residual_fn</span>(<span class="pl-smi">inputs</span>, <span class="pl-smi">outputs</span>):</td> </tr> <tr> <td id="L1122" class="blob-num js-line-number" data-line-number="1122"></td> <td id="LC1122" class="blob-code blob-code-inner js-file-line"> nest.assert_same_structure(inputs, outputs)</td> </tr> <tr> <td id="L1123" class="blob-num js-line-number" data-line-number="1123"></td> <td id="LC1123" class="blob-code blob-code-inner js-file-line"> nest.map_structure(assert_shape_match, inputs, outputs)</td> </tr> <tr> <td id="L1124" class="blob-num js-line-number" data-line-number="1124"></td> <td id="LC1124" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> nest.map_structure(<span class="pl-k">lambda</span> <span class="pl-smi">inp</span>, <span class="pl-smi">out</span>: inp <span class="pl-k">+</span> out, inputs, outputs)</td> </tr> <tr> <td id="L1125" class="blob-num js-line-number" data-line-number="1125"></td> <td id="LC1125" class="blob-code blob-code-inner js-file-line"> res_outputs <span class="pl-k">=</span> (<span class="pl-c1">self</span>._residual_fn <span class="pl-k">or</span> default_residual_fn)(inputs, outputs)</td> </tr> <tr> <td id="L1126" class="blob-num js-line-number" data-line-number="1126"></td> <td id="LC1126" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> (res_outputs, new_state)</td> </tr> <tr> <td id="L1127" class="blob-num js-line-number" data-line-number="1127"></td> <td id="LC1127" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1128" class="blob-num js-line-number" data-line-number="1128"></td> <td id="LC1128" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1129" class="blob-num js-line-number" data-line-number="1129"></td> <td id="LC1129" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.DeviceWrapper<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L1130" class="blob-num js-line-number" data-line-number="1130"></td> <td id="LC1130" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">DeviceWrapper</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L1131" class="blob-num js-line-number" data-line-number="1131"></td> <td id="LC1131" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Operator that ensures an RNNCell runs on a particular device.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1132" class="blob-num js-line-number" data-line-number="1132"></td> <td id="LC1132" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1133" class="blob-num js-line-number" data-line-number="1133"></td> <td id="LC1133" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cell</span>, <span class="pl-smi">device</span>):</td> </tr> <tr> <td id="L1134" class="blob-num js-line-number" data-line-number="1134"></td> <td id="LC1134" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Construct a `DeviceWrapper` for `cell` with device `device`.</span></td> </tr> <tr> <td id="L1135" class="blob-num js-line-number" data-line-number="1135"></td> <td id="LC1135" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1136" class="blob-num js-line-number" data-line-number="1136"></td> <td id="LC1136" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Ensures the wrapped `cell` is called with `tf.device(device)`.</span></td> </tr> <tr> <td id="L1137" class="blob-num js-line-number" data-line-number="1137"></td> <td id="LC1137" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1138" class="blob-num js-line-number" data-line-number="1138"></td> <td id="LC1138" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1139" class="blob-num js-line-number" data-line-number="1139"></td> <td id="LC1139" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell: An instance of `RNNCell`.</span></td> </tr> <tr> <td id="L1140" class="blob-num js-line-number" data-line-number="1140"></td> <td id="LC1140" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> device: A device string or function, for passing to `tf.device`.</span></td> </tr> <tr> <td id="L1141" class="blob-num js-line-number" data-line-number="1141"></td> <td id="LC1141" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1142" class="blob-num js-line-number" data-line-number="1142"></td> <td id="LC1142" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell <span class="pl-k">=</span> cell</td> </tr> <tr> <td id="L1143" class="blob-num js-line-number" data-line-number="1143"></td> <td id="LC1143" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._device <span class="pl-k">=</span> device</td> </tr> <tr> <td id="L1144" class="blob-num js-line-number" data-line-number="1144"></td> <td id="LC1144" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1145" class="blob-num js-line-number" data-line-number="1145"></td> <td id="LC1145" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1146" class="blob-num js-line-number" data-line-number="1146"></td> <td id="LC1146" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1147" class="blob-num js-line-number" data-line-number="1147"></td> <td id="LC1147" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.state_size</td> </tr> <tr> <td id="L1148" class="blob-num js-line-number" data-line-number="1148"></td> <td id="LC1148" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1149" class="blob-num js-line-number" data-line-number="1149"></td> <td id="LC1149" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1150" class="blob-num js-line-number" data-line-number="1150"></td> <td id="LC1150" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1151" class="blob-num js-line-number" data-line-number="1151"></td> <td id="LC1151" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.output_size</td> </tr> <tr> <td id="L1152" class="blob-num js-line-number" data-line-number="1152"></td> <td id="LC1152" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1153" class="blob-num js-line-number" data-line-number="1153"></td> <td id="LC1153" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L1154" class="blob-num js-line-number" data-line-number="1154"></td> <td id="LC1154" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L1155" class="blob-num js-line-number" data-line-number="1155"></td> <td id="LC1155" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.device(<span class="pl-c1">self</span>._device):</td> </tr> <tr> <td id="L1156" class="blob-num js-line-number" data-line-number="1156"></td> <td id="LC1156" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.zero_state(batch_size, dtype)</td> </tr> <tr> <td id="L1157" class="blob-num js-line-number" data-line-number="1157"></td> <td id="LC1157" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1158" class="blob-num js-line-number" data-line-number="1158"></td> <td id="LC1158" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1159" class="blob-num js-line-number" data-line-number="1159"></td> <td id="LC1159" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run the cell on specified device.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1160" class="blob-num js-line-number" data-line-number="1160"></td> <td id="LC1160" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.device(<span class="pl-c1">self</span>._device):</td> </tr> <tr> <td id="L1161" class="blob-num js-line-number" data-line-number="1161"></td> <td id="LC1161" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L1162" class="blob-num js-line-number" data-line-number="1162"></td> <td id="LC1162" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1163" class="blob-num js-line-number" data-line-number="1163"></td> <td id="LC1163" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1164" class="blob-num js-line-number" data-line-number="1164"></td> <td id="LC1164" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.MultiRNNCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L1165" class="blob-num js-line-number" data-line-number="1165"></td> <td id="LC1165" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">MultiRNNCell</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L1166" class="blob-num js-line-number" data-line-number="1166"></td> <td id="LC1166" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>RNN cell composed sequentially of multiple simple cells.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1167" class="blob-num js-line-number" data-line-number="1167"></td> <td id="LC1167" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1168" class="blob-num js-line-number" data-line-number="1168"></td> <td id="LC1168" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cells</span>, <span class="pl-smi">state_is_tuple</span><span class="pl-k">=</span><span class="pl-c1">True</span>):</td> </tr> <tr> <td id="L1169" class="blob-num js-line-number" data-line-number="1169"></td> <td id="LC1169" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Create a RNN cell composed sequentially of a number of RNNCells.</span></td> </tr> <tr> <td id="L1170" class="blob-num js-line-number" data-line-number="1170"></td> <td id="LC1170" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1171" class="blob-num js-line-number" data-line-number="1171"></td> <td id="LC1171" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1172" class="blob-num js-line-number" data-line-number="1172"></td> <td id="LC1172" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cells: list of RNNCells that will be composed in this order.</span></td> </tr> <tr> <td id="L1173" class="blob-num js-line-number" data-line-number="1173"></td> <td id="LC1173" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state_is_tuple: If True, accepted and returned states are n-tuples, where</span></td> </tr> <tr> <td id="L1174" class="blob-num js-line-number" data-line-number="1174"></td> <td id="LC1174" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `n = len(cells)`. If False, the states are all</span></td> </tr> <tr> <td id="L1175" class="blob-num js-line-number" data-line-number="1175"></td> <td id="LC1175" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> concatenated along the column axis. This latter behavior will soon be</span></td> </tr> <tr> <td id="L1176" class="blob-num js-line-number" data-line-number="1176"></td> <td id="LC1176" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> deprecated.</span></td> </tr> <tr> <td id="L1177" class="blob-num js-line-number" data-line-number="1177"></td> <td id="LC1177" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1178" class="blob-num js-line-number" data-line-number="1178"></td> <td id="LC1178" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L1179" class="blob-num js-line-number" data-line-number="1179"></td> <td id="LC1179" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if cells is empty (not allowed), or at least one of the cells</span></td> </tr> <tr> <td id="L1180" class="blob-num js-line-number" data-line-number="1180"></td> <td id="LC1180" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> returns a state tuple but the flag `state_is_tuple` is `False`.</span></td> </tr> <tr> <td id="L1181" class="blob-num js-line-number" data-line-number="1181"></td> <td id="LC1181" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1182" class="blob-num js-line-number" data-line-number="1182"></td> <td id="LC1182" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(MultiRNNCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>()</td> </tr> <tr> <td id="L1183" class="blob-num js-line-number" data-line-number="1183"></td> <td id="LC1183" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> cells:</td> </tr> <tr> <td id="L1184" class="blob-num js-line-number" data-line-number="1184"></td> <td id="LC1184" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Must specify at least one cell for MultiRNNCell.<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L1185" class="blob-num js-line-number" data-line-number="1185"></td> <td id="LC1185" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> nest.is_sequence(cells):</td> </tr> <tr> <td id="L1186" class="blob-num js-line-number" data-line-number="1186"></td> <td id="LC1186" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(</td> </tr> <tr> <td id="L1187" class="blob-num js-line-number" data-line-number="1187"></td> <td id="LC1187" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>cells must be a list or tuple, but saw: <span class="pl-c1">%s</span>.<span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> cells)</td> </tr> <tr> <td id="L1188" class="blob-num js-line-number" data-line-number="1188"></td> <td id="LC1188" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1189" class="blob-num js-line-number" data-line-number="1189"></td> <td id="LC1189" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cells <span class="pl-k">=</span> cells</td> </tr> <tr> <td id="L1190" class="blob-num js-line-number" data-line-number="1190"></td> <td id="LC1190" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">=</span> state_is_tuple</td> </tr> <tr> <td id="L1191" class="blob-num js-line-number" data-line-number="1191"></td> <td id="LC1191" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> state_is_tuple:</td> </tr> <tr> <td id="L1192" class="blob-num js-line-number" data-line-number="1192"></td> <td id="LC1192" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">any</span>(nest.is_sequence(c.state_size) <span class="pl-k">for</span> c <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells):</td> </tr> <tr> <td id="L1193" class="blob-num js-line-number" data-line-number="1193"></td> <td id="LC1193" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Some cells return tuples of states, but the flag <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L1194" class="blob-num js-line-number" data-line-number="1194"></td> <td id="LC1194" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>state_is_tuple is not set. State sizes are: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L1195" class="blob-num js-line-number" data-line-number="1195"></td> <td id="LC1195" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> <span class="pl-c1">str</span>([c.state_size <span class="pl-k">for</span> c <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells]))</td> </tr> <tr> <td id="L1196" class="blob-num js-line-number" data-line-number="1196"></td> <td id="LC1196" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1197" class="blob-num js-line-number" data-line-number="1197"></td> <td id="LC1197" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1198" class="blob-num js-line-number" data-line-number="1198"></td> <td id="LC1198" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1199" class="blob-num js-line-number" data-line-number="1199"></td> <td id="LC1199" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L1200" class="blob-num js-line-number" data-line-number="1200"></td> <td id="LC1200" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">tuple</span>(cell.state_size <span class="pl-k">for</span> cell <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells)</td> </tr> <tr> <td id="L1201" class="blob-num js-line-number" data-line-number="1201"></td> <td id="LC1201" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1202" class="blob-num js-line-number" data-line-number="1202"></td> <td id="LC1202" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">sum</span>([cell.state_size <span class="pl-k">for</span> cell <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells])</td> </tr> <tr> <td id="L1203" class="blob-num js-line-number" data-line-number="1203"></td> <td id="LC1203" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1204" class="blob-num js-line-number" data-line-number="1204"></td> <td id="LC1204" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1205" class="blob-num js-line-number" data-line-number="1205"></td> <td id="LC1205" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1206" class="blob-num js-line-number" data-line-number="1206"></td> <td id="LC1206" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cells[<span class="pl-k">-</span><span class="pl-c1">1</span>].output_size</td> </tr> <tr> <td id="L1207" class="blob-num js-line-number" data-line-number="1207"></td> <td id="LC1207" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1208" class="blob-num js-line-number" data-line-number="1208"></td> <td id="LC1208" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L1209" class="blob-num js-line-number" data-line-number="1209"></td> <td id="LC1209" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L1210" class="blob-num js-line-number" data-line-number="1210"></td> <td id="LC1210" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L1211" class="blob-num js-line-number" data-line-number="1211"></td> <td id="LC1211" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">tuple</span>(cell.zero_state(batch_size, dtype) <span class="pl-k">for</span> cell <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells)</td> </tr> <tr> <td id="L1212" class="blob-num js-line-number" data-line-number="1212"></td> <td id="LC1212" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1213" class="blob-num js-line-number" data-line-number="1213"></td> <td id="LC1213" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> We know here that state_size of each cell is not a tuple and</span></td> </tr> <tr> <td id="L1214" class="blob-num js-line-number" data-line-number="1214"></td> <td id="LC1214" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> presumably does not contain TensorArrays or anything else fancy</span></td> </tr> <tr> <td id="L1215" class="blob-num js-line-number" data-line-number="1215"></td> <td id="LC1215" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">super</span>(MultiRNNCell, <span class="pl-c1">self</span>).zero_state(batch_size, dtype)</td> </tr> <tr> <td id="L1216" class="blob-num js-line-number" data-line-number="1216"></td> <td id="LC1216" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1217" class="blob-num js-line-number" data-line-number="1217"></td> <td id="LC1217" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L1218" class="blob-num js-line-number" data-line-number="1218"></td> <td id="LC1218" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run this multi-layer cell on inputs, starting from state.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1219" class="blob-num js-line-number" data-line-number="1219"></td> <td id="LC1219" class="blob-code blob-code-inner js-file-line"> cur_state_pos <span class="pl-k">=</span> <span class="pl-c1">0</span></td> </tr> <tr> <td id="L1220" class="blob-num js-line-number" data-line-number="1220"></td> <td id="LC1220" class="blob-code blob-code-inner js-file-line"> cur_inp <span class="pl-k">=</span> inputs</td> </tr> <tr> <td id="L1221" class="blob-num js-line-number" data-line-number="1221"></td> <td id="LC1221" class="blob-code blob-code-inner js-file-line"> new_states <span class="pl-k">=</span> []</td> </tr> <tr> <td id="L1222" class="blob-num js-line-number" data-line-number="1222"></td> <td id="LC1222" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">for</span> i, cell <span class="pl-k">in</span> <span class="pl-c1">enumerate</span>(<span class="pl-c1">self</span>._cells):</td> </tr> <tr> <td id="L1223" class="blob-num js-line-number" data-line-number="1223"></td> <td id="LC1223" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> vs.variable_scope(<span class="pl-s"><span class="pl-pds">&quot;</span>cell_<span class="pl-c1">%d</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> i):</td> </tr> <tr> <td id="L1224" class="blob-num js-line-number" data-line-number="1224"></td> <td id="LC1224" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L1225" class="blob-num js-line-number" data-line-number="1225"></td> <td id="LC1225" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> nest.is_sequence(state):</td> </tr> <tr> <td id="L1226" class="blob-num js-line-number" data-line-number="1226"></td> <td id="LC1226" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(</td> </tr> <tr> <td id="L1227" class="blob-num js-line-number" data-line-number="1227"></td> <td id="LC1227" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>Expected state to be a tuple of length <span class="pl-c1">%d</span>, but received: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span></td> </tr> <tr> <td id="L1228" class="blob-num js-line-number" data-line-number="1228"></td> <td id="LC1228" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">len</span>(<span class="pl-c1">self</span>.state_size), state))</td> </tr> <tr> <td id="L1229" class="blob-num js-line-number" data-line-number="1229"></td> <td id="LC1229" class="blob-code blob-code-inner js-file-line"> cur_state <span class="pl-k">=</span> state[i]</td> </tr> <tr> <td id="L1230" class="blob-num js-line-number" data-line-number="1230"></td> <td id="LC1230" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1231" class="blob-num js-line-number" data-line-number="1231"></td> <td id="LC1231" class="blob-code blob-code-inner js-file-line"> cur_state <span class="pl-k">=</span> array_ops.slice(state, [<span class="pl-c1">0</span>, cur_state_pos],</td> </tr> <tr> <td id="L1232" class="blob-num js-line-number" data-line-number="1232"></td> <td id="LC1232" class="blob-code blob-code-inner js-file-line"> [<span class="pl-k">-</span><span class="pl-c1">1</span>, cell.state_size])</td> </tr> <tr> <td id="L1233" class="blob-num js-line-number" data-line-number="1233"></td> <td id="LC1233" class="blob-code blob-code-inner js-file-line"> cur_state_pos <span class="pl-k">+=</span> cell.state_size</td> </tr> <tr> <td id="L1234" class="blob-num js-line-number" data-line-number="1234"></td> <td id="LC1234" class="blob-code blob-code-inner js-file-line"> cur_inp, new_state <span class="pl-k">=</span> cell(cur_inp, cur_state)</td> </tr> <tr> <td id="L1235" class="blob-num js-line-number" data-line-number="1235"></td> <td id="LC1235" class="blob-code blob-code-inner js-file-line"> new_states.append(new_state)</td> </tr> <tr> <td id="L1236" class="blob-num js-line-number" data-line-number="1236"></td> <td id="LC1236" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1237" class="blob-num js-line-number" data-line-number="1237"></td> <td id="LC1237" class="blob-code blob-code-inner js-file-line"> new_states <span class="pl-k">=</span> (<span class="pl-c1">tuple</span>(new_states) <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">else</span></td> </tr> <tr> <td id="L1238" class="blob-num js-line-number" data-line-number="1238"></td> <td id="LC1238" class="blob-code blob-code-inner js-file-line"> array_ops.concat(new_states, <span class="pl-c1">1</span>))</td> </tr> <tr> <td id="L1239" class="blob-num js-line-number" data-line-number="1239"></td> <td id="LC1239" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1240" class="blob-num js-line-number" data-line-number="1240"></td> <td id="LC1240" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> cur_inp, new_states</td> </tr> <tr> <td id="L1241" class="blob-num js-line-number" data-line-number="1241"></td> <td id="LC1241" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1242" class="blob-num js-line-number" data-line-number="1242"></td> <td id="LC1242" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1243" class="blob-num js-line-number" data-line-number="1243"></td> <td id="LC1243" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">_SlimRNNCell</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L1244" class="blob-num js-line-number" data-line-number="1244"></td> <td id="LC1244" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>A simple wrapper for slim.rnn_cells.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1245" class="blob-num js-line-number" data-line-number="1245"></td> <td id="LC1245" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1246" class="blob-num js-line-number" data-line-number="1246"></td> <td id="LC1246" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cell_fn</span>):</td> </tr> <tr> <td id="L1247" class="blob-num js-line-number" data-line-number="1247"></td> <td id="LC1247" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Create a SlimRNNCell from a cell_fn.</span></td> </tr> <tr> <td id="L1248" class="blob-num js-line-number" data-line-number="1248"></td> <td id="LC1248" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1249" class="blob-num js-line-number" data-line-number="1249"></td> <td id="LC1249" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1250" class="blob-num js-line-number" data-line-number="1250"></td> <td id="LC1250" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell_fn: a function which takes (inputs, state, scope) and produces the</span></td> </tr> <tr> <td id="L1251" class="blob-num js-line-number" data-line-number="1251"></td> <td id="LC1251" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> outputs and the new_state. Additionally when called with inputs=None and</span></td> </tr> <tr> <td id="L1252" class="blob-num js-line-number" data-line-number="1252"></td> <td id="LC1252" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state=None it should return (initial_outputs, initial_state).</span></td> </tr> <tr> <td id="L1253" class="blob-num js-line-number" data-line-number="1253"></td> <td id="LC1253" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1254" class="blob-num js-line-number" data-line-number="1254"></td> <td id="LC1254" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L1255" class="blob-num js-line-number" data-line-number="1255"></td> <td id="LC1255" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> TypeError: if cell_fn is not callable</span></td> </tr> <tr> <td id="L1256" class="blob-num js-line-number" data-line-number="1256"></td> <td id="LC1256" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if cell_fn cannot produce a valid initial state.</span></td> </tr> <tr> <td id="L1257" class="blob-num js-line-number" data-line-number="1257"></td> <td id="LC1257" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1258" class="blob-num js-line-number" data-line-number="1258"></td> <td id="LC1258" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">callable</span>(cell_fn):</td> </tr> <tr> <td id="L1259" class="blob-num js-line-number" data-line-number="1259"></td> <td id="LC1259" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>cell_fn <span class="pl-c1">%s</span> needs to be callable<span class="pl-pds">&quot;</span></span>, cell_fn)</td> </tr> <tr> <td id="L1260" class="blob-num js-line-number" data-line-number="1260"></td> <td id="LC1260" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell_fn <span class="pl-k">=</span> cell_fn</td> </tr> <tr> <td id="L1261" class="blob-num js-line-number" data-line-number="1261"></td> <td id="LC1261" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell_name <span class="pl-k">=</span> cell_fn.func.<span class="pl-c1">__name__</span></td> </tr> <tr> <td id="L1262" class="blob-num js-line-number" data-line-number="1262"></td> <td id="LC1262" class="blob-code blob-code-inner js-file-line"> init_output, init_state <span class="pl-k">=</span> <span class="pl-c1">self</span>._cell_fn(<span class="pl-c1">None</span>, <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L1263" class="blob-num js-line-number" data-line-number="1263"></td> <td id="LC1263" class="blob-code blob-code-inner js-file-line"> output_shape <span class="pl-k">=</span> init_output.get_shape()</td> </tr> <tr> <td id="L1264" class="blob-num js-line-number" data-line-number="1264"></td> <td id="LC1264" class="blob-code blob-code-inner js-file-line"> state_shape <span class="pl-k">=</span> init_state.get_shape()</td> </tr> <tr> <td id="L1265" class="blob-num js-line-number" data-line-number="1265"></td> <td id="LC1265" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._output_size <span class="pl-k">=</span> output_shape.with_rank(<span class="pl-c1">2</span>)[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L1266" class="blob-num js-line-number" data-line-number="1266"></td> <td id="LC1266" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_size <span class="pl-k">=</span> state_shape.with_rank(<span class="pl-c1">2</span>)[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L1267" class="blob-num js-line-number" data-line-number="1267"></td> <td id="LC1267" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._output_size <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L1268" class="blob-num js-line-number" data-line-number="1268"></td> <td id="LC1268" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Initial output created by <span class="pl-c1">%s</span> has invalid shape <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span></td> </tr> <tr> <td id="L1269" class="blob-num js-line-number" data-line-number="1269"></td> <td id="LC1269" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">self</span>._cell_name, output_shape))</td> </tr> <tr> <td id="L1270" class="blob-num js-line-number" data-line-number="1270"></td> <td id="LC1270" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_size <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L1271" class="blob-num js-line-number" data-line-number="1271"></td> <td id="LC1271" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Initial state created by <span class="pl-c1">%s</span> has invalid shape <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span></td> </tr> <tr> <td id="L1272" class="blob-num js-line-number" data-line-number="1272"></td> <td id="LC1272" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">self</span>._cell_name, state_shape))</td> </tr> <tr> <td id="L1273" class="blob-num js-line-number" data-line-number="1273"></td> <td id="LC1273" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1274" class="blob-num js-line-number" data-line-number="1274"></td> <td id="LC1274" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1275" class="blob-num js-line-number" data-line-number="1275"></td> <td id="LC1275" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1276" class="blob-num js-line-number" data-line-number="1276"></td> <td id="LC1276" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._state_size</td> </tr> <tr> <td id="L1277" class="blob-num js-line-number" data-line-number="1277"></td> <td id="LC1277" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1278" class="blob-num js-line-number" data-line-number="1278"></td> <td id="LC1278" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1279" class="blob-num js-line-number" data-line-number="1279"></td> <td id="LC1279" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1280" class="blob-num js-line-number" data-line-number="1280"></td> <td id="LC1280" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._output_size</td> </tr> <tr> <td id="L1281" class="blob-num js-line-number" data-line-number="1281"></td> <td id="LC1281" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1282" class="blob-num js-line-number" data-line-number="1282"></td> <td id="LC1282" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1283" class="blob-num js-line-number" data-line-number="1283"></td> <td id="LC1283" class="blob-code blob-code-inner js-file-line"> scope <span class="pl-k">=</span> scope <span class="pl-k">or</span> <span class="pl-c1">self</span>._cell_name</td> </tr> <tr> <td id="L1284" class="blob-num js-line-number" data-line-number="1284"></td> <td id="LC1284" class="blob-code blob-code-inner js-file-line"> output, state <span class="pl-k">=</span> <span class="pl-c1">self</span>._cell_fn(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L1285" class="blob-num js-line-number" data-line-number="1285"></td> <td id="LC1285" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> output, state</td> </tr> </table> <div class="BlobToolbar position-absolute js-file-line-actions dropdown js-menu-container js-select-menu d-none" aria-hidden="true"> <button class="btn-octicon ml-0 px-2 p-0 bg-white border border-gray-dark rounded-1 dropdown-toggle js-menu-target" id="js-file-line-action-button" type="button" aria-expanded="false" aria-haspopup="true" aria-label="Inline file action toolbar" aria-controls="inline-file-actions"> <svg class="octicon octicon-kebab-horizontal" viewBox="0 0 13 16" version="1.1" width="13" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M1.5 9a1.5 1.5 0 1 1 0-3 1.5 1.5 0 0 1 0 3zm5 0a1.5 1.5 0 1 1 0-3 1.5 1.5 0 0 1 0 3zm5 0a1.5 1.5 0 1 1 0-3 1.5 1.5 0 0 1 0 3z"/></svg> </button> <div class="dropdown-menu-content js-menu-content" id="inline-file-actions"> <ul class="BlobToolbar-dropdown dropdown-menu dropdown-menu-se mt-2"> <li><clipboard-copy class="dropdown-item" style="cursor:pointer;" id="js-copy-lines" data-original-text="Copy lines">Copy lines</clipboard-copy></li> <li><clipboard-copy class="dropdown-item" id="js-copy-permalink" style="cursor:pointer;" data-original-text="Copy permalink">Copy permalink</clipboard-copy></li> <li><a class="dropdown-item js-update-url-with-hash" id="js-view-git-blame" href="/tensorflow/tensorflow/blame/cbc658095ae228f2f557af47e4901d552573aa15/tensorflow/python/ops/rnn_cell_impl.py">View git blame</a></li> <li><a class="dropdown-item" id="js-new-issue" href="/tensorflow/tensorflow/issues/new">Open new issue</a></li> </ul> </div> </div> </div> </div> <button type="button" data-facebox="#jump-to-line" data-facebox-class="linejump" data-hotkey="l" class="d-none">Jump to Line</button> <div id="jump-to-line" style="display:none"> <!-- '"` --><!-- </textarea></xmp> --></option></form><form class="js-jump-to-line-form" action="" accept-charset="UTF-8" method="get"><input name="utf8" type="hidden" value="&#x2713;" /> <input class="form-control linejump-input js-jump-to-line-field" type="text" placeholder="Jump to line&hellip;" aria-label="Jump to line" autofocus> <button type="submit" class="btn">Go</button> </form> </div> </div> <div class="modal-backdrop js-touch-events"></div> </div> </div> </div> </div> <div class="footer container-lg px-3" role="contentinfo"> <div class="position-relative d-flex flex-justify-between py-6 mt-6 f6 text-gray border-top border-gray-light "> <ul class="list-style-none d-flex flex-wrap "> <li class="mr-3">&copy; 2018 <span title="0.23060s from unicorn-1000010005-crk39">GitHub</span>, Inc.</li> <li class="mr-3"><a href="https://help.github.com/articles/github-terms-of-service/" data-ga-click="Footer, go to terms, text:terms">Terms</a></li> <li class="mr-3"><a href="https://github.com/site/privacy" data-ga-click="Footer, go to privacy, text:privacy">Privacy</a></li> <li class="mr-3"><a href="https://help.github.com/articles/github-security/" data-ga-click="Footer, go to security, text:security">Security</a></li> <li class="mr-3"><a href="https://status.github.com/" data-ga-click="Footer, go to status, text:status">Status</a></li> <li><a data-ga-click="Footer, go to help, text:help" href="https://help.github.com">Help</a></li> </ul> <a aria-label="Homepage" title="GitHub" class="footer-octicon" href="https://github.com"> <svg height="24" class="octicon octicon-mark-github" viewBox="0 0 16 16" version="1.1" width="24" aria-hidden="true"><path fill-rule="evenodd" d="M8 0C3.58 0 0 3.58 0 8c0 3.54 2.29 6.53 5.47 7.59.4.07.55-.17.55-.38 0-.19-.01-.82-.01-1.49-2.01.37-2.53-.49-2.69-.94-.09-.23-.48-.94-.82-1.13-.28-.15-.68-.52-.01-.53.63-.01 1.08.58 1.23.82.72 1.21 1.87.87 2.33.66.07-.52.28-.87.51-1.07-1.78-.2-3.64-.89-3.64-3.95 0-.87.31-1.59.82-2.15-.08-.2-.36-1.02.08-2.12 0 0 .67-.21 2.2.82.64-.18 1.32-.27 2-.27.68 0 1.36.09 2 .27 1.53-1.04 2.2-.82 2.2-.82.44 1.1.16 1.92.08 2.12.51.56.82 1.27.82 2.15 0 3.07-1.87 3.75-3.65 3.95.29.25.54.73.54 1.48 0 1.07-.01 1.93-.01 2.2 0 .21.15.46.55.38A8.013 8.013 0 0 0 16 8c0-4.42-3.58-8-8-8z"/></svg> </a> <ul class="list-style-none d-flex flex-wrap "> <li class="mr-3"><a data-ga-click="Footer, go to contact, text:contact" href="https://github.com/contact">Contact GitHub</a></li> <li class="mr-3"><a href="https://developer.github.com" data-ga-click="Footer, go to api, text:api">API</a></li> <li class="mr-3"><a href="https://training.github.com" data-ga-click="Footer, go to training, text:training">Training</a></li> <li class="mr-3"><a href="https://shop.github.com" data-ga-click="Footer, go to shop, text:shop">Shop</a></li> <li class="mr-3"><a data-ga-click="Footer, go to blog, text:blog" href="https://github.com/blog">Blog</a></li> <li><a data-ga-click="Footer, go to about, text:about" href="https://github.com/about">About</a></li> </ul> </div> </div> <div id="ajax-error-message" class="ajax-error-message flash flash-error"> <svg class="octicon octicon-alert" viewBox="0 0 16 16" version="1.1" width="16" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M8.865 1.52c-.18-.31-.51-.5-.87-.5s-.69.19-.87.5L.275 13.5c-.18.31-.18.69 0 1 .19.31.52.5.87.5h13.7c.36 0 .69-.19.86-.5.17-.31.18-.69.01-1L8.865 1.52zM8.995 13h-2v-2h2v2zm0-3h-2V6h2v4z"/></svg> <button type="button" class="flash-close js-ajax-error-dismiss" aria-label="Dismiss error"> <svg class="octicon octicon-x" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M7.48 8l3.75 3.75-1.48 1.48L6 9.48l-3.75 3.75-1.48-1.48L4.52 8 .77 4.25l1.48-1.48L6 6.52l3.75-3.75 1.48 1.48z"/></svg> </button> You can't perform that action at this time. </div> <script crossorigin="anonymous" type="application/javascript" src="https://assets-cdn.github.com/assets/compat-af4f19572e1d43f69746f70aa5f01dc9.js"></script> <script crossorigin="anonymous" type="application/javascript" src="https://assets-cdn.github.com/assets/frameworks-dd22b45b9495a43602787a69962f696c.js"></script> <script crossorigin="anonymous" async="async" type="application/javascript" src="https://assets-cdn.github.com/assets/github-48130667aaece5c4ce8ee28ecdb29539.js"></script> <div class="js-stale-session-flash stale-session-flash flash flash-warn flash-banner d-none"> <svg class="octicon octicon-alert" viewBox="0 0 16 16" version="1.1" width="16" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M8.865 1.52c-.18-.31-.51-.5-.87-.5s-.69.19-.87.5L.275 13.5c-.18.31-.18.69 0 1 .19.31.52.5.87.5h13.7c.36 0 .69-.19.86-.5.17-.31.18-.69.01-1L8.865 1.52zM8.995 13h-2v-2h2v2zm0-3h-2V6h2v4z"/></svg> <span class="signed-in-tab-flash">You signed in with another tab or window. <a href="">Reload</a> to refresh your session.</span> <span class="signed-out-tab-flash">You signed out in another tab or window. <a href="">Reload</a> to refresh your session.</span> </div> <div class="facebox" id="facebox" style="display:none;"> <div class="facebox-popup"> <div class="facebox-content" role="dialog" aria-labelledby="facebox-header" aria-describedby="facebox-description"> </div> <button type="button" class="facebox-close js-facebox-close" aria-label="Close modal"> <svg class="octicon octicon-x" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M7.48 8l3.75 3.75-1.48 1.48L6 9.48l-3.75 3.75-1.48-1.48L4.52 8 .77 4.25l1.48-1.48L6 6.52l3.75-3.75 1.48 1.48z"/></svg> </button> </div> </div> </body> </html>
67.591094
766
0.599302
<!DOCTYPE html> <html lang="en"> <head> <meta charset="utf-8"> <link rel="dns-prefetch" href="https://assets-cdn.github.com"> <link rel="dns-prefetch" href="https://avatars0.githubusercontent.com"> <link rel="dns-prefetch" href="https://avatars1.githubusercontent.com"> <link rel="dns-prefetch" href="https://avatars2.githubusercontent.com"> <link rel="dns-prefetch" href="https://avatars3.githubusercontent.com"> <link rel="dns-prefetch" href="https://github-cloud.s3.amazonaws.com"> <link rel="dns-prefetch" href="https://user-images.githubusercontent.com/"> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://assets-cdn.github.com/assets/frameworks-0b6eee1f89d4460d83bdbee0a4cb0020.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://assets-cdn.github.com/assets/github-c4d5af25254cfc8bb30ae5cb6f074c97.css" /> <link crossorigin="anonymous" media="all" rel="stylesheet" href="https://assets-cdn.github.com/assets/site-83dc1f7ebc9c7461fe1eab799b56c4c4.css" /> <meta name="viewport" content="width=device-width"> <title>tensorflow/rnn_cell_impl.py at r1.6 · tensorflow/tensorflow · GitHub</title> <meta name="description" content="GitHub is where people build software. More than 27 million people use GitHub to discover, fork, and contribute to over 80 million projects."> <link rel="search" type="application/opensearchdescription+xml" href="/opensearch.xml" title="GitHub"> <link rel="fluid-icon" href="https://github.com/fluidicon.png" title="GitHub"> <meta property="fb:app_id" content="1401488693436528"> <meta property="og:image" content="https://avatars2.githubusercontent.com/u/15658638?s=400&amp;v=4" /><meta property="og:site_name" content="GitHub" /><meta property="og:type" content="object" /><meta property="og:title" content="tensorflow/tensorflow" /><meta property="og:url" content="https://github.com/tensorflow/tensorflow" /><meta property="og:description" content="tensorflow - Computation using data flow graphs for scalable machine learning" /> <link rel="assets" href="https://assets-cdn.github.com/"> <meta name="pjax-timeout" content="1000"> <meta name="request-id" content="D9CE:22519:FDC93F:17A0312:5AB1D278" data-pjax-transient> <meta name="selected-link" value="repo_source" data-pjax-transient> <meta name="google-site-verification" content="KT5gs8h0wvaagLKAVWq8bbeNwnZZK1r1XQysX3xurLU"> <meta name="google-site-verification" content="ZzhVyEFwb7w3e0-uOTltm8Jsck2F5StVihD0exw2fsA"> <meta name="google-site-verification" content="GXs5KoUUkNCoaAZn7wPN-t01Pywp9M3sEjnt_3_ZWPc"> <meta name="google-analytics" content="UA-3769691-2"> <meta name="octolytics-host" content="collector.githubapp.com" /><meta name="octolytics-app-id" content="github" /><meta name="octolytics-event-url" content="https://collector.githubapp.com/github-external/browser_event" /><meta name="octolytics-dimension-request_id" content="D9CE:22519:FDC93F:17A0312:5AB1D278" /><meta name="octolytics-dimension-region_edge" content="sea" /><meta name="octolytics-dimension-region_render" content="iad" /> <meta name="hydro-events-url" content="https://github.com/hydro_browser_events" /> <meta name="analytics-location" content="/&lt;user-name&gt;/&lt;repo-name&gt;/blob/show" data-pjax-transient="true" /> <meta class="js-ga-set" name="dimension1" content="Logged Out"> <meta name="hostname" content="github.com"> <meta name="user-login" content=""> <meta name="expected-hostname" content="github.com"> <meta name="js-proxy-site-detection-payload" content="YTI3MjdkNDE4ODBjNzIwYjgyNTVlMzFjZTEwODkzZDRmYTJkNTZhMTdhZGZmNGM0ODc3MTZmNTg3MzBjNmM2Nnx7InJlbW90ZV9hZGRyZXNzIjoiMTY3LjIyMC4yMzIuMTI0IiwicmVxdWVzdF9pZCI6IkQ5Q0U6MjI1MTk6RkRDOTNGOjE3QTAzMTI6NUFCMUQyNzgiLCJ0aW1lc3RhbXAiOjE1MjE2MDMxOTMsImhvc3QiOiJnaXRodWIuY29tIn0="> <meta name="enabled-features" content="UNIVERSE_BANNER,FREE_TRIALS,MARKETPLACE_INSIGHTS,MARKETPLACE_INSIGHTS_CONVERSION_PERCENTAGES"> <meta name="html-safe-nonce" content="54cbe43f921a89e1942c69748da62cb0fab66263"> <meta http-equiv="x-pjax-version" content="f5d037ad7a1b600b21fb2f3474bd0c48"> <link href="https://github.com/tensorflow/tensorflow/commits/r1.6.atom" rel="alternate" title="Recent Commits to tensorflow:r1.6" type="application/atom+xml"> <meta name="description" content="tensorflow - Computation using data flow graphs for scalable machine learning"> <meta name="go-import" content="github.com/tensorflow/tensorflow git https://github.com/tensorflow/tensorflow.git"> <meta name="octolytics-dimension-user_id" content="15658638" /><meta name="octolytics-dimension-user_login" content="tensorflow" /><meta name="octolytics-dimension-repository_id" content="45717250" /><meta name="octolytics-dimension-repository_nwo" content="tensorflow/tensorflow" /><meta name="octolytics-dimension-repository_public" content="true" /><meta name="octolytics-dimension-repository_is_fork" content="false" /><meta name="octolytics-dimension-repository_network_root_id" content="45717250" /><meta name="octolytics-dimension-repository_network_root_nwo" content="tensorflow/tensorflow" /><meta name="octolytics-dimension-repository_explore_github_marketplace_ci_cta_shown" content="false" /> <link rel="canonical" href="https://github.com/tensorflow/tensorflow/blob/r1.6/tensorflow/python/ops/rnn_cell_impl.py" data-pjax-transient> <meta name="browser-stats-url" content="https://api.github.com/_private/browser/stats"> <meta name="browser-errors-url" content="https://api.github.com/_private/browser/errors"> <link rel="mask-icon" href="https://assets-cdn.github.com/pinned-octocat.svg" color="#000000"> <link rel="icon" type="image/x-icon" class="js-site-favicon" href="https://assets-cdn.github.com/favicon.ico"> <meta name="theme-color" content="#1e2327"> <link rel="manifest" href="/manifest.json" crossOrigin="use-credentials"> </head> <body class="logged-out env-production page-blob"> <div class="position-relative js-header-wrapper "> <a href="#start-of-content" tabindex="1" class="px-2 py-4 show-on-focus js-skip-to-content">Skip to content</a> <div id="js-pjax-loader-bar" class="pjax-loader-bar"><div class="progress"></div></div> <header class="Header header-logged-out position-relative f4 py-3" role="banner"> <div class="container-lg d-flex px-3"> <div class="d-flex flex-justify-between flex-items-center"> <a class="header-logo-invertocat my-0" href="https://github.com/" aria-label="Homepage" data-ga-click="(Logged out) Header, go to homepage, icon:logo-wordmark"> <svg height="32" class="octicon octicon-mark-github" viewBox="0 0 16 16" version="1.1" width="32" aria-hidden="true"><path fill-rule="evenodd" d="M8 0C3.58 0 0 3.58 0 8c0 3.54 2.29 6.53 5.47 7.59.4.07.55-.17.55-.38 0-.19-.01-.82-.01-1.49-2.01.37-2.53-.49-2.69-.94-.09-.23-.48-.94-.82-1.13-.28-.15-.68-.52-.01-.53.63-.01 1.08.58 1.23.82.72 1.21 1.87.87 2.33.66.07-.52.28-.87.51-1.07-1.78-.2-3.64-.89-3.64-3.95 0-.87.31-1.59.82-2.15-.08-.2-.36-1.02.08-2.12 0 0 .67-.21 2.2.82.64-.18 1.32-.27 2-.27.68 0 1.36.09 2 .27 1.53-1.04 2.2-.82 2.2-.82.44 1.1.16 1.92.08 2.12.51.56.82 1.27.82 2.15 0 3.07-1.87 3.75-3.65 3.95.29.25.54.73.54 1.48 0 1.07-.01 1.93-.01 2.2 0 .21.15.46.55.38A8.013 8.013 0 0 0 16 8c0-4.42-3.58-8-8-8z"/></svg> </a> </div> <div class="HeaderMenu HeaderMenu--bright d-flex flex-justify-between flex-auto"> <nav class="mt-0"> <ul class="d-flex list-style-none"> <li class="ml-2"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:features" data-selected-links="/features /features/project-management /features/code-review /features/project-management /features/integrations /features" href="/features"> Features </a> </li> <li class="ml-4"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:business" data-selected-links="/business /business/security /business/customers /business" href="/business"> Business </a> </li> <li class="ml-4"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:explore" data-selected-links="/explore /trending /trending/developers /integrations /integrations/feature/code /integrations/feature/collaborate /integrations/feature/ship showcases showcases_search showcases_landing /explore" href="/explore"> Explore </a> </li> <li class="ml-4"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:marketplace" data-selected-links=" /marketplace" href="/marketplace"> Marketplace </a> </li> <li class="ml-4"> <a class="js-selected-navigation-item HeaderNavlink px-0 py-2 m-0" data-ga-click="Header, click, Nav menu - item:pricing" data-selected-links="/pricing /pricing/developer /pricing/team /pricing/business-hosted /pricing/business-enterprise /pricing" href="/pricing"> Pricing </a> </li> </ul> </nav> <div class="d-flex"> <div class="d-lg-flex flex-items-center mr-3"> <div class="header-search scoped-search site-scoped-search js-site-search" role="search"> <!-- '"` --><!-- </textarea></xmp> --></option></form><form class="js-site-search-form" data-scoped-search-url="/tensorflow/tensorflow/search" data-unscoped-search-url="/search" action="/tensorflow/tensorflow/search" accept-charset="UTF-8" method="get"><input name="utf8" type="hidden" value="&#x2713;" /> <label class="form-control header-search-wrapper js-chromeless-input-container"> <a class="header-search-scope no-underline" href="/tensorflow/tensorflow/blob/r1.6/tensorflow/python/ops/rnn_cell_impl.py">This repository</a> <input type="text" class="form-control header-search-input js-site-search-focus js-site-search-field is-clearable" data-hotkey="s,/" name="q" value="" placeholder="Search" aria-label="Search this repository" data-unscoped-placeholder="Search GitHub" data-scoped-placeholder="Search" autocapitalize="off"> <input type="hidden" class="js-site-search-type-field" name="type" > </label> </form></div> </div> <span class="d-inline-block"> <div class="HeaderNavlink px-0 py-2 m-0"> <a class="text-bold text-white no-underline" href="/login?return_to=%2Ftensorflow%2Ftensorflow%2Fblob%2Fr1.6%2Ftensorflow%2Fpython%2Fops%2Frnn_cell_impl.py" data-ga-click="(Logged out) Header, clicked Sign in, text:sign-in">Sign in</a> <span class="text-gray">or</span> <a class="text-bold text-white no-underline" href="/join?source=header-repo" data-ga-click="(Logged out) Header, clicked Sign up, text:sign-up">Sign up</a> </div> </span> </div> </div> </div> </header> </div> <div id="start-of-content" class="show-on-focus"></div> <div id="js-flash-container"> </div> <div role="main" class="application-main " > <div itemscope itemtype="http://schema.org/SoftwareSourceCode" class=""> <div id="js-repo-pjax-container" data-pjax-container > <div class="pagehead repohead instapaper_ignore readability-menu experiment-repo-nav "> <div class="repohead-details-container clearfix container"> <ul class="pagehead-actions"> <li> <a href="/login?return_to=%2Ftensorflow%2Ftensorflow" class="btn btn-sm btn-with-count tooltipped tooltipped-n" aria-label="You must be signed in to watch a repository" rel="nofollow"> <svg class="octicon octicon-eye" viewBox="0 0 16 16" version="1.1" width="16" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M8.06 2C3 2 0 8 0 8s3 6 8.06 6C13 14 16 8 16 8s-3-6-7.94-6zM8 12c-2.2 0-4-1.78-4-4 0-2.2 1.8-4 4-4 2.22 0 4 1.8 4 4 0 2.22-1.78 4-4 4zm2-4c0 1.11-.89 2-2 2-1.11 0-2-.89-2-2 0-1.11.89-2 2-2 1.11 0 2 .89 2 2z"/></svg> Watch </a> <a class="social-count" href="/tensorflow/tensorflow/watchers" aria-label="7632 users are watching this repository"> 7,632 </a> </li> <li> <a href="/login?return_to=%2Ftensorflow%2Ftensorflow" class="btn btn-sm btn-with-count tooltipped tooltipped-n" aria-label="You must be signed in to star a repository" rel="nofollow"> <svg class="octicon octicon-star" viewBox="0 0 14 16" version="1.1" width="14" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M14 6l-4.9-.64L7 1 4.9 5.36 0 6l3.6 3.26L2.67 14 7 11.67 11.33 14l-.93-4.74z"/></svg> Star </a> <a class="social-count js-social-count" href="/tensorflow/tensorflow/stargazers" aria-label="93301 users starred this repository"> 93,301 </a> </li> <li> <a href="/login?return_to=%2Ftensorflow%2Ftensorflow" class="btn btn-sm btn-with-count tooltipped tooltipped-n" aria-label="You must be signed in to fork a repository" rel="nofollow"> <svg class="octicon octicon-repo-forked" viewBox="0 0 10 16" version="1.1" width="10" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M8 1a1.993 1.993 0 0 0-1 3.72V6L5 8 3 6V4.72A1.993 1.993 0 0 0 2 1a1.993 1.993 0 0 0-1 3.72V6.5l3 3v1.78A1.993 1.993 0 0 0 5 15a1.993 1.993 0 0 0 1-3.72V9.5l3-3V4.72A1.993 1.993 0 0 0 8 1zM2 4.2C1.34 4.2.8 3.65.8 3c0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2zm3 10c-.66 0-1.2-.55-1.2-1.2 0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2zm3-10c-.66 0-1.2-.55-1.2-1.2 0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2z"/></svg> Fork </a> <a href="/tensorflow/tensorflow/network" class="social-count" aria-label="59890 users forked this repository"> 59,890 </a> </li> </ul> <h1 class="public "> <svg class="octicon octicon-repo" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M4 9H3V8h1v1zm0-3H3v1h1V6zm0-2H3v1h1V4zm0-2H3v1h1V2zm8-1v12c0 .55-.45 1-1 1H6v2l-1.5-1.5L3 16v-2H1c-.55 0-1-.45-1-1V1c0-.55.45-1 1-1h10c.55 0 1 .45 1 1zm-1 10H1v2h2v-1h3v1h5v-2zm0-10H2v9h9V1z"/></svg> <span class="author" itemprop="author"><a class="url fn" rel="author" href="/tensorflow">tensorflow</a></span><!-- --><span class="path-divider">/</span><!-- --><strong itemprop="name"><a data-pjax="#js-repo-pjax-container" href="/tensorflow/tensorflow">tensorflow</a></strong> </h1> </div> <nav class="reponav js-repo-nav js-sidenav-container-pjax container" itemscope itemtype="http://schema.org/BreadcrumbList" role="navigation" data-pjax="#js-repo-pjax-container"> <span itemscope itemtype="http://schema.org/ListItem" itemprop="itemListElement"> <a class="js-selected-navigation-item selected reponav-item" itemprop="url" data-hotkey="g c" data-selected-links="repo_source repo_downloads repo_commits repo_releases repo_tags repo_branches repo_packages /tensorflow/tensorflow/tree/r1.6" href="/tensorflow/tensorflow/tree/r1.6"> <svg class="octicon octicon-code" viewBox="0 0 14 16" version="1.1" width="14" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M9.5 3L8 4.5 11.5 8 8 11.5 9.5 13 14 8 9.5 3zm-5 0L0 8l4.5 5L6 11.5 2.5 8 6 4.5 4.5 3z"/></svg> <span itemprop="name">Code</span> <meta itemprop="position" content="1"> </a> </span> <span itemscope itemtype="http://schema.org/ListItem" itemprop="itemListElement"> <a itemprop="url" data-hotkey="g i" class="js-selected-navigation-item reponav-item" data-selected-links="repo_issues repo_labels repo_milestones /tensorflow/tensorflow/issues" href="/tensorflow/tensorflow/issues"> <svg class="octicon octicon-issue-opened" viewBox="0 0 14 16" version="1.1" width="14" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M7 2.3c3.14 0 5.7 2.56 5.7 5.7s-2.56 5.7-5.7 5.7A5.71 5.71 0 0 1 1.3 8c0-3.14 2.56-5.7 5.7-5.7zM7 1C3.14 1 0 4.14 0 8s3.14 7 7 7 7-3.14 7-7-3.14-7-7-7zm1 3H6v5h2V4zm0 6H6v2h2v-2z"/></svg> <span itemprop="name">Issues</span> <span class="Counter">1,287</span> <meta itemprop="position" content="2"> </a> </span> <span itemscope itemtype="http://schema.org/ListItem" itemprop="itemListElement"> <a data-hotkey="g p" itemprop="url" class="js-selected-navigation-item reponav-item" data-selected-links="repo_pulls checks /tensorflow/tensorflow/pulls" href="/tensorflow/tensorflow/pulls"> <svg class="octicon octicon-git-pull-request" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M11 11.28V5c-.03-.78-.34-1.47-.94-2.06C9.46 2.35 8.78 2.03 8 2H7V0L4 3l3 3V4h1c.27.02.48.11.69.31.21.2.3.42.31.69v6.28A1.993 1.993 0 0 0 10 15a1.993 1.993 0 0 0 1-3.72zm-1 2.92c-.66 0-1.2-.55-1.2-1.2 0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2zM4 3c0-1.11-.89-2-2-2a1.993 1.993 0 0 0-1 3.72v6.56A1.993 1.993 0 0 0 2 15a1.993 1.993 0 0 0 1-3.72V4.72c.59-.34 1-.98 1-1.72zm-.8 10c0 .66-.55 1.2-1.2 1.2-.65 0-1.2-.55-1.2-1.2 0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2zM2 4.2C1.34 4.2.8 3.65.8 3c0-.65.55-1.2 1.2-1.2.65 0 1.2.55 1.2 1.2 0 .65-.55 1.2-1.2 1.2z"/></svg> <span itemprop="name">Pull requests</span> <span class="Counter">197</span> <meta itemprop="position" content="3"> </a> </span> <a data-hotkey="g b" class="js-selected-navigation-item reponav-item" data-selected-links="repo_projects new_repo_project repo_project /tensorflow/tensorflow/projects" href="/tensorflow/tensorflow/projects"> <svg class="octicon octicon-project" viewBox="0 0 15 16" version="1.1" width="15" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M10 12h3V2h-3v10zm-4-2h3V2H6v8zm-4 4h3V2H2v12zm-1 1h13V1H1v14zM14 0H1a1 1 0 0 0-1 1v14a1 1 0 0 0 1 1h13a1 1 0 0 0 1-1V1a1 1 0 0 0-1-1z"/></svg> Projects <span class="Counter" >0</span> </a> <a class="js-selected-navigation-item reponav-item" data-selected-links="repo_graphs repo_contributors dependency_graph pulse /tensorflow/tensorflow/pulse" href="/tensorflow/tensorflow/pulse"> <svg class="octicon octicon-graph" viewBox="0 0 16 16" version="1.1" width="16" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M16 14v1H0V0h1v14h15zM5 13H3V8h2v5zm4 0H7V3h2v10zm4 0h-2V6h2v7z"/></svg> Insights </a> </nav> </div> <div class="container new-discussion-timeline experiment-repo-nav "> <div class="repository-content "> <a class="d-none js-permalink-shortcut" data-hotkey="y" href="/tensorflow/tensorflow/blob/cbc658095ae228f2f557af47e4901d552573aa15/tensorflow/python/ops/rnn_cell_impl.py">Permalink</a> <!-- blob contrib key: blob_contributors:v21:5895d9f5b73fd7bcfb782c5d3257ea21 --> <div class="file-navigation"> <div class="select-menu branch-select-menu js-menu-container js-select-menu float-left"> <button class=" btn btn-sm select-menu-button js-menu-target css-truncate" data-hotkey="w" type="button" aria-label="Switch branches or tags" aria-expanded="false" aria-haspopup="true"> <i>Branch:</i> <span class="js-select-button css-truncate-target">r1.6</span> </button> <div class="select-menu-modal-holder js-menu-content js-navigation-container" data-pjax> <div class="select-menu-modal"> <div class="select-menu-header"> <svg class="octicon octicon-x js-menu-close" role="img" aria-label="Close" viewBox="0 0 12 16" version="1.1" width="12" height="16"><path fill-rule="evenodd" d="M7.48 8l3.75 3.75-1.48 1.48L6 9.48l-3.75 3.75-1.48-1.48L4.52 8 .77 4.25l1.48-1.48L6 6.52l3.75-3.75 1.48 1.48z"/></svg> <span class="select-menu-title">Switch branches/tags</span> </div> <div class="select-menu-filters"> <div class="select-menu-text-filter"> <input type="text" aria-label="Filter branches/tags" id="context-commitish-filter-field" class="form-control js-filterable-field js-navigation-enable" placeholder="Filter branches/tags"> </div> <div class="select-menu-tabs"> <ul> <li class="select-menu-tab"> <a href="#" data-tab-filter="branches" data-filter-placeholder="Filter branches/tags" class="js-select-menu-tab" role="tab">Branches</a> </li> <li class="select-menu-tab"> <a href="#" data-tab-filter="tags" data-filter-placeholder="Find a tag…" class="js-select-menu-tab" role="tab">Tags</a> </li> </ul> </div> </div> <div class="select-menu-list select-menu-tab-bucket js-select-menu-tab-bucket" data-tab-filter="branches" role="menu"> <div data-filterable-for="context-commitish-filter-field" data-filterable-type="substring"> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/0.6.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.6.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> 0.6.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/branch_189041421/tensorflow/python/ops/rnn_cell_impl.py" data-name="branch_189041421" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> branch_189041421 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/martinwicke-patch-1/tensorflow/python/ops/rnn_cell_impl.py" data-name="martinwicke-patch-1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> martinwicke-patch-1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/martinwicke-patch-2/tensorflow/python/ops/rnn_cell_impl.py" data-name="martinwicke-patch-2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> martinwicke-patch-2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/master/tensorflow/python/ops/rnn_cell_impl.py" data-name="master" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> master </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/mrry-patch-1/tensorflow/python/ops/rnn_cell_impl.py" data-name="mrry-patch-1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> mrry-patch-1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.7/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.7" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.7 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.8/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.8" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.8 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.9/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.9" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.9 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.10/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.10" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.10 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.11/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.11" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.11 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r0.12/tensorflow/python/ops/rnn_cell_impl.py" data-name="r0.12" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r0.12 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.2/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.3/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.3" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.3 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.4/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.4" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.4 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.5/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.5" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.5 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open selected" href="/tensorflow/tensorflow/blob/r1.6/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.6" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.6 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/r1.7/tensorflow/python/ops/rnn_cell_impl.py" data-name="r1.7" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> r1.7 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/blob/yifeif-patch-1/tensorflow/python/ops/rnn_cell_impl.py" data-name="yifeif-patch-1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target js-select-menu-filter-text"> yifeif-patch-1 </span> </a> </div> <div class="select-menu-no-results">Nothing to show</div> </div> <div class="select-menu-list select-menu-tab-bucket js-select-menu-tab-bucket" data-tab-filter="tags"> <div data-filterable-for="context-commitish-filter-field" data-filterable-type="substring"> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.7.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.7.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.7.0-rc0"> v1.7.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.6.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.6.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.6.0"> v1.6.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.6.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.6.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.6.0-rc1"> v1.6.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.6.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.6.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.6.0-rc0"> v1.6.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.5.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.5.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.5.1"> v1.5.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.5.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.5.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.5.0"> v1.5.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.5.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.5.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.5.0-rc1"> v1.5.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.5.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.5.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.5.0-rc0"> v1.5.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.4.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.4.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.4.1"> v1.4.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.4.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.4.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.4.0"> v1.4.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.4.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.4.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.4.0-rc1"> v1.4.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.4.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.4.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.4.0-rc0"> v1.4.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.1"> v1.3.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.0"> v1.3.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.0-rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.0-rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.0-rc2"> v1.3.0-rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.0-rc1"> v1.3.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.3.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.3.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.3.0-rc0"> v1.3.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.1"> v1.2.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.0"> v1.2.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.0-rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.0-rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.0-rc2"> v1.2.0-rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.0-rc1"> v1.2.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.2.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.2.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.2.0-rc0"> v1.2.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.1.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.1.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.1.0"> v1.1.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.1.0-rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.1.0-rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.1.0-rc2"> v1.1.0-rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.1.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.1.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.1.0-rc1"> v1.1.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.1.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.1.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.1.0-rc0"> v1.1.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.1"> v1.0.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0"> v1.0.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0-rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0-rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0-rc2"> v1.0.0-rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0-rc1"> v1.0.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0-rc0"> v1.0.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v1.0.0-alpha/tensorflow/python/ops/rnn_cell_impl.py" data-name="v1.0.0-alpha" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v1.0.0-alpha"> v1.0.0-alpha </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.12.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.12.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.12.0"> v0.12.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.11.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.11.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.11.0"> v0.11.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.11.0rc2/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.11.0rc2" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.11.0rc2"> v0.11.0rc2 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.11.0rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.11.0rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.11.0rc1"> v0.11.0rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.11.0rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.11.0rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.11.0rc0"> v0.11.0rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.10.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.10.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.10.0"> v0.10.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.10.0rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.10.0rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.10.0rc0"> v0.10.0rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.9.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.9.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.9.0"> v0.9.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.9.0rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.9.0rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.9.0rc0"> v0.9.0rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.8.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.8.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.8.0"> v0.8.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.8.0rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.8.0rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.8.0rc0"> v0.8.0rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.7.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.7.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.7.1"> v0.7.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.7.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.7.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.7.0"> v0.7.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/v0.6.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="v0.6.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="v0.6.0"> v0.6.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.12.1/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.12.1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.12.1"> 0.12.1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.12.0-rc1/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.12.0-rc1" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.12.0-rc1"> 0.12.0-rc1 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.12.0-rc0/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.12.0-rc0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.12.0-rc0"> 0.12.0-rc0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.6.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.6.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.6.0"> 0.6.0 </span> </a> <a class="select-menu-item js-navigation-item js-navigation-open " href="/tensorflow/tensorflow/tree/0.5.0/tensorflow/python/ops/rnn_cell_impl.py" data-name="0.5.0" data-skip-pjax="true" rel="nofollow"> <svg class="octicon octicon-check select-menu-item-icon" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M12 5l-8 8-4-4 1.5-1.5L4 10l6.5-6.5z"/></svg> <span class="select-menu-item-text css-truncate-target" title="0.5.0"> 0.5.0 </span> </a> </div> <div class="select-menu-no-results">Nothing to show</div> </div> </div> </div> </div> <div class="BtnGroup float-right"> <a href="/tensorflow/tensorflow/find/r1.6" class="js-pjax-capture-input btn btn-sm BtnGroup-item" data-pjax data-hotkey="t"> Find file </a> <clipboard-copy for="blob-path" aria-label="Copy file path to clipboard" class="btn btn-sm BtnGroup-item tooltipped tooltipped-s" data-copied-hint="Copied!"> Copy path </clipboard-copy> </div> <div id="blob-path" class="breadcrumb"> <span class="repo-root js-repo-root"><span class="js-path-segment"><a data-pjax="true" href="/tensorflow/tensorflow/tree/r1.6"><span>tensorflow</span></a></span></span><span class="separator">/</span><span class="js-path-segment"><a data-pjax="true" href="/tensorflow/tensorflow/tree/r1.6/tensorflow"><span>tensorflow</span></a></span><span class="separator">/</span><span class="js-path-segment"><a data-pjax="true" href="/tensorflow/tensorflow/tree/r1.6/tensorflow/python"><span>python</span></a></span><span class="separator">/</span><span class="js-path-segment"><a data-pjax="true" href="/tensorflow/tensorflow/tree/r1.6/tensorflow/python/ops"><span>ops</span></a></span><span class="separator">/</span><strong class="final-path">rnn_cell_impl.py</strong> </div> </div> <div class="commit-tease"> <span class="float-right"> <a class="commit-tease-sha" href="/tensorflow/tensorflow/commit/dff64beac8570d910f83774087642bb6a3fda96a" data-pjax> dff64be </a> <relative-time datetime="2018-01-23T21:46:25Z">Jan 23, 2018</relative-time> </span> <div> <img alt="@michaelkhan3" class="avatar" height="20" src="https://avatars3.githubusercontent.com/u/4420471?s=40&amp;v=4" width="20" /> <a href="/michaelkhan3" class="user-mention" rel="contributor">michaelkhan3</a> <a href="/tensorflow/tensorflow/commit/dff64beac8570d910f83774087642bb6a3fda96a" class="message" data-pjax="true" title="Add property to get cell wrapped by DropoutWrapper (#16006) * add property to get cell wrapped by dropout * updating golden files">Add property to get cell wrapped by DropoutWrapper (</a><a href="https://github.com/tensorflow/tensorflow/pull/16006" class="issue-link js-issue-link" data-error-text="Failed to load issue title" data-id="287414630" data-permission-text="Issue title is private" data-url="https://github.com/tensorflow/tensorflow/issues/16006">#16006</a><a href="/tensorflow/tensorflow/commit/dff64beac8570d910f83774087642bb6a3fda96a" class="message" data-pjax="true" title="Add property to get cell wrapped by DropoutWrapper (#16006) * add property to get cell wrapped by dropout * updating golden files">)</a> </div> <div class="commit-tease-contributors"> <button type="button" class="btn-link muted-link contributors-toggle" data-facebox="#blob_contributors_box"> <strong>13</strong> contributors </button> <a class="avatar-link tooltipped tooltipped-s" aria-label="tensorflower-gardener" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=tensorflower-gardener"><img alt="@tensorflower-gardener" class="avatar" height="20" src="https://avatars1.githubusercontent.com/u/17151892?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="ebrevdo" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=ebrevdo"><img alt="@ebrevdo" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/1794715?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="jhseu" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=jhseu"><img alt="@jhseu" class="avatar" height="20" src="https://avatars1.githubusercontent.com/u/170179?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="protoget" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=protoget"><img alt="@protoget" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/5117188?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="xiejw" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=xiejw"><img alt="@xiejw" class="avatar" height="20" src="https://avatars0.githubusercontent.com/u/1184671?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="martinwicke" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=martinwicke"><img alt="@martinwicke" class="avatar" height="20" src="https://avatars1.githubusercontent.com/u/577277?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="michaelkhan3" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=michaelkhan3"><img alt="@michaelkhan3" class="avatar" height="20" src="https://avatars3.githubusercontent.com/u/4420471?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="fchollet" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=fchollet"><img alt="@fchollet" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/710255?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="drpngx" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=drpngx"><img alt="@drpngx" class="avatar" height="20" src="https://avatars3.githubusercontent.com/u/20959853?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="caisq" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=caisq"><img alt="@caisq" class="avatar" height="20" src="https://avatars0.githubusercontent.com/u/16824702?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="alextp" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=alextp"><img alt="@alextp" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/5061?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="annarev" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=annarev"><img alt="@annarev" class="avatar" height="20" src="https://avatars0.githubusercontent.com/u/22060313?s=40&amp;v=4" width="20" /> </a> <a class="avatar-link tooltipped tooltipped-s" aria-label="kosklain" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py?author=kosklain"><img alt="@kosklain" class="avatar" height="20" src="https://avatars2.githubusercontent.com/u/1104579?s=40&amp;v=4" width="20" /> </a> </div> <div id="blob_contributors_box" style="display:none"> <h2 class="facebox-header" data-facebox-id="facebox-header">Users who have contributed to this file</h2> <ul class="facebox-user-list" data-facebox-id="facebox-description"> <li class="facebox-user-list-item"> <img alt="@tensorflower-gardener" height="24" src="https://avatars0.githubusercontent.com/u/17151892?s=48&amp;v=4" width="24" /> <a href="/tensorflower-gardener">tensorflower-gardener</a> </li> <li class="facebox-user-list-item"> <img alt="@ebrevdo" height="24" src="https://avatars3.githubusercontent.com/u/1794715?s=48&amp;v=4" width="24" /> <a href="/ebrevdo">ebrevdo</a> </li> <li class="facebox-user-list-item"> <img alt="@jhseu" height="24" src="https://avatars0.githubusercontent.com/u/170179?s=48&amp;v=4" width="24" /> <a href="/jhseu">jhseu</a> </li> <li class="facebox-user-list-item"> <img alt="@protoget" height="24" src="https://avatars3.githubusercontent.com/u/5117188?s=48&amp;v=4" width="24" /> <a href="/protoget">protoget</a> </li> <li class="facebox-user-list-item"> <img alt="@xiejw" height="24" src="https://avatars1.githubusercontent.com/u/1184671?s=48&amp;v=4" width="24" /> <a href="/xiejw">xiejw</a> </li> <li class="facebox-user-list-item"> <img alt="@martinwicke" height="24" src="https://avatars0.githubusercontent.com/u/577277?s=48&amp;v=4" width="24" /> <a href="/martinwicke">martinwicke</a> </li> <li class="facebox-user-list-item"> <img alt="@michaelkhan3" height="24" src="https://avatars2.githubusercontent.com/u/4420471?s=48&amp;v=4" width="24" /> <a href="/michaelkhan3">michaelkhan3</a> </li> <li class="facebox-user-list-item"> <img alt="@fchollet" height="24" src="https://avatars3.githubusercontent.com/u/710255?s=48&amp;v=4" width="24" /> <a href="/fchollet">fchollet</a> </li> <li class="facebox-user-list-item"> <img alt="@drpngx" height="24" src="https://avatars2.githubusercontent.com/u/20959853?s=48&amp;v=4" width="24" /> <a href="/drpngx">drpngx</a> </li> <li class="facebox-user-list-item"> <img alt="@caisq" height="24" src="https://avatars1.githubusercontent.com/u/16824702?s=48&amp;v=4" width="24" /> <a href="/caisq">caisq</a> </li> <li class="facebox-user-list-item"> <img alt="@alextp" height="24" src="https://avatars3.githubusercontent.com/u/5061?s=48&amp;v=4" width="24" /> <a href="/alextp">alextp</a> </li> <li class="facebox-user-list-item"> <img alt="@annarev" height="24" src="https://avatars1.githubusercontent.com/u/22060313?s=48&amp;v=4" width="24" /> <a href="/annarev">annarev</a> </li> <li class="facebox-user-list-item"> <img alt="@kosklain" height="24" src="https://avatars3.githubusercontent.com/u/1104579?s=48&amp;v=4" width="24" /> <a href="/kosklain">kosklain</a> </li> </ul> </div> </div> <div class="file"> <div class="file-header"> <div class="file-actions"> <div class="BtnGroup"> <a id="raw-url" class="btn btn-sm BtnGroup-item" href="/tensorflow/tensorflow/raw/r1.6/tensorflow/python/ops/rnn_cell_impl.py">Raw</a> <a class="btn btn-sm js-update-url-with-hash BtnGroup-item" data-hotkey="b" href="/tensorflow/tensorflow/blame/r1.6/tensorflow/python/ops/rnn_cell_impl.py">Blame</a> <a rel="nofollow" class="btn btn-sm BtnGroup-item" href="/tensorflow/tensorflow/commits/r1.6/tensorflow/python/ops/rnn_cell_impl.py">History</a> </div> <button type="button" class="btn-octicon disabled tooltipped tooltipped-nw" aria-label="You must be signed in to make or propose changes"> <svg class="octicon octicon-pencil" viewBox="0 0 14 16" version="1.1" width="14" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M0 12v3h3l8-8-3-3-8 8zm3 2H1v-2h1v1h1v1zm10.3-9.3L12 6 9 3l1.3-1.3a.996.996 0 0 1 1.41 0l1.59 1.59c.39.39.39 1.02 0 1.41z"/></svg> </button> <button type="button" class="btn-octicon btn-octicon-danger disabled tooltipped tooltipped-nw" aria-label="You must be signed in to make or propose changes"> <svg class="octicon octicon-trashcan" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M11 2H9c0-.55-.45-1-1-1H5c-.55 0-1 .45-1 1H2c-.55 0-1 .45-1 1v1c0 .55.45 1 1 1v9c0 .55.45 1 1 1h7c.55 0 1-.45 1-1V5c.55 0 1-.45 1-1V3c0-.55-.45-1-1-1zm-1 12H3V5h1v8h1V5h1v8h1V5h1v8h1V5h1v9zm1-10H2V3h9v1z"/></svg> </button> </div> <div class="file-info"> 1286 lines (1078 sloc) <span class="file-info-divider"></span> 49.7 KB </div> </div> <div itemprop="text" class="blob-wrapper data type-python"> <table class="highlight tab-size js-file-line-container" data-tab-size="8"> <tr> <td id="L1" class="blob-num js-line-number" data-line-number="1"></td> <td id="LC1" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> Copyright 2015 The TensorFlow Authors. All Rights Reserved.</span></td> </tr> <tr> <td id="L2" class="blob-num js-line-number" data-line-number="2"></td> <td id="LC2" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span></span></td> </tr> <tr> <td id="L3" class="blob-num js-line-number" data-line-number="3"></td> <td id="LC3" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> Licensed under the Apache License, Version 2.0 (the &quot;License&quot;);</span></td> </tr> <tr> <td id="L4" class="blob-num js-line-number" data-line-number="4"></td> <td id="LC4" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> you may not use this file except in compliance with the License.</span></td> </tr> <tr> <td id="L5" class="blob-num js-line-number" data-line-number="5"></td> <td id="LC5" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> You may obtain a copy of the License at</span></td> </tr> <tr> <td id="L6" class="blob-num js-line-number" data-line-number="6"></td> <td id="LC6" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span></span></td> </tr> <tr> <td id="L7" class="blob-num js-line-number" data-line-number="7"></td> <td id="LC7" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> http://www.apache.org/licenses/LICENSE-2.0</span></td> </tr> <tr> <td id="L8" class="blob-num js-line-number" data-line-number="8"></td> <td id="LC8" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span></span></td> </tr> <tr> <td id="L9" class="blob-num js-line-number" data-line-number="9"></td> <td id="LC9" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> Unless required by applicable law or agreed to in writing, software</span></td> </tr> <tr> <td id="L10" class="blob-num js-line-number" data-line-number="10"></td> <td id="LC10" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> distributed under the License is distributed on an &quot;AS IS&quot; BASIS,</span></td> </tr> <tr> <td id="L11" class="blob-num js-line-number" data-line-number="11"></td> <td id="LC11" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.</span></td> </tr> <tr> <td id="L12" class="blob-num js-line-number" data-line-number="12"></td> <td id="LC12" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> See the License for the specific language governing permissions and</span></td> </tr> <tr> <td id="L13" class="blob-num js-line-number" data-line-number="13"></td> <td id="LC13" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> limitations under the License.</span></td> </tr> <tr> <td id="L14" class="blob-num js-line-number" data-line-number="14"></td> <td id="LC14" class="blob-code blob-code-inner js-file-line"><span class="pl-c"><span class="pl-c">#</span> ==============================================================================</span></td> </tr> <tr> <td id="L15" class="blob-num js-line-number" data-line-number="15"></td> <td id="LC15" class="blob-code blob-code-inner js-file-line"><span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Module implementing RNN Cells.</span></td> </tr> <tr> <td id="L16" class="blob-num js-line-number" data-line-number="16"></td> <td id="LC16" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L17" class="blob-num js-line-number" data-line-number="17"></td> <td id="LC17" class="blob-code blob-code-inner js-file-line"><span class="pl-s">This module provides a number of basic commonly used RNN cells, such as LSTM</span></td> </tr> <tr> <td id="L18" class="blob-num js-line-number" data-line-number="18"></td> <td id="LC18" class="blob-code blob-code-inner js-file-line"><span class="pl-s">(Long Short Term Memory) or GRU (Gated Recurrent Unit), and a number of</span></td> </tr> <tr> <td id="L19" class="blob-num js-line-number" data-line-number="19"></td> <td id="LC19" class="blob-code blob-code-inner js-file-line"><span class="pl-s">operators that allow adding dropouts, projections, or embeddings for inputs.</span></td> </tr> <tr> <td id="L20" class="blob-num js-line-number" data-line-number="20"></td> <td id="LC20" class="blob-code blob-code-inner js-file-line"><span class="pl-s">Constructing multi-layer cells is supported by the class `MultiRNNCell`, or by</span></td> </tr> <tr> <td id="L21" class="blob-num js-line-number" data-line-number="21"></td> <td id="LC21" class="blob-code blob-code-inner js-file-line"><span class="pl-s">calling the `rnn` ops several times.</span></td> </tr> <tr> <td id="L22" class="blob-num js-line-number" data-line-number="22"></td> <td id="LC22" class="blob-code blob-code-inner js-file-line"><span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L23" class="blob-num js-line-number" data-line-number="23"></td> <td id="LC23" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-c1">__future__</span> <span class="pl-k">import</span> absolute_import</td> </tr> <tr> <td id="L24" class="blob-num js-line-number" data-line-number="24"></td> <td id="LC24" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-c1">__future__</span> <span class="pl-k">import</span> division</td> </tr> <tr> <td id="L25" class="blob-num js-line-number" data-line-number="25"></td> <td id="LC25" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> <span class="pl-c1">__future__</span> <span class="pl-k">import</span> print_function</td> </tr> <tr> <td id="L26" class="blob-num js-line-number" data-line-number="26"></td> <td id="LC26" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L27" class="blob-num js-line-number" data-line-number="27"></td> <td id="LC27" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> collections</td> </tr> <tr> <td id="L28" class="blob-num js-line-number" data-line-number="28"></td> <td id="LC28" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> hashlib</td> </tr> <tr> <td id="L29" class="blob-num js-line-number" data-line-number="29"></td> <td id="LC29" class="blob-code blob-code-inner js-file-line"><span class="pl-k">import</span> numbers</td> </tr> <tr> <td id="L30" class="blob-num js-line-number" data-line-number="30"></td> <td id="LC30" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L31" class="blob-num js-line-number" data-line-number="31"></td> <td id="LC31" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.eager <span class="pl-k">import</span> context</td> </tr> <tr> <td id="L32" class="blob-num js-line-number" data-line-number="32"></td> <td id="LC32" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> constant_op</td> </tr> <tr> <td id="L33" class="blob-num js-line-number" data-line-number="33"></td> <td id="LC33" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> dtypes</td> </tr> <tr> <td id="L34" class="blob-num js-line-number" data-line-number="34"></td> <td id="LC34" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> ops</td> </tr> <tr> <td id="L35" class="blob-num js-line-number" data-line-number="35"></td> <td id="LC35" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> tensor_shape</td> </tr> <tr> <td id="L36" class="blob-num js-line-number" data-line-number="36"></td> <td id="LC36" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.framework <span class="pl-k">import</span> tensor_util</td> </tr> <tr> <td id="L37" class="blob-num js-line-number" data-line-number="37"></td> <td id="LC37" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.layers <span class="pl-k">import</span> base <span class="pl-k">as</span> base_layer</td> </tr> <tr> <td id="L38" class="blob-num js-line-number" data-line-number="38"></td> <td id="LC38" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> array_ops</td> </tr> <tr> <td id="L39" class="blob-num js-line-number" data-line-number="39"></td> <td id="LC39" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> clip_ops</td> </tr> <tr> <td id="L40" class="blob-num js-line-number" data-line-number="40"></td> <td id="LC40" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> init_ops</td> </tr> <tr> <td id="L41" class="blob-num js-line-number" data-line-number="41"></td> <td id="LC41" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> math_ops</td> </tr> <tr> <td id="L42" class="blob-num js-line-number" data-line-number="42"></td> <td id="LC42" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> nn_ops</td> </tr> <tr> <td id="L43" class="blob-num js-line-number" data-line-number="43"></td> <td id="LC43" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> partitioned_variables</td> </tr> <tr> <td id="L44" class="blob-num js-line-number" data-line-number="44"></td> <td id="LC44" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> random_ops</td> </tr> <tr> <td id="L45" class="blob-num js-line-number" data-line-number="45"></td> <td id="LC45" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> tensor_array_ops</td> </tr> <tr> <td id="L46" class="blob-num js-line-number" data-line-number="46"></td> <td id="LC46" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> variable_scope <span class="pl-k">as</span> vs</td> </tr> <tr> <td id="L47" class="blob-num js-line-number" data-line-number="47"></td> <td id="LC47" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.ops <span class="pl-k">import</span> variables <span class="pl-k">as</span> tf_variables</td> </tr> <tr> <td id="L48" class="blob-num js-line-number" data-line-number="48"></td> <td id="LC48" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.platform <span class="pl-k">import</span> tf_logging <span class="pl-k">as</span> logging</td> </tr> <tr> <td id="L49" class="blob-num js-line-number" data-line-number="49"></td> <td id="LC49" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.util <span class="pl-k">import</span> nest</td> </tr> <tr> <td id="L50" class="blob-num js-line-number" data-line-number="50"></td> <td id="LC50" class="blob-code blob-code-inner js-file-line"><span class="pl-k">from</span> tensorflow.python.util.tf_export <span class="pl-k">import</span> tf_export</td> </tr> <tr> <td id="L51" class="blob-num js-line-number" data-line-number="51"></td> <td id="LC51" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L52" class="blob-num js-line-number" data-line-number="52"></td> <td id="LC52" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L53" class="blob-num js-line-number" data-line-number="53"></td> <td id="LC53" class="blob-code blob-code-inner js-file-line"><span class="pl-c1">_BIAS_VARIABLE_NAME</span> <span class="pl-k">=</span> <span class="pl-s"><span class="pl-pds">&quot;</span>bias<span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L54" class="blob-num js-line-number" data-line-number="54"></td> <td id="LC54" class="blob-code blob-code-inner js-file-line"><span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span> <span class="pl-k">=</span> <span class="pl-s"><span class="pl-pds">&quot;</span>kernel<span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L55" class="blob-num js-line-number" data-line-number="55"></td> <td id="LC55" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L56" class="blob-num js-line-number" data-line-number="56"></td> <td id="LC56" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L57" class="blob-num js-line-number" data-line-number="57"></td> <td id="LC57" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_like_rnncell</span>(<span class="pl-smi">cell</span>):</td> </tr> <tr> <td id="L58" class="blob-num js-line-number" data-line-number="58"></td> <td id="LC58" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Checks that a given object is an RNNCell by using duck typing.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L59" class="blob-num js-line-number" data-line-number="59"></td> <td id="LC59" class="blob-code blob-code-inner js-file-line"> conditions <span class="pl-k">=</span> [<span class="pl-c1">hasattr</span>(cell, <span class="pl-s"><span class="pl-pds">&quot;</span>output_size<span class="pl-pds">&quot;</span></span>), <span class="pl-c1">hasattr</span>(cell, <span class="pl-s"><span class="pl-pds">&quot;</span>state_size<span class="pl-pds">&quot;</span></span>),</td> </tr> <tr> <td id="L60" class="blob-num js-line-number" data-line-number="60"></td> <td id="LC60" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">hasattr</span>(cell, <span class="pl-s"><span class="pl-pds">&quot;</span>zero_state<span class="pl-pds">&quot;</span></span>), <span class="pl-c1">callable</span>(cell)]</td> </tr> <tr> <td id="L61" class="blob-num js-line-number" data-line-number="61"></td> <td id="LC61" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">all</span>(conditions)</td> </tr> <tr> <td id="L62" class="blob-num js-line-number" data-line-number="62"></td> <td id="LC62" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L63" class="blob-num js-line-number" data-line-number="63"></td> <td id="LC63" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L64" class="blob-num js-line-number" data-line-number="64"></td> <td id="LC64" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_concat</span>(<span class="pl-smi">prefix</span>, <span class="pl-smi">suffix</span>, <span class="pl-smi">static</span><span class="pl-k">=</span><span class="pl-c1">False</span>):</td> </tr> <tr> <td id="L65" class="blob-num js-line-number" data-line-number="65"></td> <td id="LC65" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Concat that enables int, Tensor, or TensorShape values.</span></td> </tr> <tr> <td id="L66" class="blob-num js-line-number" data-line-number="66"></td> <td id="LC66" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L67" class="blob-num js-line-number" data-line-number="67"></td> <td id="LC67" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> This function takes a size specification, which can be an integer, a</span></td> </tr> <tr> <td id="L68" class="blob-num js-line-number" data-line-number="68"></td> <td id="LC68" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> TensorShape, or a Tensor, and converts it into a concatenated Tensor</span></td> </tr> <tr> <td id="L69" class="blob-num js-line-number" data-line-number="69"></td> <td id="LC69" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> (if static = False) or a list of integers (if static = True).</span></td> </tr> <tr> <td id="L70" class="blob-num js-line-number" data-line-number="70"></td> <td id="LC70" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L71" class="blob-num js-line-number" data-line-number="71"></td> <td id="LC71" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L72" class="blob-num js-line-number" data-line-number="72"></td> <td id="LC72" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> prefix: The prefix; usually the batch size (and/or time step size).</span></td> </tr> <tr> <td id="L73" class="blob-num js-line-number" data-line-number="73"></td> <td id="LC73" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> (TensorShape, int, or Tensor.)</span></td> </tr> <tr> <td id="L74" class="blob-num js-line-number" data-line-number="74"></td> <td id="LC74" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> suffix: TensorShape, int, or Tensor.</span></td> </tr> <tr> <td id="L75" class="blob-num js-line-number" data-line-number="75"></td> <td id="LC75" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> static: If `True`, return a python list with possibly unknown dimensions.</span></td> </tr> <tr> <td id="L76" class="blob-num js-line-number" data-line-number="76"></td> <td id="LC76" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Otherwise return a `Tensor`.</span></td> </tr> <tr> <td id="L77" class="blob-num js-line-number" data-line-number="77"></td> <td id="LC77" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L78" class="blob-num js-line-number" data-line-number="78"></td> <td id="LC78" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L79" class="blob-num js-line-number" data-line-number="79"></td> <td id="LC79" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> shape: the concatenation of prefix and suffix.</span></td> </tr> <tr> <td id="L80" class="blob-num js-line-number" data-line-number="80"></td> <td id="LC80" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L81" class="blob-num js-line-number" data-line-number="81"></td> <td id="LC81" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L82" class="blob-num js-line-number" data-line-number="82"></td> <td id="LC82" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if `suffix` is not a scalar or vector (or TensorShape).</span></td> </tr> <tr> <td id="L83" class="blob-num js-line-number" data-line-number="83"></td> <td id="LC83" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if prefix or suffix was `None` and asked for dynamic</span></td> </tr> <tr> <td id="L84" class="blob-num js-line-number" data-line-number="84"></td> <td id="LC84" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Tensors out.</span></td> </tr> <tr> <td id="L85" class="blob-num js-line-number" data-line-number="85"></td> <td id="LC85" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L86" class="blob-num js-line-number" data-line-number="86"></td> <td id="LC86" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">isinstance</span>(prefix, ops.Tensor):</td> </tr> <tr> <td id="L87" class="blob-num js-line-number" data-line-number="87"></td> <td id="LC87" class="blob-code blob-code-inner js-file-line"> p <span class="pl-k">=</span> prefix</td> </tr> <tr> <td id="L88" class="blob-num js-line-number" data-line-number="88"></td> <td id="LC88" class="blob-code blob-code-inner js-file-line"> p_static <span class="pl-k">=</span> tensor_util.constant_value(prefix)</td> </tr> <tr> <td id="L89" class="blob-num js-line-number" data-line-number="89"></td> <td id="LC89" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> p.shape.ndims <span class="pl-k">==</span> <span class="pl-c1">0</span>:</td> </tr> <tr> <td id="L90" class="blob-num js-line-number" data-line-number="90"></td> <td id="LC90" class="blob-code blob-code-inner js-file-line"> p <span class="pl-k">=</span> array_ops.expand_dims(p, <span class="pl-c1">0</span>)</td> </tr> <tr> <td id="L91" class="blob-num js-line-number" data-line-number="91"></td> <td id="LC91" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">elif</span> p.shape.ndims <span class="pl-k">!=</span> <span class="pl-c1">1</span>:</td> </tr> <tr> <td id="L92" class="blob-num js-line-number" data-line-number="92"></td> <td id="LC92" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>prefix tensor must be either a scalar or vector, <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L93" class="blob-num js-line-number" data-line-number="93"></td> <td id="LC93" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>but saw tensor: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> p)</td> </tr> <tr> <td id="L94" class="blob-num js-line-number" data-line-number="94"></td> <td id="LC94" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L95" class="blob-num js-line-number" data-line-number="95"></td> <td id="LC95" class="blob-code blob-code-inner js-file-line"> p <span class="pl-k">=</span> tensor_shape.as_shape(prefix)</td> </tr> <tr> <td id="L96" class="blob-num js-line-number" data-line-number="96"></td> <td id="LC96" class="blob-code blob-code-inner js-file-line"> p_static <span class="pl-k">=</span> p.as_list() <span class="pl-k">if</span> p.ndims <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L97" class="blob-num js-line-number" data-line-number="97"></td> <td id="LC97" class="blob-code blob-code-inner js-file-line"> p <span class="pl-k">=</span> (constant_op.constant(p.as_list(), <span class="pl-v">dtype</span><span class="pl-k">=</span>dtypes.int32)</td> </tr> <tr> <td id="L98" class="blob-num js-line-number" data-line-number="98"></td> <td id="LC98" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> p.is_fully_defined() <span class="pl-k">else</span> <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L99" class="blob-num js-line-number" data-line-number="99"></td> <td id="LC99" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">isinstance</span>(suffix, ops.Tensor):</td> </tr> <tr> <td id="L100" class="blob-num js-line-number" data-line-number="100"></td> <td id="LC100" class="blob-code blob-code-inner js-file-line"> s <span class="pl-k">=</span> suffix</td> </tr> <tr> <td id="L101" class="blob-num js-line-number" data-line-number="101"></td> <td id="LC101" class="blob-code blob-code-inner js-file-line"> s_static <span class="pl-k">=</span> tensor_util.constant_value(suffix)</td> </tr> <tr> <td id="L102" class="blob-num js-line-number" data-line-number="102"></td> <td id="LC102" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> s.shape.ndims <span class="pl-k">==</span> <span class="pl-c1">0</span>:</td> </tr> <tr> <td id="L103" class="blob-num js-line-number" data-line-number="103"></td> <td id="LC103" class="blob-code blob-code-inner js-file-line"> s <span class="pl-k">=</span> array_ops.expand_dims(s, <span class="pl-c1">0</span>)</td> </tr> <tr> <td id="L104" class="blob-num js-line-number" data-line-number="104"></td> <td id="LC104" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">elif</span> s.shape.ndims <span class="pl-k">!=</span> <span class="pl-c1">1</span>:</td> </tr> <tr> <td id="L105" class="blob-num js-line-number" data-line-number="105"></td> <td id="LC105" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>suffix tensor must be either a scalar or vector, <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L106" class="blob-num js-line-number" data-line-number="106"></td> <td id="LC106" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>but saw tensor: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> s)</td> </tr> <tr> <td id="L107" class="blob-num js-line-number" data-line-number="107"></td> <td id="LC107" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L108" class="blob-num js-line-number" data-line-number="108"></td> <td id="LC108" class="blob-code blob-code-inner js-file-line"> s <span class="pl-k">=</span> tensor_shape.as_shape(suffix)</td> </tr> <tr> <td id="L109" class="blob-num js-line-number" data-line-number="109"></td> <td id="LC109" class="blob-code blob-code-inner js-file-line"> s_static <span class="pl-k">=</span> s.as_list() <span class="pl-k">if</span> s.ndims <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L110" class="blob-num js-line-number" data-line-number="110"></td> <td id="LC110" class="blob-code blob-code-inner js-file-line"> s <span class="pl-k">=</span> (constant_op.constant(s.as_list(), <span class="pl-v">dtype</span><span class="pl-k">=</span>dtypes.int32)</td> </tr> <tr> <td id="L111" class="blob-num js-line-number" data-line-number="111"></td> <td id="LC111" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> s.is_fully_defined() <span class="pl-k">else</span> <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L112" class="blob-num js-line-number" data-line-number="112"></td> <td id="LC112" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L113" class="blob-num js-line-number" data-line-number="113"></td> <td id="LC113" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> static:</td> </tr> <tr> <td id="L114" class="blob-num js-line-number" data-line-number="114"></td> <td id="LC114" class="blob-code blob-code-inner js-file-line"> shape <span class="pl-k">=</span> tensor_shape.as_shape(p_static).concatenate(s_static)</td> </tr> <tr> <td id="L115" class="blob-num js-line-number" data-line-number="115"></td> <td id="LC115" class="blob-code blob-code-inner js-file-line"> shape <span class="pl-k">=</span> shape.as_list() <span class="pl-k">if</span> shape.ndims <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L116" class="blob-num js-line-number" data-line-number="116"></td> <td id="LC116" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L117" class="blob-num js-line-number" data-line-number="117"></td> <td id="LC117" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> p <span class="pl-k">is</span> <span class="pl-c1">None</span> <span class="pl-k">or</span> s <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L118" class="blob-num js-line-number" data-line-number="118"></td> <td id="LC118" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Provided a prefix or suffix of None: <span class="pl-c1">%s</span> and <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L119" class="blob-num js-line-number" data-line-number="119"></td> <td id="LC119" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> (prefix, suffix))</td> </tr> <tr> <td id="L120" class="blob-num js-line-number" data-line-number="120"></td> <td id="LC120" class="blob-code blob-code-inner js-file-line"> shape <span class="pl-k">=</span> array_ops.concat((p, s), <span class="pl-c1">0</span>)</td> </tr> <tr> <td id="L121" class="blob-num js-line-number" data-line-number="121"></td> <td id="LC121" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> shape</td> </tr> <tr> <td id="L122" class="blob-num js-line-number" data-line-number="122"></td> <td id="LC122" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L123" class="blob-num js-line-number" data-line-number="123"></td> <td id="LC123" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L124" class="blob-num js-line-number" data-line-number="124"></td> <td id="LC124" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_zero_state_tensors</span>(<span class="pl-smi">state_size</span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L125" class="blob-num js-line-number" data-line-number="125"></td> <td id="LC125" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Create tensors of zeros based on state_size, batch_size, and dtype.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L126" class="blob-num js-line-number" data-line-number="126"></td> <td id="LC126" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">get_state_shape</span>(<span class="pl-smi">s</span>):</td> </tr> <tr> <td id="L127" class="blob-num js-line-number" data-line-number="127"></td> <td id="LC127" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Combine s with batch_size to get a proper tensor shape.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L128" class="blob-num js-line-number" data-line-number="128"></td> <td id="LC128" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> _concat(batch_size, s)</td> </tr> <tr> <td id="L129" class="blob-num js-line-number" data-line-number="129"></td> <td id="LC129" class="blob-code blob-code-inner js-file-line"> size <span class="pl-k">=</span> array_ops.zeros(c, <span class="pl-v">dtype</span><span class="pl-k">=</span>dtype)</td> </tr> <tr> <td id="L130" class="blob-num js-line-number" data-line-number="130"></td> <td id="LC130" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> context.in_graph_mode():</td> </tr> <tr> <td id="L131" class="blob-num js-line-number" data-line-number="131"></td> <td id="LC131" class="blob-code blob-code-inner js-file-line"> c_static <span class="pl-k">=</span> _concat(batch_size, s, <span class="pl-v">static</span><span class="pl-k">=</span><span class="pl-c1">True</span>)</td> </tr> <tr> <td id="L132" class="blob-num js-line-number" data-line-number="132"></td> <td id="LC132" class="blob-code blob-code-inner js-file-line"> size.set_shape(c_static)</td> </tr> <tr> <td id="L133" class="blob-num js-line-number" data-line-number="133"></td> <td id="LC133" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> size</td> </tr> <tr> <td id="L134" class="blob-num js-line-number" data-line-number="134"></td> <td id="LC134" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> nest.map_structure(get_state_shape, state_size)</td> </tr> <tr> <td id="L135" class="blob-num js-line-number" data-line-number="135"></td> <td id="LC135" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L136" class="blob-num js-line-number" data-line-number="136"></td> <td id="LC136" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L137" class="blob-num js-line-number" data-line-number="137"></td> <td id="LC137" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.RNNCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L138" class="blob-num js-line-number" data-line-number="138"></td> <td id="LC138" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">RNNCell</span>(<span class="pl-e">base_layer</span>.<span class="pl-e">Layer</span>):</td> </tr> <tr> <td id="L139" class="blob-num js-line-number" data-line-number="139"></td> <td id="LC139" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Abstract object representing an RNN cell.</span></td> </tr> <tr> <td id="L140" class="blob-num js-line-number" data-line-number="140"></td> <td id="LC140" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L141" class="blob-num js-line-number" data-line-number="141"></td> <td id="LC141" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Every `RNNCell` must have the properties below and implement `call` with</span></td> </tr> <tr> <td id="L142" class="blob-num js-line-number" data-line-number="142"></td> <td id="LC142" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the signature `(output, next_state) = call(input, state)`. The optional</span></td> </tr> <tr> <td id="L143" class="blob-num js-line-number" data-line-number="143"></td> <td id="LC143" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> third input argument, `scope`, is allowed for backwards compatibility</span></td> </tr> <tr> <td id="L144" class="blob-num js-line-number" data-line-number="144"></td> <td id="LC144" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> purposes; but should be left off for new subclasses.</span></td> </tr> <tr> <td id="L145" class="blob-num js-line-number" data-line-number="145"></td> <td id="LC145" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L146" class="blob-num js-line-number" data-line-number="146"></td> <td id="LC146" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> This definition of cell differs from the definition used in the literature.</span></td> </tr> <tr> <td id="L147" class="blob-num js-line-number" data-line-number="147"></td> <td id="LC147" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> In the literature, &#39;cell&#39; refers to an object with a single scalar output.</span></td> </tr> <tr> <td id="L148" class="blob-num js-line-number" data-line-number="148"></td> <td id="LC148" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> This definition refers to a horizontal array of such units.</span></td> </tr> <tr> <td id="L149" class="blob-num js-line-number" data-line-number="149"></td> <td id="LC149" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L150" class="blob-num js-line-number" data-line-number="150"></td> <td id="LC150" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> An RNN cell, in the most abstract setting, is anything that has</span></td> </tr> <tr> <td id="L151" class="blob-num js-line-number" data-line-number="151"></td> <td id="LC151" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> a state and performs some operation that takes a matrix of inputs.</span></td> </tr> <tr> <td id="L152" class="blob-num js-line-number" data-line-number="152"></td> <td id="LC152" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> This operation results in an output matrix with `self.output_size` columns.</span></td> </tr> <tr> <td id="L153" class="blob-num js-line-number" data-line-number="153"></td> <td id="LC153" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If `self.state_size` is an integer, this operation also results in a new</span></td> </tr> <tr> <td id="L154" class="blob-num js-line-number" data-line-number="154"></td> <td id="LC154" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state matrix with `self.state_size` columns. If `self.state_size` is a</span></td> </tr> <tr> <td id="L155" class="blob-num js-line-number" data-line-number="155"></td> <td id="LC155" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> (possibly nested tuple of) TensorShape object(s), then it should return a</span></td> </tr> <tr> <td id="L156" class="blob-num js-line-number" data-line-number="156"></td> <td id="LC156" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> matching structure of Tensors having shape `[batch_size].concatenate(s)`</span></td> </tr> <tr> <td id="L157" class="blob-num js-line-number" data-line-number="157"></td> <td id="LC157" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> for each `s` in `self.batch_size`.</span></td> </tr> <tr> <td id="L158" class="blob-num js-line-number" data-line-number="158"></td> <td id="LC158" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L159" class="blob-num js-line-number" data-line-number="159"></td> <td id="LC159" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L160" class="blob-num js-line-number" data-line-number="160"></td> <td id="LC160" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L161" class="blob-num js-line-number" data-line-number="161"></td> <td id="LC161" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run this RNN cell on inputs, starting from the given state.</span></td> </tr> <tr> <td id="L162" class="blob-num js-line-number" data-line-number="162"></td> <td id="LC162" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L163" class="blob-num js-line-number" data-line-number="163"></td> <td id="LC163" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L164" class="blob-num js-line-number" data-line-number="164"></td> <td id="LC164" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: `2-D` tensor with shape `[batch_size, input_size]`.</span></td> </tr> <tr> <td id="L165" class="blob-num js-line-number" data-line-number="165"></td> <td id="LC165" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: if `self.state_size` is an integer, this should be a `2-D Tensor`</span></td> </tr> <tr> <td id="L166" class="blob-num js-line-number" data-line-number="166"></td> <td id="LC166" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> with shape `[batch_size, self.state_size]`. Otherwise, if</span></td> </tr> <tr> <td id="L167" class="blob-num js-line-number" data-line-number="167"></td> <td id="LC167" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `self.state_size` is a tuple of integers, this should be a tuple</span></td> </tr> <tr> <td id="L168" class="blob-num js-line-number" data-line-number="168"></td> <td id="LC168" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> with shapes `[batch_size, s] for s in self.state_size`.</span></td> </tr> <tr> <td id="L169" class="blob-num js-line-number" data-line-number="169"></td> <td id="LC169" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> scope: VariableScope for the created subgraph; defaults to class name.</span></td> </tr> <tr> <td id="L170" class="blob-num js-line-number" data-line-number="170"></td> <td id="LC170" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L171" class="blob-num js-line-number" data-line-number="171"></td> <td id="LC171" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L172" class="blob-num js-line-number" data-line-number="172"></td> <td id="LC172" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> A pair containing:</span></td> </tr> <tr> <td id="L173" class="blob-num js-line-number" data-line-number="173"></td> <td id="LC173" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L174" class="blob-num js-line-number" data-line-number="174"></td> <td id="LC174" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - Output: A `2-D` tensor with shape `[batch_size, self.output_size]`.</span></td> </tr> <tr> <td id="L175" class="blob-num js-line-number" data-line-number="175"></td> <td id="LC175" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - New state: Either a single `2-D` tensor, or a tuple of tensors matching</span></td> </tr> <tr> <td id="L176" class="blob-num js-line-number" data-line-number="176"></td> <td id="LC176" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the arity and shapes of `state`.</span></td> </tr> <tr> <td id="L177" class="blob-num js-line-number" data-line-number="177"></td> <td id="LC177" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L178" class="blob-num js-line-number" data-line-number="178"></td> <td id="LC178" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> scope <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L179" class="blob-num js-line-number" data-line-number="179"></td> <td id="LC179" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> vs.variable_scope(scope,</td> </tr> <tr> <td id="L180" class="blob-num js-line-number" data-line-number="180"></td> <td id="LC180" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">custom_getter</span><span class="pl-k">=</span><span class="pl-c1">self</span>._rnn_get_variable) <span class="pl-k">as</span> scope:</td> </tr> <tr> <td id="L181" class="blob-num js-line-number" data-line-number="181"></td> <td id="LC181" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">super</span>(RNNCell, <span class="pl-c1">self</span>).<span class="pl-c1">__call__</span>(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L182" class="blob-num js-line-number" data-line-number="182"></td> <td id="LC182" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L183" class="blob-num js-line-number" data-line-number="183"></td> <td id="LC183" class="blob-code blob-code-inner js-file-line"> scope_attrname <span class="pl-k">=</span> <span class="pl-s"><span class="pl-pds">&quot;</span>rnncell_scope<span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L184" class="blob-num js-line-number" data-line-number="184"></td> <td id="LC184" class="blob-code blob-code-inner js-file-line"> scope <span class="pl-k">=</span> <span class="pl-c1">getattr</span>(<span class="pl-c1">self</span>, scope_attrname, <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L185" class="blob-num js-line-number" data-line-number="185"></td> <td id="LC185" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> scope <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L186" class="blob-num js-line-number" data-line-number="186"></td> <td id="LC186" class="blob-code blob-code-inner js-file-line"> scope <span class="pl-k">=</span> vs.variable_scope(vs.get_variable_scope(),</td> </tr> <tr> <td id="L187" class="blob-num js-line-number" data-line-number="187"></td> <td id="LC187" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">custom_getter</span><span class="pl-k">=</span><span class="pl-c1">self</span>._rnn_get_variable)</td> </tr> <tr> <td id="L188" class="blob-num js-line-number" data-line-number="188"></td> <td id="LC188" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">setattr</span>(<span class="pl-c1">self</span>, scope_attrname, scope)</td> </tr> <tr> <td id="L189" class="blob-num js-line-number" data-line-number="189"></td> <td id="LC189" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> scope:</td> </tr> <tr> <td id="L190" class="blob-num js-line-number" data-line-number="190"></td> <td id="LC190" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">super</span>(RNNCell, <span class="pl-c1">self</span>).<span class="pl-c1">__call__</span>(inputs, state)</td> </tr> <tr> <td id="L191" class="blob-num js-line-number" data-line-number="191"></td> <td id="LC191" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L192" class="blob-num js-line-number" data-line-number="192"></td> <td id="LC192" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_rnn_get_variable</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">getter</span>, <span class="pl-k">*</span><span class="pl-smi">args</span>, <span class="pl-k">**</span><span class="pl-smi">kwargs</span>):</td> </tr> <tr> <td id="L193" class="blob-num js-line-number" data-line-number="193"></td> <td id="LC193" class="blob-code blob-code-inner js-file-line"> variable <span class="pl-k">=</span> getter(<span class="pl-k">*</span>args, <span class="pl-k">**</span>kwargs)</td> </tr> <tr> <td id="L194" class="blob-num js-line-number" data-line-number="194"></td> <td id="LC194" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> context.in_graph_mode():</td> </tr> <tr> <td id="L195" class="blob-num js-line-number" data-line-number="195"></td> <td id="LC195" class="blob-code blob-code-inner js-file-line"> trainable <span class="pl-k">=</span> (variable <span class="pl-k">in</span> tf_variables.trainable_variables() <span class="pl-k">or</span></td> </tr> <tr> <td id="L196" class="blob-num js-line-number" data-line-number="196"></td> <td id="LC196" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">isinstance</span>(variable, tf_variables.PartitionedVariable) <span class="pl-k">and</span></td> </tr> <tr> <td id="L197" class="blob-num js-line-number" data-line-number="197"></td> <td id="LC197" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">list</span>(variable)[<span class="pl-c1">0</span>] <span class="pl-k">in</span> tf_variables.trainable_variables()))</td> </tr> <tr> <td id="L198" class="blob-num js-line-number" data-line-number="198"></td> <td id="LC198" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L199" class="blob-num js-line-number" data-line-number="199"></td> <td id="LC199" class="blob-code blob-code-inner js-file-line"> trainable <span class="pl-k">=</span> variable._trainable <span class="pl-c"><span class="pl-c">#</span> pylint: disable=protected-access</span></td> </tr> <tr> <td id="L200" class="blob-num js-line-number" data-line-number="200"></td> <td id="LC200" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> trainable <span class="pl-k">and</span> variable <span class="pl-k">not</span> <span class="pl-k">in</span> <span class="pl-c1">self</span>._trainable_weights:</td> </tr> <tr> <td id="L201" class="blob-num js-line-number" data-line-number="201"></td> <td id="LC201" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._trainable_weights.append(variable)</td> </tr> <tr> <td id="L202" class="blob-num js-line-number" data-line-number="202"></td> <td id="LC202" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">elif</span> <span class="pl-k">not</span> trainable <span class="pl-k">and</span> variable <span class="pl-k">not</span> <span class="pl-k">in</span> <span class="pl-c1">self</span>._non_trainable_weights:</td> </tr> <tr> <td id="L203" class="blob-num js-line-number" data-line-number="203"></td> <td id="LC203" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._non_trainable_weights.append(variable)</td> </tr> <tr> <td id="L204" class="blob-num js-line-number" data-line-number="204"></td> <td id="LC204" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> variable</td> </tr> <tr> <td id="L205" class="blob-num js-line-number" data-line-number="205"></td> <td id="LC205" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L206" class="blob-num js-line-number" data-line-number="206"></td> <td id="LC206" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L207" class="blob-num js-line-number" data-line-number="207"></td> <td id="LC207" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L208" class="blob-num js-line-number" data-line-number="208"></td> <td id="LC208" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>size(s) of state(s) used by this cell.</span></td> </tr> <tr> <td id="L209" class="blob-num js-line-number" data-line-number="209"></td> <td id="LC209" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L210" class="blob-num js-line-number" data-line-number="210"></td> <td id="LC210" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> It can be represented by an Integer, a TensorShape or a tuple of Integers</span></td> </tr> <tr> <td id="L211" class="blob-num js-line-number" data-line-number="211"></td> <td id="LC211" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> or TensorShapes.</span></td> </tr> <tr> <td id="L212" class="blob-num js-line-number" data-line-number="212"></td> <td id="LC212" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L213" class="blob-num js-line-number" data-line-number="213"></td> <td id="LC213" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">NotImplementedError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Abstract method<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L214" class="blob-num js-line-number" data-line-number="214"></td> <td id="LC214" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L215" class="blob-num js-line-number" data-line-number="215"></td> <td id="LC215" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L216" class="blob-num js-line-number" data-line-number="216"></td> <td id="LC216" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L217" class="blob-num js-line-number" data-line-number="217"></td> <td id="LC217" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Integer or TensorShape: size of outputs produced by this cell.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L218" class="blob-num js-line-number" data-line-number="218"></td> <td id="LC218" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">NotImplementedError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Abstract method<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L219" class="blob-num js-line-number" data-line-number="219"></td> <td id="LC219" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L220" class="blob-num js-line-number" data-line-number="220"></td> <td id="LC220" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">_</span>):</td> </tr> <tr> <td id="L221" class="blob-num js-line-number" data-line-number="221"></td> <td id="LC221" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> This tells the parent Layer object that it&#39;s OK to call</span></td> </tr> <tr> <td id="L222" class="blob-num js-line-number" data-line-number="222"></td> <td id="LC222" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> self.add_variable() inside the call() method.</span></td> </tr> <tr> <td id="L223" class="blob-num js-line-number" data-line-number="223"></td> <td id="LC223" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">pass</span></td> </tr> <tr> <td id="L224" class="blob-num js-line-number" data-line-number="224"></td> <td id="LC224" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L225" class="blob-num js-line-number" data-line-number="225"></td> <td id="LC225" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L226" class="blob-num js-line-number" data-line-number="226"></td> <td id="LC226" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Return zero-filled state tensor(s).</span></td> </tr> <tr> <td id="L227" class="blob-num js-line-number" data-line-number="227"></td> <td id="LC227" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L228" class="blob-num js-line-number" data-line-number="228"></td> <td id="LC228" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L229" class="blob-num js-line-number" data-line-number="229"></td> <td id="LC229" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> batch_size: int, float, or unit Tensor representing the batch size.</span></td> </tr> <tr> <td id="L230" class="blob-num js-line-number" data-line-number="230"></td> <td id="LC230" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> dtype: the data type to use for the state.</span></td> </tr> <tr> <td id="L231" class="blob-num js-line-number" data-line-number="231"></td> <td id="LC231" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L232" class="blob-num js-line-number" data-line-number="232"></td> <td id="LC232" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L233" class="blob-num js-line-number" data-line-number="233"></td> <td id="LC233" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If `state_size` is an int or TensorShape, then the return value is a</span></td> </tr> <tr> <td id="L234" class="blob-num js-line-number" data-line-number="234"></td> <td id="LC234" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `N-D` tensor of shape `[batch_size, state_size]` filled with zeros.</span></td> </tr> <tr> <td id="L235" class="blob-num js-line-number" data-line-number="235"></td> <td id="LC235" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L236" class="blob-num js-line-number" data-line-number="236"></td> <td id="LC236" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If `state_size` is a nested list or tuple, then the return value is</span></td> </tr> <tr> <td id="L237" class="blob-num js-line-number" data-line-number="237"></td> <td id="LC237" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> a nested list or tuple (of the same structure) of `2-D` tensors with</span></td> </tr> <tr> <td id="L238" class="blob-num js-line-number" data-line-number="238"></td> <td id="LC238" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the shapes `[batch_size, s]` for each s in `state_size`.</span></td> </tr> <tr> <td id="L239" class="blob-num js-line-number" data-line-number="239"></td> <td id="LC239" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L240" class="blob-num js-line-number" data-line-number="240"></td> <td id="LC240" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Try to use the last cached zero_state. This is done to avoid recreating</span></td> </tr> <tr> <td id="L241" class="blob-num js-line-number" data-line-number="241"></td> <td id="LC241" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> zeros, especially when eager execution is enabled.</span></td> </tr> <tr> <td id="L242" class="blob-num js-line-number" data-line-number="242"></td> <td id="LC242" class="blob-code blob-code-inner js-file-line"> state_size <span class="pl-k">=</span> <span class="pl-c1">self</span>.state_size</td> </tr> <tr> <td id="L243" class="blob-num js-line-number" data-line-number="243"></td> <td id="LC243" class="blob-code blob-code-inner js-file-line"> is_eager <span class="pl-k">=</span> context.in_eager_mode()</td> </tr> <tr> <td id="L244" class="blob-num js-line-number" data-line-number="244"></td> <td id="LC244" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> is_eager <span class="pl-k">and</span> <span class="pl-c1">hasattr</span>(<span class="pl-c1">self</span>, <span class="pl-s"><span class="pl-pds">&quot;</span>_last_zero_state<span class="pl-pds">&quot;</span></span>):</td> </tr> <tr> <td id="L245" class="blob-num js-line-number" data-line-number="245"></td> <td id="LC245" class="blob-code blob-code-inner js-file-line"> (last_state_size, last_batch_size, last_dtype,</td> </tr> <tr> <td id="L246" class="blob-num js-line-number" data-line-number="246"></td> <td id="LC246" class="blob-code blob-code-inner js-file-line"> last_output) <span class="pl-k">=</span> <span class="pl-c1">getattr</span>(<span class="pl-c1">self</span>, <span class="pl-s"><span class="pl-pds">&quot;</span>_last_zero_state<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L247" class="blob-num js-line-number" data-line-number="247"></td> <td id="LC247" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> (last_batch_size <span class="pl-k">==</span> batch_size <span class="pl-k">and</span></td> </tr> <tr> <td id="L248" class="blob-num js-line-number" data-line-number="248"></td> <td id="LC248" class="blob-code blob-code-inner js-file-line"> last_dtype <span class="pl-k">==</span> dtype <span class="pl-k">and</span></td> </tr> <tr> <td id="L249" class="blob-num js-line-number" data-line-number="249"></td> <td id="LC249" class="blob-code blob-code-inner js-file-line"> last_state_size <span class="pl-k">==</span> state_size):</td> </tr> <tr> <td id="L250" class="blob-num js-line-number" data-line-number="250"></td> <td id="LC250" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> last_output</td> </tr> <tr> <td id="L251" class="blob-num js-line-number" data-line-number="251"></td> <td id="LC251" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L252" class="blob-num js-line-number" data-line-number="252"></td> <td id="LC252" class="blob-code blob-code-inner js-file-line"> output <span class="pl-k">=</span> _zero_state_tensors(state_size, batch_size, dtype)</td> </tr> <tr> <td id="L253" class="blob-num js-line-number" data-line-number="253"></td> <td id="LC253" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> is_eager:</td> </tr> <tr> <td id="L254" class="blob-num js-line-number" data-line-number="254"></td> <td id="LC254" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._last_zero_state <span class="pl-k">=</span> (state_size, batch_size, dtype, output)</td> </tr> <tr> <td id="L255" class="blob-num js-line-number" data-line-number="255"></td> <td id="LC255" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> output</td> </tr> <tr> <td id="L256" class="blob-num js-line-number" data-line-number="256"></td> <td id="LC256" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L257" class="blob-num js-line-number" data-line-number="257"></td> <td id="LC257" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L258" class="blob-num js-line-number" data-line-number="258"></td> <td id="LC258" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">_LayerRNNCell</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L259" class="blob-num js-line-number" data-line-number="259"></td> <td id="LC259" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Subclass of RNNCells that act like proper `tf.Layer` objects.</span></td> </tr> <tr> <td id="L260" class="blob-num js-line-number" data-line-number="260"></td> <td id="LC260" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L261" class="blob-num js-line-number" data-line-number="261"></td> <td id="LC261" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> For backwards compatibility purposes, most `RNNCell` instances allow their</span></td> </tr> <tr> <td id="L262" class="blob-num js-line-number" data-line-number="262"></td> <td id="LC262" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `call` methods to instantiate variables via `tf.get_variable`. The underlying</span></td> </tr> <tr> <td id="L263" class="blob-num js-line-number" data-line-number="263"></td> <td id="LC263" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> variable scope thus keeps track of any variables, and returning cached</span></td> </tr> <tr> <td id="L264" class="blob-num js-line-number" data-line-number="264"></td> <td id="LC264" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> versions. This is atypical of `tf.layer` objects, which separate this</span></td> </tr> <tr> <td id="L265" class="blob-num js-line-number" data-line-number="265"></td> <td id="LC265" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> part of layer building into a `build` method that is only called once.</span></td> </tr> <tr> <td id="L266" class="blob-num js-line-number" data-line-number="266"></td> <td id="LC266" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L267" class="blob-num js-line-number" data-line-number="267"></td> <td id="LC267" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Here we provide a subclass for `RNNCell` objects that act exactly as</span></td> </tr> <tr> <td id="L268" class="blob-num js-line-number" data-line-number="268"></td> <td id="LC268" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `Layer` objects do. They must provide a `build` method and their</span></td> </tr> <tr> <td id="L269" class="blob-num js-line-number" data-line-number="269"></td> <td id="LC269" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `call` methods do not access Variables `tf.get_variable`.</span></td> </tr> <tr> <td id="L270" class="blob-num js-line-number" data-line-number="270"></td> <td id="LC270" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L271" class="blob-num js-line-number" data-line-number="271"></td> <td id="LC271" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L272" class="blob-num js-line-number" data-line-number="272"></td> <td id="LC272" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-k">*</span><span class="pl-smi">args</span>, <span class="pl-k">**</span><span class="pl-smi">kwargs</span>):</td> </tr> <tr> <td id="L273" class="blob-num js-line-number" data-line-number="273"></td> <td id="LC273" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run this RNN cell on inputs, starting from the given state.</span></td> </tr> <tr> <td id="L274" class="blob-num js-line-number" data-line-number="274"></td> <td id="LC274" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L275" class="blob-num js-line-number" data-line-number="275"></td> <td id="LC275" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L276" class="blob-num js-line-number" data-line-number="276"></td> <td id="LC276" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: `2-D` tensor with shape `[batch_size, input_size]`.</span></td> </tr> <tr> <td id="L277" class="blob-num js-line-number" data-line-number="277"></td> <td id="LC277" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: if `self.state_size` is an integer, this should be a `2-D Tensor`</span></td> </tr> <tr> <td id="L278" class="blob-num js-line-number" data-line-number="278"></td> <td id="LC278" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> with shape `[batch_size, self.state_size]`. Otherwise, if</span></td> </tr> <tr> <td id="L279" class="blob-num js-line-number" data-line-number="279"></td> <td id="LC279" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `self.state_size` is a tuple of integers, this should be a tuple</span></td> </tr> <tr> <td id="L280" class="blob-num js-line-number" data-line-number="280"></td> <td id="LC280" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> with shapes `[batch_size, s] for s in self.state_size`.</span></td> </tr> <tr> <td id="L281" class="blob-num js-line-number" data-line-number="281"></td> <td id="LC281" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> scope: optional cell scope.</span></td> </tr> <tr> <td id="L282" class="blob-num js-line-number" data-line-number="282"></td> <td id="LC282" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> *args: Additional positional arguments.</span></td> </tr> <tr> <td id="L283" class="blob-num js-line-number" data-line-number="283"></td> <td id="LC283" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> **kwargs: Additional keyword arguments.</span></td> </tr> <tr> <td id="L284" class="blob-num js-line-number" data-line-number="284"></td> <td id="LC284" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L285" class="blob-num js-line-number" data-line-number="285"></td> <td id="LC285" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L286" class="blob-num js-line-number" data-line-number="286"></td> <td id="LC286" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> A pair containing:</span></td> </tr> <tr> <td id="L287" class="blob-num js-line-number" data-line-number="287"></td> <td id="LC287" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L288" class="blob-num js-line-number" data-line-number="288"></td> <td id="LC288" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - Output: A `2-D` tensor with shape `[batch_size, self.output_size]`.</span></td> </tr> <tr> <td id="L289" class="blob-num js-line-number" data-line-number="289"></td> <td id="LC289" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - New state: Either a single `2-D` tensor, or a tuple of tensors matching</span></td> </tr> <tr> <td id="L290" class="blob-num js-line-number" data-line-number="290"></td> <td id="LC290" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the arity and shapes of `state`.</span></td> </tr> <tr> <td id="L291" class="blob-num js-line-number" data-line-number="291"></td> <td id="LC291" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L292" class="blob-num js-line-number" data-line-number="292"></td> <td id="LC292" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Bypass RNNCell&#39;s variable capturing semantics for LayerRNNCell.</span></td> </tr> <tr> <td id="L293" class="blob-num js-line-number" data-line-number="293"></td> <td id="LC293" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Instead, it is up to subclasses to provide a proper build</span></td> </tr> <tr> <td id="L294" class="blob-num js-line-number" data-line-number="294"></td> <td id="LC294" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> method. See the class docstring for more details.</span></td> </tr> <tr> <td id="L295" class="blob-num js-line-number" data-line-number="295"></td> <td id="LC295" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> base_layer.Layer.<span class="pl-c1">__call__</span>(<span class="pl-c1">self</span>, inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope,</td> </tr> <tr> <td id="L296" class="blob-num js-line-number" data-line-number="296"></td> <td id="LC296" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">*</span>args, <span class="pl-k">**</span>kwargs)</td> </tr> <tr> <td id="L297" class="blob-num js-line-number" data-line-number="297"></td> <td id="LC297" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L298" class="blob-num js-line-number" data-line-number="298"></td> <td id="LC298" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L299" class="blob-num js-line-number" data-line-number="299"></td> <td id="LC299" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.BasicRNNCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L300" class="blob-num js-line-number" data-line-number="300"></td> <td id="LC300" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">BasicRNNCell</span>(<span class="pl-e">_LayerRNNCell</span>):</td> </tr> <tr> <td id="L301" class="blob-num js-line-number" data-line-number="301"></td> <td id="LC301" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>The most basic RNN cell.</span></td> </tr> <tr> <td id="L302" class="blob-num js-line-number" data-line-number="302"></td> <td id="LC302" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L303" class="blob-num js-line-number" data-line-number="303"></td> <td id="LC303" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L304" class="blob-num js-line-number" data-line-number="304"></td> <td id="LC304" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units: int, The number of units in the RNN cell.</span></td> </tr> <tr> <td id="L305" class="blob-num js-line-number" data-line-number="305"></td> <td id="LC305" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> activation: Nonlinearity to use. Default: `tanh`.</span></td> </tr> <tr> <td id="L306" class="blob-num js-line-number" data-line-number="306"></td> <td id="LC306" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reuse: (optional) Python boolean describing whether to reuse variables</span></td> </tr> <tr> <td id="L307" class="blob-num js-line-number" data-line-number="307"></td> <td id="LC307" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in an existing scope. If not `True`, and the existing scope already has</span></td> </tr> <tr> <td id="L308" class="blob-num js-line-number" data-line-number="308"></td> <td id="LC308" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the given variables, an error is raised.</span></td> </tr> <tr> <td id="L309" class="blob-num js-line-number" data-line-number="309"></td> <td id="LC309" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> name: String, the name of the layer. Layers with the same name will</span></td> </tr> <tr> <td id="L310" class="blob-num js-line-number" data-line-number="310"></td> <td id="LC310" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> share weights, but to avoid mistakes we require reuse=True in such</span></td> </tr> <tr> <td id="L311" class="blob-num js-line-number" data-line-number="311"></td> <td id="LC311" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cases.</span></td> </tr> <tr> <td id="L312" class="blob-num js-line-number" data-line-number="312"></td> <td id="LC312" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L313" class="blob-num js-line-number" data-line-number="313"></td> <td id="LC313" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L314" class="blob-num js-line-number" data-line-number="314"></td> <td id="LC314" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">num_units</span>, <span class="pl-smi">activation</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">reuse</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">name</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L315" class="blob-num js-line-number" data-line-number="315"></td> <td id="LC315" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(BasicRNNCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>(<span class="pl-v">_reuse</span><span class="pl-k">=</span>reuse, <span class="pl-v">name</span><span class="pl-k">=</span>name)</td> </tr> <tr> <td id="L316" class="blob-num js-line-number" data-line-number="316"></td> <td id="LC316" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L317" class="blob-num js-line-number" data-line-number="317"></td> <td id="LC317" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Inputs must be 2-dimensional.</span></td> </tr> <tr> <td id="L318" class="blob-num js-line-number" data-line-number="318"></td> <td id="LC318" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.input_spec <span class="pl-k">=</span> base_layer.InputSpec(<span class="pl-v">ndim</span><span class="pl-k">=</span><span class="pl-c1">2</span>)</td> </tr> <tr> <td id="L319" class="blob-num js-line-number" data-line-number="319"></td> <td id="LC319" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L320" class="blob-num js-line-number" data-line-number="320"></td> <td id="LC320" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_units <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L321" class="blob-num js-line-number" data-line-number="321"></td> <td id="LC321" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation <span class="pl-k">=</span> activation <span class="pl-k">or</span> math_ops.tanh</td> </tr> <tr> <td id="L322" class="blob-num js-line-number" data-line-number="322"></td> <td id="LC322" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L323" class="blob-num js-line-number" data-line-number="323"></td> <td id="LC323" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L324" class="blob-num js-line-number" data-line-number="324"></td> <td id="LC324" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L325" class="blob-num js-line-number" data-line-number="325"></td> <td id="LC325" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L326" class="blob-num js-line-number" data-line-number="326"></td> <td id="LC326" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L327" class="blob-num js-line-number" data-line-number="327"></td> <td id="LC327" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L328" class="blob-num js-line-number" data-line-number="328"></td> <td id="LC328" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L329" class="blob-num js-line-number" data-line-number="329"></td> <td id="LC329" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L330" class="blob-num js-line-number" data-line-number="330"></td> <td id="LC330" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L331" class="blob-num js-line-number" data-line-number="331"></td> <td id="LC331" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs_shape</span>):</td> </tr> <tr> <td id="L332" class="blob-num js-line-number" data-line-number="332"></td> <td id="LC332" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> inputs_shape[<span class="pl-c1">1</span>].value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L333" class="blob-num js-line-number" data-line-number="333"></td> <td id="LC333" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Expected inputs.shape[-1] to be known, saw shape: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L334" class="blob-num js-line-number" data-line-number="334"></td> <td id="LC334" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> inputs_shape)</td> </tr> <tr> <td id="L335" class="blob-num js-line-number" data-line-number="335"></td> <td id="LC335" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L336" class="blob-num js-line-number" data-line-number="336"></td> <td id="LC336" class="blob-code blob-code-inner js-file-line"> input_depth <span class="pl-k">=</span> inputs_shape[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L337" class="blob-num js-line-number" data-line-number="337"></td> <td id="LC337" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L338" class="blob-num js-line-number" data-line-number="338"></td> <td id="LC338" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L339" class="blob-num js-line-number" data-line-number="339"></td> <td id="LC339" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> <span class="pl-c1">self</span>._num_units, <span class="pl-c1">self</span>._num_units])</td> </tr> <tr> <td id="L340" class="blob-num js-line-number" data-line-number="340"></td> <td id="LC340" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L341" class="blob-num js-line-number" data-line-number="341"></td> <td id="LC341" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L342" class="blob-num js-line-number" data-line-number="342"></td> <td id="LC342" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L343" class="blob-num js-line-number" data-line-number="343"></td> <td id="LC343" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>init_ops.zeros_initializer(<span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype))</td> </tr> <tr> <td id="L344" class="blob-num js-line-number" data-line-number="344"></td> <td id="LC344" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L345" class="blob-num js-line-number" data-line-number="345"></td> <td id="LC345" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.built <span class="pl-k">=</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L346" class="blob-num js-line-number" data-line-number="346"></td> <td id="LC346" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L347" class="blob-num js-line-number" data-line-number="347"></td> <td id="LC347" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L348" class="blob-num js-line-number" data-line-number="348"></td> <td id="LC348" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Most basic RNN: output = new_state = act(W * input + U * state + B).<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L349" class="blob-num js-line-number" data-line-number="349"></td> <td id="LC349" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L350" class="blob-num js-line-number" data-line-number="350"></td> <td id="LC350" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L351" class="blob-num js-line-number" data-line-number="351"></td> <td id="LC351" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, state], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._kernel)</td> </tr> <tr> <td id="L352" class="blob-num js-line-number" data-line-number="352"></td> <td id="LC352" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> nn_ops.bias_add(gate_inputs, <span class="pl-c1">self</span>._bias)</td> </tr> <tr> <td id="L353" class="blob-num js-line-number" data-line-number="353"></td> <td id="LC353" class="blob-code blob-code-inner js-file-line"> output <span class="pl-k">=</span> <span class="pl-c1">self</span>._activation(gate_inputs)</td> </tr> <tr> <td id="L354" class="blob-num js-line-number" data-line-number="354"></td> <td id="LC354" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> output, output</td> </tr> <tr> <td id="L355" class="blob-num js-line-number" data-line-number="355"></td> <td id="LC355" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L356" class="blob-num js-line-number" data-line-number="356"></td> <td id="LC356" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L357" class="blob-num js-line-number" data-line-number="357"></td> <td id="LC357" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.GRUCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L358" class="blob-num js-line-number" data-line-number="358"></td> <td id="LC358" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">GRUCell</span>(<span class="pl-e">_LayerRNNCell</span>):</td> </tr> <tr> <td id="L359" class="blob-num js-line-number" data-line-number="359"></td> <td id="LC359" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Gated Recurrent Unit cell (cf. http://arxiv.org/abs/1406.1078).</span></td> </tr> <tr> <td id="L360" class="blob-num js-line-number" data-line-number="360"></td> <td id="LC360" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L361" class="blob-num js-line-number" data-line-number="361"></td> <td id="LC361" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L362" class="blob-num js-line-number" data-line-number="362"></td> <td id="LC362" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units: int, The number of units in the GRU cell.</span></td> </tr> <tr> <td id="L363" class="blob-num js-line-number" data-line-number="363"></td> <td id="LC363" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> activation: Nonlinearity to use. Default: `tanh`.</span></td> </tr> <tr> <td id="L364" class="blob-num js-line-number" data-line-number="364"></td> <td id="LC364" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reuse: (optional) Python boolean describing whether to reuse variables</span></td> </tr> <tr> <td id="L365" class="blob-num js-line-number" data-line-number="365"></td> <td id="LC365" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in an existing scope. If not `True`, and the existing scope already has</span></td> </tr> <tr> <td id="L366" class="blob-num js-line-number" data-line-number="366"></td> <td id="LC366" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the given variables, an error is raised.</span></td> </tr> <tr> <td id="L367" class="blob-num js-line-number" data-line-number="367"></td> <td id="LC367" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> kernel_initializer: (optional) The initializer to use for the weight and</span></td> </tr> <tr> <td id="L368" class="blob-num js-line-number" data-line-number="368"></td> <td id="LC368" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> projection matrices.</span></td> </tr> <tr> <td id="L369" class="blob-num js-line-number" data-line-number="369"></td> <td id="LC369" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> bias_initializer: (optional) The initializer to use for the bias.</span></td> </tr> <tr> <td id="L370" class="blob-num js-line-number" data-line-number="370"></td> <td id="LC370" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> name: String, the name of the layer. Layers with the same name will</span></td> </tr> <tr> <td id="L371" class="blob-num js-line-number" data-line-number="371"></td> <td id="LC371" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> share weights, but to avoid mistakes we require reuse=True in such</span></td> </tr> <tr> <td id="L372" class="blob-num js-line-number" data-line-number="372"></td> <td id="LC372" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cases.</span></td> </tr> <tr> <td id="L373" class="blob-num js-line-number" data-line-number="373"></td> <td id="LC373" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L374" class="blob-num js-line-number" data-line-number="374"></td> <td id="LC374" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L375" class="blob-num js-line-number" data-line-number="375"></td> <td id="LC375" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>,</td> </tr> <tr> <td id="L376" class="blob-num js-line-number" data-line-number="376"></td> <td id="LC376" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">num_units</span>,</td> </tr> <tr> <td id="L377" class="blob-num js-line-number" data-line-number="377"></td> <td id="LC377" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">activation</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L378" class="blob-num js-line-number" data-line-number="378"></td> <td id="LC378" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">reuse</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L379" class="blob-num js-line-number" data-line-number="379"></td> <td id="LC379" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">kernel_initializer</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L380" class="blob-num js-line-number" data-line-number="380"></td> <td id="LC380" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">bias_initializer</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L381" class="blob-num js-line-number" data-line-number="381"></td> <td id="LC381" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">name</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L382" class="blob-num js-line-number" data-line-number="382"></td> <td id="LC382" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(GRUCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>(<span class="pl-v">_reuse</span><span class="pl-k">=</span>reuse, <span class="pl-v">name</span><span class="pl-k">=</span>name)</td> </tr> <tr> <td id="L383" class="blob-num js-line-number" data-line-number="383"></td> <td id="LC383" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L384" class="blob-num js-line-number" data-line-number="384"></td> <td id="LC384" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Inputs must be 2-dimensional.</span></td> </tr> <tr> <td id="L385" class="blob-num js-line-number" data-line-number="385"></td> <td id="LC385" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.input_spec <span class="pl-k">=</span> base_layer.InputSpec(<span class="pl-v">ndim</span><span class="pl-k">=</span><span class="pl-c1">2</span>)</td> </tr> <tr> <td id="L386" class="blob-num js-line-number" data-line-number="386"></td> <td id="LC386" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L387" class="blob-num js-line-number" data-line-number="387"></td> <td id="LC387" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_units <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L388" class="blob-num js-line-number" data-line-number="388"></td> <td id="LC388" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation <span class="pl-k">=</span> activation <span class="pl-k">or</span> math_ops.tanh</td> </tr> <tr> <td id="L389" class="blob-num js-line-number" data-line-number="389"></td> <td id="LC389" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._kernel_initializer <span class="pl-k">=</span> kernel_initializer</td> </tr> <tr> <td id="L390" class="blob-num js-line-number" data-line-number="390"></td> <td id="LC390" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias_initializer <span class="pl-k">=</span> bias_initializer</td> </tr> <tr> <td id="L391" class="blob-num js-line-number" data-line-number="391"></td> <td id="LC391" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L392" class="blob-num js-line-number" data-line-number="392"></td> <td id="LC392" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L393" class="blob-num js-line-number" data-line-number="393"></td> <td id="LC393" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L394" class="blob-num js-line-number" data-line-number="394"></td> <td id="LC394" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L395" class="blob-num js-line-number" data-line-number="395"></td> <td id="LC395" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L396" class="blob-num js-line-number" data-line-number="396"></td> <td id="LC396" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L397" class="blob-num js-line-number" data-line-number="397"></td> <td id="LC397" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L398" class="blob-num js-line-number" data-line-number="398"></td> <td id="LC398" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L399" class="blob-num js-line-number" data-line-number="399"></td> <td id="LC399" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L400" class="blob-num js-line-number" data-line-number="400"></td> <td id="LC400" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs_shape</span>):</td> </tr> <tr> <td id="L401" class="blob-num js-line-number" data-line-number="401"></td> <td id="LC401" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> inputs_shape[<span class="pl-c1">1</span>].value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L402" class="blob-num js-line-number" data-line-number="402"></td> <td id="LC402" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Expected inputs.shape[-1] to be known, saw shape: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L403" class="blob-num js-line-number" data-line-number="403"></td> <td id="LC403" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> inputs_shape)</td> </tr> <tr> <td id="L404" class="blob-num js-line-number" data-line-number="404"></td> <td id="LC404" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L405" class="blob-num js-line-number" data-line-number="405"></td> <td id="LC405" class="blob-code blob-code-inner js-file-line"> input_depth <span class="pl-k">=</span> inputs_shape[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L406" class="blob-num js-line-number" data-line-number="406"></td> <td id="LC406" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._gate_kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L407" class="blob-num js-line-number" data-line-number="407"></td> <td id="LC407" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>gates/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L408" class="blob-num js-line-number" data-line-number="408"></td> <td id="LC408" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> <span class="pl-c1">self</span>._num_units, <span class="pl-c1">2</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L409" class="blob-num js-line-number" data-line-number="409"></td> <td id="LC409" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._kernel_initializer)</td> </tr> <tr> <td id="L410" class="blob-num js-line-number" data-line-number="410"></td> <td id="LC410" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._gate_bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L411" class="blob-num js-line-number" data-line-number="411"></td> <td id="LC411" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>gates/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L412" class="blob-num js-line-number" data-line-number="412"></td> <td id="LC412" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">2</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L413" class="blob-num js-line-number" data-line-number="413"></td> <td id="LC413" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>(</td> </tr> <tr> <td id="L414" class="blob-num js-line-number" data-line-number="414"></td> <td id="LC414" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias_initializer</td> </tr> <tr> <td id="L415" class="blob-num js-line-number" data-line-number="415"></td> <td id="LC415" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._bias_initializer <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L416" class="blob-num js-line-number" data-line-number="416"></td> <td id="LC416" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span> init_ops.constant_initializer(<span class="pl-c1">1.0</span>, <span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype)))</td> </tr> <tr> <td id="L417" class="blob-num js-line-number" data-line-number="417"></td> <td id="LC417" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._candidate_kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L418" class="blob-num js-line-number" data-line-number="418"></td> <td id="LC418" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>candidate/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L419" class="blob-num js-line-number" data-line-number="419"></td> <td id="LC419" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> <span class="pl-c1">self</span>._num_units, <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L420" class="blob-num js-line-number" data-line-number="420"></td> <td id="LC420" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._kernel_initializer)</td> </tr> <tr> <td id="L421" class="blob-num js-line-number" data-line-number="421"></td> <td id="LC421" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._candidate_bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L422" class="blob-num js-line-number" data-line-number="422"></td> <td id="LC422" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>candidate/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L423" class="blob-num js-line-number" data-line-number="423"></td> <td id="LC423" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L424" class="blob-num js-line-number" data-line-number="424"></td> <td id="LC424" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>(</td> </tr> <tr> <td id="L425" class="blob-num js-line-number" data-line-number="425"></td> <td id="LC425" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias_initializer</td> </tr> <tr> <td id="L426" class="blob-num js-line-number" data-line-number="426"></td> <td id="LC426" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._bias_initializer <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L427" class="blob-num js-line-number" data-line-number="427"></td> <td id="LC427" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span> init_ops.zeros_initializer(<span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype)))</td> </tr> <tr> <td id="L428" class="blob-num js-line-number" data-line-number="428"></td> <td id="LC428" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L429" class="blob-num js-line-number" data-line-number="429"></td> <td id="LC429" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.built <span class="pl-k">=</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L430" class="blob-num js-line-number" data-line-number="430"></td> <td id="LC430" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L431" class="blob-num js-line-number" data-line-number="431"></td> <td id="LC431" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L432" class="blob-num js-line-number" data-line-number="432"></td> <td id="LC432" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Gated recurrent unit (GRU) with nunits cells.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L433" class="blob-num js-line-number" data-line-number="433"></td> <td id="LC433" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L434" class="blob-num js-line-number" data-line-number="434"></td> <td id="LC434" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L435" class="blob-num js-line-number" data-line-number="435"></td> <td id="LC435" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, state], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._gate_kernel)</td> </tr> <tr> <td id="L436" class="blob-num js-line-number" data-line-number="436"></td> <td id="LC436" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> nn_ops.bias_add(gate_inputs, <span class="pl-c1">self</span>._gate_bias)</td> </tr> <tr> <td id="L437" class="blob-num js-line-number" data-line-number="437"></td> <td id="LC437" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L438" class="blob-num js-line-number" data-line-number="438"></td> <td id="LC438" class="blob-code blob-code-inner js-file-line"> value <span class="pl-k">=</span> math_ops.sigmoid(gate_inputs)</td> </tr> <tr> <td id="L439" class="blob-num js-line-number" data-line-number="439"></td> <td id="LC439" class="blob-code blob-code-inner js-file-line"> r, u <span class="pl-k">=</span> array_ops.split(<span class="pl-v">value</span><span class="pl-k">=</span>value, <span class="pl-v">num_or_size_splits</span><span class="pl-k">=</span><span class="pl-c1">2</span>, <span class="pl-v">axis</span><span class="pl-k">=</span><span class="pl-c1">1</span>)</td> </tr> <tr> <td id="L440" class="blob-num js-line-number" data-line-number="440"></td> <td id="LC440" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L441" class="blob-num js-line-number" data-line-number="441"></td> <td id="LC441" class="blob-code blob-code-inner js-file-line"> r_state <span class="pl-k">=</span> r <span class="pl-k">*</span> state</td> </tr> <tr> <td id="L442" class="blob-num js-line-number" data-line-number="442"></td> <td id="LC442" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L443" class="blob-num js-line-number" data-line-number="443"></td> <td id="LC443" class="blob-code blob-code-inner js-file-line"> candidate <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L444" class="blob-num js-line-number" data-line-number="444"></td> <td id="LC444" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, r_state], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._candidate_kernel)</td> </tr> <tr> <td id="L445" class="blob-num js-line-number" data-line-number="445"></td> <td id="LC445" class="blob-code blob-code-inner js-file-line"> candidate <span class="pl-k">=</span> nn_ops.bias_add(candidate, <span class="pl-c1">self</span>._candidate_bias)</td> </tr> <tr> <td id="L446" class="blob-num js-line-number" data-line-number="446"></td> <td id="LC446" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L447" class="blob-num js-line-number" data-line-number="447"></td> <td id="LC447" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> <span class="pl-c1">self</span>._activation(candidate)</td> </tr> <tr> <td id="L448" class="blob-num js-line-number" data-line-number="448"></td> <td id="LC448" class="blob-code blob-code-inner js-file-line"> new_h <span class="pl-k">=</span> u <span class="pl-k">*</span> state <span class="pl-k">+</span> (<span class="pl-c1">1</span> <span class="pl-k">-</span> u) <span class="pl-k">*</span> c</td> </tr> <tr> <td id="L449" class="blob-num js-line-number" data-line-number="449"></td> <td id="LC449" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> new_h, new_h</td> </tr> <tr> <td id="L450" class="blob-num js-line-number" data-line-number="450"></td> <td id="LC450" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L451" class="blob-num js-line-number" data-line-number="451"></td> <td id="LC451" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L452" class="blob-num js-line-number" data-line-number="452"></td> <td id="LC452" class="blob-code blob-code-inner js-file-line">_LSTMStateTuple <span class="pl-k">=</span> collections.namedtuple(<span class="pl-s"><span class="pl-pds">&quot;</span>LSTMStateTuple<span class="pl-pds">&quot;</span></span>, (<span class="pl-s"><span class="pl-pds">&quot;</span>c<span class="pl-pds">&quot;</span></span>, <span class="pl-s"><span class="pl-pds">&quot;</span>h<span class="pl-pds">&quot;</span></span>))</td> </tr> <tr> <td id="L453" class="blob-num js-line-number" data-line-number="453"></td> <td id="LC453" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L454" class="blob-num js-line-number" data-line-number="454"></td> <td id="LC454" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L455" class="blob-num js-line-number" data-line-number="455"></td> <td id="LC455" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.LSTMStateTuple<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L456" class="blob-num js-line-number" data-line-number="456"></td> <td id="LC456" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">LSTMStateTuple</span>(<span class="pl-e">_LSTMStateTuple</span>):</td> </tr> <tr> <td id="L457" class="blob-num js-line-number" data-line-number="457"></td> <td id="LC457" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Tuple used by LSTM Cells for `state_size`, `zero_state`, and output state.</span></td> </tr> <tr> <td id="L458" class="blob-num js-line-number" data-line-number="458"></td> <td id="LC458" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L459" class="blob-num js-line-number" data-line-number="459"></td> <td id="LC459" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Stores two elements: `(c, h)`, in that order. Where `c` is the hidden state</span></td> </tr> <tr> <td id="L460" class="blob-num js-line-number" data-line-number="460"></td> <td id="LC460" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> and `h` is the output.</span></td> </tr> <tr> <td id="L461" class="blob-num js-line-number" data-line-number="461"></td> <td id="LC461" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L462" class="blob-num js-line-number" data-line-number="462"></td> <td id="LC462" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Only used when `state_is_tuple=True`.</span></td> </tr> <tr> <td id="L463" class="blob-num js-line-number" data-line-number="463"></td> <td id="LC463" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L464" class="blob-num js-line-number" data-line-number="464"></td> <td id="LC464" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">__slots__</span> <span class="pl-k">=</span> ()</td> </tr> <tr> <td id="L465" class="blob-num js-line-number" data-line-number="465"></td> <td id="LC465" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L466" class="blob-num js-line-number" data-line-number="466"></td> <td id="LC466" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L467" class="blob-num js-line-number" data-line-number="467"></td> <td id="LC467" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">dtype</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L468" class="blob-num js-line-number" data-line-number="468"></td> <td id="LC468" class="blob-code blob-code-inner js-file-line"> (c, h) <span class="pl-k">=</span> <span class="pl-c1">self</span></td> </tr> <tr> <td id="L469" class="blob-num js-line-number" data-line-number="469"></td> <td id="LC469" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> c.dtype <span class="pl-k">!=</span> h.dtype:</td> </tr> <tr> <td id="L470" class="blob-num js-line-number" data-line-number="470"></td> <td id="LC470" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Inconsistent internal state: <span class="pl-c1">%s</span> vs <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span></td> </tr> <tr> <td id="L471" class="blob-num js-line-number" data-line-number="471"></td> <td id="LC471" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">str</span>(c.dtype), <span class="pl-c1">str</span>(h.dtype)))</td> </tr> <tr> <td id="L472" class="blob-num js-line-number" data-line-number="472"></td> <td id="LC472" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> c.dtype</td> </tr> <tr> <td id="L473" class="blob-num js-line-number" data-line-number="473"></td> <td id="LC473" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L474" class="blob-num js-line-number" data-line-number="474"></td> <td id="LC474" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L475" class="blob-num js-line-number" data-line-number="475"></td> <td id="LC475" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.BasicLSTMCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L476" class="blob-num js-line-number" data-line-number="476"></td> <td id="LC476" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">BasicLSTMCell</span>(<span class="pl-e">_LayerRNNCell</span>):</td> </tr> <tr> <td id="L477" class="blob-num js-line-number" data-line-number="477"></td> <td id="LC477" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Basic LSTM recurrent network cell.</span></td> </tr> <tr> <td id="L478" class="blob-num js-line-number" data-line-number="478"></td> <td id="LC478" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L479" class="blob-num js-line-number" data-line-number="479"></td> <td id="LC479" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> The implementation is based on: http://arxiv.org/abs/1409.2329.</span></td> </tr> <tr> <td id="L480" class="blob-num js-line-number" data-line-number="480"></td> <td id="LC480" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L481" class="blob-num js-line-number" data-line-number="481"></td> <td id="LC481" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> We add forget_bias (default: 1) to the biases of the forget gate in order to</span></td> </tr> <tr> <td id="L482" class="blob-num js-line-number" data-line-number="482"></td> <td id="LC482" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reduce the scale of forgetting in the beginning of the training.</span></td> </tr> <tr> <td id="L483" class="blob-num js-line-number" data-line-number="483"></td> <td id="LC483" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L484" class="blob-num js-line-number" data-line-number="484"></td> <td id="LC484" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> It does not allow cell clipping, a projection layer, and does not</span></td> </tr> <tr> <td id="L485" class="blob-num js-line-number" data-line-number="485"></td> <td id="LC485" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> use peep-hole connections: it is the basic baseline.</span></td> </tr> <tr> <td id="L486" class="blob-num js-line-number" data-line-number="486"></td> <td id="LC486" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L487" class="blob-num js-line-number" data-line-number="487"></td> <td id="LC487" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> For advanced models, please use the full @{tf.nn.rnn_cell.LSTMCell}</span></td> </tr> <tr> <td id="L488" class="blob-num js-line-number" data-line-number="488"></td> <td id="LC488" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> that follows.</span></td> </tr> <tr> <td id="L489" class="blob-num js-line-number" data-line-number="489"></td> <td id="LC489" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L490" class="blob-num js-line-number" data-line-number="490"></td> <td id="LC490" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L491" class="blob-num js-line-number" data-line-number="491"></td> <td id="LC491" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">num_units</span>, <span class="pl-smi">forget_bias</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>,</td> </tr> <tr> <td id="L492" class="blob-num js-line-number" data-line-number="492"></td> <td id="LC492" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">state_is_tuple</span><span class="pl-k">=</span><span class="pl-c1">True</span>, <span class="pl-smi">activation</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">reuse</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">name</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L493" class="blob-num js-line-number" data-line-number="493"></td> <td id="LC493" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Initialize the basic LSTM cell.</span></td> </tr> <tr> <td id="L494" class="blob-num js-line-number" data-line-number="494"></td> <td id="LC494" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L495" class="blob-num js-line-number" data-line-number="495"></td> <td id="LC495" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L496" class="blob-num js-line-number" data-line-number="496"></td> <td id="LC496" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units: int, The number of units in the LSTM cell.</span></td> </tr> <tr> <td id="L497" class="blob-num js-line-number" data-line-number="497"></td> <td id="LC497" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> forget_bias: float, The bias added to forget gates (see above).</span></td> </tr> <tr> <td id="L498" class="blob-num js-line-number" data-line-number="498"></td> <td id="LC498" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Must set to `0.0` manually when restoring from CudnnLSTM-trained</span></td> </tr> <tr> <td id="L499" class="blob-num js-line-number" data-line-number="499"></td> <td id="LC499" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> checkpoints.</span></td> </tr> <tr> <td id="L500" class="blob-num js-line-number" data-line-number="500"></td> <td id="LC500" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state_is_tuple: If True, accepted and returned states are 2-tuples of</span></td> </tr> <tr> <td id="L501" class="blob-num js-line-number" data-line-number="501"></td> <td id="LC501" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the `c_state` and `m_state`. If False, they are concatenated</span></td> </tr> <tr> <td id="L502" class="blob-num js-line-number" data-line-number="502"></td> <td id="LC502" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> along the column axis. The latter behavior will soon be deprecated.</span></td> </tr> <tr> <td id="L503" class="blob-num js-line-number" data-line-number="503"></td> <td id="LC503" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> activation: Activation function of the inner states. Default: `tanh`.</span></td> </tr> <tr> <td id="L504" class="blob-num js-line-number" data-line-number="504"></td> <td id="LC504" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reuse: (optional) Python boolean describing whether to reuse variables</span></td> </tr> <tr> <td id="L505" class="blob-num js-line-number" data-line-number="505"></td> <td id="LC505" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in an existing scope. If not `True`, and the existing scope already has</span></td> </tr> <tr> <td id="L506" class="blob-num js-line-number" data-line-number="506"></td> <td id="LC506" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the given variables, an error is raised.</span></td> </tr> <tr> <td id="L507" class="blob-num js-line-number" data-line-number="507"></td> <td id="LC507" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> name: String, the name of the layer. Layers with the same name will</span></td> </tr> <tr> <td id="L508" class="blob-num js-line-number" data-line-number="508"></td> <td id="LC508" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> share weights, but to avoid mistakes we require reuse=True in such</span></td> </tr> <tr> <td id="L509" class="blob-num js-line-number" data-line-number="509"></td> <td id="LC509" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cases.</span></td> </tr> <tr> <td id="L510" class="blob-num js-line-number" data-line-number="510"></td> <td id="LC510" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L511" class="blob-num js-line-number" data-line-number="511"></td> <td id="LC511" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> When restoring from CudnnLSTM-trained checkpoints, must use</span></td> </tr> <tr> <td id="L512" class="blob-num js-line-number" data-line-number="512"></td> <td id="LC512" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `CudnnCompatibleLSTMCell` instead.</span></td> </tr> <tr> <td id="L513" class="blob-num js-line-number" data-line-number="513"></td> <td id="LC513" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L514" class="blob-num js-line-number" data-line-number="514"></td> <td id="LC514" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(BasicLSTMCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>(<span class="pl-v">_reuse</span><span class="pl-k">=</span>reuse, <span class="pl-v">name</span><span class="pl-k">=</span>name)</td> </tr> <tr> <td id="L515" class="blob-num js-line-number" data-line-number="515"></td> <td id="LC515" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> state_is_tuple:</td> </tr> <tr> <td id="L516" class="blob-num js-line-number" data-line-number="516"></td> <td id="LC516" class="blob-code blob-code-inner js-file-line"> logging.warn(<span class="pl-s"><span class="pl-pds">&quot;</span><span class="pl-c1">%s</span>: Using a concatenated state is slower and will soon be <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L517" class="blob-num js-line-number" data-line-number="517"></td> <td id="LC517" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>deprecated. Use state_is_tuple=True.<span class="pl-pds">&quot;</span></span>, <span class="pl-c1">self</span>)</td> </tr> <tr> <td id="L518" class="blob-num js-line-number" data-line-number="518"></td> <td id="LC518" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L519" class="blob-num js-line-number" data-line-number="519"></td> <td id="LC519" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Inputs must be 2-dimensional.</span></td> </tr> <tr> <td id="L520" class="blob-num js-line-number" data-line-number="520"></td> <td id="LC520" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.input_spec <span class="pl-k">=</span> base_layer.InputSpec(<span class="pl-v">ndim</span><span class="pl-k">=</span><span class="pl-c1">2</span>)</td> </tr> <tr> <td id="L521" class="blob-num js-line-number" data-line-number="521"></td> <td id="LC521" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L522" class="blob-num js-line-number" data-line-number="522"></td> <td id="LC522" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_units <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L523" class="blob-num js-line-number" data-line-number="523"></td> <td id="LC523" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._forget_bias <span class="pl-k">=</span> forget_bias</td> </tr> <tr> <td id="L524" class="blob-num js-line-number" data-line-number="524"></td> <td id="LC524" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">=</span> state_is_tuple</td> </tr> <tr> <td id="L525" class="blob-num js-line-number" data-line-number="525"></td> <td id="LC525" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation <span class="pl-k">=</span> activation <span class="pl-k">or</span> math_ops.tanh</td> </tr> <tr> <td id="L526" class="blob-num js-line-number" data-line-number="526"></td> <td id="LC526" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L527" class="blob-num js-line-number" data-line-number="527"></td> <td id="LC527" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L528" class="blob-num js-line-number" data-line-number="528"></td> <td id="LC528" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L529" class="blob-num js-line-number" data-line-number="529"></td> <td id="LC529" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> (LSTMStateTuple(<span class="pl-c1">self</span>._num_units, <span class="pl-c1">self</span>._num_units)</td> </tr> <tr> <td id="L530" class="blob-num js-line-number" data-line-number="530"></td> <td id="LC530" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">else</span> <span class="pl-c1">2</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units)</td> </tr> <tr> <td id="L531" class="blob-num js-line-number" data-line-number="531"></td> <td id="LC531" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L532" class="blob-num js-line-number" data-line-number="532"></td> <td id="LC532" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L533" class="blob-num js-line-number" data-line-number="533"></td> <td id="LC533" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L534" class="blob-num js-line-number" data-line-number="534"></td> <td id="LC534" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L535" class="blob-num js-line-number" data-line-number="535"></td> <td id="LC535" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L536" class="blob-num js-line-number" data-line-number="536"></td> <td id="LC536" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs_shape</span>):</td> </tr> <tr> <td id="L537" class="blob-num js-line-number" data-line-number="537"></td> <td id="LC537" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> inputs_shape[<span class="pl-c1">1</span>].value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L538" class="blob-num js-line-number" data-line-number="538"></td> <td id="LC538" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Expected inputs.shape[-1] to be known, saw shape: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L539" class="blob-num js-line-number" data-line-number="539"></td> <td id="LC539" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> inputs_shape)</td> </tr> <tr> <td id="L540" class="blob-num js-line-number" data-line-number="540"></td> <td id="LC540" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L541" class="blob-num js-line-number" data-line-number="541"></td> <td id="LC541" class="blob-code blob-code-inner js-file-line"> input_depth <span class="pl-k">=</span> inputs_shape[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L542" class="blob-num js-line-number" data-line-number="542"></td> <td id="LC542" class="blob-code blob-code-inner js-file-line"> h_depth <span class="pl-k">=</span> <span class="pl-c1">self</span>._num_units</td> </tr> <tr> <td id="L543" class="blob-num js-line-number" data-line-number="543"></td> <td id="LC543" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L544" class="blob-num js-line-number" data-line-number="544"></td> <td id="LC544" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L545" class="blob-num js-line-number" data-line-number="545"></td> <td id="LC545" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> h_depth, <span class="pl-c1">4</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units])</td> </tr> <tr> <td id="L546" class="blob-num js-line-number" data-line-number="546"></td> <td id="LC546" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L547" class="blob-num js-line-number" data-line-number="547"></td> <td id="LC547" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L548" class="blob-num js-line-number" data-line-number="548"></td> <td id="LC548" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">4</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L549" class="blob-num js-line-number" data-line-number="549"></td> <td id="LC549" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>init_ops.zeros_initializer(<span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype))</td> </tr> <tr> <td id="L550" class="blob-num js-line-number" data-line-number="550"></td> <td id="LC550" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L551" class="blob-num js-line-number" data-line-number="551"></td> <td id="LC551" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.built <span class="pl-k">=</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L552" class="blob-num js-line-number" data-line-number="552"></td> <td id="LC552" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L553" class="blob-num js-line-number" data-line-number="553"></td> <td id="LC553" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L554" class="blob-num js-line-number" data-line-number="554"></td> <td id="LC554" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Long short-term memory cell (LSTM).</span></td> </tr> <tr> <td id="L555" class="blob-num js-line-number" data-line-number="555"></td> <td id="LC555" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L556" class="blob-num js-line-number" data-line-number="556"></td> <td id="LC556" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L557" class="blob-num js-line-number" data-line-number="557"></td> <td id="LC557" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: `2-D` tensor with shape `[batch_size, input_size]`.</span></td> </tr> <tr> <td id="L558" class="blob-num js-line-number" data-line-number="558"></td> <td id="LC558" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: An `LSTMStateTuple` of state tensors, each shaped</span></td> </tr> <tr> <td id="L559" class="blob-num js-line-number" data-line-number="559"></td> <td id="LC559" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `[batch_size, self.state_size]`, if `state_is_tuple` has been set to</span></td> </tr> <tr> <td id="L560" class="blob-num js-line-number" data-line-number="560"></td> <td id="LC560" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `True`. Otherwise, a `Tensor` shaped</span></td> </tr> <tr> <td id="L561" class="blob-num js-line-number" data-line-number="561"></td> <td id="LC561" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `[batch_size, 2 * self.state_size]`.</span></td> </tr> <tr> <td id="L562" class="blob-num js-line-number" data-line-number="562"></td> <td id="LC562" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L563" class="blob-num js-line-number" data-line-number="563"></td> <td id="LC563" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L564" class="blob-num js-line-number" data-line-number="564"></td> <td id="LC564" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> A pair containing the new hidden state, and the new state (either a</span></td> </tr> <tr> <td id="L565" class="blob-num js-line-number" data-line-number="565"></td> <td id="LC565" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `LSTMStateTuple` or a concatenated state, depending on</span></td> </tr> <tr> <td id="L566" class="blob-num js-line-number" data-line-number="566"></td> <td id="LC566" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `state_is_tuple`).</span></td> </tr> <tr> <td id="L567" class="blob-num js-line-number" data-line-number="567"></td> <td id="LC567" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L568" class="blob-num js-line-number" data-line-number="568"></td> <td id="LC568" class="blob-code blob-code-inner js-file-line"> sigmoid <span class="pl-k">=</span> math_ops.sigmoid</td> </tr> <tr> <td id="L569" class="blob-num js-line-number" data-line-number="569"></td> <td id="LC569" class="blob-code blob-code-inner js-file-line"> one <span class="pl-k">=</span> constant_op.constant(<span class="pl-c1">1</span>, <span class="pl-v">dtype</span><span class="pl-k">=</span>dtypes.int32)</td> </tr> <tr> <td id="L570" class="blob-num js-line-number" data-line-number="570"></td> <td id="LC570" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Parameters of gates are concatenated into one multiply for efficiency.</span></td> </tr> <tr> <td id="L571" class="blob-num js-line-number" data-line-number="571"></td> <td id="LC571" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L572" class="blob-num js-line-number" data-line-number="572"></td> <td id="LC572" class="blob-code blob-code-inner js-file-line"> c, h <span class="pl-k">=</span> state</td> </tr> <tr> <td id="L573" class="blob-num js-line-number" data-line-number="573"></td> <td id="LC573" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L574" class="blob-num js-line-number" data-line-number="574"></td> <td id="LC574" class="blob-code blob-code-inner js-file-line"> c, h <span class="pl-k">=</span> array_ops.split(<span class="pl-v">value</span><span class="pl-k">=</span>state, <span class="pl-v">num_or_size_splits</span><span class="pl-k">=</span><span class="pl-c1">2</span>, <span class="pl-v">axis</span><span class="pl-k">=</span>one)</td> </tr> <tr> <td id="L575" class="blob-num js-line-number" data-line-number="575"></td> <td id="LC575" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L576" class="blob-num js-line-number" data-line-number="576"></td> <td id="LC576" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L577" class="blob-num js-line-number" data-line-number="577"></td> <td id="LC577" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, h], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._kernel)</td> </tr> <tr> <td id="L578" class="blob-num js-line-number" data-line-number="578"></td> <td id="LC578" class="blob-code blob-code-inner js-file-line"> gate_inputs <span class="pl-k">=</span> nn_ops.bias_add(gate_inputs, <span class="pl-c1">self</span>._bias)</td> </tr> <tr> <td id="L579" class="blob-num js-line-number" data-line-number="579"></td> <td id="LC579" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L580" class="blob-num js-line-number" data-line-number="580"></td> <td id="LC580" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> i = input_gate, j = new_input, f = forget_gate, o = output_gate</span></td> </tr> <tr> <td id="L581" class="blob-num js-line-number" data-line-number="581"></td> <td id="LC581" class="blob-code blob-code-inner js-file-line"> i, j, f, o <span class="pl-k">=</span> array_ops.split(</td> </tr> <tr> <td id="L582" class="blob-num js-line-number" data-line-number="582"></td> <td id="LC582" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">value</span><span class="pl-k">=</span>gate_inputs, <span class="pl-v">num_or_size_splits</span><span class="pl-k">=</span><span class="pl-c1">4</span>, <span class="pl-v">axis</span><span class="pl-k">=</span>one)</td> </tr> <tr> <td id="L583" class="blob-num js-line-number" data-line-number="583"></td> <td id="LC583" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L584" class="blob-num js-line-number" data-line-number="584"></td> <td id="LC584" class="blob-code blob-code-inner js-file-line"> forget_bias_tensor <span class="pl-k">=</span> constant_op.constant(<span class="pl-c1">self</span>._forget_bias, <span class="pl-v">dtype</span><span class="pl-k">=</span>f.dtype)</td> </tr> <tr> <td id="L585" class="blob-num js-line-number" data-line-number="585"></td> <td id="LC585" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Note that using `add` and `multiply` instead of `+` and `*` gives a</span></td> </tr> <tr> <td id="L586" class="blob-num js-line-number" data-line-number="586"></td> <td id="LC586" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> performance improvement. So using those at the cost of readability.</span></td> </tr> <tr> <td id="L587" class="blob-num js-line-number" data-line-number="587"></td> <td id="LC587" class="blob-code blob-code-inner js-file-line"> add <span class="pl-k">=</span> math_ops.add</td> </tr> <tr> <td id="L588" class="blob-num js-line-number" data-line-number="588"></td> <td id="LC588" class="blob-code blob-code-inner js-file-line"> multiply <span class="pl-k">=</span> math_ops.multiply</td> </tr> <tr> <td id="L589" class="blob-num js-line-number" data-line-number="589"></td> <td id="LC589" class="blob-code blob-code-inner js-file-line"> new_c <span class="pl-k">=</span> add(multiply(c, sigmoid(add(f, forget_bias_tensor))),</td> </tr> <tr> <td id="L590" class="blob-num js-line-number" data-line-number="590"></td> <td id="LC590" class="blob-code blob-code-inner js-file-line"> multiply(sigmoid(i), <span class="pl-c1">self</span>._activation(j)))</td> </tr> <tr> <td id="L591" class="blob-num js-line-number" data-line-number="591"></td> <td id="LC591" class="blob-code blob-code-inner js-file-line"> new_h <span class="pl-k">=</span> multiply(<span class="pl-c1">self</span>._activation(new_c), sigmoid(o))</td> </tr> <tr> <td id="L592" class="blob-num js-line-number" data-line-number="592"></td> <td id="LC592" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L593" class="blob-num js-line-number" data-line-number="593"></td> <td id="LC593" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L594" class="blob-num js-line-number" data-line-number="594"></td> <td id="LC594" class="blob-code blob-code-inner js-file-line"> new_state <span class="pl-k">=</span> LSTMStateTuple(new_c, new_h)</td> </tr> <tr> <td id="L595" class="blob-num js-line-number" data-line-number="595"></td> <td id="LC595" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L596" class="blob-num js-line-number" data-line-number="596"></td> <td id="LC596" class="blob-code blob-code-inner js-file-line"> new_state <span class="pl-k">=</span> array_ops.concat([new_c, new_h], <span class="pl-c1">1</span>)</td> </tr> <tr> <td id="L597" class="blob-num js-line-number" data-line-number="597"></td> <td id="LC597" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> new_h, new_state</td> </tr> <tr> <td id="L598" class="blob-num js-line-number" data-line-number="598"></td> <td id="LC598" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L599" class="blob-num js-line-number" data-line-number="599"></td> <td id="LC599" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L600" class="blob-num js-line-number" data-line-number="600"></td> <td id="LC600" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.LSTMCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L601" class="blob-num js-line-number" data-line-number="601"></td> <td id="LC601" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">LSTMCell</span>(<span class="pl-e">_LayerRNNCell</span>):</td> </tr> <tr> <td id="L602" class="blob-num js-line-number" data-line-number="602"></td> <td id="LC602" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Long short-term memory unit (LSTM) recurrent network cell.</span></td> </tr> <tr> <td id="L603" class="blob-num js-line-number" data-line-number="603"></td> <td id="LC603" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L604" class="blob-num js-line-number" data-line-number="604"></td> <td id="LC604" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> The default non-peephole implementation is based on:</span></td> </tr> <tr> <td id="L605" class="blob-num js-line-number" data-line-number="605"></td> <td id="LC605" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L606" class="blob-num js-line-number" data-line-number="606"></td> <td id="LC606" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> http://www.bioinf.jku.at/publications/older/2604.pdf</span></td> </tr> <tr> <td id="L607" class="blob-num js-line-number" data-line-number="607"></td> <td id="LC607" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L608" class="blob-num js-line-number" data-line-number="608"></td> <td id="LC608" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> S. Hochreiter and J. Schmidhuber.</span></td> </tr> <tr> <td id="L609" class="blob-num js-line-number" data-line-number="609"></td> <td id="LC609" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> &quot;Long Short-Term Memory&quot;. Neural Computation, 9(8):1735-1780, 1997.</span></td> </tr> <tr> <td id="L610" class="blob-num js-line-number" data-line-number="610"></td> <td id="LC610" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L611" class="blob-num js-line-number" data-line-number="611"></td> <td id="LC611" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> The peephole implementation is based on:</span></td> </tr> <tr> <td id="L612" class="blob-num js-line-number" data-line-number="612"></td> <td id="LC612" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L613" class="blob-num js-line-number" data-line-number="613"></td> <td id="LC613" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> https://research.google.com/pubs/archive/43905.pdf</span></td> </tr> <tr> <td id="L614" class="blob-num js-line-number" data-line-number="614"></td> <td id="LC614" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L615" class="blob-num js-line-number" data-line-number="615"></td> <td id="LC615" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Hasim Sak, Andrew Senior, and Francoise Beaufays.</span></td> </tr> <tr> <td id="L616" class="blob-num js-line-number" data-line-number="616"></td> <td id="LC616" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> &quot;Long short-term memory recurrent neural network architectures for</span></td> </tr> <tr> <td id="L617" class="blob-num js-line-number" data-line-number="617"></td> <td id="LC617" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> large scale acoustic modeling.&quot; INTERSPEECH, 2014.</span></td> </tr> <tr> <td id="L618" class="blob-num js-line-number" data-line-number="618"></td> <td id="LC618" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L619" class="blob-num js-line-number" data-line-number="619"></td> <td id="LC619" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> The class uses optional peep-hole connections, optional cell clipping, and</span></td> </tr> <tr> <td id="L620" class="blob-num js-line-number" data-line-number="620"></td> <td id="LC620" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> an optional projection layer.</span></td> </tr> <tr> <td id="L621" class="blob-num js-line-number" data-line-number="621"></td> <td id="LC621" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L622" class="blob-num js-line-number" data-line-number="622"></td> <td id="LC622" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L623" class="blob-num js-line-number" data-line-number="623"></td> <td id="LC623" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">num_units</span>,</td> </tr> <tr> <td id="L624" class="blob-num js-line-number" data-line-number="624"></td> <td id="LC624" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">use_peepholes</span><span class="pl-k">=</span><span class="pl-c1">False</span>, <span class="pl-smi">cell_clip</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L625" class="blob-num js-line-number" data-line-number="625"></td> <td id="LC625" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">initializer</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">num_proj</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">proj_clip</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L626" class="blob-num js-line-number" data-line-number="626"></td> <td id="LC626" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">num_unit_shards</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">num_proj_shards</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L627" class="blob-num js-line-number" data-line-number="627"></td> <td id="LC627" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">forget_bias</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>, <span class="pl-smi">state_is_tuple</span><span class="pl-k">=</span><span class="pl-c1">True</span>,</td> </tr> <tr> <td id="L628" class="blob-num js-line-number" data-line-number="628"></td> <td id="LC628" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">activation</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">reuse</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">name</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L629" class="blob-num js-line-number" data-line-number="629"></td> <td id="LC629" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Initialize the parameters for an LSTM cell.</span></td> </tr> <tr> <td id="L630" class="blob-num js-line-number" data-line-number="630"></td> <td id="LC630" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L631" class="blob-num js-line-number" data-line-number="631"></td> <td id="LC631" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L632" class="blob-num js-line-number" data-line-number="632"></td> <td id="LC632" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units: int, The number of units in the LSTM cell.</span></td> </tr> <tr> <td id="L633" class="blob-num js-line-number" data-line-number="633"></td> <td id="LC633" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> use_peepholes: bool, set True to enable diagonal/peephole connections.</span></td> </tr> <tr> <td id="L634" class="blob-num js-line-number" data-line-number="634"></td> <td id="LC634" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell_clip: (optional) A float value, if provided the cell state is clipped</span></td> </tr> <tr> <td id="L635" class="blob-num js-line-number" data-line-number="635"></td> <td id="LC635" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> by this value prior to the cell output activation.</span></td> </tr> <tr> <td id="L636" class="blob-num js-line-number" data-line-number="636"></td> <td id="LC636" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> initializer: (optional) The initializer to use for the weight and</span></td> </tr> <tr> <td id="L637" class="blob-num js-line-number" data-line-number="637"></td> <td id="LC637" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> projection matrices.</span></td> </tr> <tr> <td id="L638" class="blob-num js-line-number" data-line-number="638"></td> <td id="LC638" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_proj: (optional) int, The output dimensionality for the projection</span></td> </tr> <tr> <td id="L639" class="blob-num js-line-number" data-line-number="639"></td> <td id="LC639" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> matrices. If None, no projection is performed.</span></td> </tr> <tr> <td id="L640" class="blob-num js-line-number" data-line-number="640"></td> <td id="LC640" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> proj_clip: (optional) A float value. If `num_proj &gt; 0` and `proj_clip` is</span></td> </tr> <tr> <td id="L641" class="blob-num js-line-number" data-line-number="641"></td> <td id="LC641" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> provided, then the projected values are clipped elementwise to within</span></td> </tr> <tr> <td id="L642" class="blob-num js-line-number" data-line-number="642"></td> <td id="LC642" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `[-proj_clip, proj_clip]`.</span></td> </tr> <tr> <td id="L643" class="blob-num js-line-number" data-line-number="643"></td> <td id="LC643" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_unit_shards: Deprecated, will be removed by Jan. 2017.</span></td> </tr> <tr> <td id="L644" class="blob-num js-line-number" data-line-number="644"></td> <td id="LC644" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Use a variable_scope partitioner instead.</span></td> </tr> <tr> <td id="L645" class="blob-num js-line-number" data-line-number="645"></td> <td id="LC645" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_proj_shards: Deprecated, will be removed by Jan. 2017.</span></td> </tr> <tr> <td id="L646" class="blob-num js-line-number" data-line-number="646"></td> <td id="LC646" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Use a variable_scope partitioner instead.</span></td> </tr> <tr> <td id="L647" class="blob-num js-line-number" data-line-number="647"></td> <td id="LC647" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> forget_bias: Biases of the forget gate are initialized by default to 1</span></td> </tr> <tr> <td id="L648" class="blob-num js-line-number" data-line-number="648"></td> <td id="LC648" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in order to reduce the scale of forgetting at the beginning of</span></td> </tr> <tr> <td id="L649" class="blob-num js-line-number" data-line-number="649"></td> <td id="LC649" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the training. Must set it manually to `0.0` when restoring from</span></td> </tr> <tr> <td id="L650" class="blob-num js-line-number" data-line-number="650"></td> <td id="LC650" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> CudnnLSTM trained checkpoints.</span></td> </tr> <tr> <td id="L651" class="blob-num js-line-number" data-line-number="651"></td> <td id="LC651" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state_is_tuple: If True, accepted and returned states are 2-tuples of</span></td> </tr> <tr> <td id="L652" class="blob-num js-line-number" data-line-number="652"></td> <td id="LC652" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the `c_state` and `m_state`. If False, they are concatenated</span></td> </tr> <tr> <td id="L653" class="blob-num js-line-number" data-line-number="653"></td> <td id="LC653" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> along the column axis. This latter behavior will soon be deprecated.</span></td> </tr> <tr> <td id="L654" class="blob-num js-line-number" data-line-number="654"></td> <td id="LC654" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> activation: Activation function of the inner states. Default: `tanh`.</span></td> </tr> <tr> <td id="L655" class="blob-num js-line-number" data-line-number="655"></td> <td id="LC655" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> reuse: (optional) Python boolean describing whether to reuse variables</span></td> </tr> <tr> <td id="L656" class="blob-num js-line-number" data-line-number="656"></td> <td id="LC656" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> in an existing scope. If not `True`, and the existing scope already has</span></td> </tr> <tr> <td id="L657" class="blob-num js-line-number" data-line-number="657"></td> <td id="LC657" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the given variables, an error is raised.</span></td> </tr> <tr> <td id="L658" class="blob-num js-line-number" data-line-number="658"></td> <td id="LC658" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> name: String, the name of the layer. Layers with the same name will</span></td> </tr> <tr> <td id="L659" class="blob-num js-line-number" data-line-number="659"></td> <td id="LC659" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> share weights, but to avoid mistakes we require reuse=True in such</span></td> </tr> <tr> <td id="L660" class="blob-num js-line-number" data-line-number="660"></td> <td id="LC660" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cases.</span></td> </tr> <tr> <td id="L661" class="blob-num js-line-number" data-line-number="661"></td> <td id="LC661" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L662" class="blob-num js-line-number" data-line-number="662"></td> <td id="LC662" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> When restoring from CudnnLSTM-trained checkpoints, use</span></td> </tr> <tr> <td id="L663" class="blob-num js-line-number" data-line-number="663"></td> <td id="LC663" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `CudnnCompatibleLSTMCell` instead.</span></td> </tr> <tr> <td id="L664" class="blob-num js-line-number" data-line-number="664"></td> <td id="LC664" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L665" class="blob-num js-line-number" data-line-number="665"></td> <td id="LC665" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(LSTMCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>(<span class="pl-v">_reuse</span><span class="pl-k">=</span>reuse, <span class="pl-v">name</span><span class="pl-k">=</span>name)</td> </tr> <tr> <td id="L666" class="blob-num js-line-number" data-line-number="666"></td> <td id="LC666" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> state_is_tuple:</td> </tr> <tr> <td id="L667" class="blob-num js-line-number" data-line-number="667"></td> <td id="LC667" class="blob-code blob-code-inner js-file-line"> logging.warn(<span class="pl-s"><span class="pl-pds">&quot;</span><span class="pl-c1">%s</span>: Using a concatenated state is slower and will soon be <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L668" class="blob-num js-line-number" data-line-number="668"></td> <td id="LC668" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>deprecated. Use state_is_tuple=True.<span class="pl-pds">&quot;</span></span>, <span class="pl-c1">self</span>)</td> </tr> <tr> <td id="L669" class="blob-num js-line-number" data-line-number="669"></td> <td id="LC669" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> num_unit_shards <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span> <span class="pl-k">or</span> num_proj_shards <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L670" class="blob-num js-line-number" data-line-number="670"></td> <td id="LC670" class="blob-code blob-code-inner js-file-line"> logging.warn(</td> </tr> <tr> <td id="L671" class="blob-num js-line-number" data-line-number="671"></td> <td id="LC671" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span><span class="pl-c1">%s</span>: The num_unit_shards and proj_unit_shards parameters are <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L672" class="blob-num js-line-number" data-line-number="672"></td> <td id="LC672" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>deprecated and will be removed in Jan 2017. <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L673" class="blob-num js-line-number" data-line-number="673"></td> <td id="LC673" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>Use a variable scope with a partitioner instead.<span class="pl-pds">&quot;</span></span>, <span class="pl-c1">self</span>)</td> </tr> <tr> <td id="L674" class="blob-num js-line-number" data-line-number="674"></td> <td id="LC674" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L675" class="blob-num js-line-number" data-line-number="675"></td> <td id="LC675" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Inputs must be 2-dimensional.</span></td> </tr> <tr> <td id="L676" class="blob-num js-line-number" data-line-number="676"></td> <td id="LC676" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.input_spec <span class="pl-k">=</span> base_layer.InputSpec(<span class="pl-v">ndim</span><span class="pl-k">=</span><span class="pl-c1">2</span>)</td> </tr> <tr> <td id="L677" class="blob-num js-line-number" data-line-number="677"></td> <td id="LC677" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L678" class="blob-num js-line-number" data-line-number="678"></td> <td id="LC678" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_units <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L679" class="blob-num js-line-number" data-line-number="679"></td> <td id="LC679" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._use_peepholes <span class="pl-k">=</span> use_peepholes</td> </tr> <tr> <td id="L680" class="blob-num js-line-number" data-line-number="680"></td> <td id="LC680" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell_clip <span class="pl-k">=</span> cell_clip</td> </tr> <tr> <td id="L681" class="blob-num js-line-number" data-line-number="681"></td> <td id="LC681" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._initializer <span class="pl-k">=</span> initializer</td> </tr> <tr> <td id="L682" class="blob-num js-line-number" data-line-number="682"></td> <td id="LC682" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_proj <span class="pl-k">=</span> num_proj</td> </tr> <tr> <td id="L683" class="blob-num js-line-number" data-line-number="683"></td> <td id="LC683" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._proj_clip <span class="pl-k">=</span> proj_clip</td> </tr> <tr> <td id="L684" class="blob-num js-line-number" data-line-number="684"></td> <td id="LC684" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_unit_shards <span class="pl-k">=</span> num_unit_shards</td> </tr> <tr> <td id="L685" class="blob-num js-line-number" data-line-number="685"></td> <td id="LC685" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._num_proj_shards <span class="pl-k">=</span> num_proj_shards</td> </tr> <tr> <td id="L686" class="blob-num js-line-number" data-line-number="686"></td> <td id="LC686" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._forget_bias <span class="pl-k">=</span> forget_bias</td> </tr> <tr> <td id="L687" class="blob-num js-line-number" data-line-number="687"></td> <td id="LC687" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">=</span> state_is_tuple</td> </tr> <tr> <td id="L688" class="blob-num js-line-number" data-line-number="688"></td> <td id="LC688" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation <span class="pl-k">=</span> activation <span class="pl-k">or</span> math_ops.tanh</td> </tr> <tr> <td id="L689" class="blob-num js-line-number" data-line-number="689"></td> <td id="LC689" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L690" class="blob-num js-line-number" data-line-number="690"></td> <td id="LC690" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> num_proj:</td> </tr> <tr> <td id="L691" class="blob-num js-line-number" data-line-number="691"></td> <td id="LC691" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_size <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L692" class="blob-num js-line-number" data-line-number="692"></td> <td id="LC692" class="blob-code blob-code-inner js-file-line"> LSTMStateTuple(num_units, num_proj)</td> </tr> <tr> <td id="L693" class="blob-num js-line-number" data-line-number="693"></td> <td id="LC693" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> state_is_tuple <span class="pl-k">else</span> num_units <span class="pl-k">+</span> num_proj)</td> </tr> <tr> <td id="L694" class="blob-num js-line-number" data-line-number="694"></td> <td id="LC694" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._output_size <span class="pl-k">=</span> num_proj</td> </tr> <tr> <td id="L695" class="blob-num js-line-number" data-line-number="695"></td> <td id="LC695" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L696" class="blob-num js-line-number" data-line-number="696"></td> <td id="LC696" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_size <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L697" class="blob-num js-line-number" data-line-number="697"></td> <td id="LC697" class="blob-code blob-code-inner js-file-line"> LSTMStateTuple(num_units, num_units)</td> </tr> <tr> <td id="L698" class="blob-num js-line-number" data-line-number="698"></td> <td id="LC698" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> state_is_tuple <span class="pl-k">else</span> <span class="pl-c1">2</span> <span class="pl-k">*</span> num_units)</td> </tr> <tr> <td id="L699" class="blob-num js-line-number" data-line-number="699"></td> <td id="LC699" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._output_size <span class="pl-k">=</span> num_units</td> </tr> <tr> <td id="L700" class="blob-num js-line-number" data-line-number="700"></td> <td id="LC700" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L701" class="blob-num js-line-number" data-line-number="701"></td> <td id="LC701" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L702" class="blob-num js-line-number" data-line-number="702"></td> <td id="LC702" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L703" class="blob-num js-line-number" data-line-number="703"></td> <td id="LC703" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._state_size</td> </tr> <tr> <td id="L704" class="blob-num js-line-number" data-line-number="704"></td> <td id="LC704" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L705" class="blob-num js-line-number" data-line-number="705"></td> <td id="LC705" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L706" class="blob-num js-line-number" data-line-number="706"></td> <td id="LC706" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L707" class="blob-num js-line-number" data-line-number="707"></td> <td id="LC707" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._output_size</td> </tr> <tr> <td id="L708" class="blob-num js-line-number" data-line-number="708"></td> <td id="LC708" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L709" class="blob-num js-line-number" data-line-number="709"></td> <td id="LC709" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">build</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs_shape</span>):</td> </tr> <tr> <td id="L710" class="blob-num js-line-number" data-line-number="710"></td> <td id="LC710" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> inputs_shape[<span class="pl-c1">1</span>].value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L711" class="blob-num js-line-number" data-line-number="711"></td> <td id="LC711" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Expected inputs.shape[-1] to be known, saw shape: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L712" class="blob-num js-line-number" data-line-number="712"></td> <td id="LC712" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> inputs_shape)</td> </tr> <tr> <td id="L713" class="blob-num js-line-number" data-line-number="713"></td> <td id="LC713" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L714" class="blob-num js-line-number" data-line-number="714"></td> <td id="LC714" class="blob-code blob-code-inner js-file-line"> input_depth <span class="pl-k">=</span> inputs_shape[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L715" class="blob-num js-line-number" data-line-number="715"></td> <td id="LC715" class="blob-code blob-code-inner js-file-line"> h_depth <span class="pl-k">=</span> <span class="pl-c1">self</span>._num_units <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj <span class="pl-k">is</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">self</span>._num_proj</td> </tr> <tr> <td id="L716" class="blob-num js-line-number" data-line-number="716"></td> <td id="LC716" class="blob-code blob-code-inner js-file-line"> maybe_partitioner <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L717" class="blob-num js-line-number" data-line-number="717"></td> <td id="LC717" class="blob-code blob-code-inner js-file-line"> partitioned_variables.fixed_size_partitioner(<span class="pl-c1">self</span>._num_unit_shards)</td> </tr> <tr> <td id="L718" class="blob-num js-line-number" data-line-number="718"></td> <td id="LC718" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_unit_shards <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L719" class="blob-num js-line-number" data-line-number="719"></td> <td id="LC719" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span> <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L720" class="blob-num js-line-number" data-line-number="720"></td> <td id="LC720" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L721" class="blob-num js-line-number" data-line-number="721"></td> <td id="LC721" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L722" class="blob-num js-line-number" data-line-number="722"></td> <td id="LC722" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[input_depth <span class="pl-k">+</span> h_depth, <span class="pl-c1">4</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L723" class="blob-num js-line-number" data-line-number="723"></td> <td id="LC723" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer,</td> </tr> <tr> <td id="L724" class="blob-num js-line-number" data-line-number="724"></td> <td id="LC724" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">partitioner</span><span class="pl-k">=</span>maybe_partitioner)</td> </tr> <tr> <td id="L725" class="blob-num js-line-number" data-line-number="725"></td> <td id="LC725" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._bias <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L726" class="blob-num js-line-number" data-line-number="726"></td> <td id="LC726" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">_BIAS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L727" class="blob-num js-line-number" data-line-number="727"></td> <td id="LC727" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">4</span> <span class="pl-k">*</span> <span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L728" class="blob-num js-line-number" data-line-number="728"></td> <td id="LC728" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span>init_ops.zeros_initializer(<span class="pl-v">dtype</span><span class="pl-k">=</span><span class="pl-c1">self</span>.dtype))</td> </tr> <tr> <td id="L729" class="blob-num js-line-number" data-line-number="729"></td> <td id="LC729" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._use_peepholes:</td> </tr> <tr> <td id="L730" class="blob-num js-line-number" data-line-number="730"></td> <td id="LC730" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._w_f_diag <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(<span class="pl-s"><span class="pl-pds">&quot;</span>w_f_diag<span class="pl-pds">&quot;</span></span>, <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L731" class="blob-num js-line-number" data-line-number="731"></td> <td id="LC731" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer)</td> </tr> <tr> <td id="L732" class="blob-num js-line-number" data-line-number="732"></td> <td id="LC732" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._w_i_diag <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(<span class="pl-s"><span class="pl-pds">&quot;</span>w_i_diag<span class="pl-pds">&quot;</span></span>, <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L733" class="blob-num js-line-number" data-line-number="733"></td> <td id="LC733" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer)</td> </tr> <tr> <td id="L734" class="blob-num js-line-number" data-line-number="734"></td> <td id="LC734" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._w_o_diag <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(<span class="pl-s"><span class="pl-pds">&quot;</span>w_o_diag<span class="pl-pds">&quot;</span></span>, <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units],</td> </tr> <tr> <td id="L735" class="blob-num js-line-number" data-line-number="735"></td> <td id="LC735" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer)</td> </tr> <tr> <td id="L736" class="blob-num js-line-number" data-line-number="736"></td> <td id="LC736" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L737" class="blob-num js-line-number" data-line-number="737"></td> <td id="LC737" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L738" class="blob-num js-line-number" data-line-number="738"></td> <td id="LC738" class="blob-code blob-code-inner js-file-line"> maybe_proj_partitioner <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L739" class="blob-num js-line-number" data-line-number="739"></td> <td id="LC739" class="blob-code blob-code-inner js-file-line"> partitioned_variables.fixed_size_partitioner(<span class="pl-c1">self</span>._num_proj_shards)</td> </tr> <tr> <td id="L740" class="blob-num js-line-number" data-line-number="740"></td> <td id="LC740" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj_shards <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L741" class="blob-num js-line-number" data-line-number="741"></td> <td id="LC741" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span> <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L742" class="blob-num js-line-number" data-line-number="742"></td> <td id="LC742" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._proj_kernel <span class="pl-k">=</span> <span class="pl-c1">self</span>.add_variable(</td> </tr> <tr> <td id="L743" class="blob-num js-line-number" data-line-number="743"></td> <td id="LC743" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>projection/<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> <span class="pl-c1">_WEIGHTS_VARIABLE_NAME</span>,</td> </tr> <tr> <td id="L744" class="blob-num js-line-number" data-line-number="744"></td> <td id="LC744" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">shape</span><span class="pl-k">=</span>[<span class="pl-c1">self</span>._num_units, <span class="pl-c1">self</span>._num_proj],</td> </tr> <tr> <td id="L745" class="blob-num js-line-number" data-line-number="745"></td> <td id="LC745" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">initializer</span><span class="pl-k">=</span><span class="pl-c1">self</span>._initializer,</td> </tr> <tr> <td id="L746" class="blob-num js-line-number" data-line-number="746"></td> <td id="LC746" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">partitioner</span><span class="pl-k">=</span>maybe_proj_partitioner)</td> </tr> <tr> <td id="L747" class="blob-num js-line-number" data-line-number="747"></td> <td id="LC747" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L748" class="blob-num js-line-number" data-line-number="748"></td> <td id="LC748" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>.built <span class="pl-k">=</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L749" class="blob-num js-line-number" data-line-number="749"></td> <td id="LC749" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L750" class="blob-num js-line-number" data-line-number="750"></td> <td id="LC750" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L751" class="blob-num js-line-number" data-line-number="751"></td> <td id="LC751" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run one step of LSTM.</span></td> </tr> <tr> <td id="L752" class="blob-num js-line-number" data-line-number="752"></td> <td id="LC752" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L753" class="blob-num js-line-number" data-line-number="753"></td> <td id="LC753" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L754" class="blob-num js-line-number" data-line-number="754"></td> <td id="LC754" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: input Tensor, 2D, `[batch, num_units].</span></td> </tr> <tr> <td id="L755" class="blob-num js-line-number" data-line-number="755"></td> <td id="LC755" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: if `state_is_tuple` is False, this must be a state Tensor,</span></td> </tr> <tr> <td id="L756" class="blob-num js-line-number" data-line-number="756"></td> <td id="LC756" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `2-D, [batch, state_size]`. If `state_is_tuple` is True, this must be a</span></td> </tr> <tr> <td id="L757" class="blob-num js-line-number" data-line-number="757"></td> <td id="LC757" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> tuple of state Tensors, both `2-D`, with column sizes `c_state` and</span></td> </tr> <tr> <td id="L758" class="blob-num js-line-number" data-line-number="758"></td> <td id="LC758" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `m_state`.</span></td> </tr> <tr> <td id="L759" class="blob-num js-line-number" data-line-number="759"></td> <td id="LC759" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L760" class="blob-num js-line-number" data-line-number="760"></td> <td id="LC760" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L761" class="blob-num js-line-number" data-line-number="761"></td> <td id="LC761" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> A tuple containing:</span></td> </tr> <tr> <td id="L762" class="blob-num js-line-number" data-line-number="762"></td> <td id="LC762" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L763" class="blob-num js-line-number" data-line-number="763"></td> <td id="LC763" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - A `2-D, [batch, output_dim]`, Tensor representing the output of the</span></td> </tr> <tr> <td id="L764" class="blob-num js-line-number" data-line-number="764"></td> <td id="LC764" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> LSTM after reading `inputs` when previous state was `state`.</span></td> </tr> <tr> <td id="L765" class="blob-num js-line-number" data-line-number="765"></td> <td id="LC765" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Here output_dim is:</span></td> </tr> <tr> <td id="L766" class="blob-num js-line-number" data-line-number="766"></td> <td id="LC766" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_proj if num_proj was set,</span></td> </tr> <tr> <td id="L767" class="blob-num js-line-number" data-line-number="767"></td> <td id="LC767" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> num_units otherwise.</span></td> </tr> <tr> <td id="L768" class="blob-num js-line-number" data-line-number="768"></td> <td id="LC768" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> - Tensor(s) representing the new state of LSTM after reading `inputs` when</span></td> </tr> <tr> <td id="L769" class="blob-num js-line-number" data-line-number="769"></td> <td id="LC769" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the previous state was `state`. Same type and shape(s) as `state`.</span></td> </tr> <tr> <td id="L770" class="blob-num js-line-number" data-line-number="770"></td> <td id="LC770" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L771" class="blob-num js-line-number" data-line-number="771"></td> <td id="LC771" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L772" class="blob-num js-line-number" data-line-number="772"></td> <td id="LC772" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: If input size cannot be inferred from inputs via</span></td> </tr> <tr> <td id="L773" class="blob-num js-line-number" data-line-number="773"></td> <td id="LC773" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> static shape inference.</span></td> </tr> <tr> <td id="L774" class="blob-num js-line-number" data-line-number="774"></td> <td id="LC774" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L775" class="blob-num js-line-number" data-line-number="775"></td> <td id="LC775" class="blob-code blob-code-inner js-file-line"> num_proj <span class="pl-k">=</span> <span class="pl-c1">self</span>._num_units <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj <span class="pl-k">is</span> <span class="pl-c1">None</span> <span class="pl-k">else</span> <span class="pl-c1">self</span>._num_proj</td> </tr> <tr> <td id="L776" class="blob-num js-line-number" data-line-number="776"></td> <td id="LC776" class="blob-code blob-code-inner js-file-line"> sigmoid <span class="pl-k">=</span> math_ops.sigmoid</td> </tr> <tr> <td id="L777" class="blob-num js-line-number" data-line-number="777"></td> <td id="LC777" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L778" class="blob-num js-line-number" data-line-number="778"></td> <td id="LC778" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L779" class="blob-num js-line-number" data-line-number="779"></td> <td id="LC779" class="blob-code blob-code-inner js-file-line"> (c_prev, m_prev) <span class="pl-k">=</span> state</td> </tr> <tr> <td id="L780" class="blob-num js-line-number" data-line-number="780"></td> <td id="LC780" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L781" class="blob-num js-line-number" data-line-number="781"></td> <td id="LC781" class="blob-code blob-code-inner js-file-line"> c_prev <span class="pl-k">=</span> array_ops.slice(state, [<span class="pl-c1">0</span>, <span class="pl-c1">0</span>], [<span class="pl-k">-</span><span class="pl-c1">1</span>, <span class="pl-c1">self</span>._num_units])</td> </tr> <tr> <td id="L782" class="blob-num js-line-number" data-line-number="782"></td> <td id="LC782" class="blob-code blob-code-inner js-file-line"> m_prev <span class="pl-k">=</span> array_ops.slice(state, [<span class="pl-c1">0</span>, <span class="pl-c1">self</span>._num_units], [<span class="pl-k">-</span><span class="pl-c1">1</span>, num_proj])</td> </tr> <tr> <td id="L783" class="blob-num js-line-number" data-line-number="783"></td> <td id="LC783" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L784" class="blob-num js-line-number" data-line-number="784"></td> <td id="LC784" class="blob-code blob-code-inner js-file-line"> input_size <span class="pl-k">=</span> inputs.get_shape().with_rank(<span class="pl-c1">2</span>)[<span class="pl-c1">1</span>]</td> </tr> <tr> <td id="L785" class="blob-num js-line-number" data-line-number="785"></td> <td id="LC785" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> input_size.value <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L786" class="blob-num js-line-number" data-line-number="786"></td> <td id="LC786" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Could not infer input size from inputs.get_shape()[-1]<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L787" class="blob-num js-line-number" data-line-number="787"></td> <td id="LC787" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L788" class="blob-num js-line-number" data-line-number="788"></td> <td id="LC788" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> i = input_gate, j = new_input, f = forget_gate, o = output_gate</span></td> </tr> <tr> <td id="L789" class="blob-num js-line-number" data-line-number="789"></td> <td id="LC789" class="blob-code blob-code-inner js-file-line"> lstm_matrix <span class="pl-k">=</span> math_ops.matmul(</td> </tr> <tr> <td id="L790" class="blob-num js-line-number" data-line-number="790"></td> <td id="LC790" class="blob-code blob-code-inner js-file-line"> array_ops.concat([inputs, m_prev], <span class="pl-c1">1</span>), <span class="pl-c1">self</span>._kernel)</td> </tr> <tr> <td id="L791" class="blob-num js-line-number" data-line-number="791"></td> <td id="LC791" class="blob-code blob-code-inner js-file-line"> lstm_matrix <span class="pl-k">=</span> nn_ops.bias_add(lstm_matrix, <span class="pl-c1">self</span>._bias)</td> </tr> <tr> <td id="L792" class="blob-num js-line-number" data-line-number="792"></td> <td id="LC792" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L793" class="blob-num js-line-number" data-line-number="793"></td> <td id="LC793" class="blob-code blob-code-inner js-file-line"> i, j, f, o <span class="pl-k">=</span> array_ops.split(</td> </tr> <tr> <td id="L794" class="blob-num js-line-number" data-line-number="794"></td> <td id="LC794" class="blob-code blob-code-inner js-file-line"> <span class="pl-v">value</span><span class="pl-k">=</span>lstm_matrix, <span class="pl-v">num_or_size_splits</span><span class="pl-k">=</span><span class="pl-c1">4</span>, <span class="pl-v">axis</span><span class="pl-k">=</span><span class="pl-c1">1</span>)</td> </tr> <tr> <td id="L795" class="blob-num js-line-number" data-line-number="795"></td> <td id="LC795" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Diagonal connections</span></td> </tr> <tr> <td id="L796" class="blob-num js-line-number" data-line-number="796"></td> <td id="LC796" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._use_peepholes:</td> </tr> <tr> <td id="L797" class="blob-num js-line-number" data-line-number="797"></td> <td id="LC797" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> (sigmoid(f <span class="pl-k">+</span> <span class="pl-c1">self</span>._forget_bias <span class="pl-k">+</span> <span class="pl-c1">self</span>._w_f_diag <span class="pl-k">*</span> c_prev) <span class="pl-k">*</span> c_prev <span class="pl-k">+</span></td> </tr> <tr> <td id="L798" class="blob-num js-line-number" data-line-number="798"></td> <td id="LC798" class="blob-code blob-code-inner js-file-line"> sigmoid(i <span class="pl-k">+</span> <span class="pl-c1">self</span>._w_i_diag <span class="pl-k">*</span> c_prev) <span class="pl-k">*</span> <span class="pl-c1">self</span>._activation(j))</td> </tr> <tr> <td id="L799" class="blob-num js-line-number" data-line-number="799"></td> <td id="LC799" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L800" class="blob-num js-line-number" data-line-number="800"></td> <td id="LC800" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> (sigmoid(f <span class="pl-k">+</span> <span class="pl-c1">self</span>._forget_bias) <span class="pl-k">*</span> c_prev <span class="pl-k">+</span> sigmoid(i) <span class="pl-k">*</span></td> </tr> <tr> <td id="L801" class="blob-num js-line-number" data-line-number="801"></td> <td id="LC801" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._activation(j))</td> </tr> <tr> <td id="L802" class="blob-num js-line-number" data-line-number="802"></td> <td id="LC802" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L803" class="blob-num js-line-number" data-line-number="803"></td> <td id="LC803" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._cell_clip <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L804" class="blob-num js-line-number" data-line-number="804"></td> <td id="LC804" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> pylint: disable=invalid-unary-operand-type</span></td> </tr> <tr> <td id="L805" class="blob-num js-line-number" data-line-number="805"></td> <td id="LC805" class="blob-code blob-code-inner js-file-line"> c <span class="pl-k">=</span> clip_ops.clip_by_value(c, <span class="pl-k">-</span><span class="pl-c1">self</span>._cell_clip, <span class="pl-c1">self</span>._cell_clip)</td> </tr> <tr> <td id="L806" class="blob-num js-line-number" data-line-number="806"></td> <td id="LC806" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> pylint: enable=invalid-unary-operand-type</span></td> </tr> <tr> <td id="L807" class="blob-num js-line-number" data-line-number="807"></td> <td id="LC807" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._use_peepholes:</td> </tr> <tr> <td id="L808" class="blob-num js-line-number" data-line-number="808"></td> <td id="LC808" class="blob-code blob-code-inner js-file-line"> m <span class="pl-k">=</span> sigmoid(o <span class="pl-k">+</span> <span class="pl-c1">self</span>._w_o_diag <span class="pl-k">*</span> c) <span class="pl-k">*</span> <span class="pl-c1">self</span>._activation(c)</td> </tr> <tr> <td id="L809" class="blob-num js-line-number" data-line-number="809"></td> <td id="LC809" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L810" class="blob-num js-line-number" data-line-number="810"></td> <td id="LC810" class="blob-code blob-code-inner js-file-line"> m <span class="pl-k">=</span> sigmoid(o) <span class="pl-k">*</span> <span class="pl-c1">self</span>._activation(c)</td> </tr> <tr> <td id="L811" class="blob-num js-line-number" data-line-number="811"></td> <td id="LC811" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L812" class="blob-num js-line-number" data-line-number="812"></td> <td id="LC812" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._num_proj <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L813" class="blob-num js-line-number" data-line-number="813"></td> <td id="LC813" class="blob-code blob-code-inner js-file-line"> m <span class="pl-k">=</span> math_ops.matmul(m, <span class="pl-c1">self</span>._proj_kernel)</td> </tr> <tr> <td id="L814" class="blob-num js-line-number" data-line-number="814"></td> <td id="LC814" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L815" class="blob-num js-line-number" data-line-number="815"></td> <td id="LC815" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._proj_clip <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L816" class="blob-num js-line-number" data-line-number="816"></td> <td id="LC816" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> pylint: disable=invalid-unary-operand-type</span></td> </tr> <tr> <td id="L817" class="blob-num js-line-number" data-line-number="817"></td> <td id="LC817" class="blob-code blob-code-inner js-file-line"> m <span class="pl-k">=</span> clip_ops.clip_by_value(m, <span class="pl-k">-</span><span class="pl-c1">self</span>._proj_clip, <span class="pl-c1">self</span>._proj_clip)</td> </tr> <tr> <td id="L818" class="blob-num js-line-number" data-line-number="818"></td> <td id="LC818" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> pylint: enable=invalid-unary-operand-type</span></td> </tr> <tr> <td id="L819" class="blob-num js-line-number" data-line-number="819"></td> <td id="LC819" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L820" class="blob-num js-line-number" data-line-number="820"></td> <td id="LC820" class="blob-code blob-code-inner js-file-line"> new_state <span class="pl-k">=</span> (LSTMStateTuple(c, m) <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">else</span></td> </tr> <tr> <td id="L821" class="blob-num js-line-number" data-line-number="821"></td> <td id="LC821" class="blob-code blob-code-inner js-file-line"> array_ops.concat([c, m], <span class="pl-c1">1</span>))</td> </tr> <tr> <td id="L822" class="blob-num js-line-number" data-line-number="822"></td> <td id="LC822" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> m, new_state</td> </tr> <tr> <td id="L823" class="blob-num js-line-number" data-line-number="823"></td> <td id="LC823" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L824" class="blob-num js-line-number" data-line-number="824"></td> <td id="LC824" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L825" class="blob-num js-line-number" data-line-number="825"></td> <td id="LC825" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_enumerated_map_structure_up_to</span>(<span class="pl-smi">shallow_structure</span>, <span class="pl-smi">map_fn</span>, <span class="pl-k">*</span><span class="pl-smi">args</span>, <span class="pl-k">**</span><span class="pl-smi">kwargs</span>):</td> </tr> <tr> <td id="L826" class="blob-num js-line-number" data-line-number="826"></td> <td id="LC826" class="blob-code blob-code-inner js-file-line"> ix <span class="pl-k">=</span> [<span class="pl-c1">0</span>]</td> </tr> <tr> <td id="L827" class="blob-num js-line-number" data-line-number="827"></td> <td id="LC827" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">enumerated_fn</span>(<span class="pl-k">*</span><span class="pl-smi">inner_args</span>, <span class="pl-k">**</span><span class="pl-smi">inner_kwargs</span>):</td> </tr> <tr> <td id="L828" class="blob-num js-line-number" data-line-number="828"></td> <td id="LC828" class="blob-code blob-code-inner js-file-line"> r <span class="pl-k">=</span> map_fn(ix[<span class="pl-c1">0</span>], <span class="pl-k">*</span>inner_args, <span class="pl-k">**</span>inner_kwargs)</td> </tr> <tr> <td id="L829" class="blob-num js-line-number" data-line-number="829"></td> <td id="LC829" class="blob-code blob-code-inner js-file-line"> ix[<span class="pl-c1">0</span>] <span class="pl-k">+=</span> <span class="pl-c1">1</span></td> </tr> <tr> <td id="L830" class="blob-num js-line-number" data-line-number="830"></td> <td id="LC830" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> r</td> </tr> <tr> <td id="L831" class="blob-num js-line-number" data-line-number="831"></td> <td id="LC831" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> nest.map_structure_up_to(shallow_structure,</td> </tr> <tr> <td id="L832" class="blob-num js-line-number" data-line-number="832"></td> <td id="LC832" class="blob-code blob-code-inner js-file-line"> enumerated_fn, <span class="pl-k">*</span>args, <span class="pl-k">**</span>kwargs)</td> </tr> <tr> <td id="L833" class="blob-num js-line-number" data-line-number="833"></td> <td id="LC833" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L834" class="blob-num js-line-number" data-line-number="834"></td> <td id="LC834" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L835" class="blob-num js-line-number" data-line-number="835"></td> <td id="LC835" class="blob-code blob-code-inner js-file-line"><span class="pl-k">def</span> <span class="pl-en">_default_dropout_state_filter_visitor</span>(<span class="pl-smi">substate</span>):</td> </tr> <tr> <td id="L836" class="blob-num js-line-number" data-line-number="836"></td> <td id="LC836" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">isinstance</span>(substate, LSTMStateTuple):</td> </tr> <tr> <td id="L837" class="blob-num js-line-number" data-line-number="837"></td> <td id="LC837" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Do not perform dropout on the memory state.</span></td> </tr> <tr> <td id="L838" class="blob-num js-line-number" data-line-number="838"></td> <td id="LC838" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> LSTMStateTuple(<span class="pl-v">c</span><span class="pl-k">=</span><span class="pl-c1">False</span>, <span class="pl-v">h</span><span class="pl-k">=</span><span class="pl-c1">True</span>)</td> </tr> <tr> <td id="L839" class="blob-num js-line-number" data-line-number="839"></td> <td id="LC839" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">elif</span> <span class="pl-c1">isinstance</span>(substate, tensor_array_ops.TensorArray):</td> </tr> <tr> <td id="L840" class="blob-num js-line-number" data-line-number="840"></td> <td id="LC840" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">False</span></td> </tr> <tr> <td id="L841" class="blob-num js-line-number" data-line-number="841"></td> <td id="LC841" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">True</span></td> </tr> <tr> <td id="L842" class="blob-num js-line-number" data-line-number="842"></td> <td id="LC842" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L843" class="blob-num js-line-number" data-line-number="843"></td> <td id="LC843" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L844" class="blob-num js-line-number" data-line-number="844"></td> <td id="LC844" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.DropoutWrapper<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L845" class="blob-num js-line-number" data-line-number="845"></td> <td id="LC845" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">DropoutWrapper</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L846" class="blob-num js-line-number" data-line-number="846"></td> <td id="LC846" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Operator adding dropout to inputs and outputs of the given cell.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L847" class="blob-num js-line-number" data-line-number="847"></td> <td id="LC847" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L848" class="blob-num js-line-number" data-line-number="848"></td> <td id="LC848" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cell</span>, <span class="pl-smi">input_keep_prob</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>, <span class="pl-smi">output_keep_prob</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>,</td> </tr> <tr> <td id="L849" class="blob-num js-line-number" data-line-number="849"></td> <td id="LC849" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">state_keep_prob</span><span class="pl-k">=</span><span class="pl-c1">1.0</span>, <span class="pl-smi">variational_recurrent</span><span class="pl-k">=</span><span class="pl-c1">False</span>,</td> </tr> <tr> <td id="L850" class="blob-num js-line-number" data-line-number="850"></td> <td id="LC850" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">input_size</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">dtype</span><span class="pl-k">=</span><span class="pl-c1">None</span>, <span class="pl-smi">seed</span><span class="pl-k">=</span><span class="pl-c1">None</span>,</td> </tr> <tr> <td id="L851" class="blob-num js-line-number" data-line-number="851"></td> <td id="LC851" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">dropout_state_filter_visitor</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L852" class="blob-num js-line-number" data-line-number="852"></td> <td id="LC852" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Create a cell with added input, state, and/or output dropout.</span></td> </tr> <tr> <td id="L853" class="blob-num js-line-number" data-line-number="853"></td> <td id="LC853" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L854" class="blob-num js-line-number" data-line-number="854"></td> <td id="LC854" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If `variational_recurrent` is set to `True` (**NOT** the default behavior),</span></td> </tr> <tr> <td id="L855" class="blob-num js-line-number" data-line-number="855"></td> <td id="LC855" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> then the same dropout mask is applied at every step, as described in:</span></td> </tr> <tr> <td id="L856" class="blob-num js-line-number" data-line-number="856"></td> <td id="LC856" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L857" class="blob-num js-line-number" data-line-number="857"></td> <td id="LC857" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Y. Gal, Z Ghahramani. &quot;A Theoretically Grounded Application of Dropout in</span></td> </tr> <tr> <td id="L858" class="blob-num js-line-number" data-line-number="858"></td> <td id="LC858" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Recurrent Neural Networks&quot;. https://arxiv.org/abs/1512.05287</span></td> </tr> <tr> <td id="L859" class="blob-num js-line-number" data-line-number="859"></td> <td id="LC859" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L860" class="blob-num js-line-number" data-line-number="860"></td> <td id="LC860" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Otherwise a different dropout mask is applied at every time step.</span></td> </tr> <tr> <td id="L861" class="blob-num js-line-number" data-line-number="861"></td> <td id="LC861" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L862" class="blob-num js-line-number" data-line-number="862"></td> <td id="LC862" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Note, by default (unless a custom `dropout_state_filter` is provided),</span></td> </tr> <tr> <td id="L863" class="blob-num js-line-number" data-line-number="863"></td> <td id="LC863" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the memory state (`c` component of any `LSTMStateTuple`) passing through</span></td> </tr> <tr> <td id="L864" class="blob-num js-line-number" data-line-number="864"></td> <td id="LC864" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> a `DropoutWrapper` is never modified. This behavior is described in the</span></td> </tr> <tr> <td id="L865" class="blob-num js-line-number" data-line-number="865"></td> <td id="LC865" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> above article.</span></td> </tr> <tr> <td id="L866" class="blob-num js-line-number" data-line-number="866"></td> <td id="LC866" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L867" class="blob-num js-line-number" data-line-number="867"></td> <td id="LC867" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L868" class="blob-num js-line-number" data-line-number="868"></td> <td id="LC868" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell: an RNNCell, a projection to output_size is added to it.</span></td> </tr> <tr> <td id="L869" class="blob-num js-line-number" data-line-number="869"></td> <td id="LC869" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> input_keep_prob: unit Tensor or float between 0 and 1, input keep</span></td> </tr> <tr> <td id="L870" class="blob-num js-line-number" data-line-number="870"></td> <td id="LC870" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> probability; if it is constant and 1, no input dropout will be added.</span></td> </tr> <tr> <td id="L871" class="blob-num js-line-number" data-line-number="871"></td> <td id="LC871" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> output_keep_prob: unit Tensor or float between 0 and 1, output keep</span></td> </tr> <tr> <td id="L872" class="blob-num js-line-number" data-line-number="872"></td> <td id="LC872" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> probability; if it is constant and 1, no output dropout will be added.</span></td> </tr> <tr> <td id="L873" class="blob-num js-line-number" data-line-number="873"></td> <td id="LC873" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state_keep_prob: unit Tensor or float between 0 and 1, output keep</span></td> </tr> <tr> <td id="L874" class="blob-num js-line-number" data-line-number="874"></td> <td id="LC874" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> probability; if it is constant and 1, no output dropout will be added.</span></td> </tr> <tr> <td id="L875" class="blob-num js-line-number" data-line-number="875"></td> <td id="LC875" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> State dropout is performed on the outgoing states of the cell.</span></td> </tr> <tr> <td id="L876" class="blob-num js-line-number" data-line-number="876"></td> <td id="LC876" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> **Note** the state components to which dropout is applied when</span></td> </tr> <tr> <td id="L877" class="blob-num js-line-number" data-line-number="877"></td> <td id="LC877" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `state_keep_prob` is in `(0, 1)` are also determined by</span></td> </tr> <tr> <td id="L878" class="blob-num js-line-number" data-line-number="878"></td> <td id="LC878" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the argument `dropout_state_filter_visitor` (e.g. by default dropout</span></td> </tr> <tr> <td id="L879" class="blob-num js-line-number" data-line-number="879"></td> <td id="LC879" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> is never applied to the `c` component of an `LSTMStateTuple`).</span></td> </tr> <tr> <td id="L880" class="blob-num js-line-number" data-line-number="880"></td> <td id="LC880" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> variational_recurrent: Python bool. If `True`, then the same</span></td> </tr> <tr> <td id="L881" class="blob-num js-line-number" data-line-number="881"></td> <td id="LC881" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> dropout pattern is applied across all time steps per run call.</span></td> </tr> <tr> <td id="L882" class="blob-num js-line-number" data-line-number="882"></td> <td id="LC882" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> If this parameter is set, `input_size` **must** be provided.</span></td> </tr> <tr> <td id="L883" class="blob-num js-line-number" data-line-number="883"></td> <td id="LC883" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> input_size: (optional) (possibly nested tuple of) `TensorShape` objects</span></td> </tr> <tr> <td id="L884" class="blob-num js-line-number" data-line-number="884"></td> <td id="LC884" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> containing the depth(s) of the input tensors expected to be passed in to</span></td> </tr> <tr> <td id="L885" class="blob-num js-line-number" data-line-number="885"></td> <td id="LC885" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the `DropoutWrapper`. Required and used **iff**</span></td> </tr> <tr> <td id="L886" class="blob-num js-line-number" data-line-number="886"></td> <td id="LC886" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `variational_recurrent = True` and `input_keep_prob &lt; 1`.</span></td> </tr> <tr> <td id="L887" class="blob-num js-line-number" data-line-number="887"></td> <td id="LC887" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> dtype: (optional) The `dtype` of the input, state, and output tensors.</span></td> </tr> <tr> <td id="L888" class="blob-num js-line-number" data-line-number="888"></td> <td id="LC888" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Required and used **iff** `variational_recurrent = True`.</span></td> </tr> <tr> <td id="L889" class="blob-num js-line-number" data-line-number="889"></td> <td id="LC889" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> seed: (optional) integer, the randomness seed.</span></td> </tr> <tr> <td id="L890" class="blob-num js-line-number" data-line-number="890"></td> <td id="LC890" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> dropout_state_filter_visitor: (optional), default: (see below). Function</span></td> </tr> <tr> <td id="L891" class="blob-num js-line-number" data-line-number="891"></td> <td id="LC891" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> that takes any hierarchical level of the state and returns</span></td> </tr> <tr> <td id="L892" class="blob-num js-line-number" data-line-number="892"></td> <td id="LC892" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> a scalar or depth=1 structure of Python booleans describing</span></td> </tr> <tr> <td id="L893" class="blob-num js-line-number" data-line-number="893"></td> <td id="LC893" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> which terms in the state should be dropped out. In addition, if the</span></td> </tr> <tr> <td id="L894" class="blob-num js-line-number" data-line-number="894"></td> <td id="LC894" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> function returns `True`, dropout is applied across this sublevel. If</span></td> </tr> <tr> <td id="L895" class="blob-num js-line-number" data-line-number="895"></td> <td id="LC895" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> the function returns `False`, dropout is not applied across this entire</span></td> </tr> <tr> <td id="L896" class="blob-num js-line-number" data-line-number="896"></td> <td id="LC896" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> sublevel.</span></td> </tr> <tr> <td id="L897" class="blob-num js-line-number" data-line-number="897"></td> <td id="LC897" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Default behavior: perform dropout on all terms except the memory (`c`)</span></td> </tr> <tr> <td id="L898" class="blob-num js-line-number" data-line-number="898"></td> <td id="LC898" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state of `LSTMCellState` objects, and don&#39;t try to apply dropout to</span></td> </tr> <tr> <td id="L899" class="blob-num js-line-number" data-line-number="899"></td> <td id="LC899" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `TensorArray` objects:</span></td> </tr> <tr> <td id="L900" class="blob-num js-line-number" data-line-number="900"></td> <td id="LC900" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ```</span></td> </tr> <tr> <td id="L901" class="blob-num js-line-number" data-line-number="901"></td> <td id="LC901" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> def dropout_state_filter_visitor(s):</span></td> </tr> <tr> <td id="L902" class="blob-num js-line-number" data-line-number="902"></td> <td id="LC902" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> if isinstance(s, LSTMCellState):</span></td> </tr> <tr> <td id="L903" class="blob-num js-line-number" data-line-number="903"></td> <td id="LC903" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> # Never perform dropout on the c state.</span></td> </tr> <tr> <td id="L904" class="blob-num js-line-number" data-line-number="904"></td> <td id="LC904" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> return LSTMCellState(c=False, h=True)</span></td> </tr> <tr> <td id="L905" class="blob-num js-line-number" data-line-number="905"></td> <td id="LC905" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> elif isinstance(s, TensorArray):</span></td> </tr> <tr> <td id="L906" class="blob-num js-line-number" data-line-number="906"></td> <td id="LC906" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> return False</span></td> </tr> <tr> <td id="L907" class="blob-num js-line-number" data-line-number="907"></td> <td id="LC907" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> return True</span></td> </tr> <tr> <td id="L908" class="blob-num js-line-number" data-line-number="908"></td> <td id="LC908" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ```</span></td> </tr> <tr> <td id="L909" class="blob-num js-line-number" data-line-number="909"></td> <td id="LC909" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L910" class="blob-num js-line-number" data-line-number="910"></td> <td id="LC910" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L911" class="blob-num js-line-number" data-line-number="911"></td> <td id="LC911" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> TypeError: if `cell` is not an `RNNCell`, or `keep_state_fn` is provided</span></td> </tr> <tr> <td id="L912" class="blob-num js-line-number" data-line-number="912"></td> <td id="LC912" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> but not `callable`.</span></td> </tr> <tr> <td id="L913" class="blob-num js-line-number" data-line-number="913"></td> <td id="LC913" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if any of the keep_probs are not between 0 and 1.</span></td> </tr> <tr> <td id="L914" class="blob-num js-line-number" data-line-number="914"></td> <td id="LC914" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L915" class="blob-num js-line-number" data-line-number="915"></td> <td id="LC915" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> _like_rnncell(cell):</td> </tr> <tr> <td id="L916" class="blob-num js-line-number" data-line-number="916"></td> <td id="LC916" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>The parameter cell is not a RNNCell.<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L917" class="blob-num js-line-number" data-line-number="917"></td> <td id="LC917" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> (dropout_state_filter_visitor <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L918" class="blob-num js-line-number" data-line-number="918"></td> <td id="LC918" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">and</span> <span class="pl-k">not</span> <span class="pl-c1">callable</span>(dropout_state_filter_visitor)):</td> </tr> <tr> <td id="L919" class="blob-num js-line-number" data-line-number="919"></td> <td id="LC919" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>dropout_state_filter_visitor must be callable<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L920" class="blob-num js-line-number" data-line-number="920"></td> <td id="LC920" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._dropout_state_filter <span class="pl-k">=</span> (</td> </tr> <tr> <td id="L921" class="blob-num js-line-number" data-line-number="921"></td> <td id="LC921" class="blob-code blob-code-inner js-file-line"> dropout_state_filter_visitor <span class="pl-k">or</span> _default_dropout_state_filter_visitor)</td> </tr> <tr> <td id="L922" class="blob-num js-line-number" data-line-number="922"></td> <td id="LC922" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-s"><span class="pl-pds">&quot;</span>DropoutWrapperInit<span class="pl-pds">&quot;</span></span>):</td> </tr> <tr> <td id="L923" class="blob-num js-line-number" data-line-number="923"></td> <td id="LC923" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">tensor_and_const_value</span>(<span class="pl-smi">v</span>):</td> </tr> <tr> <td id="L924" class="blob-num js-line-number" data-line-number="924"></td> <td id="LC924" class="blob-code blob-code-inner js-file-line"> tensor_value <span class="pl-k">=</span> ops.convert_to_tensor(v)</td> </tr> <tr> <td id="L925" class="blob-num js-line-number" data-line-number="925"></td> <td id="LC925" class="blob-code blob-code-inner js-file-line"> const_value <span class="pl-k">=</span> tensor_util.constant_value(tensor_value)</td> </tr> <tr> <td id="L926" class="blob-num js-line-number" data-line-number="926"></td> <td id="LC926" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> (tensor_value, const_value)</td> </tr> <tr> <td id="L927" class="blob-num js-line-number" data-line-number="927"></td> <td id="LC927" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">for</span> prob, attr <span class="pl-k">in</span> [(input_keep_prob, <span class="pl-s"><span class="pl-pds">&quot;</span>input_keep_prob<span class="pl-pds">&quot;</span></span>),</td> </tr> <tr> <td id="L928" class="blob-num js-line-number" data-line-number="928"></td> <td id="LC928" class="blob-code blob-code-inner js-file-line"> (state_keep_prob, <span class="pl-s"><span class="pl-pds">&quot;</span>state_keep_prob<span class="pl-pds">&quot;</span></span>),</td> </tr> <tr> <td id="L929" class="blob-num js-line-number" data-line-number="929"></td> <td id="LC929" class="blob-code blob-code-inner js-file-line"> (output_keep_prob, <span class="pl-s"><span class="pl-pds">&quot;</span>output_keep_prob<span class="pl-pds">&quot;</span></span>)]:</td> </tr> <tr> <td id="L930" class="blob-num js-line-number" data-line-number="930"></td> <td id="LC930" class="blob-code blob-code-inner js-file-line"> tensor_prob, const_prob <span class="pl-k">=</span> tensor_and_const_value(prob)</td> </tr> <tr> <td id="L931" class="blob-num js-line-number" data-line-number="931"></td> <td id="LC931" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> const_prob <span class="pl-k">is</span> <span class="pl-k">not</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L932" class="blob-num js-line-number" data-line-number="932"></td> <td id="LC932" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> const_prob <span class="pl-k">&lt;</span> <span class="pl-c1">0</span> <span class="pl-k">or</span> const_prob <span class="pl-k">&gt;</span> <span class="pl-c1">1</span>:</td> </tr> <tr> <td id="L933" class="blob-num js-line-number" data-line-number="933"></td> <td id="LC933" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Parameter <span class="pl-c1">%s</span> must be between 0 and 1: <span class="pl-c1">%d</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L934" class="blob-num js-line-number" data-line-number="934"></td> <td id="LC934" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> (attr, const_prob))</td> </tr> <tr> <td id="L935" class="blob-num js-line-number" data-line-number="935"></td> <td id="LC935" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">setattr</span>(<span class="pl-c1">self</span>, <span class="pl-s"><span class="pl-pds">&quot;</span>_<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> attr, <span class="pl-c1">float</span>(const_prob))</td> </tr> <tr> <td id="L936" class="blob-num js-line-number" data-line-number="936"></td> <td id="LC936" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L937" class="blob-num js-line-number" data-line-number="937"></td> <td id="LC937" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">setattr</span>(<span class="pl-c1">self</span>, <span class="pl-s"><span class="pl-pds">&quot;</span>_<span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> attr, tensor_prob)</td> </tr> <tr> <td id="L938" class="blob-num js-line-number" data-line-number="938"></td> <td id="LC938" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L939" class="blob-num js-line-number" data-line-number="939"></td> <td id="LC939" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Set cell, variational_recurrent, seed before running the code below</span></td> </tr> <tr> <td id="L940" class="blob-num js-line-number" data-line-number="940"></td> <td id="LC940" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell <span class="pl-k">=</span> cell</td> </tr> <tr> <td id="L941" class="blob-num js-line-number" data-line-number="941"></td> <td id="LC941" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._variational_recurrent <span class="pl-k">=</span> variational_recurrent</td> </tr> <tr> <td id="L942" class="blob-num js-line-number" data-line-number="942"></td> <td id="LC942" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._seed <span class="pl-k">=</span> seed</td> </tr> <tr> <td id="L943" class="blob-num js-line-number" data-line-number="943"></td> <td id="LC943" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L944" class="blob-num js-line-number" data-line-number="944"></td> <td id="LC944" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_input_noise <span class="pl-k">=</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L945" class="blob-num js-line-number" data-line-number="945"></td> <td id="LC945" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_state_noise <span class="pl-k">=</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L946" class="blob-num js-line-number" data-line-number="946"></td> <td id="LC946" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_output_noise <span class="pl-k">=</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L947" class="blob-num js-line-number" data-line-number="947"></td> <td id="LC947" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L948" class="blob-num js-line-number" data-line-number="948"></td> <td id="LC948" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> variational_recurrent:</td> </tr> <tr> <td id="L949" class="blob-num js-line-number" data-line-number="949"></td> <td id="LC949" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> dtype <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L950" class="blob-num js-line-number" data-line-number="950"></td> <td id="LC950" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(</td> </tr> <tr> <td id="L951" class="blob-num js-line-number" data-line-number="951"></td> <td id="LC951" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>When variational_recurrent=True, dtype must be provided<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L952" class="blob-num js-line-number" data-line-number="952"></td> <td id="LC952" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L953" class="blob-num js-line-number" data-line-number="953"></td> <td id="LC953" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">convert_to_batch_shape</span>(<span class="pl-smi">s</span>):</td> </tr> <tr> <td id="L954" class="blob-num js-line-number" data-line-number="954"></td> <td id="LC954" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Prepend a 1 for the batch dimension; for recurrent</span></td> </tr> <tr> <td id="L955" class="blob-num js-line-number" data-line-number="955"></td> <td id="LC955" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> variational dropout we use the same dropout mask for all</span></td> </tr> <tr> <td id="L956" class="blob-num js-line-number" data-line-number="956"></td> <td id="LC956" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> batch elements.</span></td> </tr> <tr> <td id="L957" class="blob-num js-line-number" data-line-number="957"></td> <td id="LC957" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> array_ops.concat(</td> </tr> <tr> <td id="L958" class="blob-num js-line-number" data-line-number="958"></td> <td id="LC958" class="blob-code blob-code-inner js-file-line"> ([<span class="pl-c1">1</span>], tensor_shape.TensorShape(s).as_list()), <span class="pl-c1">0</span>)</td> </tr> <tr> <td id="L959" class="blob-num js-line-number" data-line-number="959"></td> <td id="LC959" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L960" class="blob-num js-line-number" data-line-number="960"></td> <td id="LC960" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">batch_noise</span>(<span class="pl-smi">s</span>, <span class="pl-smi">inner_seed</span>):</td> </tr> <tr> <td id="L961" class="blob-num js-line-number" data-line-number="961"></td> <td id="LC961" class="blob-code blob-code-inner js-file-line"> shape <span class="pl-k">=</span> convert_to_batch_shape(s)</td> </tr> <tr> <td id="L962" class="blob-num js-line-number" data-line-number="962"></td> <td id="LC962" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> random_ops.random_uniform(shape, <span class="pl-v">seed</span><span class="pl-k">=</span>inner_seed, <span class="pl-v">dtype</span><span class="pl-k">=</span>dtype)</td> </tr> <tr> <td id="L963" class="blob-num js-line-number" data-line-number="963"></td> <td id="LC963" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L964" class="blob-num js-line-number" data-line-number="964"></td> <td id="LC964" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> (<span class="pl-k">not</span> <span class="pl-c1">isinstance</span>(<span class="pl-c1">self</span>._input_keep_prob, numbers.Real) <span class="pl-k">or</span></td> </tr> <tr> <td id="L965" class="blob-num js-line-number" data-line-number="965"></td> <td id="LC965" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._input_keep_prob <span class="pl-k">&lt;</span> <span class="pl-c1">1.0</span>):</td> </tr> <tr> <td id="L966" class="blob-num js-line-number" data-line-number="966"></td> <td id="LC966" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> input_size <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L967" class="blob-num js-line-number" data-line-number="967"></td> <td id="LC967" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(</td> </tr> <tr> <td id="L968" class="blob-num js-line-number" data-line-number="968"></td> <td id="LC968" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>When variational_recurrent=True and input_keep_prob &lt; 1.0 or <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L969" class="blob-num js-line-number" data-line-number="969"></td> <td id="LC969" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>is unknown, input_size must be provided<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L970" class="blob-num js-line-number" data-line-number="970"></td> <td id="LC970" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_input_noise <span class="pl-k">=</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L971" class="blob-num js-line-number" data-line-number="971"></td> <td id="LC971" class="blob-code blob-code-inner js-file-line"> input_size,</td> </tr> <tr> <td id="L972" class="blob-num js-line-number" data-line-number="972"></td> <td id="LC972" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">lambda</span> <span class="pl-smi">i</span>, <span class="pl-smi">s</span>: batch_noise(s, <span class="pl-v">inner_seed</span><span class="pl-k">=</span><span class="pl-c1">self</span>._gen_seed(<span class="pl-s"><span class="pl-pds">&quot;</span>input<span class="pl-pds">&quot;</span></span>, i)),</td> </tr> <tr> <td id="L973" class="blob-num js-line-number" data-line-number="973"></td> <td id="LC973" class="blob-code blob-code-inner js-file-line"> input_size)</td> </tr> <tr> <td id="L974" class="blob-num js-line-number" data-line-number="974"></td> <td id="LC974" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_state_noise <span class="pl-k">=</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L975" class="blob-num js-line-number" data-line-number="975"></td> <td id="LC975" class="blob-code blob-code-inner js-file-line"> cell.state_size,</td> </tr> <tr> <td id="L976" class="blob-num js-line-number" data-line-number="976"></td> <td id="LC976" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">lambda</span> <span class="pl-smi">i</span>, <span class="pl-smi">s</span>: batch_noise(s, <span class="pl-v">inner_seed</span><span class="pl-k">=</span><span class="pl-c1">self</span>._gen_seed(<span class="pl-s"><span class="pl-pds">&quot;</span>state<span class="pl-pds">&quot;</span></span>, i)),</td> </tr> <tr> <td id="L977" class="blob-num js-line-number" data-line-number="977"></td> <td id="LC977" class="blob-code blob-code-inner js-file-line"> cell.state_size)</td> </tr> <tr> <td id="L978" class="blob-num js-line-number" data-line-number="978"></td> <td id="LC978" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_output_noise <span class="pl-k">=</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L979" class="blob-num js-line-number" data-line-number="979"></td> <td id="LC979" class="blob-code blob-code-inner js-file-line"> cell.output_size,</td> </tr> <tr> <td id="L980" class="blob-num js-line-number" data-line-number="980"></td> <td id="LC980" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">lambda</span> <span class="pl-smi">i</span>, <span class="pl-smi">s</span>: batch_noise(s, <span class="pl-v">inner_seed</span><span class="pl-k">=</span><span class="pl-c1">self</span>._gen_seed(<span class="pl-s"><span class="pl-pds">&quot;</span>output<span class="pl-pds">&quot;</span></span>, i)),</td> </tr> <tr> <td id="L981" class="blob-num js-line-number" data-line-number="981"></td> <td id="LC981" class="blob-code blob-code-inner js-file-line"> cell.output_size)</td> </tr> <tr> <td id="L982" class="blob-num js-line-number" data-line-number="982"></td> <td id="LC982" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L983" class="blob-num js-line-number" data-line-number="983"></td> <td id="LC983" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_gen_seed</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">salt_prefix</span>, <span class="pl-smi">index</span>):</td> </tr> <tr> <td id="L984" class="blob-num js-line-number" data-line-number="984"></td> <td id="LC984" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._seed <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L985" class="blob-num js-line-number" data-line-number="985"></td> <td id="LC985" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">None</span></td> </tr> <tr> <td id="L986" class="blob-num js-line-number" data-line-number="986"></td> <td id="LC986" class="blob-code blob-code-inner js-file-line"> salt <span class="pl-k">=</span> <span class="pl-s"><span class="pl-pds">&quot;</span><span class="pl-c1">%s</span>_<span class="pl-c1">%d</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> (salt_prefix, index)</td> </tr> <tr> <td id="L987" class="blob-num js-line-number" data-line-number="987"></td> <td id="LC987" class="blob-code blob-code-inner js-file-line"> string <span class="pl-k">=</span> (<span class="pl-c1">str</span>(<span class="pl-c1">self</span>._seed) <span class="pl-k">+</span> salt).encode(<span class="pl-s"><span class="pl-pds">&quot;</span>utf-8<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L988" class="blob-num js-line-number" data-line-number="988"></td> <td id="LC988" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">int</span>(hashlib.md5(string).hexdigest()[:<span class="pl-c1">8</span>], <span class="pl-c1">16</span>) <span class="pl-k">&amp;</span> <span class="pl-c1"><span class="pl-k">0x</span>7FFFFFFF</span></td> </tr> <tr> <td id="L989" class="blob-num js-line-number" data-line-number="989"></td> <td id="LC989" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L990" class="blob-num js-line-number" data-line-number="990"></td> <td id="LC990" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L991" class="blob-num js-line-number" data-line-number="991"></td> <td id="LC991" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">wrapped_cell</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L992" class="blob-num js-line-number" data-line-number="992"></td> <td id="LC992" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell</td> </tr> <tr> <td id="L993" class="blob-num js-line-number" data-line-number="993"></td> <td id="LC993" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L994" class="blob-num js-line-number" data-line-number="994"></td> <td id="LC994" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L995" class="blob-num js-line-number" data-line-number="995"></td> <td id="LC995" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L996" class="blob-num js-line-number" data-line-number="996"></td> <td id="LC996" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.state_size</td> </tr> <tr> <td id="L997" class="blob-num js-line-number" data-line-number="997"></td> <td id="LC997" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L998" class="blob-num js-line-number" data-line-number="998"></td> <td id="LC998" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L999" class="blob-num js-line-number" data-line-number="999"></td> <td id="LC999" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1000" class="blob-num js-line-number" data-line-number="1000"></td> <td id="LC1000" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.output_size</td> </tr> <tr> <td id="L1001" class="blob-num js-line-number" data-line-number="1001"></td> <td id="LC1001" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1002" class="blob-num js-line-number" data-line-number="1002"></td> <td id="LC1002" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L1003" class="blob-num js-line-number" data-line-number="1003"></td> <td id="LC1003" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L1004" class="blob-num js-line-number" data-line-number="1004"></td> <td id="LC1004" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.zero_state(batch_size, dtype)</td> </tr> <tr> <td id="L1005" class="blob-num js-line-number" data-line-number="1005"></td> <td id="LC1005" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1006" class="blob-num js-line-number" data-line-number="1006"></td> <td id="LC1006" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_variational_recurrent_dropout_value</span>(</td> </tr> <tr> <td id="L1007" class="blob-num js-line-number" data-line-number="1007"></td> <td id="LC1007" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">index</span>, <span class="pl-smi">value</span>, <span class="pl-smi">noise</span>, <span class="pl-smi">keep_prob</span>):</td> </tr> <tr> <td id="L1008" class="blob-num js-line-number" data-line-number="1008"></td> <td id="LC1008" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Performs dropout given the pre-calculated noise tensor.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1009" class="blob-num js-line-number" data-line-number="1009"></td> <td id="LC1009" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> uniform [keep_prob, 1.0 + keep_prob)</span></td> </tr> <tr> <td id="L1010" class="blob-num js-line-number" data-line-number="1010"></td> <td id="LC1010" class="blob-code blob-code-inner js-file-line"> random_tensor <span class="pl-k">=</span> keep_prob <span class="pl-k">+</span> noise</td> </tr> <tr> <td id="L1011" class="blob-num js-line-number" data-line-number="1011"></td> <td id="LC1011" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1012" class="blob-num js-line-number" data-line-number="1012"></td> <td id="LC1012" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> 0. if [keep_prob, 1.0) and 1. if [1.0, 1.0 + keep_prob)</span></td> </tr> <tr> <td id="L1013" class="blob-num js-line-number" data-line-number="1013"></td> <td id="LC1013" class="blob-code blob-code-inner js-file-line"> binary_tensor <span class="pl-k">=</span> math_ops.floor(random_tensor)</td> </tr> <tr> <td id="L1014" class="blob-num js-line-number" data-line-number="1014"></td> <td id="LC1014" class="blob-code blob-code-inner js-file-line"> ret <span class="pl-k">=</span> math_ops.div(value, keep_prob) <span class="pl-k">*</span> binary_tensor</td> </tr> <tr> <td id="L1015" class="blob-num js-line-number" data-line-number="1015"></td> <td id="LC1015" class="blob-code blob-code-inner js-file-line"> ret.set_shape(value.get_shape())</td> </tr> <tr> <td id="L1016" class="blob-num js-line-number" data-line-number="1016"></td> <td id="LC1016" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> ret</td> </tr> <tr> <td id="L1017" class="blob-num js-line-number" data-line-number="1017"></td> <td id="LC1017" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1018" class="blob-num js-line-number" data-line-number="1018"></td> <td id="LC1018" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_dropout</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">values</span>, <span class="pl-smi">salt_prefix</span>, <span class="pl-smi">recurrent_noise</span>, <span class="pl-smi">keep_prob</span>,</td> </tr> <tr> <td id="L1019" class="blob-num js-line-number" data-line-number="1019"></td> <td id="LC1019" class="blob-code blob-code-inner js-file-line"> <span class="pl-smi">shallow_filtered_substructure</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1020" class="blob-num js-line-number" data-line-number="1020"></td> <td id="LC1020" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Decides whether to perform standard dropout or recurrent dropout.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1021" class="blob-num js-line-number" data-line-number="1021"></td> <td id="LC1021" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1022" class="blob-num js-line-number" data-line-number="1022"></td> <td id="LC1022" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> shallow_filtered_substructure <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L1023" class="blob-num js-line-number" data-line-number="1023"></td> <td id="LC1023" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Put something so we traverse the entire structure; inside the</span></td> </tr> <tr> <td id="L1024" class="blob-num js-line-number" data-line-number="1024"></td> <td id="LC1024" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> dropout function we check to see if leafs of this are bool or not.</span></td> </tr> <tr> <td id="L1025" class="blob-num js-line-number" data-line-number="1025"></td> <td id="LC1025" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure <span class="pl-k">=</span> values</td> </tr> <tr> <td id="L1026" class="blob-num js-line-number" data-line-number="1026"></td> <td id="LC1026" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1027" class="blob-num js-line-number" data-line-number="1027"></td> <td id="LC1027" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">self</span>._variational_recurrent:</td> </tr> <tr> <td id="L1028" class="blob-num js-line-number" data-line-number="1028"></td> <td id="LC1028" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">dropout</span>(<span class="pl-smi">i</span>, <span class="pl-smi">do_dropout</span>, <span class="pl-smi">v</span>):</td> </tr> <tr> <td id="L1029" class="blob-num js-line-number" data-line-number="1029"></td> <td id="LC1029" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">isinstance</span>(do_dropout, <span class="pl-c1">bool</span>) <span class="pl-k">or</span> do_dropout:</td> </tr> <tr> <td id="L1030" class="blob-num js-line-number" data-line-number="1030"></td> <td id="LC1030" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> nn_ops.dropout(</td> </tr> <tr> <td id="L1031" class="blob-num js-line-number" data-line-number="1031"></td> <td id="LC1031" class="blob-code blob-code-inner js-file-line"> v, <span class="pl-v">keep_prob</span><span class="pl-k">=</span>keep_prob, <span class="pl-v">seed</span><span class="pl-k">=</span><span class="pl-c1">self</span>._gen_seed(salt_prefix, i))</td> </tr> <tr> <td id="L1032" class="blob-num js-line-number" data-line-number="1032"></td> <td id="LC1032" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1033" class="blob-num js-line-number" data-line-number="1033"></td> <td id="LC1033" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> v</td> </tr> <tr> <td id="L1034" class="blob-num js-line-number" data-line-number="1034"></td> <td id="LC1034" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L1035" class="blob-num js-line-number" data-line-number="1035"></td> <td id="LC1035" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure, dropout,</td> </tr> <tr> <td id="L1036" class="blob-num js-line-number" data-line-number="1036"></td> <td id="LC1036" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">*</span>[shallow_filtered_substructure, values])</td> </tr> <tr> <td id="L1037" class="blob-num js-line-number" data-line-number="1037"></td> <td id="LC1037" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1038" class="blob-num js-line-number" data-line-number="1038"></td> <td id="LC1038" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">dropout</span>(<span class="pl-smi">i</span>, <span class="pl-smi">do_dropout</span>, <span class="pl-smi">v</span>, <span class="pl-smi">n</span>):</td> </tr> <tr> <td id="L1039" class="blob-num js-line-number" data-line-number="1039"></td> <td id="LC1039" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">isinstance</span>(do_dropout, <span class="pl-c1">bool</span>) <span class="pl-k">or</span> do_dropout:</td> </tr> <tr> <td id="L1040" class="blob-num js-line-number" data-line-number="1040"></td> <td id="LC1040" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._variational_recurrent_dropout_value(i, v, n, keep_prob)</td> </tr> <tr> <td id="L1041" class="blob-num js-line-number" data-line-number="1041"></td> <td id="LC1041" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1042" class="blob-num js-line-number" data-line-number="1042"></td> <td id="LC1042" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> v</td> </tr> <tr> <td id="L1043" class="blob-num js-line-number" data-line-number="1043"></td> <td id="LC1043" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> _enumerated_map_structure_up_to(</td> </tr> <tr> <td id="L1044" class="blob-num js-line-number" data-line-number="1044"></td> <td id="LC1044" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure, dropout,</td> </tr> <tr> <td id="L1045" class="blob-num js-line-number" data-line-number="1045"></td> <td id="LC1045" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">*</span>[shallow_filtered_substructure, values, recurrent_noise])</td> </tr> <tr> <td id="L1046" class="blob-num js-line-number" data-line-number="1046"></td> <td id="LC1046" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1047" class="blob-num js-line-number" data-line-number="1047"></td> <td id="LC1047" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1048" class="blob-num js-line-number" data-line-number="1048"></td> <td id="LC1048" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run the cell with the declared dropouts.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1049" class="blob-num js-line-number" data-line-number="1049"></td> <td id="LC1049" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">_should_dropout</span>(<span class="pl-smi">p</span>):</td> </tr> <tr> <td id="L1050" class="blob-num js-line-number" data-line-number="1050"></td> <td id="LC1050" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> (<span class="pl-k">not</span> <span class="pl-c1">isinstance</span>(p, <span class="pl-c1">float</span>)) <span class="pl-k">or</span> p <span class="pl-k">&lt;</span> <span class="pl-c1">1</span></td> </tr> <tr> <td id="L1051" class="blob-num js-line-number" data-line-number="1051"></td> <td id="LC1051" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1052" class="blob-num js-line-number" data-line-number="1052"></td> <td id="LC1052" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> _should_dropout(<span class="pl-c1">self</span>._input_keep_prob):</td> </tr> <tr> <td id="L1053" class="blob-num js-line-number" data-line-number="1053"></td> <td id="LC1053" class="blob-code blob-code-inner js-file-line"> inputs <span class="pl-k">=</span> <span class="pl-c1">self</span>._dropout(inputs, <span class="pl-s"><span class="pl-pds">&quot;</span>input<span class="pl-pds">&quot;</span></span>,</td> </tr> <tr> <td id="L1054" class="blob-num js-line-number" data-line-number="1054"></td> <td id="LC1054" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_input_noise,</td> </tr> <tr> <td id="L1055" class="blob-num js-line-number" data-line-number="1055"></td> <td id="LC1055" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._input_keep_prob)</td> </tr> <tr> <td id="L1056" class="blob-num js-line-number" data-line-number="1056"></td> <td id="LC1056" class="blob-code blob-code-inner js-file-line"> output, new_state <span class="pl-k">=</span> <span class="pl-c1">self</span>._cell(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L1057" class="blob-num js-line-number" data-line-number="1057"></td> <td id="LC1057" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> _should_dropout(<span class="pl-c1">self</span>._state_keep_prob):</td> </tr> <tr> <td id="L1058" class="blob-num js-line-number" data-line-number="1058"></td> <td id="LC1058" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Identify which subsets of the state to perform dropout on and</span></td> </tr> <tr> <td id="L1059" class="blob-num js-line-number" data-line-number="1059"></td> <td id="LC1059" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> which ones to keep.</span></td> </tr> <tr> <td id="L1060" class="blob-num js-line-number" data-line-number="1060"></td> <td id="LC1060" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure <span class="pl-k">=</span> nest.get_traverse_shallow_structure(</td> </tr> <tr> <td id="L1061" class="blob-num js-line-number" data-line-number="1061"></td> <td id="LC1061" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._dropout_state_filter, new_state)</td> </tr> <tr> <td id="L1062" class="blob-num js-line-number" data-line-number="1062"></td> <td id="LC1062" class="blob-code blob-code-inner js-file-line"> new_state <span class="pl-k">=</span> <span class="pl-c1">self</span>._dropout(new_state, <span class="pl-s"><span class="pl-pds">&quot;</span>state<span class="pl-pds">&quot;</span></span>,</td> </tr> <tr> <td id="L1063" class="blob-num js-line-number" data-line-number="1063"></td> <td id="LC1063" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_state_noise,</td> </tr> <tr> <td id="L1064" class="blob-num js-line-number" data-line-number="1064"></td> <td id="LC1064" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_keep_prob,</td> </tr> <tr> <td id="L1065" class="blob-num js-line-number" data-line-number="1065"></td> <td id="LC1065" class="blob-code blob-code-inner js-file-line"> shallow_filtered_substructure)</td> </tr> <tr> <td id="L1066" class="blob-num js-line-number" data-line-number="1066"></td> <td id="LC1066" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> _should_dropout(<span class="pl-c1">self</span>._output_keep_prob):</td> </tr> <tr> <td id="L1067" class="blob-num js-line-number" data-line-number="1067"></td> <td id="LC1067" class="blob-code blob-code-inner js-file-line"> output <span class="pl-k">=</span> <span class="pl-c1">self</span>._dropout(output, <span class="pl-s"><span class="pl-pds">&quot;</span>output<span class="pl-pds">&quot;</span></span>,</td> </tr> <tr> <td id="L1068" class="blob-num js-line-number" data-line-number="1068"></td> <td id="LC1068" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._recurrent_output_noise,</td> </tr> <tr> <td id="L1069" class="blob-num js-line-number" data-line-number="1069"></td> <td id="LC1069" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._output_keep_prob)</td> </tr> <tr> <td id="L1070" class="blob-num js-line-number" data-line-number="1070"></td> <td id="LC1070" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> output, new_state</td> </tr> <tr> <td id="L1071" class="blob-num js-line-number" data-line-number="1071"></td> <td id="LC1071" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1072" class="blob-num js-line-number" data-line-number="1072"></td> <td id="LC1072" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1073" class="blob-num js-line-number" data-line-number="1073"></td> <td id="LC1073" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.ResidualWrapper<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L1074" class="blob-num js-line-number" data-line-number="1074"></td> <td id="LC1074" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">ResidualWrapper</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L1075" class="blob-num js-line-number" data-line-number="1075"></td> <td id="LC1075" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>RNNCell wrapper that ensures cell inputs are added to the outputs.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1076" class="blob-num js-line-number" data-line-number="1076"></td> <td id="LC1076" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1077" class="blob-num js-line-number" data-line-number="1077"></td> <td id="LC1077" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cell</span>, <span class="pl-smi">residual_fn</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1078" class="blob-num js-line-number" data-line-number="1078"></td> <td id="LC1078" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Constructs a `ResidualWrapper` for `cell`.</span></td> </tr> <tr> <td id="L1079" class="blob-num js-line-number" data-line-number="1079"></td> <td id="LC1079" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1080" class="blob-num js-line-number" data-line-number="1080"></td> <td id="LC1080" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1081" class="blob-num js-line-number" data-line-number="1081"></td> <td id="LC1081" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell: An instance of `RNNCell`.</span></td> </tr> <tr> <td id="L1082" class="blob-num js-line-number" data-line-number="1082"></td> <td id="LC1082" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> residual_fn: (Optional) The function to map raw cell inputs and raw cell</span></td> </tr> <tr> <td id="L1083" class="blob-num js-line-number" data-line-number="1083"></td> <td id="LC1083" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> outputs to the actual cell outputs of the residual network.</span></td> </tr> <tr> <td id="L1084" class="blob-num js-line-number" data-line-number="1084"></td> <td id="LC1084" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Defaults to calling nest.map_structure on (lambda i, o: i + o), inputs</span></td> </tr> <tr> <td id="L1085" class="blob-num js-line-number" data-line-number="1085"></td> <td id="LC1085" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> and outputs.</span></td> </tr> <tr> <td id="L1086" class="blob-num js-line-number" data-line-number="1086"></td> <td id="LC1086" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1087" class="blob-num js-line-number" data-line-number="1087"></td> <td id="LC1087" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell <span class="pl-k">=</span> cell</td> </tr> <tr> <td id="L1088" class="blob-num js-line-number" data-line-number="1088"></td> <td id="LC1088" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._residual_fn <span class="pl-k">=</span> residual_fn</td> </tr> <tr> <td id="L1089" class="blob-num js-line-number" data-line-number="1089"></td> <td id="LC1089" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1090" class="blob-num js-line-number" data-line-number="1090"></td> <td id="LC1090" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1091" class="blob-num js-line-number" data-line-number="1091"></td> <td id="LC1091" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1092" class="blob-num js-line-number" data-line-number="1092"></td> <td id="LC1092" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.state_size</td> </tr> <tr> <td id="L1093" class="blob-num js-line-number" data-line-number="1093"></td> <td id="LC1093" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1094" class="blob-num js-line-number" data-line-number="1094"></td> <td id="LC1094" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1095" class="blob-num js-line-number" data-line-number="1095"></td> <td id="LC1095" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1096" class="blob-num js-line-number" data-line-number="1096"></td> <td id="LC1096" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.output_size</td> </tr> <tr> <td id="L1097" class="blob-num js-line-number" data-line-number="1097"></td> <td id="LC1097" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1098" class="blob-num js-line-number" data-line-number="1098"></td> <td id="LC1098" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L1099" class="blob-num js-line-number" data-line-number="1099"></td> <td id="LC1099" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L1100" class="blob-num js-line-number" data-line-number="1100"></td> <td id="LC1100" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.zero_state(batch_size, dtype)</td> </tr> <tr> <td id="L1101" class="blob-num js-line-number" data-line-number="1101"></td> <td id="LC1101" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1102" class="blob-num js-line-number" data-line-number="1102"></td> <td id="LC1102" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1103" class="blob-num js-line-number" data-line-number="1103"></td> <td id="LC1103" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run the cell and then apply the residual_fn on its inputs to its outputs.</span></td> </tr> <tr> <td id="L1104" class="blob-num js-line-number" data-line-number="1104"></td> <td id="LC1104" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1105" class="blob-num js-line-number" data-line-number="1105"></td> <td id="LC1105" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1106" class="blob-num js-line-number" data-line-number="1106"></td> <td id="LC1106" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> inputs: cell inputs.</span></td> </tr> <tr> <td id="L1107" class="blob-num js-line-number" data-line-number="1107"></td> <td id="LC1107" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state: cell state.</span></td> </tr> <tr> <td id="L1108" class="blob-num js-line-number" data-line-number="1108"></td> <td id="LC1108" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> scope: optional cell scope.</span></td> </tr> <tr> <td id="L1109" class="blob-num js-line-number" data-line-number="1109"></td> <td id="LC1109" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1110" class="blob-num js-line-number" data-line-number="1110"></td> <td id="LC1110" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Returns:</span></td> </tr> <tr> <td id="L1111" class="blob-num js-line-number" data-line-number="1111"></td> <td id="LC1111" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Tuple of cell outputs and new state.</span></td> </tr> <tr> <td id="L1112" class="blob-num js-line-number" data-line-number="1112"></td> <td id="LC1112" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1113" class="blob-num js-line-number" data-line-number="1113"></td> <td id="LC1113" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L1114" class="blob-num js-line-number" data-line-number="1114"></td> <td id="LC1114" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> TypeError: If cell inputs and outputs have different structure (type).</span></td> </tr> <tr> <td id="L1115" class="blob-num js-line-number" data-line-number="1115"></td> <td id="LC1115" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: If cell inputs and outputs have different structure (value).</span></td> </tr> <tr> <td id="L1116" class="blob-num js-line-number" data-line-number="1116"></td> <td id="LC1116" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1117" class="blob-num js-line-number" data-line-number="1117"></td> <td id="LC1117" class="blob-code blob-code-inner js-file-line"> outputs, new_state <span class="pl-k">=</span> <span class="pl-c1">self</span>._cell(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L1118" class="blob-num js-line-number" data-line-number="1118"></td> <td id="LC1118" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> Ensure shapes match</span></td> </tr> <tr> <td id="L1119" class="blob-num js-line-number" data-line-number="1119"></td> <td id="LC1119" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">assert_shape_match</span>(<span class="pl-smi">inp</span>, <span class="pl-smi">out</span>):</td> </tr> <tr> <td id="L1120" class="blob-num js-line-number" data-line-number="1120"></td> <td id="LC1120" class="blob-code blob-code-inner js-file-line"> inp.get_shape().assert_is_compatible_with(out.get_shape())</td> </tr> <tr> <td id="L1121" class="blob-num js-line-number" data-line-number="1121"></td> <td id="LC1121" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">default_residual_fn</span>(<span class="pl-smi">inputs</span>, <span class="pl-smi">outputs</span>):</td> </tr> <tr> <td id="L1122" class="blob-num js-line-number" data-line-number="1122"></td> <td id="LC1122" class="blob-code blob-code-inner js-file-line"> nest.assert_same_structure(inputs, outputs)</td> </tr> <tr> <td id="L1123" class="blob-num js-line-number" data-line-number="1123"></td> <td id="LC1123" class="blob-code blob-code-inner js-file-line"> nest.map_structure(assert_shape_match, inputs, outputs)</td> </tr> <tr> <td id="L1124" class="blob-num js-line-number" data-line-number="1124"></td> <td id="LC1124" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> nest.map_structure(<span class="pl-k">lambda</span> <span class="pl-smi">inp</span>, <span class="pl-smi">out</span>: inp <span class="pl-k">+</span> out, inputs, outputs)</td> </tr> <tr> <td id="L1125" class="blob-num js-line-number" data-line-number="1125"></td> <td id="LC1125" class="blob-code blob-code-inner js-file-line"> res_outputs <span class="pl-k">=</span> (<span class="pl-c1">self</span>._residual_fn <span class="pl-k">or</span> default_residual_fn)(inputs, outputs)</td> </tr> <tr> <td id="L1126" class="blob-num js-line-number" data-line-number="1126"></td> <td id="LC1126" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> (res_outputs, new_state)</td> </tr> <tr> <td id="L1127" class="blob-num js-line-number" data-line-number="1127"></td> <td id="LC1127" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1128" class="blob-num js-line-number" data-line-number="1128"></td> <td id="LC1128" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1129" class="blob-num js-line-number" data-line-number="1129"></td> <td id="LC1129" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.DeviceWrapper<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L1130" class="blob-num js-line-number" data-line-number="1130"></td> <td id="LC1130" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">DeviceWrapper</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L1131" class="blob-num js-line-number" data-line-number="1131"></td> <td id="LC1131" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Operator that ensures an RNNCell runs on a particular device.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1132" class="blob-num js-line-number" data-line-number="1132"></td> <td id="LC1132" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1133" class="blob-num js-line-number" data-line-number="1133"></td> <td id="LC1133" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cell</span>, <span class="pl-smi">device</span>):</td> </tr> <tr> <td id="L1134" class="blob-num js-line-number" data-line-number="1134"></td> <td id="LC1134" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Construct a `DeviceWrapper` for `cell` with device `device`.</span></td> </tr> <tr> <td id="L1135" class="blob-num js-line-number" data-line-number="1135"></td> <td id="LC1135" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1136" class="blob-num js-line-number" data-line-number="1136"></td> <td id="LC1136" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Ensures the wrapped `cell` is called with `tf.device(device)`.</span></td> </tr> <tr> <td id="L1137" class="blob-num js-line-number" data-line-number="1137"></td> <td id="LC1137" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1138" class="blob-num js-line-number" data-line-number="1138"></td> <td id="LC1138" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1139" class="blob-num js-line-number" data-line-number="1139"></td> <td id="LC1139" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell: An instance of `RNNCell`.</span></td> </tr> <tr> <td id="L1140" class="blob-num js-line-number" data-line-number="1140"></td> <td id="LC1140" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> device: A device string or function, for passing to `tf.device`.</span></td> </tr> <tr> <td id="L1141" class="blob-num js-line-number" data-line-number="1141"></td> <td id="LC1141" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1142" class="blob-num js-line-number" data-line-number="1142"></td> <td id="LC1142" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell <span class="pl-k">=</span> cell</td> </tr> <tr> <td id="L1143" class="blob-num js-line-number" data-line-number="1143"></td> <td id="LC1143" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._device <span class="pl-k">=</span> device</td> </tr> <tr> <td id="L1144" class="blob-num js-line-number" data-line-number="1144"></td> <td id="LC1144" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1145" class="blob-num js-line-number" data-line-number="1145"></td> <td id="LC1145" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1146" class="blob-num js-line-number" data-line-number="1146"></td> <td id="LC1146" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1147" class="blob-num js-line-number" data-line-number="1147"></td> <td id="LC1147" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.state_size</td> </tr> <tr> <td id="L1148" class="blob-num js-line-number" data-line-number="1148"></td> <td id="LC1148" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1149" class="blob-num js-line-number" data-line-number="1149"></td> <td id="LC1149" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1150" class="blob-num js-line-number" data-line-number="1150"></td> <td id="LC1150" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1151" class="blob-num js-line-number" data-line-number="1151"></td> <td id="LC1151" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.output_size</td> </tr> <tr> <td id="L1152" class="blob-num js-line-number" data-line-number="1152"></td> <td id="LC1152" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1153" class="blob-num js-line-number" data-line-number="1153"></td> <td id="LC1153" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L1154" class="blob-num js-line-number" data-line-number="1154"></td> <td id="LC1154" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L1155" class="blob-num js-line-number" data-line-number="1155"></td> <td id="LC1155" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.device(<span class="pl-c1">self</span>._device):</td> </tr> <tr> <td id="L1156" class="blob-num js-line-number" data-line-number="1156"></td> <td id="LC1156" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell.zero_state(batch_size, dtype)</td> </tr> <tr> <td id="L1157" class="blob-num js-line-number" data-line-number="1157"></td> <td id="LC1157" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1158" class="blob-num js-line-number" data-line-number="1158"></td> <td id="LC1158" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1159" class="blob-num js-line-number" data-line-number="1159"></td> <td id="LC1159" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run the cell on specified device.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1160" class="blob-num js-line-number" data-line-number="1160"></td> <td id="LC1160" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.device(<span class="pl-c1">self</span>._device):</td> </tr> <tr> <td id="L1161" class="blob-num js-line-number" data-line-number="1161"></td> <td id="LC1161" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cell(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L1162" class="blob-num js-line-number" data-line-number="1162"></td> <td id="LC1162" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1163" class="blob-num js-line-number" data-line-number="1163"></td> <td id="LC1163" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1164" class="blob-num js-line-number" data-line-number="1164"></td> <td id="LC1164" class="blob-code blob-code-inner js-file-line"><span class="pl-en">@tf_export</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>nn.rnn_cell.MultiRNNCell<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L1165" class="blob-num js-line-number" data-line-number="1165"></td> <td id="LC1165" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">MultiRNNCell</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L1166" class="blob-num js-line-number" data-line-number="1166"></td> <td id="LC1166" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>RNN cell composed sequentially of multiple simple cells.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1167" class="blob-num js-line-number" data-line-number="1167"></td> <td id="LC1167" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1168" class="blob-num js-line-number" data-line-number="1168"></td> <td id="LC1168" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cells</span>, <span class="pl-smi">state_is_tuple</span><span class="pl-k">=</span><span class="pl-c1">True</span>):</td> </tr> <tr> <td id="L1169" class="blob-num js-line-number" data-line-number="1169"></td> <td id="LC1169" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Create a RNN cell composed sequentially of a number of RNNCells.</span></td> </tr> <tr> <td id="L1170" class="blob-num js-line-number" data-line-number="1170"></td> <td id="LC1170" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1171" class="blob-num js-line-number" data-line-number="1171"></td> <td id="LC1171" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1172" class="blob-num js-line-number" data-line-number="1172"></td> <td id="LC1172" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cells: list of RNNCells that will be composed in this order.</span></td> </tr> <tr> <td id="L1173" class="blob-num js-line-number" data-line-number="1173"></td> <td id="LC1173" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state_is_tuple: If True, accepted and returned states are n-tuples, where</span></td> </tr> <tr> <td id="L1174" class="blob-num js-line-number" data-line-number="1174"></td> <td id="LC1174" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> `n = len(cells)`. If False, the states are all</span></td> </tr> <tr> <td id="L1175" class="blob-num js-line-number" data-line-number="1175"></td> <td id="LC1175" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> concatenated along the column axis. This latter behavior will soon be</span></td> </tr> <tr> <td id="L1176" class="blob-num js-line-number" data-line-number="1176"></td> <td id="LC1176" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> deprecated.</span></td> </tr> <tr> <td id="L1177" class="blob-num js-line-number" data-line-number="1177"></td> <td id="LC1177" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1178" class="blob-num js-line-number" data-line-number="1178"></td> <td id="LC1178" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L1179" class="blob-num js-line-number" data-line-number="1179"></td> <td id="LC1179" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if cells is empty (not allowed), or at least one of the cells</span></td> </tr> <tr> <td id="L1180" class="blob-num js-line-number" data-line-number="1180"></td> <td id="LC1180" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> returns a state tuple but the flag `state_is_tuple` is `False`.</span></td> </tr> <tr> <td id="L1181" class="blob-num js-line-number" data-line-number="1181"></td> <td id="LC1181" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1182" class="blob-num js-line-number" data-line-number="1182"></td> <td id="LC1182" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">super</span>(MultiRNNCell, <span class="pl-c1">self</span>).<span class="pl-c1">__init__</span>()</td> </tr> <tr> <td id="L1183" class="blob-num js-line-number" data-line-number="1183"></td> <td id="LC1183" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> cells:</td> </tr> <tr> <td id="L1184" class="blob-num js-line-number" data-line-number="1184"></td> <td id="LC1184" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Must specify at least one cell for MultiRNNCell.<span class="pl-pds">&quot;</span></span>)</td> </tr> <tr> <td id="L1185" class="blob-num js-line-number" data-line-number="1185"></td> <td id="LC1185" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> nest.is_sequence(cells):</td> </tr> <tr> <td id="L1186" class="blob-num js-line-number" data-line-number="1186"></td> <td id="LC1186" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(</td> </tr> <tr> <td id="L1187" class="blob-num js-line-number" data-line-number="1187"></td> <td id="LC1187" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>cells must be a list or tuple, but saw: <span class="pl-c1">%s</span>.<span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> cells)</td> </tr> <tr> <td id="L1188" class="blob-num js-line-number" data-line-number="1188"></td> <td id="LC1188" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1189" class="blob-num js-line-number" data-line-number="1189"></td> <td id="LC1189" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cells <span class="pl-k">=</span> cells</td> </tr> <tr> <td id="L1190" class="blob-num js-line-number" data-line-number="1190"></td> <td id="LC1190" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">=</span> state_is_tuple</td> </tr> <tr> <td id="L1191" class="blob-num js-line-number" data-line-number="1191"></td> <td id="LC1191" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> state_is_tuple:</td> </tr> <tr> <td id="L1192" class="blob-num js-line-number" data-line-number="1192"></td> <td id="LC1192" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">any</span>(nest.is_sequence(c.state_size) <span class="pl-k">for</span> c <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells):</td> </tr> <tr> <td id="L1193" class="blob-num js-line-number" data-line-number="1193"></td> <td id="LC1193" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Some cells return tuples of states, but the flag <span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L1194" class="blob-num js-line-number" data-line-number="1194"></td> <td id="LC1194" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>state_is_tuple is not set. State sizes are: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span></td> </tr> <tr> <td id="L1195" class="blob-num js-line-number" data-line-number="1195"></td> <td id="LC1195" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">%</span> <span class="pl-c1">str</span>([c.state_size <span class="pl-k">for</span> c <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells]))</td> </tr> <tr> <td id="L1196" class="blob-num js-line-number" data-line-number="1196"></td> <td id="LC1196" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1197" class="blob-num js-line-number" data-line-number="1197"></td> <td id="LC1197" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1198" class="blob-num js-line-number" data-line-number="1198"></td> <td id="LC1198" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1199" class="blob-num js-line-number" data-line-number="1199"></td> <td id="LC1199" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L1200" class="blob-num js-line-number" data-line-number="1200"></td> <td id="LC1200" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">tuple</span>(cell.state_size <span class="pl-k">for</span> cell <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells)</td> </tr> <tr> <td id="L1201" class="blob-num js-line-number" data-line-number="1201"></td> <td id="LC1201" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1202" class="blob-num js-line-number" data-line-number="1202"></td> <td id="LC1202" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">sum</span>([cell.state_size <span class="pl-k">for</span> cell <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells])</td> </tr> <tr> <td id="L1203" class="blob-num js-line-number" data-line-number="1203"></td> <td id="LC1203" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1204" class="blob-num js-line-number" data-line-number="1204"></td> <td id="LC1204" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1205" class="blob-num js-line-number" data-line-number="1205"></td> <td id="LC1205" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1206" class="blob-num js-line-number" data-line-number="1206"></td> <td id="LC1206" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._cells[<span class="pl-k">-</span><span class="pl-c1">1</span>].output_size</td> </tr> <tr> <td id="L1207" class="blob-num js-line-number" data-line-number="1207"></td> <td id="LC1207" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1208" class="blob-num js-line-number" data-line-number="1208"></td> <td id="LC1208" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">zero_state</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">batch_size</span>, <span class="pl-smi">dtype</span>):</td> </tr> <tr> <td id="L1209" class="blob-num js-line-number" data-line-number="1209"></td> <td id="LC1209" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> ops.name_scope(<span class="pl-c1">type</span>(<span class="pl-c1">self</span>).<span class="pl-c1">__name__</span> <span class="pl-k">+</span> <span class="pl-s"><span class="pl-pds">&quot;</span>ZeroState<span class="pl-pds">&quot;</span></span>, <span class="pl-v">values</span><span class="pl-k">=</span>[batch_size]):</td> </tr> <tr> <td id="L1210" class="blob-num js-line-number" data-line-number="1210"></td> <td id="LC1210" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L1211" class="blob-num js-line-number" data-line-number="1211"></td> <td id="LC1211" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">tuple</span>(cell.zero_state(batch_size, dtype) <span class="pl-k">for</span> cell <span class="pl-k">in</span> <span class="pl-c1">self</span>._cells)</td> </tr> <tr> <td id="L1212" class="blob-num js-line-number" data-line-number="1212"></td> <td id="LC1212" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1213" class="blob-num js-line-number" data-line-number="1213"></td> <td id="LC1213" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> We know here that state_size of each cell is not a tuple and</span></td> </tr> <tr> <td id="L1214" class="blob-num js-line-number" data-line-number="1214"></td> <td id="LC1214" class="blob-code blob-code-inner js-file-line"> <span class="pl-c"><span class="pl-c">#</span> presumably does not contain TensorArrays or anything else fancy</span></td> </tr> <tr> <td id="L1215" class="blob-num js-line-number" data-line-number="1215"></td> <td id="LC1215" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">super</span>(MultiRNNCell, <span class="pl-c1">self</span>).zero_state(batch_size, dtype)</td> </tr> <tr> <td id="L1216" class="blob-num js-line-number" data-line-number="1216"></td> <td id="LC1216" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1217" class="blob-num js-line-number" data-line-number="1217"></td> <td id="LC1217" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">call</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>):</td> </tr> <tr> <td id="L1218" class="blob-num js-line-number" data-line-number="1218"></td> <td id="LC1218" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Run this multi-layer cell on inputs, starting from state.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1219" class="blob-num js-line-number" data-line-number="1219"></td> <td id="LC1219" class="blob-code blob-code-inner js-file-line"> cur_state_pos <span class="pl-k">=</span> <span class="pl-c1">0</span></td> </tr> <tr> <td id="L1220" class="blob-num js-line-number" data-line-number="1220"></td> <td id="LC1220" class="blob-code blob-code-inner js-file-line"> cur_inp <span class="pl-k">=</span> inputs</td> </tr> <tr> <td id="L1221" class="blob-num js-line-number" data-line-number="1221"></td> <td id="LC1221" class="blob-code blob-code-inner js-file-line"> new_states <span class="pl-k">=</span> []</td> </tr> <tr> <td id="L1222" class="blob-num js-line-number" data-line-number="1222"></td> <td id="LC1222" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">for</span> i, cell <span class="pl-k">in</span> <span class="pl-c1">enumerate</span>(<span class="pl-c1">self</span>._cells):</td> </tr> <tr> <td id="L1223" class="blob-num js-line-number" data-line-number="1223"></td> <td id="LC1223" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">with</span> vs.variable_scope(<span class="pl-s"><span class="pl-pds">&quot;</span>cell_<span class="pl-c1">%d</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span> i):</td> </tr> <tr> <td id="L1224" class="blob-num js-line-number" data-line-number="1224"></td> <td id="LC1224" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple:</td> </tr> <tr> <td id="L1225" class="blob-num js-line-number" data-line-number="1225"></td> <td id="LC1225" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> nest.is_sequence(state):</td> </tr> <tr> <td id="L1226" class="blob-num js-line-number" data-line-number="1226"></td> <td id="LC1226" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(</td> </tr> <tr> <td id="L1227" class="blob-num js-line-number" data-line-number="1227"></td> <td id="LC1227" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;</span>Expected state to be a tuple of length <span class="pl-c1">%d</span>, but received: <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span></td> </tr> <tr> <td id="L1228" class="blob-num js-line-number" data-line-number="1228"></td> <td id="LC1228" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">len</span>(<span class="pl-c1">self</span>.state_size), state))</td> </tr> <tr> <td id="L1229" class="blob-num js-line-number" data-line-number="1229"></td> <td id="LC1229" class="blob-code blob-code-inner js-file-line"> cur_state <span class="pl-k">=</span> state[i]</td> </tr> <tr> <td id="L1230" class="blob-num js-line-number" data-line-number="1230"></td> <td id="LC1230" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">else</span>:</td> </tr> <tr> <td id="L1231" class="blob-num js-line-number" data-line-number="1231"></td> <td id="LC1231" class="blob-code blob-code-inner js-file-line"> cur_state <span class="pl-k">=</span> array_ops.slice(state, [<span class="pl-c1">0</span>, cur_state_pos],</td> </tr> <tr> <td id="L1232" class="blob-num js-line-number" data-line-number="1232"></td> <td id="LC1232" class="blob-code blob-code-inner js-file-line"> [<span class="pl-k">-</span><span class="pl-c1">1</span>, cell.state_size])</td> </tr> <tr> <td id="L1233" class="blob-num js-line-number" data-line-number="1233"></td> <td id="LC1233" class="blob-code blob-code-inner js-file-line"> cur_state_pos <span class="pl-k">+=</span> cell.state_size</td> </tr> <tr> <td id="L1234" class="blob-num js-line-number" data-line-number="1234"></td> <td id="LC1234" class="blob-code blob-code-inner js-file-line"> cur_inp, new_state <span class="pl-k">=</span> cell(cur_inp, cur_state)</td> </tr> <tr> <td id="L1235" class="blob-num js-line-number" data-line-number="1235"></td> <td id="LC1235" class="blob-code blob-code-inner js-file-line"> new_states.append(new_state)</td> </tr> <tr> <td id="L1236" class="blob-num js-line-number" data-line-number="1236"></td> <td id="LC1236" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1237" class="blob-num js-line-number" data-line-number="1237"></td> <td id="LC1237" class="blob-code blob-code-inner js-file-line"> new_states <span class="pl-k">=</span> (<span class="pl-c1">tuple</span>(new_states) <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_is_tuple <span class="pl-k">else</span></td> </tr> <tr> <td id="L1238" class="blob-num js-line-number" data-line-number="1238"></td> <td id="LC1238" class="blob-code blob-code-inner js-file-line"> array_ops.concat(new_states, <span class="pl-c1">1</span>))</td> </tr> <tr> <td id="L1239" class="blob-num js-line-number" data-line-number="1239"></td> <td id="LC1239" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1240" class="blob-num js-line-number" data-line-number="1240"></td> <td id="LC1240" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> cur_inp, new_states</td> </tr> <tr> <td id="L1241" class="blob-num js-line-number" data-line-number="1241"></td> <td id="LC1241" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1242" class="blob-num js-line-number" data-line-number="1242"></td> <td id="LC1242" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1243" class="blob-num js-line-number" data-line-number="1243"></td> <td id="LC1243" class="blob-code blob-code-inner js-file-line"><span class="pl-k">class</span> <span class="pl-en">_SlimRNNCell</span>(<span class="pl-e">RNNCell</span>):</td> </tr> <tr> <td id="L1244" class="blob-num js-line-number" data-line-number="1244"></td> <td id="LC1244" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>A simple wrapper for slim.rnn_cells.<span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1245" class="blob-num js-line-number" data-line-number="1245"></td> <td id="LC1245" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1246" class="blob-num js-line-number" data-line-number="1246"></td> <td id="LC1246" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__init__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">cell_fn</span>):</td> </tr> <tr> <td id="L1247" class="blob-num js-line-number" data-line-number="1247"></td> <td id="LC1247" class="blob-code blob-code-inner js-file-line"> <span class="pl-s"><span class="pl-pds">&quot;&quot;&quot;</span>Create a SlimRNNCell from a cell_fn.</span></td> </tr> <tr> <td id="L1248" class="blob-num js-line-number" data-line-number="1248"></td> <td id="LC1248" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1249" class="blob-num js-line-number" data-line-number="1249"></td> <td id="LC1249" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Args:</span></td> </tr> <tr> <td id="L1250" class="blob-num js-line-number" data-line-number="1250"></td> <td id="LC1250" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> cell_fn: a function which takes (inputs, state, scope) and produces the</span></td> </tr> <tr> <td id="L1251" class="blob-num js-line-number" data-line-number="1251"></td> <td id="LC1251" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> outputs and the new_state. Additionally when called with inputs=None and</span></td> </tr> <tr> <td id="L1252" class="blob-num js-line-number" data-line-number="1252"></td> <td id="LC1252" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> state=None it should return (initial_outputs, initial_state).</span></td> </tr> <tr> <td id="L1253" class="blob-num js-line-number" data-line-number="1253"></td> <td id="LC1253" class="blob-code blob-code-inner js-file-line"><span class="pl-s"></span></td> </tr> <tr> <td id="L1254" class="blob-num js-line-number" data-line-number="1254"></td> <td id="LC1254" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> Raises:</span></td> </tr> <tr> <td id="L1255" class="blob-num js-line-number" data-line-number="1255"></td> <td id="LC1255" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> TypeError: if cell_fn is not callable</span></td> </tr> <tr> <td id="L1256" class="blob-num js-line-number" data-line-number="1256"></td> <td id="LC1256" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> ValueError: if cell_fn cannot produce a valid initial state.</span></td> </tr> <tr> <td id="L1257" class="blob-num js-line-number" data-line-number="1257"></td> <td id="LC1257" class="blob-code blob-code-inner js-file-line"><span class="pl-s"> <span class="pl-pds">&quot;&quot;&quot;</span></span></td> </tr> <tr> <td id="L1258" class="blob-num js-line-number" data-line-number="1258"></td> <td id="LC1258" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-k">not</span> <span class="pl-c1">callable</span>(cell_fn):</td> </tr> <tr> <td id="L1259" class="blob-num js-line-number" data-line-number="1259"></td> <td id="LC1259" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">TypeError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>cell_fn <span class="pl-c1">%s</span> needs to be callable<span class="pl-pds">&quot;</span></span>, cell_fn)</td> </tr> <tr> <td id="L1260" class="blob-num js-line-number" data-line-number="1260"></td> <td id="LC1260" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell_fn <span class="pl-k">=</span> cell_fn</td> </tr> <tr> <td id="L1261" class="blob-num js-line-number" data-line-number="1261"></td> <td id="LC1261" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._cell_name <span class="pl-k">=</span> cell_fn.func.<span class="pl-c1">__name__</span></td> </tr> <tr> <td id="L1262" class="blob-num js-line-number" data-line-number="1262"></td> <td id="LC1262" class="blob-code blob-code-inner js-file-line"> init_output, init_state <span class="pl-k">=</span> <span class="pl-c1">self</span>._cell_fn(<span class="pl-c1">None</span>, <span class="pl-c1">None</span>)</td> </tr> <tr> <td id="L1263" class="blob-num js-line-number" data-line-number="1263"></td> <td id="LC1263" class="blob-code blob-code-inner js-file-line"> output_shape <span class="pl-k">=</span> init_output.get_shape()</td> </tr> <tr> <td id="L1264" class="blob-num js-line-number" data-line-number="1264"></td> <td id="LC1264" class="blob-code blob-code-inner js-file-line"> state_shape <span class="pl-k">=</span> init_state.get_shape()</td> </tr> <tr> <td id="L1265" class="blob-num js-line-number" data-line-number="1265"></td> <td id="LC1265" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._output_size <span class="pl-k">=</span> output_shape.with_rank(<span class="pl-c1">2</span>)[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L1266" class="blob-num js-line-number" data-line-number="1266"></td> <td id="LC1266" class="blob-code blob-code-inner js-file-line"> <span class="pl-c1">self</span>._state_size <span class="pl-k">=</span> state_shape.with_rank(<span class="pl-c1">2</span>)[<span class="pl-c1">1</span>].value</td> </tr> <tr> <td id="L1267" class="blob-num js-line-number" data-line-number="1267"></td> <td id="LC1267" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._output_size <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L1268" class="blob-num js-line-number" data-line-number="1268"></td> <td id="LC1268" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Initial output created by <span class="pl-c1">%s</span> has invalid shape <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span></td> </tr> <tr> <td id="L1269" class="blob-num js-line-number" data-line-number="1269"></td> <td id="LC1269" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">self</span>._cell_name, output_shape))</td> </tr> <tr> <td id="L1270" class="blob-num js-line-number" data-line-number="1270"></td> <td id="LC1270" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">if</span> <span class="pl-c1">self</span>._state_size <span class="pl-k">is</span> <span class="pl-c1">None</span>:</td> </tr> <tr> <td id="L1271" class="blob-num js-line-number" data-line-number="1271"></td> <td id="LC1271" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">raise</span> <span class="pl-c1">ValueError</span>(<span class="pl-s"><span class="pl-pds">&quot;</span>Initial state created by <span class="pl-c1">%s</span> has invalid shape <span class="pl-c1">%s</span><span class="pl-pds">&quot;</span></span> <span class="pl-k">%</span></td> </tr> <tr> <td id="L1272" class="blob-num js-line-number" data-line-number="1272"></td> <td id="LC1272" class="blob-code blob-code-inner js-file-line"> (<span class="pl-c1">self</span>._cell_name, state_shape))</td> </tr> <tr> <td id="L1273" class="blob-num js-line-number" data-line-number="1273"></td> <td id="LC1273" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1274" class="blob-num js-line-number" data-line-number="1274"></td> <td id="LC1274" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1275" class="blob-num js-line-number" data-line-number="1275"></td> <td id="LC1275" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">state_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1276" class="blob-num js-line-number" data-line-number="1276"></td> <td id="LC1276" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._state_size</td> </tr> <tr> <td id="L1277" class="blob-num js-line-number" data-line-number="1277"></td> <td id="LC1277" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1278" class="blob-num js-line-number" data-line-number="1278"></td> <td id="LC1278" class="blob-code blob-code-inner js-file-line"> <span class="pl-en">@</span><span class="pl-c1">property</span></td> </tr> <tr> <td id="L1279" class="blob-num js-line-number" data-line-number="1279"></td> <td id="LC1279" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-en">output_size</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>):</td> </tr> <tr> <td id="L1280" class="blob-num js-line-number" data-line-number="1280"></td> <td id="LC1280" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> <span class="pl-c1">self</span>._output_size</td> </tr> <tr> <td id="L1281" class="blob-num js-line-number" data-line-number="1281"></td> <td id="LC1281" class="blob-code blob-code-inner js-file-line"> </td> </tr> <tr> <td id="L1282" class="blob-num js-line-number" data-line-number="1282"></td> <td id="LC1282" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">def</span> <span class="pl-c1">__call__</span>(<span class="pl-smi"><span class="pl-smi">self</span></span>, <span class="pl-smi">inputs</span>, <span class="pl-smi">state</span>, <span class="pl-smi">scope</span><span class="pl-k">=</span><span class="pl-c1">None</span>):</td> </tr> <tr> <td id="L1283" class="blob-num js-line-number" data-line-number="1283"></td> <td id="LC1283" class="blob-code blob-code-inner js-file-line"> scope <span class="pl-k">=</span> scope <span class="pl-k">or</span> <span class="pl-c1">self</span>._cell_name</td> </tr> <tr> <td id="L1284" class="blob-num js-line-number" data-line-number="1284"></td> <td id="LC1284" class="blob-code blob-code-inner js-file-line"> output, state <span class="pl-k">=</span> <span class="pl-c1">self</span>._cell_fn(inputs, state, <span class="pl-v">scope</span><span class="pl-k">=</span>scope)</td> </tr> <tr> <td id="L1285" class="blob-num js-line-number" data-line-number="1285"></td> <td id="LC1285" class="blob-code blob-code-inner js-file-line"> <span class="pl-k">return</span> output, state</td> </tr> </table> <div class="BlobToolbar position-absolute js-file-line-actions dropdown js-menu-container js-select-menu d-none" aria-hidden="true"> <button class="btn-octicon ml-0 px-2 p-0 bg-white border border-gray-dark rounded-1 dropdown-toggle js-menu-target" id="js-file-line-action-button" type="button" aria-expanded="false" aria-haspopup="true" aria-label="Inline file action toolbar" aria-controls="inline-file-actions"> <svg class="octicon octicon-kebab-horizontal" viewBox="0 0 13 16" version="1.1" width="13" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M1.5 9a1.5 1.5 0 1 1 0-3 1.5 1.5 0 0 1 0 3zm5 0a1.5 1.5 0 1 1 0-3 1.5 1.5 0 0 1 0 3zm5 0a1.5 1.5 0 1 1 0-3 1.5 1.5 0 0 1 0 3z"/></svg> </button> <div class="dropdown-menu-content js-menu-content" id="inline-file-actions"> <ul class="BlobToolbar-dropdown dropdown-menu dropdown-menu-se mt-2"> <li><clipboard-copy class="dropdown-item" style="cursor:pointer;" id="js-copy-lines" data-original-text="Copy lines">Copy lines</clipboard-copy></li> <li><clipboard-copy class="dropdown-item" id="js-copy-permalink" style="cursor:pointer;" data-original-text="Copy permalink">Copy permalink</clipboard-copy></li> <li><a class="dropdown-item js-update-url-with-hash" id="js-view-git-blame" href="/tensorflow/tensorflow/blame/cbc658095ae228f2f557af47e4901d552573aa15/tensorflow/python/ops/rnn_cell_impl.py">View git blame</a></li> <li><a class="dropdown-item" id="js-new-issue" href="/tensorflow/tensorflow/issues/new">Open new issue</a></li> </ul> </div> </div> </div> </div> <button type="button" data-facebox="#jump-to-line" data-facebox-class="linejump" data-hotkey="l" class="d-none">Jump to Line</button> <div id="jump-to-line" style="display:none"> <!-- '"` --><!-- </textarea></xmp> --></option></form><form class="js-jump-to-line-form" action="" accept-charset="UTF-8" method="get"><input name="utf8" type="hidden" value="&#x2713;" /> <input class="form-control linejump-input js-jump-to-line-field" type="text" placeholder="Jump to line&hellip;" aria-label="Jump to line" autofocus> <button type="submit" class="btn">Go</button> </form> </div> </div> <div class="modal-backdrop js-touch-events"></div> </div> </div> </div> </div> <div class="footer container-lg px-3" role="contentinfo"> <div class="position-relative d-flex flex-justify-between py-6 mt-6 f6 text-gray border-top border-gray-light "> <ul class="list-style-none d-flex flex-wrap "> <li class="mr-3">&copy; 2018 <span title="0.23060s from unicorn-1000010005-crk39">GitHub</span>, Inc.</li> <li class="mr-3"><a href="https://help.github.com/articles/github-terms-of-service/" data-ga-click="Footer, go to terms, text:terms">Terms</a></li> <li class="mr-3"><a href="https://github.com/site/privacy" data-ga-click="Footer, go to privacy, text:privacy">Privacy</a></li> <li class="mr-3"><a href="https://help.github.com/articles/github-security/" data-ga-click="Footer, go to security, text:security">Security</a></li> <li class="mr-3"><a href="https://status.github.com/" data-ga-click="Footer, go to status, text:status">Status</a></li> <li><a data-ga-click="Footer, go to help, text:help" href="https://help.github.com">Help</a></li> </ul> <a aria-label="Homepage" title="GitHub" class="footer-octicon" href="https://github.com"> <svg height="24" class="octicon octicon-mark-github" viewBox="0 0 16 16" version="1.1" width="24" aria-hidden="true"><path fill-rule="evenodd" d="M8 0C3.58 0 0 3.58 0 8c0 3.54 2.29 6.53 5.47 7.59.4.07.55-.17.55-.38 0-.19-.01-.82-.01-1.49-2.01.37-2.53-.49-2.69-.94-.09-.23-.48-.94-.82-1.13-.28-.15-.68-.52-.01-.53.63-.01 1.08.58 1.23.82.72 1.21 1.87.87 2.33.66.07-.52.28-.87.51-1.07-1.78-.2-3.64-.89-3.64-3.95 0-.87.31-1.59.82-2.15-.08-.2-.36-1.02.08-2.12 0 0 .67-.21 2.2.82.64-.18 1.32-.27 2-.27.68 0 1.36.09 2 .27 1.53-1.04 2.2-.82 2.2-.82.44 1.1.16 1.92.08 2.12.51.56.82 1.27.82 2.15 0 3.07-1.87 3.75-3.65 3.95.29.25.54.73.54 1.48 0 1.07-.01 1.93-.01 2.2 0 .21.15.46.55.38A8.013 8.013 0 0 0 16 8c0-4.42-3.58-8-8-8z"/></svg> </a> <ul class="list-style-none d-flex flex-wrap "> <li class="mr-3"><a data-ga-click="Footer, go to contact, text:contact" href="https://github.com/contact">Contact GitHub</a></li> <li class="mr-3"><a href="https://developer.github.com" data-ga-click="Footer, go to api, text:api">API</a></li> <li class="mr-3"><a href="https://training.github.com" data-ga-click="Footer, go to training, text:training">Training</a></li> <li class="mr-3"><a href="https://shop.github.com" data-ga-click="Footer, go to shop, text:shop">Shop</a></li> <li class="mr-3"><a data-ga-click="Footer, go to blog, text:blog" href="https://github.com/blog">Blog</a></li> <li><a data-ga-click="Footer, go to about, text:about" href="https://github.com/about">About</a></li> </ul> </div> </div> <div id="ajax-error-message" class="ajax-error-message flash flash-error"> <svg class="octicon octicon-alert" viewBox="0 0 16 16" version="1.1" width="16" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M8.865 1.52c-.18-.31-.51-.5-.87-.5s-.69.19-.87.5L.275 13.5c-.18.31-.18.69 0 1 .19.31.52.5.87.5h13.7c.36 0 .69-.19.86-.5.17-.31.18-.69.01-1L8.865 1.52zM8.995 13h-2v-2h2v2zm0-3h-2V6h2v4z"/></svg> <button type="button" class="flash-close js-ajax-error-dismiss" aria-label="Dismiss error"> <svg class="octicon octicon-x" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M7.48 8l3.75 3.75-1.48 1.48L6 9.48l-3.75 3.75-1.48-1.48L4.52 8 .77 4.25l1.48-1.48L6 6.52l3.75-3.75 1.48 1.48z"/></svg> </button> You can't perform that action at this time. </div> <script crossorigin="anonymous" type="application/javascript" src="https://assets-cdn.github.com/assets/compat-af4f19572e1d43f69746f70aa5f01dc9.js"></script> <script crossorigin="anonymous" type="application/javascript" src="https://assets-cdn.github.com/assets/frameworks-dd22b45b9495a43602787a69962f696c.js"></script> <script crossorigin="anonymous" async="async" type="application/javascript" src="https://assets-cdn.github.com/assets/github-48130667aaece5c4ce8ee28ecdb29539.js"></script> <div class="js-stale-session-flash stale-session-flash flash flash-warn flash-banner d-none"> <svg class="octicon octicon-alert" viewBox="0 0 16 16" version="1.1" width="16" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M8.865 1.52c-.18-.31-.51-.5-.87-.5s-.69.19-.87.5L.275 13.5c-.18.31-.18.69 0 1 .19.31.52.5.87.5h13.7c.36 0 .69-.19.86-.5.17-.31.18-.69.01-1L8.865 1.52zM8.995 13h-2v-2h2v2zm0-3h-2V6h2v4z"/></svg> <span class="signed-in-tab-flash">You signed in with another tab or window. <a href="">Reload</a> to refresh your session.</span> <span class="signed-out-tab-flash">You signed out in another tab or window. <a href="">Reload</a> to refresh your session.</span> </div> <div class="facebox" id="facebox" style="display:none;"> <div class="facebox-popup"> <div class="facebox-content" role="dialog" aria-labelledby="facebox-header" aria-describedby="facebox-description"> </div> <button type="button" class="facebox-close js-facebox-close" aria-label="Close modal"> <svg class="octicon octicon-x" viewBox="0 0 12 16" version="1.1" width="12" height="16" aria-hidden="true"><path fill-rule="evenodd" d="M7.48 8l3.75 3.75-1.48 1.48L6 9.48l-3.75 3.75-1.48-1.48L4.52 8 .77 4.25l1.48-1.48L6 6.52l3.75-3.75 1.48 1.48z"/></svg> </button> </div> </div> </body> </html>
false
true
f70eab9a72dba0a78b746376f282d170ea2df7b0
5,527
py
Python
loader_2.py
wangyan9411/entity
8d5601b8061b580da5395c46a52440e85eeff934
[ "Apache-2.0" ]
null
null
null
loader_2.py
wangyan9411/entity
8d5601b8061b580da5395c46a52440e85eeff934
[ "Apache-2.0" ]
null
null
null
loader_2.py
wangyan9411/entity
8d5601b8061b580da5395c46a52440e85eeff934
[ "Apache-2.0" ]
null
null
null
import os import re import codecs from utils import create_dico, create_mapping, zero_digits from utils import iob2, iob_iobes def load_sentences(path, lower, zeros): """ Load sentences. A line must contain at least a word and its tag. Sentences are separated by empty lines. """ sentences = [] sentence = [] for line in codecs.open(path, 'r', 'utf8'): line = zero_digits(line.rstrip()) if zeros else line.rstrip() if not line: if len(sentence) > 0: if 'DOCSTART' not in sentence[0][0]: sentences.append(sentence) sentence = [] else: word = line.split() assert len(word) >= 2 sentence.append(word) if len(sentence) > 0: if 'DOCSTART' not in sentence[0][0]: sentences.append(sentence) return sentences def update_tag_scheme(sentences, tag_scheme): """ Check and update sentences tagging scheme to IOB2. Only IOB1 and IOB2 schemes are accepted. """ for i, s in enumerate(sentences): tags = [w[-1] for w in s] # Check that tags are given in the IOB format if not iob2(tags): s_str = '\n'.join(' '.join(w) for w in s) raise Exception('Sentences should be given in IOB format! ' + 'Please check sentence %i:\n%s' % (i, s_str)) if tag_scheme == 'iob': # If format was IOB1, we convert to IOB2 for word, new_tag in zip(s, tags): word[-1] = new_tag elif tag_scheme == 'iobes': new_tags = iob_iobes(tags) for word, new_tag in zip(s, new_tags): word[-1] = new_tag else: raise Exception('Unknown tagging scheme!') def word_mapping(sentences, lower): """ Create a dictionary and a mapping of words, sorted by frequency. """ words = [[x[0].lower() if lower else x[0] for x in s] for s in sentences] dico = create_dico(words) dico['<UNK>'] = 10000000 word_to_id, id_to_word = create_mapping(dico) print "Found %i unique words (%i in total)" % ( len(dico), sum(len(x) for x in words) ) return dico, word_to_id, id_to_word def char_mapping(sentences): """ Create a dictionary and mapping of characters, sorted by frequency. """ chars = ["".join([w[0] for w in s]) for s in sentences] dico = create_dico(chars) char_to_id, id_to_char = create_mapping(dico) print "Found %i unique characters" % len(dico) return dico, char_to_id, id_to_char def tag_mapping(sentences): """ Create a dictionary and a mapping of tags, sorted by frequency. """ tags = [[word[-1] for word in s] for s in sentences] dico = create_dico(tags) tag_to_id, id_to_tag = create_mapping(dico) print "Found %i unique named entity tags" % len(dico) return dico, tag_to_id, id_to_tag def cap_feature(s): """ Capitalization feature: 0 = low caps 1 = all caps 2 = first letter caps 3 = one capital (not first letter) """ if s.lower() == s: return 0 elif s.upper() == s: return 1 elif s[0].upper() == s[0]: return 2 else: return 3 def prepare_dataset(sentences, word_to_id, char_to_id, tag_to_id, lower=False): """ Prepare the dataset. Return a list of lists of dictionaries containing: - word indexes - word char indexes - tag indexes """ def f(x): return x.lower() if lower else x data = [] for s in sentences: str_words = [w[0] for w in s] words = [word_to_id[f(w) if f(w) in word_to_id else '<UNK>'] for w in str_words] # Skip characters that are not in the training set chars = [[char_to_id[c] for c in w if c in char_to_id] for w in str_words] caps = [cap_feature(w) for w in str_words] tags = [tag_to_id[w[-1]] for w in s] data.append({ 'str_words': str_words, 'words': words, 'chars': chars, 'caps': caps, 'tags': tags, }) return data def augment_with_pretrained(dictionary, ext_emb_path, words): """ Augment the dictionary with words that have a pretrained embedding. If `words` is None, we add every word that has a pretrained embedding to the dictionary, otherwise, we only add the words that are given by `words` (typically the words in the development and test sets.) """ print 'Loading pretrained embeddings from %s...' % ext_emb_path assert os.path.isfile(ext_emb_path) # Load pretrained embeddings from file pretrained = set([ line.rstrip().split()[0].strip() for line in codecs.open(ext_emb_path, 'r', 'utf-8') if len(ext_emb_path) > 0 ]) # We either add every word in the pretrained file, # or only words given in the `words` list to which # we can assign a pretrained embedding if words is None: for word in pretrained: if word not in dictionary: dictionary[word] = 0 else: for word in words: if any(x in pretrained for x in [ word, word.lower(), re.sub('\d', '0', word.lower()) ]) and word not in dictionary: dictionary[word] = 0 word_to_id, id_to_word = create_mapping(dictionary) return dictionary, word_to_id, id_to_word
31.947977
79
0.588746
import os import re import codecs from utils import create_dico, create_mapping, zero_digits from utils import iob2, iob_iobes def load_sentences(path, lower, zeros): """ Load sentences. A line must contain at least a word and its tag. Sentences are separated by empty lines. """ sentences = [] sentence = [] for line in codecs.open(path, 'r', 'utf8'): line = zero_digits(line.rstrip()) if zeros else line.rstrip() if not line: if len(sentence) > 0: if 'DOCSTART' not in sentence[0][0]: sentences.append(sentence) sentence = [] else: word = line.split() assert len(word) >= 2 sentence.append(word) if len(sentence) > 0: if 'DOCSTART' not in sentence[0][0]: sentences.append(sentence) return sentences def update_tag_scheme(sentences, tag_scheme): """ Check and update sentences tagging scheme to IOB2. Only IOB1 and IOB2 schemes are accepted. """ for i, s in enumerate(sentences): tags = [w[-1] for w in s] if not iob2(tags): s_str = '\n'.join(' '.join(w) for w in s) raise Exception('Sentences should be given in IOB format! ' + 'Please check sentence %i:\n%s' % (i, s_str)) if tag_scheme == 'iob': for word, new_tag in zip(s, tags): word[-1] = new_tag elif tag_scheme == 'iobes': new_tags = iob_iobes(tags) for word, new_tag in zip(s, new_tags): word[-1] = new_tag else: raise Exception('Unknown tagging scheme!') def word_mapping(sentences, lower): """ Create a dictionary and a mapping of words, sorted by frequency. """ words = [[x[0].lower() if lower else x[0] for x in s] for s in sentences] dico = create_dico(words) dico['<UNK>'] = 10000000 word_to_id, id_to_word = create_mapping(dico) print "Found %i unique words (%i in total)" % ( len(dico), sum(len(x) for x in words) ) return dico, word_to_id, id_to_word def char_mapping(sentences): """ Create a dictionary and mapping of characters, sorted by frequency. """ chars = ["".join([w[0] for w in s]) for s in sentences] dico = create_dico(chars) char_to_id, id_to_char = create_mapping(dico) print "Found %i unique characters" % len(dico) return dico, char_to_id, id_to_char def tag_mapping(sentences): """ Create a dictionary and a mapping of tags, sorted by frequency. """ tags = [[word[-1] for word in s] for s in sentences] dico = create_dico(tags) tag_to_id, id_to_tag = create_mapping(dico) print "Found %i unique named entity tags" % len(dico) return dico, tag_to_id, id_to_tag def cap_feature(s): """ Capitalization feature: 0 = low caps 1 = all caps 2 = first letter caps 3 = one capital (not first letter) """ if s.lower() == s: return 0 elif s.upper() == s: return 1 elif s[0].upper() == s[0]: return 2 else: return 3 def prepare_dataset(sentences, word_to_id, char_to_id, tag_to_id, lower=False): """ Prepare the dataset. Return a list of lists of dictionaries containing: - word indexes - word char indexes - tag indexes """ def f(x): return x.lower() if lower else x data = [] for s in sentences: str_words = [w[0] for w in s] words = [word_to_id[f(w) if f(w) in word_to_id else '<UNK>'] for w in str_words] chars = [[char_to_id[c] for c in w if c in char_to_id] for w in str_words] caps = [cap_feature(w) for w in str_words] tags = [tag_to_id[w[-1]] for w in s] data.append({ 'str_words': str_words, 'words': words, 'chars': chars, 'caps': caps, 'tags': tags, }) return data def augment_with_pretrained(dictionary, ext_emb_path, words): """ Augment the dictionary with words that have a pretrained embedding. If `words` is None, we add every word that has a pretrained embedding to the dictionary, otherwise, we only add the words that are given by `words` (typically the words in the development and test sets.) """ print 'Loading pretrained embeddings from %s...' % ext_emb_path assert os.path.isfile(ext_emb_path) pretrained = set([ line.rstrip().split()[0].strip() for line in codecs.open(ext_emb_path, 'r', 'utf-8') if len(ext_emb_path) > 0 ]) if words is None: for word in pretrained: if word not in dictionary: dictionary[word] = 0 else: for word in words: if any(x in pretrained for x in [ word, word.lower(), re.sub('\d', '0', word.lower()) ]) and word not in dictionary: dictionary[word] = 0 word_to_id, id_to_word = create_mapping(dictionary) return dictionary, word_to_id, id_to_word
false
true
f70eac78971c0c7d00f18c3d2c1b7bbe17c6918c
11,789
py
Python
napari/_vispy/overlays/axes.py
kolibril13/napari
b39647d94e587f0255b0d4cc3087855e160a8929
[ "BSD-3-Clause" ]
null
null
null
napari/_vispy/overlays/axes.py
kolibril13/napari
b39647d94e587f0255b0d4cc3087855e160a8929
[ "BSD-3-Clause" ]
null
null
null
napari/_vispy/overlays/axes.py
kolibril13/napari
b39647d94e587f0255b0d4cc3087855e160a8929
[ "BSD-3-Clause" ]
null
null
null
import numpy as np from vispy.scene.visuals import Compound, Line, Mesh, Text from vispy.visuals.transforms import STTransform from ...layers.shapes._shapes_utils import triangulate_ellipse from ...utils.colormaps.standardize_color import transform_color from ...utils.theme import get_theme from ...utils.translations import trans def make_dashed_line(num_dashes, axis): """Make a dashed line. Parameters ---------- num_dashes : int Number of dashes in the line. axis : int Axis which is dashed. Returns ------- np.ndarray Dashed line, of shape (num_dashes, 3) with zeros in the non dashed axes and line segments in the dashed axis. """ dashes = np.linspace(0, 1, num_dashes * 2) dashed_line_ends = np.concatenate( [[dashes[2 * i], dashes[2 * i + 1]] for i in range(num_dashes)], axis=0 ) dashed_line = np.zeros((2 * num_dashes, 3)) dashed_line[:, axis] = np.array(dashed_line_ends) return dashed_line def make_arrow_head(num_segments, axis): """Make an arrowhead line. Parameters ---------- num_segments : int Number of segments in the arrowhead. axis Arrowhead direction. Returns ------- np.ndarray, np.ndarray Vertices and faces of the arrowhead. """ corners = np.array([[-1, -1], [-1, 1], [1, 1], [1, -1]]) * 0.1 vertices, faces = triangulate_ellipse(corners, num_segments) full_vertices = np.zeros((num_segments + 1, 3)) inds = list(range(3)) inds.pop(axis) full_vertices[:, inds] = vertices full_vertices[:, axis] = 0.9 full_vertices[0, axis] = 1.02 return full_vertices, faces def color_lines(colors): if len(colors) == 2: return np.concatenate( [[colors[0]] * 2, [colors[1]] * 2], axis=0, ) elif len(colors) == 3: return np.concatenate( [[colors[0]] * 2, [colors[1]] * 2, [colors[2]] * 2], axis=0, ) else: return ValueError( trans._( 'Either 2 or 3 colors must be provided, got {number}.', deferred=True, number=len(colors), ) ) def color_dashed_lines(colors): if len(colors) == 2: return np.concatenate( [[colors[0]] * 2, [colors[1]] * 4 * 2], axis=0, ) elif len(colors) == 3: return np.concatenate( [[colors[0]] * 2, [colors[1]] * 4 * 2, [colors[2]] * 8 * 2], axis=0, ) else: return ValueError( trans._( 'Either 2 or 3 colors must be provided, got {number}.', deferred=True, number=len(colors), ) ) def color_arrowheads(colors, num_segments): if len(colors) == 2: return np.concatenate( [[colors[0]] * num_segments, [colors[1]] * num_segments], axis=0, ) elif len(colors) == 3: return np.concatenate( [ [colors[0]] * num_segments, [colors[1]] * num_segments, [colors[2]] * num_segments, ], axis=0, ) else: return ValueError( trans._( 'Either 2 or 3 colors must be provided, got {number}.', deferred=True, number=len(colors), ) ) class VispyAxesOverlay: """Axes indicating world coordinate origin and orientation.""" _NUM_SEGMENTS_ARROWHEAD = 100 def __init__(self, viewer, parent=None, order=0): self._viewer = viewer self._scale = 1 # Target axes length in canvas pixels self._target_length = 80 # CMYRGB for 6 axes data in x, y, z, ... ordering self._default_color = [ [0, 1, 1, 1], [1, 0, 1, 1], [1, 1, 0, 1], [1, 0, 0, 1], [0, 1, 0, 1], [0, 0, 1, 1], ] # Text offset from line end position self._text_offsets = 0.1 * np.array([1, 1, 1]) # note order is x, y, z for VisPy self._line_data2D = np.array( [[0, 0, 0], [1, 0, 0], [0, 0, 0], [0, 1, 0]] ) self._line_data3D = np.array( [[0, 0, 0], [1, 0, 0], [0, 0, 0], [0, 1, 0], [0, 0, 0], [0, 0, 1]] ) # note order is x, y, z for VisPy self._dashed_line_data2D = np.concatenate( [[[1, 0, 0], [0, 0, 0]], make_dashed_line(4, axis=1)], axis=0, ) self._dashed_line_data3D = np.concatenate( [ [[1, 0, 0], [0, 0, 0]], make_dashed_line(4, axis=1), make_dashed_line(8, axis=2), ], axis=0, ) # note order is x, y, z for VisPy vertices = np.empty((0, 3)) faces = np.empty((0, 3)) for axis in range(2): v, f = make_arrow_head(self._NUM_SEGMENTS_ARROWHEAD, axis) faces = np.concatenate([faces, f + len(vertices)], axis=0) vertices = np.concatenate([vertices, v], axis=0) self._default_arrow_vertices2D = vertices self._default_arrow_faces2D = faces.astype(int) vertices = np.empty((0, 3)) faces = np.empty((0, 3)) for axis in range(3): v, f = make_arrow_head(self._NUM_SEGMENTS_ARROWHEAD, axis) faces = np.concatenate([faces, f + len(vertices)], axis=0) vertices = np.concatenate([vertices, v], axis=0) self._default_arrow_vertices3D = vertices self._default_arrow_faces3D = faces.astype(int) self.node = Compound( [Line(connect='segments', method='gl', width=3), Mesh(), Text()], parent=parent, ) self.node.transform = STTransform() self.node.order = order # Add a text node to display axes labels self.text_node = self.node._subvisuals[2] self.text_node.font_size = 10 self.text_node.anchors = ('center', 'center') self.text_node.text = f'{1}' self.node.canvas._backend.destroyed.connect(self._set_canvas_none) # End Note self._viewer.events.theme.connect(self._on_data_change) self._viewer.axes.events.visible.connect(self._on_visible_change) self._viewer.axes.events.colored.connect(self._on_data_change) self._viewer.axes.events.dashed.connect(self._on_data_change) self._viewer.axes.events.labels.connect(self._on_data_change) self._viewer.axes.events.arrows.connect(self._on_data_change) self._viewer.dims.events.order.connect(self._on_data_change) self._viewer.dims.events.range.connect(self._on_data_change) self._viewer.dims.events.ndisplay.connect(self._on_data_change) self._viewer.dims.events.axis_labels.connect(self._on_data_change) self._viewer.camera.events.zoom.connect(self._on_zoom_change) self._on_visible_change(None) self._on_data_change(None) def _set_canvas_none(self): self.node._set_canvas(None) self.text_node._set_canvas(None) def _on_visible_change(self, event): """Change visibiliy of axes.""" self.node.visible = self._viewer.axes.visible self._on_zoom_change(event) self._on_data_change(event) def _on_data_change(self, event): """Change style of axes.""" if not self._viewer.axes.visible: return # Determine which axes are displayed axes = self._viewer.dims.displayed # Actual number of displayed dims ndisplay = len(self._viewer.dims.displayed) # Determine the labels of those axes axes_labels = [self._viewer.dims.axis_labels[a] for a in axes[::-1]] # Counting backwards from total number of dimensions # determine axes positions. This is done as by default # the last NumPy axis corresponds to the first Vispy axis reversed_axes = [self._viewer.dims.ndim - 1 - a for a in axes[::-1]] # Determine colors of axes based on reverse position if self._viewer.axes.colored: axes_colors = [ self._default_color[ra % len(self._default_color)] for ra in reversed_axes ] else: # the reason for using the `as_hex` here is to avoid # `UserWarning` which is emitted when RGB values are above 1 background_color = get_theme( self._viewer.theme, False ).canvas.as_hex() background_color = transform_color(background_color)[0] color = np.subtract(1, background_color) color[-1] = background_color[-1] axes_colors = [color] * ndisplay # Determine data based on number of displayed dimensions and # axes visualization parameters if self._viewer.axes.dashed and ndisplay == 2: data = self._dashed_line_data2D color = color_dashed_lines(axes_colors) text_data = self._line_data2D[1::2] elif self._viewer.axes.dashed and ndisplay == 3: data = self._dashed_line_data3D color = color_dashed_lines(axes_colors) text_data = self._line_data3D[1::2] elif not self._viewer.axes.dashed and ndisplay == 2: data = self._line_data2D color = color_lines(axes_colors) text_data = self._line_data2D[1::2] elif not self._viewer.axes.dashed and ndisplay == 3: data = self._line_data3D color = color_lines(axes_colors) text_data = self._line_data3D[1::2] else: raise ValueError( trans._( 'Axes dash status and ndisplay combination not supported', deferred=True, ) ) if self._viewer.axes.arrows and ndisplay == 2: arrow_vertices = self._default_arrow_vertices2D arrow_faces = self._default_arrow_faces2D arrow_color = color_arrowheads( axes_colors, self._NUM_SEGMENTS_ARROWHEAD ) elif self._viewer.axes.arrows and ndisplay == 3: arrow_vertices = self._default_arrow_vertices3D arrow_faces = self._default_arrow_faces3D arrow_color = color_arrowheads( axes_colors, self._NUM_SEGMENTS_ARROWHEAD ) else: arrow_vertices = np.zeros((3, 3)) arrow_faces = np.array([[0, 1, 2]]) arrow_color = [[0, 0, 0, 0]] self.node._subvisuals[0].set_data(data, color) self.node._subvisuals[1].set_data( vertices=arrow_vertices, faces=arrow_faces, face_colors=arrow_color, ) # Set visibility status of text self.text_node.visible = ( self._viewer.axes.visible and self._viewer.axes.labels ) self.text_node.text = axes_labels self.text_node.color = axes_colors self.text_node.pos = text_data + self._text_offsets def _on_zoom_change(self, event): """Update axes length based on zoom scale.""" if not self._viewer.axes.visible: return scale = 1 / self._viewer.camera.zoom # If scale has not changed, do not redraw if abs(np.log10(self._scale) - np.log10(scale)) < 1e-4: return self._scale = scale scale_canvas2world = self._scale target_canvas_pixels = self._target_length scale = target_canvas_pixels * scale_canvas2world # Update axes scale self.node.transform.scale = [scale, scale, scale, 1]
34.270349
79
0.573925
import numpy as np from vispy.scene.visuals import Compound, Line, Mesh, Text from vispy.visuals.transforms import STTransform from ...layers.shapes._shapes_utils import triangulate_ellipse from ...utils.colormaps.standardize_color import transform_color from ...utils.theme import get_theme from ...utils.translations import trans def make_dashed_line(num_dashes, axis): dashes = np.linspace(0, 1, num_dashes * 2) dashed_line_ends = np.concatenate( [[dashes[2 * i], dashes[2 * i + 1]] for i in range(num_dashes)], axis=0 ) dashed_line = np.zeros((2 * num_dashes, 3)) dashed_line[:, axis] = np.array(dashed_line_ends) return dashed_line def make_arrow_head(num_segments, axis): corners = np.array([[-1, -1], [-1, 1], [1, 1], [1, -1]]) * 0.1 vertices, faces = triangulate_ellipse(corners, num_segments) full_vertices = np.zeros((num_segments + 1, 3)) inds = list(range(3)) inds.pop(axis) full_vertices[:, inds] = vertices full_vertices[:, axis] = 0.9 full_vertices[0, axis] = 1.02 return full_vertices, faces def color_lines(colors): if len(colors) == 2: return np.concatenate( [[colors[0]] * 2, [colors[1]] * 2], axis=0, ) elif len(colors) == 3: return np.concatenate( [[colors[0]] * 2, [colors[1]] * 2, [colors[2]] * 2], axis=0, ) else: return ValueError( trans._( 'Either 2 or 3 colors must be provided, got {number}.', deferred=True, number=len(colors), ) ) def color_dashed_lines(colors): if len(colors) == 2: return np.concatenate( [[colors[0]] * 2, [colors[1]] * 4 * 2], axis=0, ) elif len(colors) == 3: return np.concatenate( [[colors[0]] * 2, [colors[1]] * 4 * 2, [colors[2]] * 8 * 2], axis=0, ) else: return ValueError( trans._( 'Either 2 or 3 colors must be provided, got {number}.', deferred=True, number=len(colors), ) ) def color_arrowheads(colors, num_segments): if len(colors) == 2: return np.concatenate( [[colors[0]] * num_segments, [colors[1]] * num_segments], axis=0, ) elif len(colors) == 3: return np.concatenate( [ [colors[0]] * num_segments, [colors[1]] * num_segments, [colors[2]] * num_segments, ], axis=0, ) else: return ValueError( trans._( 'Either 2 or 3 colors must be provided, got {number}.', deferred=True, number=len(colors), ) ) class VispyAxesOverlay: _NUM_SEGMENTS_ARROWHEAD = 100 def __init__(self, viewer, parent=None, order=0): self._viewer = viewer self._scale = 1 self._target_length = 80 self._default_color = [ [0, 1, 1, 1], [1, 0, 1, 1], [1, 1, 0, 1], [1, 0, 0, 1], [0, 1, 0, 1], [0, 0, 1, 1], ] self._text_offsets = 0.1 * np.array([1, 1, 1]) self._line_data2D = np.array( [[0, 0, 0], [1, 0, 0], [0, 0, 0], [0, 1, 0]] ) self._line_data3D = np.array( [[0, 0, 0], [1, 0, 0], [0, 0, 0], [0, 1, 0], [0, 0, 0], [0, 0, 1]] ) self._dashed_line_data2D = np.concatenate( [[[1, 0, 0], [0, 0, 0]], make_dashed_line(4, axis=1)], axis=0, ) self._dashed_line_data3D = np.concatenate( [ [[1, 0, 0], [0, 0, 0]], make_dashed_line(4, axis=1), make_dashed_line(8, axis=2), ], axis=0, ) vertices = np.empty((0, 3)) faces = np.empty((0, 3)) for axis in range(2): v, f = make_arrow_head(self._NUM_SEGMENTS_ARROWHEAD, axis) faces = np.concatenate([faces, f + len(vertices)], axis=0) vertices = np.concatenate([vertices, v], axis=0) self._default_arrow_vertices2D = vertices self._default_arrow_faces2D = faces.astype(int) vertices = np.empty((0, 3)) faces = np.empty((0, 3)) for axis in range(3): v, f = make_arrow_head(self._NUM_SEGMENTS_ARROWHEAD, axis) faces = np.concatenate([faces, f + len(vertices)], axis=0) vertices = np.concatenate([vertices, v], axis=0) self._default_arrow_vertices3D = vertices self._default_arrow_faces3D = faces.astype(int) self.node = Compound( [Line(connect='segments', method='gl', width=3), Mesh(), Text()], parent=parent, ) self.node.transform = STTransform() self.node.order = order self.text_node = self.node._subvisuals[2] self.text_node.font_size = 10 self.text_node.anchors = ('center', 'center') self.text_node.text = f'{1}' self.node.canvas._backend.destroyed.connect(self._set_canvas_none) self._viewer.events.theme.connect(self._on_data_change) self._viewer.axes.events.visible.connect(self._on_visible_change) self._viewer.axes.events.colored.connect(self._on_data_change) self._viewer.axes.events.dashed.connect(self._on_data_change) self._viewer.axes.events.labels.connect(self._on_data_change) self._viewer.axes.events.arrows.connect(self._on_data_change) self._viewer.dims.events.order.connect(self._on_data_change) self._viewer.dims.events.range.connect(self._on_data_change) self._viewer.dims.events.ndisplay.connect(self._on_data_change) self._viewer.dims.events.axis_labels.connect(self._on_data_change) self._viewer.camera.events.zoom.connect(self._on_zoom_change) self._on_visible_change(None) self._on_data_change(None) def _set_canvas_none(self): self.node._set_canvas(None) self.text_node._set_canvas(None) def _on_visible_change(self, event): self.node.visible = self._viewer.axes.visible self._on_zoom_change(event) self._on_data_change(event) def _on_data_change(self, event): if not self._viewer.axes.visible: return axes = self._viewer.dims.displayed ndisplay = len(self._viewer.dims.displayed) axes_labels = [self._viewer.dims.axis_labels[a] for a in axes[::-1]] reversed_axes = [self._viewer.dims.ndim - 1 - a for a in axes[::-1]] if self._viewer.axes.colored: axes_colors = [ self._default_color[ra % len(self._default_color)] for ra in reversed_axes ] else: background_color = get_theme( self._viewer.theme, False ).canvas.as_hex() background_color = transform_color(background_color)[0] color = np.subtract(1, background_color) color[-1] = background_color[-1] axes_colors = [color] * ndisplay if self._viewer.axes.dashed and ndisplay == 2: data = self._dashed_line_data2D color = color_dashed_lines(axes_colors) text_data = self._line_data2D[1::2] elif self._viewer.axes.dashed and ndisplay == 3: data = self._dashed_line_data3D color = color_dashed_lines(axes_colors) text_data = self._line_data3D[1::2] elif not self._viewer.axes.dashed and ndisplay == 2: data = self._line_data2D color = color_lines(axes_colors) text_data = self._line_data2D[1::2] elif not self._viewer.axes.dashed and ndisplay == 3: data = self._line_data3D color = color_lines(axes_colors) text_data = self._line_data3D[1::2] else: raise ValueError( trans._( 'Axes dash status and ndisplay combination not supported', deferred=True, ) ) if self._viewer.axes.arrows and ndisplay == 2: arrow_vertices = self._default_arrow_vertices2D arrow_faces = self._default_arrow_faces2D arrow_color = color_arrowheads( axes_colors, self._NUM_SEGMENTS_ARROWHEAD ) elif self._viewer.axes.arrows and ndisplay == 3: arrow_vertices = self._default_arrow_vertices3D arrow_faces = self._default_arrow_faces3D arrow_color = color_arrowheads( axes_colors, self._NUM_SEGMENTS_ARROWHEAD ) else: arrow_vertices = np.zeros((3, 3)) arrow_faces = np.array([[0, 1, 2]]) arrow_color = [[0, 0, 0, 0]] self.node._subvisuals[0].set_data(data, color) self.node._subvisuals[1].set_data( vertices=arrow_vertices, faces=arrow_faces, face_colors=arrow_color, ) self.text_node.visible = ( self._viewer.axes.visible and self._viewer.axes.labels ) self.text_node.text = axes_labels self.text_node.color = axes_colors self.text_node.pos = text_data + self._text_offsets def _on_zoom_change(self, event): if not self._viewer.axes.visible: return scale = 1 / self._viewer.camera.zoom if abs(np.log10(self._scale) - np.log10(scale)) < 1e-4: return self._scale = scale scale_canvas2world = self._scale target_canvas_pixels = self._target_length scale = target_canvas_pixels * scale_canvas2world self.node.transform.scale = [scale, scale, scale, 1]
true
true
f70eacd9831e30421747847c4c2092590d7cb3b4
4,179
py
Python
dataloader.py
PaperCodeReview/MoCo-TF
1ea01b2d005de3e030229f79a37135468fa1631e
[ "MIT" ]
22
2020-10-01T10:14:36.000Z
2022-02-02T12:20:42.000Z
dataloader.py
PaperCodeReview/MoCo-TF
1ea01b2d005de3e030229f79a37135468fa1631e
[ "MIT" ]
2
2021-06-25T06:06:50.000Z
2021-11-08T23:43:38.000Z
dataloader.py
PaperCodeReview/MoCo-TF
1ea01b2d005de3e030229f79a37135468fa1631e
[ "MIT" ]
4
2021-03-03T06:19:45.000Z
2021-05-20T08:07:50.000Z
import os import random import numpy as np import pandas as pd import tensorflow as tf from augment import Augment AUTO = tf.data.experimental.AUTOTUNE def set_dataset(task, data_path): trainset = pd.read_csv( os.path.join( data_path, 'imagenet_trainset.csv' )).values.tolist() trainset = [[os.path.join(data_path, t[0]), t[1]] for t in trainset] if task == 'lincls': valset = pd.read_csv( os.path.join( data_path, 'imagenet_valset.csv' )).values.tolist() valset = [[os.path.join(data_path, t[0]), t[1]] for t in valset] return np.array(trainset, dtype='object'), np.array(valset, dtype='object') return np.array(trainset, dtype='object') class DataLoader: def __init__(self, args, mode, datalist, batch_size, num_workers=1, shuffle=True): self.args = args self.mode = mode self.datalist = datalist self.batch_size = batch_size self.num_workers = num_workers self.shuffle = shuffle self.dataloader = self._dataloader() def __len__(self): return len(self.datalist) def fetch_dataset(self, path, y=None): x = tf.io.read_file(path) if y is not None: return tf.data.Dataset.from_tensors((x, y)) return tf.data.Dataset.from_tensors(x) def augmentation(self, img, shape): augset = Augment(self.args, self.mode) if self.args.task in ['v1', 'v2']: img_list = [] for _ in range(2): # query, key aug_img = tf.identity(img) if self.args.task == 'v1': aug_img = augset._augmentv1(aug_img, shape) # moco v1 else: radius = np.random.choice([3, 5]) aug_img = augset._augmentv2(aug_img, shape, (radius, radius)) # moco v2 img_list.append(aug_img) return img_list else: return augset._augment_lincls(img, shape) def dataset_parser(self, value, label=None): shape = tf.image.extract_jpeg_shape(value) img = tf.io.decode_jpeg(value, channels=3) if label is None: # moco query, key = self.augmentation(img, shape) inputs = {'query': query, 'key': key} labels = tf.zeros([]) else: # lincls inputs = self.augmentation(img, shape) labels = tf.one_hot(label, self.args.classes) return (inputs, labels) def shuffle_BN(self, value, labels): if self.num_workers > 1: pre_shuffle = [(i, value['key'][i]) for i in range(self.batch_size)] random.shuffle(pre_shuffle) shuffle_idx = [] value_temp = [] for vv in pre_shuffle: shuffle_idx.append(vv[0]) value_temp.append(tf.expand_dims(vv[1], axis=0)) value['key'] = tf.concat(value_temp, axis=0) unshuffle_idx = np.array(shuffle_idx).argsort().tolist() value.update({'unshuffle': unshuffle_idx}) return (value, labels) def _dataloader(self): self.imglist = self.datalist[:,0].tolist() if self.args.task in ['v1', 'v2']: dataset = tf.data.Dataset.from_tensor_slices(self.imglist) else: self.labellist = self.datalist[:,1].tolist() dataset = tf.data.Dataset.from_tensor_slices((self.imglist, self.labellist)) dataset = dataset.repeat() if self.shuffle: dataset = dataset.shuffle(len(self.datalist)) dataset = dataset.interleave(self.fetch_dataset, num_parallel_calls=AUTO) dataset = dataset.map(self.dataset_parser, num_parallel_calls=AUTO) dataset = dataset.batch(self.batch_size) dataset = dataset.prefetch(AUTO) if self.args.shuffle_bn and self.args.task in ['v1', 'v2']: # only moco dataset = dataset.map(self.shuffle_BN, num_parallel_calls=AUTO) return dataset
36.657895
92
0.569275
import os import random import numpy as np import pandas as pd import tensorflow as tf from augment import Augment AUTO = tf.data.experimental.AUTOTUNE def set_dataset(task, data_path): trainset = pd.read_csv( os.path.join( data_path, 'imagenet_trainset.csv' )).values.tolist() trainset = [[os.path.join(data_path, t[0]), t[1]] for t in trainset] if task == 'lincls': valset = pd.read_csv( os.path.join( data_path, 'imagenet_valset.csv' )).values.tolist() valset = [[os.path.join(data_path, t[0]), t[1]] for t in valset] return np.array(trainset, dtype='object'), np.array(valset, dtype='object') return np.array(trainset, dtype='object') class DataLoader: def __init__(self, args, mode, datalist, batch_size, num_workers=1, shuffle=True): self.args = args self.mode = mode self.datalist = datalist self.batch_size = batch_size self.num_workers = num_workers self.shuffle = shuffle self.dataloader = self._dataloader() def __len__(self): return len(self.datalist) def fetch_dataset(self, path, y=None): x = tf.io.read_file(path) if y is not None: return tf.data.Dataset.from_tensors((x, y)) return tf.data.Dataset.from_tensors(x) def augmentation(self, img, shape): augset = Augment(self.args, self.mode) if self.args.task in ['v1', 'v2']: img_list = [] for _ in range(2): aug_img = tf.identity(img) if self.args.task == 'v1': aug_img = augset._augmentv1(aug_img, shape) else: radius = np.random.choice([3, 5]) aug_img = augset._augmentv2(aug_img, shape, (radius, radius)) img_list.append(aug_img) return img_list else: return augset._augment_lincls(img, shape) def dataset_parser(self, value, label=None): shape = tf.image.extract_jpeg_shape(value) img = tf.io.decode_jpeg(value, channels=3) if label is None: query, key = self.augmentation(img, shape) inputs = {'query': query, 'key': key} labels = tf.zeros([]) else: inputs = self.augmentation(img, shape) labels = tf.one_hot(label, self.args.classes) return (inputs, labels) def shuffle_BN(self, value, labels): if self.num_workers > 1: pre_shuffle = [(i, value['key'][i]) for i in range(self.batch_size)] random.shuffle(pre_shuffle) shuffle_idx = [] value_temp = [] for vv in pre_shuffle: shuffle_idx.append(vv[0]) value_temp.append(tf.expand_dims(vv[1], axis=0)) value['key'] = tf.concat(value_temp, axis=0) unshuffle_idx = np.array(shuffle_idx).argsort().tolist() value.update({'unshuffle': unshuffle_idx}) return (value, labels) def _dataloader(self): self.imglist = self.datalist[:,0].tolist() if self.args.task in ['v1', 'v2']: dataset = tf.data.Dataset.from_tensor_slices(self.imglist) else: self.labellist = self.datalist[:,1].tolist() dataset = tf.data.Dataset.from_tensor_slices((self.imglist, self.labellist)) dataset = dataset.repeat() if self.shuffle: dataset = dataset.shuffle(len(self.datalist)) dataset = dataset.interleave(self.fetch_dataset, num_parallel_calls=AUTO) dataset = dataset.map(self.dataset_parser, num_parallel_calls=AUTO) dataset = dataset.batch(self.batch_size) dataset = dataset.prefetch(AUTO) if self.args.shuffle_bn and self.args.task in ['v1', 'v2']: dataset = dataset.map(self.shuffle_BN, num_parallel_calls=AUTO) return dataset
true
true
f70eacda302b793da33c4fb5beb6b45c6cc933ec
1,315
py
Python
user_modeling.py
leeamen/k_means
dfa9cad22033c108e3988a99f4d58c685eb06921
[ "MIT" ]
null
null
null
user_modeling.py
leeamen/k_means
dfa9cad22033c108e3988a99f4d58c685eb06921
[ "MIT" ]
null
null
null
user_modeling.py
leeamen/k_means
dfa9cad22033c108e3988a99f4d58c685eb06921
[ "MIT" ]
null
null
null
#!/usr/bin/python #coding:utf-8 import numpy as np import logging import mylog import mykmeans as ml logger = logging.getLogger(__name__) logger.setLevel(logging.ERROR) def str2num(s): a = ['very_low', 'Low', 'Middle', 'High'] for i in range(0, len(a)): if a[i] == s: return float(i) if __name__ == '__main__': filename = './data/data_user_modeling.txt' train_data = np.loadtxt(filename, delimiter = ',', converters = {5:str2num}) logger.debug(train_data) logger.debug(train_data.shape) train_x = train_data[:,0:-1] train_y = train_data[:,-1] logger.debug(train_x) logger.debug(train_y) param = {} param['use_random_for_k'] = 1 param['k'] = [i for i in range(0, 258, 1)] param['n_clusters'] = 4 param['max_iter'] = 100 kmeans = ml.Kmeans(param) kmeans.Fit(train_x) # logger.debug(kmeans) pred = kmeans.Predict(train_x) logger.info('train_y:%s', train_y) logger.info(' pred:%s', pred) # logger.info('k-means准确率:%f', 1.0*sum(pred == train_y)/len(train_y)) # ml.PickingRightK(train_x, param) import myplot myplot.Figure() ml.FitMulti(train_x, param, 100) ml.BisectingFitMulti(train_x, param, 100) myplot.Legend(['k-means','bisecting']) myplot.Title('user modeling') myplot.Show()
28.586957
108
0.641825
import numpy as np import logging import mylog import mykmeans as ml logger = logging.getLogger(__name__) logger.setLevel(logging.ERROR) def str2num(s): a = ['very_low', 'Low', 'Middle', 'High'] for i in range(0, len(a)): if a[i] == s: return float(i) if __name__ == '__main__': filename = './data/data_user_modeling.txt' train_data = np.loadtxt(filename, delimiter = ',', converters = {5:str2num}) logger.debug(train_data) logger.debug(train_data.shape) train_x = train_data[:,0:-1] train_y = train_data[:,-1] logger.debug(train_x) logger.debug(train_y) param = {} param['use_random_for_k'] = 1 param['k'] = [i for i in range(0, 258, 1)] param['n_clusters'] = 4 param['max_iter'] = 100 kmeans = ml.Kmeans(param) kmeans.Fit(train_x) pred = kmeans.Predict(train_x) logger.info('train_y:%s', train_y) logger.info(' pred:%s', pred) import myplot myplot.Figure() ml.FitMulti(train_x, param, 100) ml.BisectingFitMulti(train_x, param, 100) myplot.Legend(['k-means','bisecting']) myplot.Title('user modeling') myplot.Show()
true
true
f70eaf3e78788501749a3d605244c7e2c2fa2be1
448
py
Python
creme/linear_model/__init__.py
Raul9595/creme
39cec7ac27ccd40ff0a7bdd6bceaf7ce25c1a8da
[ "BSD-3-Clause" ]
1
2020-07-27T03:06:46.000Z
2020-07-27T03:06:46.000Z
creme/linear_model/__init__.py
2torus/creme
bcc5e2a0155663a1f0ba779c68f23456695bcb54
[ "BSD-3-Clause" ]
null
null
null
creme/linear_model/__init__.py
2torus/creme
bcc5e2a0155663a1f0ba779c68f23456695bcb54
[ "BSD-3-Clause" ]
null
null
null
""" Generalized linear models optimized with online gradient descent from :mod:`creme.optim`. """ from .fm import FMRegressor from .lin_reg import LinearRegression from .log_reg import LogisticRegression from .pa import PAClassifier from .pa import PARegressor from .softmax import SoftmaxRegression __all__ = [ 'FMRegressor', 'LinearRegression', 'LogisticRegression', 'PAClassifier', 'PARegressor', 'SoftmaxRegression' ]
22.4
89
0.754464
from .fm import FMRegressor from .lin_reg import LinearRegression from .log_reg import LogisticRegression from .pa import PAClassifier from .pa import PARegressor from .softmax import SoftmaxRegression __all__ = [ 'FMRegressor', 'LinearRegression', 'LogisticRegression', 'PAClassifier', 'PARegressor', 'SoftmaxRegression' ]
true
true
f70eafdd54bf8b3c012d6d488772e18beb4c652b
8,654
py
Python
glue/plugins/export_d3po.py
bsipocz/glue
7b7e4879b4c746b2419a0eca2a17c2d07a3fded3
[ "BSD-3-Clause" ]
null
null
null
glue/plugins/export_d3po.py
bsipocz/glue
7b7e4879b4c746b2419a0eca2a17c2d07a3fded3
[ "BSD-3-Clause" ]
null
null
null
glue/plugins/export_d3po.py
bsipocz/glue
7b7e4879b4c746b2419a0eca2a17c2d07a3fded3
[ "BSD-3-Clause" ]
null
null
null
import json import os from astropy.table import Table, Column from ..config import exporters from ..qt.widgets import ScatterWidget, HistogramWidget from ..core import Subset def save_page(page, page_number, label, subset): """ Convert a tab of a glue session into a D3PO page :param page: Tuple of data viewers to save :param label: Tab label """ result = {} # layout settings result['grid'] = {'nRows': 1, 'nColumns': len(page)} result['name'] = str(label) result['caption'] = 'Generated by Glue' # style settings d = page[0]._data[0] unselected = dict(opacity=d.style.alpha, size=d.style.markersize / 2, color=d.style.color) result['markerStyle'] = dict(unselected=unselected) if subset is not None: s = subset.style selected = dict(opacity=s.alpha, size=s.markersize / 2, color=s.color) result['markerStyle']['selected'] = selected result['selection'] = {'type': 'booleanColumn', 'columnName': 'selection_%i' % page_number} result['histogramStyle'] = result['markerStyle'] # save each plot result['plots'] = map(save_plot, page, range(len(page))) return result def save_plot_base(plot, index): result = {} result['gridPosition'] = [0, index] return result def save_plot(plot, index): dispatch = {ScatterWidget: save_scatter, HistogramWidget: save_histogram} typ = type(plot) return dispatch[typ](plot, index) def save_scatter(plot, index): """ Convert a single glue scatter plot to a D3PO plot :param plot: Glue scatter plot :class:`~glue.qt.widgets.scatter_widget.ScatterWidget` :param index: 1D index of plot on the page :type index: int :rtype: json-serializable dict """ result = save_plot_base(plot, index) props = plot.properties result['type'] = 'scatter' result['xAxis'] = dict(columnName=props['xatt'].label, range=[props['xmin'], props['xmax']]) result['yAxis'] = dict(columnName=props['yatt'].label, range=[props['ymin'], props['ymax']]) # XXX log scales return result def save_histogram(plot, index): """ Convert a single histogram to a D3PO plot :param plot: Glue histogram :type plot: :class:`~glue.qt.widgets.histogram_widget.HistogramWidget` :param index: 1D index of plot on the page :type index: int :rtype: json-serializable dict """ result = save_plot_base(plot, index) props = plot.properties result['type'] = 'histogram' result['xAxis'] = dict(columnName=props['component'].label, bins=props['nbins'], range=[props['xmin'], props['xmax']]) # XXX normed, cumultive, log return result def stage_subsets(application): """ Return a tuple of the subset to use for each stage/tab, or None if the tab has no subset If more than one subset is used per stage/tab, returns None """ result = [] for page in application.viewers: subset = None for viewer in page: for layer_artist in viewer.layers: if not layer_artist.visible: continue s = layer_artist.layer if not isinstance(s, Subset): continue if subset is not None and s is not subset: return None if subset is None: subset = s result.append(subset) return tuple(result) def can_save_d3po(application): """ Check whether an application can be exported to D3PO. Raises an exception if not """ dc = application.session.data_collection if len(dc) != 1: raise ValueError("D3PO Export only supports a single dataset") data = dc[0] for tab in application.viewers: for viewer in tab: if not isinstance(viewer, (ScatterWidget, HistogramWidget)): raise ValueError("D3PO Export only supports scatter " "and histogram plots") if sum(len(tab) for tab in application.viewers) == 0: raise ValueError("D3PO Export requires at least one scatterplot " "or histogram") if stage_subsets(application) is None: raise ValueError("D3PO Export restricted to 0 or 1 subsets visible " "in each tab") def make_data_file(data, subsets, path): """ Create the data.csv file, given Data and tuple of subsets """ data_path = os.path.join(path, 'data.csv') t = Table([data[c] for c in data.components], names=[c.label for c in data.components]) for i, subset in enumerate(subsets): if subset is None: continue c = Column(data=subset.to_mask().astype('i'), name='selection_%i' % i) t.add_column(c) t.write(data_path, format='ascii', delimiter=',') def save_d3po(application, path): """Save a Glue session to a D3PO bundle. Currently, this has the following restrictions: - The Glue session must have only one dataset open, and 0 or 1 subsets - Only scatter plots or histograms are present - At least one plot is present :param application: Glue appication to save :param path: Path to directory to save in. Will be created if needed """ if os.path.exists(path) and not os.path.isdir(path): os.unlink(path) if not os.path.exists(path): os.mkdir(path) data = application.session.data_collection[0] subsets = stage_subsets(application) viewers = application.viewers # data.csv make_data_file(data, subsets, path) # states.json result = {} result['filename'] = 'data.csv' # XXX don't think this is needed? result['title'] = "Glue export of %s" % data.label result['states'] = map(save_page, application.viewers, range(len(viewers)), application.tab_names, subsets) state_path = os.path.join(path, 'states.json') with open(state_path, 'w') as outfile: json.dump(result, outfile, indent=2) # index.html html_path = os.path.join(path, 'index.html') with open(html_path, 'w') as outfile: outfile.write(HTML) # show the result launch(path) def launch(path): """Start a server to view an exported D3PO bundle, and open a browser. :param path: The TLD of the bundle """ from SocketServer import TCPServer from SimpleHTTPServer import SimpleHTTPRequestHandler from random import randrange from socket import error import webbrowser from threading import Thread os.chdir(path) while True: try: PORT = randrange(8000, 9000) server = TCPServer(("", PORT), SimpleHTTPRequestHandler, False) server.allow_reuse_address = True server.server_bind() break except error: # port already taken pass print 'Serving D3PO on port 0.0.0.0:%i' % PORT server.server_activate() thread = Thread(target=server.serve_forever) thread.setDaemon(True) # do not prevent shutdown thread.start() webbrowser.open('http://0.0.0.0:%i' % PORT) exporters.add('D3PO', save_d3po, can_save_d3po, outmode='directory') HTML = """ <!DOCTYPE html> <html> <head> <meta charset="utf-8" /> <link rel="stylesheet" type="text/css" href="http://d3po.org/static/css/style.css"> <link rel="stylesheet" type="text/css" href="http://d3po.org/static/css/d3po.css"> <link href='http://fonts.googleapis.com/css?family=Source+Sans+Pro:100,200,300,400,700' rel='stylesheet' type='text/css'> <style> #footer { position: fixed; bottom: 0; right: 0; } </style> <!-- not to be confused with Planet Telex --> <!-- Javscript dependencies --> <script src="http://d3js.org/d3.v3.min.js" charset="utf-8"></script> <script src="http://d3po.org/static/js/util.js"></script> <script src="//ajax.googleapis.com/ajax/libs/jquery/1.10.2/jquery.min.js"></script> <script src="http://d3po.org/static/js/d3po.js"></script> <script src="http://d3po.org/static/js/d3po.init.js"></script> </head> <body> <div id="svg"><svg></svg></div> <div id="controls"> <ul class="navigation"> </ul> </div> <div id="caption"></div> <div id="footer"> More information: <a href="http://d3po.org">d3po.org</a> </div> <script type="text/javascript"> $(document).ready(function() { initialize('states.json', 'data.csv'); } ); </script> </body> </html> """
28.846667
121
0.619829
import json import os from astropy.table import Table, Column from ..config import exporters from ..qt.widgets import ScatterWidget, HistogramWidget from ..core import Subset def save_page(page, page_number, label, subset): """ Convert a tab of a glue session into a D3PO page :param page: Tuple of data viewers to save :param label: Tab label """ result = {} result['grid'] = {'nRows': 1, 'nColumns': len(page)} result['name'] = str(label) result['caption'] = 'Generated by Glue' d = page[0]._data[0] unselected = dict(opacity=d.style.alpha, size=d.style.markersize / 2, color=d.style.color) result['markerStyle'] = dict(unselected=unselected) if subset is not None: s = subset.style selected = dict(opacity=s.alpha, size=s.markersize / 2, color=s.color) result['markerStyle']['selected'] = selected result['selection'] = {'type': 'booleanColumn', 'columnName': 'selection_%i' % page_number} result['histogramStyle'] = result['markerStyle'] result['plots'] = map(save_plot, page, range(len(page))) return result def save_plot_base(plot, index): result = {} result['gridPosition'] = [0, index] return result def save_plot(plot, index): dispatch = {ScatterWidget: save_scatter, HistogramWidget: save_histogram} typ = type(plot) return dispatch[typ](plot, index) def save_scatter(plot, index): """ Convert a single glue scatter plot to a D3PO plot :param plot: Glue scatter plot :class:`~glue.qt.widgets.scatter_widget.ScatterWidget` :param index: 1D index of plot on the page :type index: int :rtype: json-serializable dict """ result = save_plot_base(plot, index) props = plot.properties result['type'] = 'scatter' result['xAxis'] = dict(columnName=props['xatt'].label, range=[props['xmin'], props['xmax']]) result['yAxis'] = dict(columnName=props['yatt'].label, range=[props['ymin'], props['ymax']]) return result def save_histogram(plot, index): """ Convert a single histogram to a D3PO plot :param plot: Glue histogram :type plot: :class:`~glue.qt.widgets.histogram_widget.HistogramWidget` :param index: 1D index of plot on the page :type index: int :rtype: json-serializable dict """ result = save_plot_base(plot, index) props = plot.properties result['type'] = 'histogram' result['xAxis'] = dict(columnName=props['component'].label, bins=props['nbins'], range=[props['xmin'], props['xmax']]) return result def stage_subsets(application): """ Return a tuple of the subset to use for each stage/tab, or None if the tab has no subset If more than one subset is used per stage/tab, returns None """ result = [] for page in application.viewers: subset = None for viewer in page: for layer_artist in viewer.layers: if not layer_artist.visible: continue s = layer_artist.layer if not isinstance(s, Subset): continue if subset is not None and s is not subset: return None if subset is None: subset = s result.append(subset) return tuple(result) def can_save_d3po(application): """ Check whether an application can be exported to D3PO. Raises an exception if not """ dc = application.session.data_collection if len(dc) != 1: raise ValueError("D3PO Export only supports a single dataset") data = dc[0] for tab in application.viewers: for viewer in tab: if not isinstance(viewer, (ScatterWidget, HistogramWidget)): raise ValueError("D3PO Export only supports scatter " "and histogram plots") if sum(len(tab) for tab in application.viewers) == 0: raise ValueError("D3PO Export requires at least one scatterplot " "or histogram") if stage_subsets(application) is None: raise ValueError("D3PO Export restricted to 0 or 1 subsets visible " "in each tab") def make_data_file(data, subsets, path): """ Create the data.csv file, given Data and tuple of subsets """ data_path = os.path.join(path, 'data.csv') t = Table([data[c] for c in data.components], names=[c.label for c in data.components]) for i, subset in enumerate(subsets): if subset is None: continue c = Column(data=subset.to_mask().astype('i'), name='selection_%i' % i) t.add_column(c) t.write(data_path, format='ascii', delimiter=',') def save_d3po(application, path): """Save a Glue session to a D3PO bundle. Currently, this has the following restrictions: - The Glue session must have only one dataset open, and 0 or 1 subsets - Only scatter plots or histograms are present - At least one plot is present :param application: Glue appication to save :param path: Path to directory to save in. Will be created if needed """ if os.path.exists(path) and not os.path.isdir(path): os.unlink(path) if not os.path.exists(path): os.mkdir(path) data = application.session.data_collection[0] subsets = stage_subsets(application) viewers = application.viewers make_data_file(data, subsets, path) result = {} result['filename'] = 'data.csv' result['title'] = "Glue export of %s" % data.label result['states'] = map(save_page, application.viewers, range(len(viewers)), application.tab_names, subsets) state_path = os.path.join(path, 'states.json') with open(state_path, 'w') as outfile: json.dump(result, outfile, indent=2) # index.html html_path = os.path.join(path, 'index.html') with open(html_path, 'w') as outfile: outfile.write(HTML) # show the result launch(path) def launch(path): """Start a server to view an exported D3PO bundle, and open a browser. :param path: The TLD of the bundle """ from SocketServer import TCPServer from SimpleHTTPServer import SimpleHTTPRequestHandler from random import randrange from socket import error import webbrowser from threading import Thread os.chdir(path) while True: try: PORT = randrange(8000, 9000) server = TCPServer(("", PORT), SimpleHTTPRequestHandler, False) server.allow_reuse_address = True server.server_bind() break except error: # port already taken pass print 'Serving D3PO on port 0.0.0.0:%i' % PORT server.server_activate() thread = Thread(target=server.serve_forever) thread.setDaemon(True) # do not prevent shutdown thread.start() webbrowser.open('http://0.0.0.0:%i' % PORT) exporters.add('D3PO', save_d3po, can_save_d3po, outmode='directory') HTML = """ <!DOCTYPE html> <html> <head> <meta charset="utf-8" /> <link rel="stylesheet" type="text/css" href="http://d3po.org/static/css/style.css"> <link rel="stylesheet" type="text/css" href="http://d3po.org/static/css/d3po.css"> <link href='http://fonts.googleapis.com/css?family=Source+Sans+Pro:100,200,300,400,700' rel='stylesheet' type='text/css'> <style> #footer { position: fixed; bottom: 0; right: 0; } </style> <!-- not to be confused with Planet Telex --> <!-- Javscript dependencies --> <script src="http://d3js.org/d3.v3.min.js" charset="utf-8"></script> <script src="http://d3po.org/static/js/util.js"></script> <script src="//ajax.googleapis.com/ajax/libs/jquery/1.10.2/jquery.min.js"></script> <script src="http://d3po.org/static/js/d3po.js"></script> <script src="http://d3po.org/static/js/d3po.init.js"></script> </head> <body> <div id="svg"><svg></svg></div> <div id="controls"> <ul class="navigation"> </ul> </div> <div id="caption"></div> <div id="footer"> More information: <a href="http://d3po.org">d3po.org</a> </div> <script type="text/javascript"> $(document).ready(function() { initialize('states.json', 'data.csv'); } ); </script> </body> </html> """
false
true
f70eb08931175461e6046102ac021119a7ea22c2
8,490
py
Python
docs/conf.py
daherman/dedupe
053d373aaed47201f720c5b6d1a568fc49742cc3
[ "MIT" ]
1
2020-06-26T18:03:34.000Z
2020-06-26T18:03:34.000Z
docs/conf.py
daherman/dedupe
053d373aaed47201f720c5b6d1a568fc49742cc3
[ "MIT" ]
null
null
null
docs/conf.py
daherman/dedupe
053d373aaed47201f720c5b6d1a568fc49742cc3
[ "MIT" ]
null
null
null
# -*- coding: utf-8 -*- # # dedupe documentation build configuration file, created by # sphinx-quickstart on Thu Apr 10 11:27:59 2014. # # This file is execfile()d with the current directory set to its # containing dir. # # Note that not all possible configuration values are present in this # autogenerated file. # # All configuration values have a default; values that are commented out # serve to show the default. import sys import os # If extensions (or modules to document with autodoc) are in another directory, # add these directories to sys.path here. If the directory is relative to the # documentation root, use os.path.abspath to make it absolute, like shown here. #sys.path.insert(0, os.path.abspath('.')) # -- General configuration ------------------------------------------------ # If your documentation needs a minimal Sphinx version, state it here. #needs_sphinx = '1.0' # Add any Sphinx extension module names here, as strings. They can be # extensions coming with Sphinx (named 'sphinx.ext.*') or your custom # ones. extensions = [ 'sphinx.ext.todo', 'sphinx.ext.coverage', 'sphinx.ext.mathjax', 'sphinx.ext.viewcode', ] # Add any paths that contain templates here, relative to this directory. templates_path = ['_templates'] # The suffix of source filenames. source_suffix = '.rst' # The encoding of source files. #source_encoding = 'utf-8-sig' # The master toctree document. master_doc = 'index' # General information about the project. project = u'dedupe' copyright = u'2018, Forest Gregg, Derek Eder, and contributors' # The version info for the project you're documenting, acts as replacement for # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = '1.9.4' # The full version, including alpha/beta/rc tags. release = '1.9.4' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. #language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: #today = '' # Else, today_fmt is used as the format for a strftime call. #today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = ['_build', 'common_*.rst'] # The reST default role (used for this markup: `text`) to use for all # documents. #default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. #add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). #add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. #show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. #modindex_common_prefix = [] # If true, keep warnings as "system message" paragraphs in the built documents. #keep_warnings = False # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'sphinx_rtd_theme' # Custom stylesheet # html_style = 'css/custom.css' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. # html_theme_options = { # 'canonical_url': 'https://docs.dedupe.io/' # } # Add any paths that contain custom themes here, relative to this directory. # html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". #html_title = None # A shorter title for the navigation bar. Default is the same as html_title. #html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. #html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. #html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. #html_extra_path = [] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. #html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. #html_use_smartypants = True # Custom sidebar templates, maps document names to template names. #html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. #html_additional_pages = {} # If false, no module index is generated. #html_domain_indices = True # If false, no index is generated. #html_use_index = True # If true, the index is split into individual pages for each letter. #html_split_index = False # If true, links to the reST sources are added to the pages. #html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. #html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. #html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. #html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). #html_file_suffix = None # Output file base name for HTML help builder. htmlhelp_basename = 'dedupedoc' # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). #'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). #'pointsize': '10pt', # Additional stuff for the LaTeX preamble. #'preamble': '', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ ('index', 'dedupe.tex', u'dedupe Documentation', u'Forest Gregg, Derek Eder, and contributors', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. #latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. #latex_use_parts = False # If true, show page references after internal links. #latex_show_pagerefs = False # If true, show URL addresses after external links. #latex_show_urls = False # Documents to append as an appendix to all manuals. #latex_appendices = [] # If false, no module index is generated. #latex_domain_indices = True # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ ('index', 'dedupe', u'dedupe Documentation', [u'Forest Gregg, Derek Eder, and contributors'], 1) ] # If true, show URL addresses after external links. #man_show_urls = False # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ('index', 'dedupe', u'dedupe Documentation', u'Forest Gregg, Derek Eder, and contributors', 'dedupe', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. #texinfo_appendices = [] # If false, no module index is generated. #texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. #texinfo_show_urls = 'footnote' # If true, do not generate a @detailmenu in the "Top" node's menu. #texinfo_no_detailmenu = False
31.444444
95
0.715901
import sys import os extensions = [ 'sphinx.ext.todo', 'sphinx.ext.coverage', 'sphinx.ext.mathjax', 'sphinx.ext.viewcode', ] templates_path = ['_templates'] source_suffix = '.rst' master_doc = 'index' project = u'dedupe' copyright = u'2018, Forest Gregg, Derek Eder, and contributors' # |version| and |release|, also used in various other places throughout the # built documents. # # The short X.Y version. version = '1.9.4' # The full version, including alpha/beta/rc tags. release = '1.9.4' # The language for content autogenerated by Sphinx. Refer to documentation # for a list of supported languages. #language = None # There are two options for replacing |today|: either, you set today to some # non-false value, then it is used: #today = '' # Else, today_fmt is used as the format for a strftime call. #today_fmt = '%B %d, %Y' # List of patterns, relative to source directory, that match files and # directories to ignore when looking for source files. exclude_patterns = ['_build', 'common_*.rst'] # The reST default role (used for this markup: `text`) to use for all # documents. #default_role = None # If true, '()' will be appended to :func: etc. cross-reference text. #add_function_parentheses = True # If true, the current module name will be prepended to all description # unit titles (such as .. function::). #add_module_names = True # If true, sectionauthor and moduleauthor directives will be shown in the # output. They are ignored by default. #show_authors = False # The name of the Pygments (syntax highlighting) style to use. pygments_style = 'sphinx' # A list of ignored prefixes for module index sorting. #modindex_common_prefix = [] # If true, keep warnings as "system message" paragraphs in the built documents. #keep_warnings = False # -- Options for HTML output ---------------------------------------------- # The theme to use for HTML and HTML Help pages. See the documentation for # a list of builtin themes. html_theme = 'sphinx_rtd_theme' # Custom stylesheet # html_style = 'css/custom.css' # Theme options are theme-specific and customize the look and feel of a theme # further. For a list of options available for each theme, see the # documentation. # html_theme_options = { # 'canonical_url': 'https://docs.dedupe.io/' # } # Add any paths that contain custom themes here, relative to this directory. # html_theme_path = [] # The name for this set of Sphinx documents. If None, it defaults to # "<project> v<release> documentation". #html_title = None # A shorter title for the navigation bar. Default is the same as html_title. #html_short_title = None # The name of an image file (relative to this directory) to place at the top # of the sidebar. #html_logo = None # The name of an image file (within the static path) to use as favicon of the # docs. This file should be a Windows icon file (.ico) being 16x16 or 32x32 # pixels large. #html_favicon = None # Add any paths that contain custom static files (such as style sheets) here, # relative to this directory. They are copied after the builtin static files, # so a file named "default.css" will overwrite the builtin "default.css". html_static_path = ['_static'] # Add any extra paths that contain custom files (such as robots.txt or # .htaccess) here, relative to this directory. These files are copied # directly to the root of the documentation. #html_extra_path = [] # If not '', a 'Last updated on:' timestamp is inserted at every page bottom, # using the given strftime format. #html_last_updated_fmt = '%b %d, %Y' # If true, SmartyPants will be used to convert quotes and dashes to # typographically correct entities. #html_use_smartypants = True # Custom sidebar templates, maps document names to template names. #html_sidebars = {} # Additional templates that should be rendered to pages, maps page names to # template names. #html_additional_pages = {} # If false, no module index is generated. #html_domain_indices = True # If false, no index is generated. #html_use_index = True # If true, the index is split into individual pages for each letter. #html_split_index = False # If true, links to the reST sources are added to the pages. #html_show_sourcelink = True # If true, "Created using Sphinx" is shown in the HTML footer. Default is True. #html_show_sphinx = True # If true, "(C) Copyright ..." is shown in the HTML footer. Default is True. #html_show_copyright = True # If true, an OpenSearch description file will be output, and all pages will # contain a <link> tag referring to it. The value of this option must be the # base URL from which the finished HTML is served. #html_use_opensearch = '' # This is the file name suffix for HTML files (e.g. ".xhtml"). #html_file_suffix = None # Output file base name for HTML help builder. htmlhelp_basename = 'dedupedoc' # -- Options for LaTeX output --------------------------------------------- latex_elements = { # The paper size ('letterpaper' or 'a4paper'). #'papersize': 'letterpaper', # The font size ('10pt', '11pt' or '12pt'). #'pointsize': '10pt', # Additional stuff for the LaTeX preamble. #'preamble': '', } # Grouping the document tree into LaTeX files. List of tuples # (source start file, target name, title, # author, documentclass [howto, manual, or own class]). latex_documents = [ ('index', 'dedupe.tex', u'dedupe Documentation', u'Forest Gregg, Derek Eder, and contributors', 'manual'), ] # The name of an image file (relative to this directory) to place at the top of # the title page. #latex_logo = None # For "manual" documents, if this is true, then toplevel headings are parts, # not chapters. #latex_use_parts = False # If true, show page references after internal links. #latex_show_pagerefs = False # If true, show URL addresses after external links. #latex_show_urls = False # Documents to append as an appendix to all manuals. #latex_appendices = [] # If false, no module index is generated. #latex_domain_indices = True # -- Options for manual page output --------------------------------------- # One entry per manual page. List of tuples # (source start file, name, description, authors, manual section). man_pages = [ ('index', 'dedupe', u'dedupe Documentation', [u'Forest Gregg, Derek Eder, and contributors'], 1) ] # If true, show URL addresses after external links. #man_show_urls = False # -- Options for Texinfo output ------------------------------------------- # Grouping the document tree into Texinfo files. List of tuples # (source start file, target name, title, author, # dir menu entry, description, category) texinfo_documents = [ ('index', 'dedupe', u'dedupe Documentation', u'Forest Gregg, Derek Eder, and contributors', 'dedupe', 'One line description of project.', 'Miscellaneous'), ] # Documents to append as an appendix to all manuals. #texinfo_appendices = [] # If false, no module index is generated. #texinfo_domain_indices = True # How to display URL addresses: 'footnote', 'no', or 'inline'. #texinfo_show_urls = 'footnote' # If true, do not generate a @detailmenu in the "Top" node's menu.
true
true
f70eb0922b08398271e740b375a058194fb88a63
10,033
py
Python
auth-api/src/auth_api/models/user.py
vysakh-menon-aot/sbc-auth
b5c18df3b7586cb7d9761f7fc0809cb2cbb3b096
[ "Apache-2.0" ]
null
null
null
auth-api/src/auth_api/models/user.py
vysakh-menon-aot/sbc-auth
b5c18df3b7586cb7d9761f7fc0809cb2cbb3b096
[ "Apache-2.0" ]
null
null
null
auth-api/src/auth_api/models/user.py
vysakh-menon-aot/sbc-auth
b5c18df3b7586cb7d9761f7fc0809cb2cbb3b096
[ "Apache-2.0" ]
null
null
null
# Copyright © 2019 Province of British Columbia # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """This manages a User record in the Auth service. A User stores basic information from a KeyCloak user (including the KeyCloak GUID). """ import datetime from flask import current_app from sqlalchemy import Boolean, Column, DateTime, ForeignKey, Integer, String, and_, or_ from sqlalchemy.dialects.postgresql import UUID from sqlalchemy.orm import relationship from auth_api.utils.enums import AccessType, LoginSource, Status, UserStatus from auth_api.utils.roles import Role from auth_api.utils.user_context import UserContext, user_context from .base_model import BaseModel from .db import db from .membership import Membership as MembershipModel from .org import Org as OrgModel from .user_status_code import UserStatusCode class User(BaseModel): """This is the model for a User.""" __tablename__ = 'users' __versioned__ = { 'exclude': ['modified', 'modified_by_id', 'modified_by', 'created'] } id = Column(Integer, primary_key=True) username = Column('username', String(100), index=True) firstname = Column('first_name', String(200), index=True) lastname = Column('last_name', String(200), index=True) email = Column('email', String(200), index=True) keycloak_guid = Column( 'keycloak_guid', UUID(as_uuid=True), unique=True, nullable=True # bcros users comes with no guid ) is_terms_of_use_accepted = Column(Boolean(), default=False, nullable=True) terms_of_use_accepted_version = Column( ForeignKey('documents.version_id'), nullable=True ) # a type for the user to identify what kind of user it is..ie anonymous , bcsc etc ..similar to login source type = Column('type', String(200), nullable=True) status = Column(ForeignKey('user_status_codes.id')) idp_userid = Column('idp_userid', String(256), index=True) login_source = Column('login_source', String(200), nullable=True) login_time = Column(DateTime, default=None, nullable=True) contacts = relationship('ContactLink', primaryjoin='User.id == ContactLink.user_id', lazy='select') orgs = relationship('Membership', primaryjoin='and_(User.id == Membership.user_id, or_(Membership.status == ' + str( Status.ACTIVE.value) + ', Membership.status == ' + str( Status.PENDING_APPROVAL.value) + '))', lazy='select') # noqa:E127 terms_of_use_version = relationship('Documents', foreign_keys=[terms_of_use_accepted_version], uselist=False, lazy='select') user_status = relationship('UserStatusCode', foreign_keys=[status], lazy='subquery') @classmethod def find_by_username(cls, username): """Return the first user with the provided username.""" return cls.query.filter_by(username=username).first() @classmethod @user_context def find_by_jwt_token(cls, **kwargs): """Find an existing user by the keycloak GUID and (idpUserId is null or from token) in the provided token.""" user_from_context: UserContext = kwargs['user_context'] return db.session.query(User).filter( and_(User.keycloak_guid == user_from_context.sub, or_(User.idp_userid == user_from_context.token_info.get('idp_userid', None), User.idp_userid.is_(None)))).one_or_none() @classmethod @user_context def create_from_jwt_token(cls, first_name: str, last_name: str, **kwargs): """Create a User from the provided JWT.""" user_from_context: UserContext = kwargs['user_context'] token = user_from_context.token_info if token: user = User( username=user_from_context.user_name, firstname=first_name, lastname=last_name, email=token.get('email', None), keycloak_guid=user_from_context.sub, created=datetime.datetime.now(), login_source=user_from_context.login_source, status=UserStatusCode.get_default_type(), idp_userid=token.get('idp_userid', None), login_time=datetime.datetime.now(), type=cls._get_type(user_from_context=user_from_context) ) current_app.logger.debug( 'Creating user from JWT:{}; User:{}'.format(token, user) ) user.save() return user return None @classmethod @user_context def update_from_jwt_token(cls, user, # pylint:disable=too-many-arguments first_name: str, last_name: str, is_login: bool = False, **kwargs): """Update a User from the provided JWT.""" user_from_context: UserContext = kwargs['user_context'] token = user_from_context.token_info if not token or not user: return None # Do not save if nothing has been changed # pylint: disable=too-many-boolean-expressions if not is_login \ and (user.username == user_from_context.user_name or user.username) \ and user.firstname == first_name \ and user.lastname == last_name \ and user.email == token.get('email', user.email) \ and (str(user.keycloak_guid) == user_from_context.sub or user.keycloak_guid) \ and user.status == UserStatus.ACTIVE.value \ and (user.login_source == user_from_context.login_source or user.login_source) \ and user.idp_userid == token.get('idp_userid', None): return user current_app.logger.debug( 'Updating user from JWT:{}; User:{}'.format(token, user) ) user.username = user_from_context.user_name or user.username user.firstname = first_name user.lastname = last_name user.email = token.get('email', user.email) user.modified = datetime.datetime.now() if token.get('accessType', None) == AccessType.ANONYMOUS.value: # update kcguid for anonymous users user.keycloak_guid = user_from_context.sub or user.keycloak_guid # If this user is marked as Inactive, this login will re-activate them user.status = UserStatus.ACTIVE.value user.login_source = user_from_context.login_source or user.login_source user.type = cls._get_type(user_from_context) # If this is a request during login, update login_time if is_login: user.login_time = datetime.datetime.now() user.idp_userid = token.get('idp_userid') cls.commit() return user @classmethod def find_users(cls, first_name, last_name, email): """Return a set of users with either the given username or the given email.""" # TODO: This needs to be improved for scalability. Paging large datasets etc. if first_name == '' and last_name == '' and email == '': return cls.query.all() return cls.query.filter(or_(cls.firstname == first_name, cls.lastname == last_name, cls.email == email)).all() @classmethod @user_context def update_terms_of_use(cls, is_terms_accepted, terms_of_use_version, **kwargs): """Update the terms of service for the user.""" user_from_context: UserContext = kwargs['user_context'] if user_from_context.token_info: user = cls.find_by_jwt_token() user.is_terms_of_use_accepted = is_terms_accepted user.terms_of_use_accepted_version = terms_of_use_version current_app.logger.debug( 'Updating users Terms of use is_terms_accepted:{}; terms_of_use_version:{}'.format( is_terms_accepted, terms_of_use_version) ) cls.save(user) return user return None @classmethod def find_users_by_org_id_by_status_by_roles(cls, org_id, roles, status=Status.ACTIVE.value): """Find all members of the org with a status.""" return db.session.query(User). \ join(MembershipModel, (User.id == MembershipModel.user_id) & (MembershipModel.status == status) & (MembershipModel.membership_type_code.in_(roles))). \ join(OrgModel).filter(OrgModel.id == org_id).all() def delete(self): """Users cannot be deleted so intercept the ORM by just returning.""" return self @classmethod def _get_type(cls, user_from_context: UserContext) -> str: """Return type of the user from the token info.""" user_type: str = None if user_from_context.roles: if Role.ANONYMOUS_USER.value in user_from_context.roles \ or user_from_context.login_source == LoginSource.BCROS.value: user_type = Role.ANONYMOUS_USER.name elif Role.GOV_ACCOUNT_USER.value in user_from_context.roles: user_type = Role.GOV_ACCOUNT_USER.name elif Role.PUBLIC_USER.value in user_from_context.roles \ or user_from_context.login_source in [LoginSource.BCEID.value, LoginSource.BCSC.value]: user_type = Role.PUBLIC_USER.name elif user_from_context.is_staff(): user_type = Role.STAFF.name elif user_from_context.is_system(): user_type = Role.SYSTEM.name return user_type
43.4329
118
0.653344
import datetime from flask import current_app from sqlalchemy import Boolean, Column, DateTime, ForeignKey, Integer, String, and_, or_ from sqlalchemy.dialects.postgresql import UUID from sqlalchemy.orm import relationship from auth_api.utils.enums import AccessType, LoginSource, Status, UserStatus from auth_api.utils.roles import Role from auth_api.utils.user_context import UserContext, user_context from .base_model import BaseModel from .db import db from .membership import Membership as MembershipModel from .org import Org as OrgModel from .user_status_code import UserStatusCode class User(BaseModel): __tablename__ = 'users' __versioned__ = { 'exclude': ['modified', 'modified_by_id', 'modified_by', 'created'] } id = Column(Integer, primary_key=True) username = Column('username', String(100), index=True) firstname = Column('first_name', String(200), index=True) lastname = Column('last_name', String(200), index=True) email = Column('email', String(200), index=True) keycloak_guid = Column( 'keycloak_guid', UUID(as_uuid=True), unique=True, nullable=True ) is_terms_of_use_accepted = Column(Boolean(), default=False, nullable=True) terms_of_use_accepted_version = Column( ForeignKey('documents.version_id'), nullable=True ) type = Column('type', String(200), nullable=True) status = Column(ForeignKey('user_status_codes.id')) idp_userid = Column('idp_userid', String(256), index=True) login_source = Column('login_source', String(200), nullable=True) login_time = Column(DateTime, default=None, nullable=True) contacts = relationship('ContactLink', primaryjoin='User.id == ContactLink.user_id', lazy='select') orgs = relationship('Membership', primaryjoin='and_(User.id == Membership.user_id, or_(Membership.status == ' + str( Status.ACTIVE.value) + ', Membership.status == ' + str( Status.PENDING_APPROVAL.value) + '))', lazy='select') terms_of_use_version = relationship('Documents', foreign_keys=[terms_of_use_accepted_version], uselist=False, lazy='select') user_status = relationship('UserStatusCode', foreign_keys=[status], lazy='subquery') @classmethod def find_by_username(cls, username): return cls.query.filter_by(username=username).first() @classmethod @user_context def find_by_jwt_token(cls, **kwargs): user_from_context: UserContext = kwargs['user_context'] return db.session.query(User).filter( and_(User.keycloak_guid == user_from_context.sub, or_(User.idp_userid == user_from_context.token_info.get('idp_userid', None), User.idp_userid.is_(None)))).one_or_none() @classmethod @user_context def create_from_jwt_token(cls, first_name: str, last_name: str, **kwargs): user_from_context: UserContext = kwargs['user_context'] token = user_from_context.token_info if token: user = User( username=user_from_context.user_name, firstname=first_name, lastname=last_name, email=token.get('email', None), keycloak_guid=user_from_context.sub, created=datetime.datetime.now(), login_source=user_from_context.login_source, status=UserStatusCode.get_default_type(), idp_userid=token.get('idp_userid', None), login_time=datetime.datetime.now(), type=cls._get_type(user_from_context=user_from_context) ) current_app.logger.debug( 'Creating user from JWT:{}; User:{}'.format(token, user) ) user.save() return user return None @classmethod @user_context def update_from_jwt_token(cls, user, first_name: str, last_name: str, is_login: bool = False, **kwargs): user_from_context: UserContext = kwargs['user_context'] token = user_from_context.token_info if not token or not user: return None if not is_login \ and (user.username == user_from_context.user_name or user.username) \ and user.firstname == first_name \ and user.lastname == last_name \ and user.email == token.get('email', user.email) \ and (str(user.keycloak_guid) == user_from_context.sub or user.keycloak_guid) \ and user.status == UserStatus.ACTIVE.value \ and (user.login_source == user_from_context.login_source or user.login_source) \ and user.idp_userid == token.get('idp_userid', None): return user current_app.logger.debug( 'Updating user from JWT:{}; User:{}'.format(token, user) ) user.username = user_from_context.user_name or user.username user.firstname = first_name user.lastname = last_name user.email = token.get('email', user.email) user.modified = datetime.datetime.now() if token.get('accessType', None) == AccessType.ANONYMOUS.value: user.keycloak_guid = user_from_context.sub or user.keycloak_guid user.status = UserStatus.ACTIVE.value user.login_source = user_from_context.login_source or user.login_source user.type = cls._get_type(user_from_context) if is_login: user.login_time = datetime.datetime.now() user.idp_userid = token.get('idp_userid') cls.commit() return user @classmethod def find_users(cls, first_name, last_name, email): if first_name == '' and last_name == '' and email == '': return cls.query.all() return cls.query.filter(or_(cls.firstname == first_name, cls.lastname == last_name, cls.email == email)).all() @classmethod @user_context def update_terms_of_use(cls, is_terms_accepted, terms_of_use_version, **kwargs): user_from_context: UserContext = kwargs['user_context'] if user_from_context.token_info: user = cls.find_by_jwt_token() user.is_terms_of_use_accepted = is_terms_accepted user.terms_of_use_accepted_version = terms_of_use_version current_app.logger.debug( 'Updating users Terms of use is_terms_accepted:{}; terms_of_use_version:{}'.format( is_terms_accepted, terms_of_use_version) ) cls.save(user) return user return None @classmethod def find_users_by_org_id_by_status_by_roles(cls, org_id, roles, status=Status.ACTIVE.value): return db.session.query(User). \ join(MembershipModel, (User.id == MembershipModel.user_id) & (MembershipModel.status == status) & (MembershipModel.membership_type_code.in_(roles))). \ join(OrgModel).filter(OrgModel.id == org_id).all() def delete(self): return self @classmethod def _get_type(cls, user_from_context: UserContext) -> str: user_type: str = None if user_from_context.roles: if Role.ANONYMOUS_USER.value in user_from_context.roles \ or user_from_context.login_source == LoginSource.BCROS.value: user_type = Role.ANONYMOUS_USER.name elif Role.GOV_ACCOUNT_USER.value in user_from_context.roles: user_type = Role.GOV_ACCOUNT_USER.name elif Role.PUBLIC_USER.value in user_from_context.roles \ or user_from_context.login_source in [LoginSource.BCEID.value, LoginSource.BCSC.value]: user_type = Role.PUBLIC_USER.name elif user_from_context.is_staff(): user_type = Role.STAFF.name elif user_from_context.is_system(): user_type = Role.SYSTEM.name return user_type
true
true
f70eb14a7a606f28ca7a0378c580389cd746f583
213
py
Python
tests/test.py
JamesRunnalls/netcdf2geotiff
69ed4232998ce45bcb47035a077009ad0882f5ec
[ "MIT" ]
null
null
null
tests/test.py
JamesRunnalls/netcdf2geotiff
69ed4232998ce45bcb47035a077009ad0882f5ec
[ "MIT" ]
null
null
null
tests/test.py
JamesRunnalls/netcdf2geotiff
69ed4232998ce45bcb47035a077009ad0882f5ec
[ "MIT" ]
null
null
null
from netcdf2geotiff import rgb_geotiff, singleband_geotiff rgb_geotiff("test3.nc", "test3.tif", "RED", "GREEN", "BLUE", "lat", "lon") singleband_geotiff("test3.nc", "tests3.tif", "IDEPIX_SNOW_ICE", "lat", "lon")
42.6
77
0.713615
from netcdf2geotiff import rgb_geotiff, singleband_geotiff rgb_geotiff("test3.nc", "test3.tif", "RED", "GREEN", "BLUE", "lat", "lon") singleband_geotiff("test3.nc", "tests3.tif", "IDEPIX_SNOW_ICE", "lat", "lon")
true
true
f70eb1feadb7d4ea843f887e4935a9b29163f734
1,071
py
Python
python/src/aoc/year2015/day6.py
ocirne/adventofcode
ea9b5f1b48a04284521e85c96b420ed54adf55f0
[ "Unlicense" ]
1
2021-02-16T21:30:04.000Z
2021-02-16T21:30:04.000Z
python/src/aoc/year2015/day6.py
ocirne/adventofcode
ea9b5f1b48a04284521e85c96b420ed54adf55f0
[ "Unlicense" ]
null
null
null
python/src/aoc/year2015/day6.py
ocirne/adventofcode
ea9b5f1b48a04284521e85c96b420ed54adf55f0
[ "Unlicense" ]
null
null
null
from collections import defaultdict from aoc.util import load_input def turn(d, fun, sxy, exy): sx, sy = map(int, sxy.split(",")) ex, ey = map(int, exy.split(",")) for x in range(sx, ex + 1): for y in range(sy, ey + 1): d[(x, y)] = fun(d[(x, y)]) def run(data, toggle, turn_on, turn_off): grid = defaultdict(lambda: 0) for line in data: token = line.split() if line.startswith("toggle"): turn(grid, toggle, token[1], token[3]) elif line.startswith("turn on"): turn(grid, turn_on, token[2], token[4]) elif line.startswith("turn off"): turn(grid, turn_off, token[2], token[4]) else: raise Exception return sum(grid.values()) def part1(lines): return run(lines, lambda v: not v, lambda _: True, lambda _: False) def part2(lines): return run(lines, lambda x: x + 2, lambda x: x + 1, lambda x: max(0, x - 1)) if __name__ == "__main__": data = load_input(__file__, 2015, "6") print(part1(data)) print(part2(data))
26.121951
80
0.573296
from collections import defaultdict from aoc.util import load_input def turn(d, fun, sxy, exy): sx, sy = map(int, sxy.split(",")) ex, ey = map(int, exy.split(",")) for x in range(sx, ex + 1): for y in range(sy, ey + 1): d[(x, y)] = fun(d[(x, y)]) def run(data, toggle, turn_on, turn_off): grid = defaultdict(lambda: 0) for line in data: token = line.split() if line.startswith("toggle"): turn(grid, toggle, token[1], token[3]) elif line.startswith("turn on"): turn(grid, turn_on, token[2], token[4]) elif line.startswith("turn off"): turn(grid, turn_off, token[2], token[4]) else: raise Exception return sum(grid.values()) def part1(lines): return run(lines, lambda v: not v, lambda _: True, lambda _: False) def part2(lines): return run(lines, lambda x: x + 2, lambda x: x + 1, lambda x: max(0, x - 1)) if __name__ == "__main__": data = load_input(__file__, 2015, "6") print(part1(data)) print(part2(data))
true
true
f70eb2beb7c9126a39c5a492ce778317fcf5af57
73,671
py
Python
lexer_parser/domas_parser.py
franzbarron/proyecto-compis
29b3b9a588a92f5257bc3b1bb646734b7ace985b
[ "MIT" ]
null
null
null
lexer_parser/domas_parser.py
franzbarron/proyecto-compis
29b3b9a588a92f5257bc3b1bb646734b7ace985b
[ "MIT" ]
null
null
null
lexer_parser/domas_parser.py
franzbarron/proyecto-compis
29b3b9a588a92f5257bc3b1bb646734b7ace985b
[ "MIT" ]
null
null
null
from sly import Parser from sly.yacc import _decorator as _ from .domas_lexer import DomasLexer from .domas_quadruples import Quadruple from .domas_errors import * from . import domas_semantic_cube as sm import json # to debug only import os import copy os.system('color') class DomasParser(Parser): # Parser directives tokens = DomasLexer.tokens # debugfile = 'parser.out' start = 'programa' # Tables function_table = {} class_table = {} constant_table = {'int': [], 'float': [], 'string': [], 'bool': []} # Stacks stack_of_stacks = [[], []] # operands, operators !important stack_vars = [] last_arr_t = [] displacements = [] for_var_dir = [] break_stack = [] # Lists quadruples = [] jumps = [] # Counters quad_counter = 1 param_counter = 0 temp_counter = 0 attr_counter = 1 # Aux vars current_class = None last_arr_id = None last_type = None last_func_added = None has_returned = False found_errors = False types = ['int', 'float', 'string', 'bool', 'void'] operators = ['+', '-', '*', '/', '<', '>', '<=', '>=', '==', '<>', '&', '|'] # Add a function to the function table def add_to_func_table(self, id, return_type): self.function_table[id] = { 'return_type': return_type, 'vars': {}, 'num_types': '0\u001f' * len(self.types), 'params': '', 'num_temps': '0\u001f' * len(self.types) } # Checks if a variable exists def check_variable_exists(self, var): if self.current_class != None: return var in self.function_table[self.curr_scope]['vars'] or var in self.class_table[self.current_class]['vars'] return var in self.function_table[self.curr_scope]['vars'] or var in self.function_table[self.program_name]['vars'] # Returns the type of a variable if it exists def get_var_type(self, var, tok): if not self.check_variable_exists(var): self.found_errors = True print('ERROR: No variable\033[1m', var, '\033[0mwas found.') print(' Missing reference found on line', tok.lineno) return None if self.current_class != None: if var in self.function_table[self.curr_scope]['vars']: return self.function_table[self.curr_scope]['vars'][var]['type'] return self.class_table[self.current_class]['vars'][var]['type'] if var in self.function_table[self.curr_scope]['vars']: return self.function_table[self.curr_scope]['vars'][var]['type'] return self.function_table[self.program_name]['vars'][var]['type'] # Updates the amount of temporals used in fucntions. def update_num_temps(self, func_num_temps, type_idx, quantity=1): lst = func_num_temps.split('\u001f') lst[type_idx] = str(int(lst[type_idx]) + quantity) return '\u001f'.join(lst) # Cheks if a var is an array by finding its first dimension def check_var_is_array(self, var): if not var: return if var['dir'] >= 4500 and var['dir'] < 6000: return False if not self.check_variable_exists(var['value']): return False if self.current_class != None: if var['value'] in self.function_table[self.curr_scope]['vars']: return 'd1' in self.function_table[self.curr_scope]['vars'][var['value']] else: return 'd1' in self.class_table[self.current_class]['vars'][var['value']] elif var['value'] in self.function_table[self.curr_scope]['vars']: return 'd1' in self.function_table[self.curr_scope]['vars'][var['value']] else: return 'd1' in self.function_table[self.program_name]['vars'][var['value']] # Makes quadruples for arithemtic operators and pushes them into the quadruple stack def make_and_push_quad(self): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() if not ro or not lo: raise SystemError("Reached unsolvable state") r_type = sm.checkOperation(lo['type'], ro['type'], op) self.last_type = r_type idx = self.types.index(r_type) num_temps = self.function_table[self.curr_scope]['num_temps'] self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.stack_of_stacks[-2].append( {'value': 't' + str(self.temp_counter), 'type': r_type, 'dir': t_dir}) if self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 @_('PROGRAM ID pro1 SEMI pro0 declarations') def programa(self, p): if self.found_errors: raise CompilationError() # func_dir_out = open('debug/funcdir.out', 'w') # class_dir_out = open('debug/classdir.out', 'w') # func_dir_out.write(json.dumps( # self.function_table, indent=2)) # class_dir_out.write(json.dumps( # self.class_table, indent=2)) return (self.program_name, self.function_table, self.class_table, self.constant_table, self.quadruples) # Creates goto main quadruple and appends it to the quadruple list @_('') def pro0(self, p): self.quadruples.append(Quadruple(-1, -1, 'goto', -1)) self.quad_counter += 1 # creates function table @_('') def pro1(self, p): self.program_name = p[-1] self.curr_scope = p[-1] self.function_table[p[-1]] = { 'return_type': None, 'vars': {}, 'num_types': '0\u001f0\u001f0\u001f0\u001f0\u001f' } @ _('class_declaration out_class var_declaration function_definition main') def declarations(self, p): return 'declarations' @ _('''CLASS ID cd1 inherits LCURL ATTRIBUTES attribute_declaration METHODS method_definition RCURL class_declaration''', 'empty') def class_declaration(self, p): return 'class_declaration' # Adds class to class table @ _('') def cd1(self, p): if p[-1] in self.class_table: self.found_errors = True print( 'ERROR: A class with the name\033[1m', p[-1], '\033[0mhas already been defined') print(' Redefinition found on line', self.symstack[-1].lineno) else: self.class_table[p[-1]] = { 'vars': {}, 'num_types': '0\u001f0\u001f0\u001f0\u001f0\u001f' } self.current_class = p[-1] @ _('INHERITS ID cd3', 'empty') def inherits(self, p): return 'inherits' # Copies the information from parent class to its child. @ _('') def cd3(self, p): if not p[-1] in self.class_table: self.found_errors = True print('ERROR: Id\033[1m', p[-1], '\033[0mis not defined as a class') print(' Missing reference found on line', self.symstack[-1].lineno) else: self.class_table[self.current_class] = copy.deepcopy( self.class_table[p[-1]]) @ _('VAR ID ad1 attr_vector', 'VAR ID ad1 attr_simple_var', 'empty') def attribute_declaration(self, p): return 'attribute_declaration' # Appends declared variable to the stack of variables if it doesn't exist @ _('') def ad1(self, p): if p[-1] in self.class_table[self.current_class]['vars']: self.found_errors = True print('ERROR: An attribute\033[1m', p[-1], '\033[0mhas already been defined') print(' Redefinition found on line', self.symstack[-1].lineno) else: self.stack_vars.append(p[-1]) @ _('''LBRACKET CTE_I ad2 attr_multidim RBRACKET COLON simple_type ad4 SEMI attribute_declaration''') def attr_vector(self, p): return 'vector' # Adds the first dimension of an array to the information of the variable @ _('') def ad2(self, p): self.latest_var = self.stack_vars.pop() if self.class_table[self.current_class]['vars']: self.class_table[self.current_class]['vars'][self.latest_var] = { 'd1': p[-1]} else: self.class_table[self.current_class]['vars'] = { self.latest_var: {'d1': p[-1]}} @ _('COMMA CTE_I ad3', 'empty') def attr_multidim(self, p): return 'attr_multidim' # Adds the second dimension of an array to the information of the variable. @ _('') def ad3(self, p): self.class_table[self.current_class]['vars'][self.latest_var]['d2'] = p[-1] # Adds the type and direction to the information of the variable declared and updated the amount of types used in the class. @ _('') def ad4(self, p): idx = self.types.index(p[-1]) num_types = self.class_table[self.current_class]['num_types'] if 'd1' in self.class_table[self.current_class]['vars'][self.latest_var]: q = self.class_table[self.current_class]['vars'][self.latest_var]['d1'] if 'd2' in self.class_table[self.current_class]['vars'][self.latest_var]: q *= self.class_table[self.current_class]['vars'][self.latest_var]['d2'] self.class_table[self.current_class]['vars'][self.latest_var]['dir'] = 6000 + idx * \ 300 + int(num_types.split('\u001f')[idx]) self.class_table[self.current_class]['vars'][self.latest_var]['type'] = p[-1] self.class_table[self.current_class]['num_types'] = self.update_num_temps( num_types, idx, q) @ _('attr_var_list COLON simple_type ad5 SEMI attribute_declaration') def attr_simple_var(self, p): return 'attr_simple_var' @ _('COMMA ID ad1 attr_var_list', 'empty') def attr_var_list(self, p): return 'attr_var_list' # Pops the stack of variables and giving each their corresponding type. @ _('') def ad5(self, p): while len(self.stack_vars) > 0: curr_var = self.stack_vars.pop() if curr_var in self.class_table[self.current_class]['vars']: self.found_errors = True print('ERROR: An attribute\033[1m', curr_var, '\033[0mhas already been defined in class', self.current_class) print(' Redefinition found on line', self.symstack[-2].lineno) idx = self.types.index(p[-1]) num_types = self.class_table[self.current_class]['num_types'] self.class_table[self.current_class]['num_types'] = self.update_num_temps( num_types, idx) self.class_table[self.current_class]['vars'][curr_var] = { 'type': p[-1], 'dir': 6000 + idx * 300 + int(num_types.split('\u001f')[idx])} @ _('''def_type fd1 FUNCTION ID md3 LPAREN m_parameters RPAREN LCURL fd4 var_declaration statements RCURL fd5 fd6 method_definition''', 'empty') def method_definition(self, p): return 'method_definition' # Adds the id with the name of the class prefixed to the function table. @ _('') def md3(self, p): if p[-1] in self.class_table[self.current_class]['vars']: self.found_errors = True print('ERROR: An attribute or method\033[1m', p[-1], '\033[0mhas already been defined in class', self.current_class) print(' Redefinition found on line', self.symstack[-1].lineno) else: self.add_to_func_table( self.current_class + '.' + p[-1], self.curr_func_type) self.last_func_added = self.current_class + '.' + p[-1] self.curr_scope = self.last_func_added idx = self.types.index(self.curr_func_type) num_types = self.function_table[self.program_name]['num_types'] self.function_table[self.program_name]['num_types'] = self.update_num_temps( num_types, idx) self.function_table[self.program_name]['vars'][self.current_class + '.' + p[-1]] = { 'type': self.curr_func_type, 'dir': 0, 'real_dir': idx * 300 + int(num_types.split('\u001f')[idx])} @_('ID p1 COLON simple_type m2 m_param_choose', 'empty') def m_parameters(self, p): return 'parameters' # Adds the id and type of the parameter to the table of variables of the current method. @ _('') def m2(self, p): if self.latest_var in self.function_table[self.curr_scope]['vars']: self.found_errors = True print('ERROR: A parameter\033[1m', self.latest_var, '\033[0mhas already been declared for method', self.last_func_added.split('.')[1], 'in class', self.current_class) print(' Redefinition found on line', self.symstack[-2].lineno) else: idx = self.types.index(p[-1]) self.function_table[self.curr_scope]['params'] += str( idx) num_types = self.function_table[self.curr_scope]['num_types'] self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx) self.function_table[self.curr_scope]['vars'][self.latest_var] = { 'type': p[-1], 'dir': 1500 + idx * 300 + int(num_types.split('\u001f')[idx]) } @ _('COMMA m_parameters', 'empty') def m_param_choose(self, p): return 'm_param_choose' @ _('') def out_class(self, p): self.current_class = None self.curr_scope = self.program_name @ _('VAR ID gvd1 vector', 'VAR ID gvd1 simple_var', 'empty') def var_declaration(self, p): return p[0] # Appends the current var to the stack of variables of it doesn't exist. @ _('') def gvd1(self, p): if p[-1] in self.function_table: self.found_errors = True print('ERROR: A function with ID\033[1m', p[-1], '\033[0mhas already been declared. Variables may not share name with functions') print(' Redefinition found on line', self.symstack[-1].lineno) # elif self.current_class != None and p[-1] in self.class_table[self.current_class]: # raise RedefinitionError(p[-1]) else: self.stack_vars.append(p[-1]) @ _('LBRACKET CTE_I gvd2 multidim RBRACKET COLON simple_type gvd4 SEMI var_declaration') def vector(self, p): return 'vector' # Adds the variable and its first dimension to the variable table of the current scope. @ _('') def gvd2(self, p): self.latest_var = self.stack_vars.pop() if self.function_table[self.curr_scope]['vars']: self.function_table[self.curr_scope]['vars'][self.latest_var] = { 'd1': p[-1] } else: self.function_table[self.curr_scope]['vars'] = { self.latest_var: {'d1': p[-1]} } @ _('COMMA CTE_I gvd3', 'empty') def multidim(self, p): return 'multidim' @ _('var_list COLON composite_type gvd6 SEMI var_declaration', 'var_list COLON simple_type gvd5 SEMI var_declaration') def simple_var(self, p): return 'simple_var' @ _('COMMA ID gvd1 var_list', 'empty') def var_list(self, p): return 'var_list' # Adds the second dimension to the latest variable in the current scope @ _('') def gvd3(self, p): self.function_table[self.curr_scope]['vars'][self.latest_var]['d2'] = p[-1] # Adds the type and address to the information of the variable declared and updated the amount of types used in the class. @ _('') def gvd4(self, p): idx = self.types.index(p[-1]) num_types = self.function_table[self.curr_scope]['num_types'] offset = 1500 if self.curr_scope != self.program_name else 0 if 'd1' in self.function_table[self.curr_scope]['vars'][self.latest_var]: q = self.function_table[self.curr_scope]['vars'][self.latest_var]['d1'] if 'd2' in self.function_table[self.curr_scope]['vars'][self.latest_var]: q *= self.function_table[self.curr_scope]['vars'][self.latest_var]['d2'] self.function_table[self.curr_scope]['vars'][self.latest_var]['dir'] = idx * \ 300 + int(num_types.split('\u001f')[idx]) + offset self.function_table[self.curr_scope]['vars'][self.latest_var]['type'] = p[-1] self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx, q) # Pops the var stack, adding their ids and types to the variable directory of the current scope @ _('') def gvd5(self, p): while len(self.stack_vars) > 0: curr_var = self.stack_vars.pop() if curr_var in self.function_table[self.curr_scope]['vars']: self.found_errors = True print('ERROR: A variable\033[1m', curr_var, '\033[0mhas already been declared.') print(' Redefinition found on line', self.symstack[-5].lineno) idx = self.types.index(p[-1]) num_types = self.function_table[self.curr_scope]['num_types'] offset = 1500 if self.curr_scope != self.program_name else 0 self.function_table[self.curr_scope]['vars'][curr_var] = { 'type': p[-1], 'dir': idx * 300 + int(num_types.split('\u001f')[idx]) + offset} self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx) # Same as gvd5 but it also adds the variables of the class in question to the table of the current scope @ _('') def gvd6(self, p): while len(self.stack_vars) > 0: var_id = self.stack_vars.pop() if var_id in self.function_table[self.curr_scope]['vars']: self.found_errors = True print('ERROR: A variable\033[1m', var_id, '\033[0mhas already been declared.') print(' Redefinition found on line', self.symstack[-5].lineno) offset = 1500 if self.curr_scope != self.program_name else 0 num_types = self.function_table[self.curr_scope]['num_types'] base_addrs = [int(n) for n in num_types.split('\u001f')[:-1]] if not p[-1] in self.class_table: for i in range(1, len(self.symstack)): if hasattr(self.symstack[i * -1], 'lineno'): lineno = self.symstack[i * -1].lineno break print('ERROR: No class\033[1m', p[-1], '\033[0mwas found.') print(' Missing reference found on line', lineno) return for attr in self.class_table[p[-1]]['vars']: attr_type = self.class_table[p[-1]]['vars'][attr]['type'] idx = self.types.index(attr_type) num_types = self.function_table[self.curr_scope]['num_types'] q = 1 self.function_table[self.curr_scope]['vars'][var_id + '.' + attr] = { 'type': attr_type, 'dir': base_addrs[idx] + self.class_table[p[-1]]['vars'][attr]['dir'] - 6000 + offset } if 'd1' in self.class_table[p[-1]]['vars'][attr]: q = self.class_table[p[-1]]['vars'][attr]['d1'] self.function_table[self.curr_scope]['vars'][var_id + '.' + attr]['d1'] = q if 'd2' in self.class_table[p[-1]]['vars'][attr]: d2 = self.class_table[p[-1]]['vars'][attr]['d2'] q *= d2 self.function_table[self.curr_scope]['vars'][var_id + '.' + attr]['d2'] = d2 self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx, q) self.function_table[self.curr_scope]['vars'][var_id] = { 'type': p[-1]} @ _('def_type fd1 FUNCTION ID fd3 LPAREN parameters RPAREN LCURL fd4 var_declaration statements RCURL fd5 fd6 function_definition', 'empty') def function_definition(self, p): return 'function_definition' # saves thee current function type in a variable @ _('') def fd1(self, p): self.curr_func_type = p[-1] # Adds the id of the function to the function table @ _('') def fd3(self, p): if p[-1] in self.function_table: self.found_errors = True print('ERROR: A function\033[1m', p[-1], '\033[0mhas already been defined.') print(' Redefinition found on line', self.symstack[-1].lineno) elif p[-1] in self.function_table[self.program_name]['vars']: self.found_errors = True print('ERROR: A global variable\033[1m', p[-1], '\033[0mhas been declared. Functions may not share names with global variables') print(' Redefinition found on line', self.symstack[-1].lineno) else: self.add_to_func_table(p[-1], self.curr_func_type) self.last_func_added = p[-1] self.curr_scope = self.last_func_added idx = self.types.index(self.curr_func_type) num_types = self.function_table[self.program_name]['num_types'] self.function_table[self.program_name]['num_types'] = self.update_num_temps( num_types, idx) self.function_table[self.program_name]['vars'][p[-1]] = { 'type': self.curr_func_type, 'dir': 0, 'real_dir': idx * 300 + int(num_types.split('\u001f')[idx])} # Adds the start of the quadruples related to the current function to the ts information in the function table @ _('') def fd4(self, p): if not self.last_func_added: return self.function_table[self.last_func_added]['start'] = self.quad_counter # Deletes the variable table of the current scope @ _('') def fd5(self, p): if not self.last_func_added: return del self.function_table[self.last_func_added]['vars'] # Creates and appends the end_func quadruple to the quadruple stack @ _('') def fd6(self, p): if self.curr_func_type != 'void' and self.has_returned == False: self.found_errors = True print('ERROR: Function\033[1m', self.curr_scope, '\033[0mis missing a return statement') print(' Non-void functions must have a return statement.') self.quadruples.append(Quadruple(-1, -1, 'end_func', -1)) self.quad_counter += 1 self.temp_counter = 1 self.has_returned = False @ _('statement statements', 'empty') def statements(self, p): return 'statements' @ _('simple_type', 'VOID') def def_type(self, p): return p[0] @ _('INT', 'FLOAT', 'STRING', 'BOOL') def simple_type(self, p): return p[0] @ _('ID') def composite_type(self, p): return p[0] @ _('ID p1 COLON simple_type p2 param_choose', 'empty') def parameters(self, p): return 'parameters' @ _('COMMA parameters', 'empty') def param_choose(self, p): return 'param_choose' # Saves the ID of the parameter ina variable @ _('') def p1(self, p): self.latest_var = p[-1] # Adds the type of the parameter to its information in the variable table of the current scope @ _('') def p2(self, p): if self.latest_var in self.function_table[self.curr_scope]['vars']: self.found_errors = True print('ERROR: A parameter\033[1m', self.latest_var, '\033[0mhas already been declared for function', self.last_func_added) print(' Redefinition found on line', self.symstack[-2].lineno) idx = self.types.index(p[-1]) self.function_table[self.curr_scope]['params'] += str(idx) num_types = self.function_table[self.curr_scope]['num_types'] offset = 1500 if self.curr_scope != self.program_name else 0 self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx) self.function_table[self.curr_scope]['vars'][self.latest_var] = { 'type': p[-1], 'dir': idx * 300 + int(num_types.split('\u001f')[idx]) + offset} @_('assignment', 'call_to_void_function', 'function_returns', 'read', 'print', 'decision_statement', 'repetition_statement', 'BREAK br0 SEMI') def statement(self, p): return 'statement' # Adds the quadruple counter to the break stack @_('') def br0(self, p): if len(self.jumps) == 0: self.found_errors = True print('ERROR: break statement on line', self.symstack[-1].lineno, 'used outside a loop') self.quadruples.append(Quadruple(-1, -1, 'goto', None)) self.break_stack.append(self.quad_counter) self.quad_counter += 1 @_('variable ass1 EQUALS expression ass2 SEMI') def assignment(self, p): return 'assignment' # Save the id in a variable @_('') def ass1(self, p): self.latest_var = p[-1] # Generate the quadruple containing an '=' as its operator. Get the addresses of the left and res @_('') def ass2(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): self.make_and_push_quad() made_quad = True lo = self.stack_of_stacks[-2].pop() if not lo: return v_type = self.get_var_type(self.latest_var, self.symstack[-2]) if not v_type: return self.last_type = sm.checkOperation(v_type, lo['type'], '=') if not made_quad and self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if self.current_class != None: if self.latest_var in self.class_table[self.current_class]['vars']: if 'd1' in self.class_table[self.current_class]['vars'][self.latest_var]: if not self.last_arr_t: return var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = self.class_table[self.current_class]['vars'][self.latest_var]['dir'] else: if 'd1' in self.function_table[self.curr_scope]['vars'][self.latest_var]: if not self.last_arr_t: return var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = self.function_table[self.curr_scope]['vars'][self.latest_var]['dir'] elif self.latest_var in self.function_table[self.curr_scope]['vars']: if 'd1' in self.function_table[self.curr_scope]['vars'][self.latest_var]: if not self.last_arr_t: return var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = self.function_table[self.curr_scope]['vars'][self.latest_var]['dir'] else: if 'd1' in self.function_table[self.program_name]['vars'][self.latest_var]: if not self.last_arr_t: return var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = self.function_table[self.program_name]['vars'][self.latest_var]['dir'] q = Quadruple(lo_dir, -1, '=', var_dir) self.quadruples.append(q) self.quad_counter += 1 @_('id_or_attribute', 'id_or_attribute v0 LBRACKET expression v1 RBRACKET', 'id_or_attribute v0 LBRACKET expression v2 COMMA v4 expression v3 RBRACKET') def variable(self, p): return p[0] # Checks that the variable exists and is an array @_('') def v0(self, p): self.check_variable_exists(p[-1]) if self.current_class != None: if not 'd1' in self.class_table[self.current_class]['vars'][p[-1]]: self.found_errors = True print('ERROR: Variable\033[1m', p[-1], '\033[0mis not an array or matrix.') self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) return elif p[-1] in self.function_table[self.curr_scope]['vars']: if not 'd1' in self.function_table[self.curr_scope]['vars'][p[-1]]: self.found_errors = True print('ERROR: Variable\033[1m', p[-1], '\033[0mis not an array or matrix.') self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) return elif not 'd1' in self.function_table[self.program_name]['vars'][p[-1]]: self.found_errors = True print('ERROR: Variable\033[1m', p[-1], '\033[0mis not an array or matrix.') self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) return self.last_arr_id = p[-1] self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) # Checks that the variable is a matrix @_('') def v4(self, p): self.check_variable_exists(self.last_arr_id) if self.current_class != None: if not 'd2' in self.class_table[self.current_class]['vars'][self.last_arr_id]: self.found_errors = True print('ERROR: Variable\033[1m', self.last_arr_id, '\033[0mis not a matrix.') elif self.last_arr_id in self.function_table[self.curr_scope]['vars']: if not 'd2' in self.function_table[self.curr_scope]['vars'][self.last_arr_id]: self.found_errors = True print('ERROR: Variable\033[1m', self.last_arr_id, '\033[0mis not a matrix.') else: if not 'd2' in self.function_table[self.program_name]['vars'][self.last_arr_id]: self.found_errors = True print('ERROR: Variable\033[1m', self.last_arr_id, '\033[0mis not a matrix.') self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) # Calculates the address of the array index @_('') def v1(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) self.quadruples.append( Quadruple(lo['dir'], ro['dir'], op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if self.current_class != None and self.last_arr_id in self.class_table[self.current_class]['vars']: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.class_table[self.current_class]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.class_table[self.current_class]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '+', t_dir)) self.quad_counter += 1 self.last_arr_t.append(t_dir) elif self.last_arr_id in self.function_table[self.curr_scope]['vars']: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '+', t_dir)) self.quad_counter += 1 self.last_arr_t.append(t_dir) else: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.program_name]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.function_table[self.program_name]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '+', t_dir)) self.quad_counter += 1 self.last_arr_t.append(t_dir) self.stack_of_stacks.pop() self.stack_of_stacks.pop() # Calculate the address of the matrix index @_('') def v2(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) self.quadruples.append( Quadruple(lo['dir'], ro['dir'], op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if self.current_class != None: pass elif self.last_arr_id in self.function_table[self.curr_scope]['vars']: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 d2 = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['d2'] if not d2 in self.constant_table['int']: self.constant_table['int'].append(d2) cons_dir = self.constant_table['int'].index(d2) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '*', t_dir)) self.quad_counter += 1 self.displacements.append(t_dir) else: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.program_name]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 d2 = self.function_table[self.program_name]['vars'][self.last_arr_id]['d2'] if not d2 in self.constant_table['int']: self.constant_table['int'].append(d2) cons_dir = self.constant_table['int'].index(d2) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '*', t_dir)) self.quad_counter += 1 self.displacements.append(t_dir) self.stack_of_stacks.pop() self.stack_of_stacks.pop() # Calculate s1*d2 @_('') def v3(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) self.quadruples.append( Quadruple(lo['dir'], ro['dir'], op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if self.current_class != None: pass elif self.last_arr_id in self.function_table[self.curr_scope]['vars']: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['d2'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0, 2) self.quadruples.append( Quadruple(self.displacements.pop(), t_addr, '+', t_dir)) self.quadruples.append(Quadruple(cons_dir, t_dir, '+', t_dir + 1)) self.quad_counter += 2 self.last_arr_t.append(t_dir + 1) else: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.program_name]['vars'][self.last_arr_id]['d2'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.function_table[self.program_name]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0, 2) self.quadruples.append( Quadruple(self.displacements.pop(), t_addr, '+', t_dir)) self.quadruples.append(Quadruple(cons_dir, t_dir, '+', t_dir + 1)) self.quad_counter += 2 self.last_arr_t.append(t_dir + 1) self.stack_of_stacks.pop() self.stack_of_stacks.pop() @_('ID', 'ID DOT ID') def id_or_attribute(self, p): if len(p) > 1: return p[0] + p[1] + p[2] return p[0] # Returns value, type and address of the constant @_('variable', 'CTE_I', 'CTE_F', 'CTE_STRING', 'cte_bool', 'call_to_function') def var_cte(self, p): offset = 4500 if hasattr(p, 'CTE_I'): cte_type = 'int' if not p[0] in self.constant_table['int']: self.constant_table['int'].append(p[0]) cons_dir = self.constant_table['int'].index(p[0]) + offset elif hasattr(p, 'CTE_F'): cte_type = 'float' if not p[0] in self.constant_table['float']: self.constant_table['float'].append(p[0]) cons_dir = self.constant_table['float'].index(p[0]) + offset + 300 elif hasattr(p, 'CTE_STRING'): cte_type = 'string' if not p[0] in self.constant_table['string']: self.constant_table['string'].append(p[0]) cons_dir = self.constant_table['string'].index(p[0]) + offset + 600 elif hasattr(p, 'cte_bool'): cte_type = 'bool' if not p[0] in self.constant_table['bool']: self.constant_table['bool'].append(p[0]) cons_dir = self.constant_table['bool'].index(p[0]) + offset + 900 elif hasattr(p, 'call_to_function'): return p[0] else: if not self.check_variable_exists(p[0]): for i in range(1, len(self.symstack)): if hasattr(self.symstack[i * -1], 'lineno'): lineno = self.symstack[i * -1].lineno break self.found_errors = True print('ERROR: No variable\033[1m', p[0], '\033[0mwas found.') print(' Missing reference found on line', lineno) return if self.current_class != None and p[0] in self.class_table[self.current_class]['vars']: cte_type = self.class_table[self.current_class]['vars'][p[0]]['type'] cons_dir = self.class_table[self.current_class]['vars'][p[0]]['dir'] else: cte_type = self.get_var_type(p[0], self.symstack[-2]) if p[0] in self.function_table[self.curr_scope]['vars']: cons_dir = self.function_table[self.curr_scope]['vars'][p[0]]['dir'] else: cons_dir = self.function_table[self.program_name]['vars'][p[0]]['dir'] return {'value': p[0], 'type': cte_type, 'dir': cons_dir} @_('constant e2 operator e3 expression', 'constant e2', 'LPAREN e1 expression RPAREN e4', 'LPAREN e1 expression RPAREN e4 operator e3 expression') def expression(self, p): if hasattr(p, 'LPAREN'): return p[2] return p[0] # Append the open parenthesis to the stack of operators in the stacks_of_stacks @_('') def e1(self, p): self.stack_of_stacks[-1].append('(') # Appende the operand to the stack of operands in the stacks_of _stacks @_('') def e2(self, p): self.stack_of_stacks[-2].append(p[-1]) # Makes the quadruple of operation @_('') def e3(self, p): if len(self.stack_of_stacks[-1]) == 0 or self.stack_of_stacks[-1][-1] == '(': self.stack_of_stacks[-1].append(p[-1]) elif self.stack_of_stacks[-1][-1] == '*' or self.stack_of_stacks[-1][-1] == '/': self.make_and_push_quad() if (self.stack_of_stacks[-1] and (self.stack_of_stacks[-1][-1] == '+' or self.stack_of_stacks[-1][-1] == '-')) and (p[-1] == '+' or p[-1] == '-'): self.make_and_push_quad() self.stack_of_stacks[-1].append(p[-1]) elif p[-1] == '*' or p[-1] == '/': self.stack_of_stacks[-1].append(p[-1]) elif self.stack_of_stacks[-1][-1] == '+' or self.stack_of_stacks[-1][-1] == '-': self.make_and_push_quad() self.stack_of_stacks[-1].append(p[-1]) elif p[-1] == '+' or p[-1] == '-': self.stack_of_stacks[-1].append(p[-1]) elif self.stack_of_stacks[-1][-1] in sm.comparison_ops or self.stack_of_stacks[-1][-1] in sm.equality_ops: self.make_and_push_quad() self.stack_of_stacks[-1].append(p[-1]) elif p[-1] in sm.comparison_ops or p[-1] in sm.equality_ops: self.stack_of_stacks[-1].append(p[-1]) elif self.stack_of_stacks[-1][-1] in sm.logic_ops: self.make_and_push_quad() self.stack_of_stacks[-1].append(p[-1]) elif p[-1] in sm.logic_ops: self.stack_of_stacks[-1].append(p[-1]) # Pops the operator stack and makes quads until an open parenthesis is found @_('') def e4(self, p): while(self.stack_of_stacks[-1][-1] != '('): self.make_and_push_quad() self.stack_of_stacks[-1].pop() @_('AND', 'OR') def logical_operator(self, p): return p[-1] @_('LT', 'GT', 'SAME', 'GEQ', 'LEQ', 'NEQ') def relational_operator(self, p): return p[0] @_('PLUS', 'MINUS', 'MULTIPLY', 'DIVIDE') def arithmetic_operator(self, p): return p[0] @_('logical_operator', 'relational_operator', 'arithmetic_operator') def operator(self, p): return p[0] @_('PLUS var_cte', 'MINUS var_cte', 'var_cte') def constant(self, p): if len(p) > 1 and p[1] == '-': return -p.var_cte else: return p.var_cte @_('READ LPAREN read_h') def read(self, p): return 'read' @_('variable r1 COMMA read_h', 'variable r1 RPAREN SEMI') def read_h(self, p): return 'read_h' # Makes the read quadruple wiht res being the address prefixed with a dollar sign ($) @_('') def r1(self, p): if self.current_class != None and p[-1] in self.class_table[self.current_class]['vars']: if 'd1' in self.class_table[self.current_class]['vars'][p[-1]]: var_addr = '$' + str(self.last_arr_t.pop()) else: var_addr = self.class_table[self.current_class]['vars'][p[-1]]['dir'] elif p[-1] in self.function_table[self.curr_scope]['vars']: if 'd1' in self.function_table[self.curr_scope]['vars'][p[-1]]: var_addr = '$' + str(self.last_arr_t.pop()) else: var_addr = self.function_table[self.curr_scope]['vars'][p[-1]]['dir'] elif p[-1] in self.function_table[self.program_name]['vars']: if 'd1' in self.function_table[self.program_name]['vars'][p[-1]]: var_addr = '$' + str(self.last_arr_t.pop()) else: var_addr = self.function_table[self.program_name]['vars'][p[-1]]['dir'] else: raise UndeclaredIdError(p[-1]) self.quadruples.append(Quadruple(-1, -1, 'read', var_addr)) self.quad_counter += 1 @_('function_or_method vf0 ctf2 LPAREN func_params RPAREN fp2 fp3 ctf0 ctf3') def call_to_function(self, p): if not self.check_variable_exists(self.called_func): # self.found_errors = True # print('ERROR: No function\033[1m', # self.called_func, '\033[0mwas found.') # print(' Missing reference found on line', # self.symstack[-5].lineno) return func_dir = self.function_table[self.program_name]['vars'][self.called_func]['dir'] func_type = self.function_table[self.called_func]['return_type'] return {'value': 't' + str(self.temp_counter - 1), 'type': func_type, 'dir': func_dir} # Append two empty stacks to the stack_of_stacks @_('') def ctf2(self, p): self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) # Pops the staks from the stacks_of_stacks @_('') def ctf3(self, p): self.stack_of_stacks.pop() self.stack_of_stacks.pop() # Call the return value in the address of the corresponding variable @_('') def ctf0(self, p): if not self.check_variable_exists(self.called_func): self.found_errors = True print('ERROR: No function\033[1m', self.called_func, '\033[0mwas found.') print(' Missing reference found on line', self.symstack[-3].lineno) return func_dir = self.function_table[self.program_name]['vars'][self.called_func]['real_dir'] func_type = self.function_table[self.program_name]['vars'][self.called_func]['type'] idx = self.types.index(func_type) num_temps = self.function_table[self.curr_scope]['num_temps'] # self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( # num_temps, idx) t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) self.quadruples.append( Quadruple(func_dir, -1, '=', t_dir)) self.function_table[self.program_name]['vars'][self.called_func]['dir'] = t_dir self.quad_counter += 1 self.temp_counter += 1 # @_('ID ctf1', 'ID DOT ID') def function_or_method(self, p): if(len(p) == 2): return (p[0], None) else: var_type = self.get_var_type(p[0], self.symstack[-1]) quads = [] for attr in self.class_table[var_type]['vars']: var_dir = self.function_table[self.curr_scope]['vars'][p[0]+'.'+attr]['dir'] attr_dir = self.class_table[var_type]['vars'][attr]['dir'] quads.append(Quadruple(var_dir, -2, '=', attr_dir)) # self.quad_counter += 1 return (var_type + p[1] + p[2], quads) # Check if name of the function being called exists in the function table @_('') def ctf1(self, p): if self.current_class != None: if not self.current_class + '.' + p[-1] in self.function_table: self.found_errors = True print('ERROR: No function\033[1m', p[-1], '\033[0mwas found.') print(' Missing reference found on line', self.symstack[-1].lineno) elif not p[-1] in self.function_table: self.found_errors = True print('ERROR: No function\033[1m', p[-1], '\033[0mwas found.') print(' Missing reference found on line', self.symstack[-1].lineno) @_('COMMA expression fp1 param_list', 'empty') def param_list(self, p): return 'param_list' @_('PRINT LPAREN res_write RPAREN SEMI') def print(self, p): return 'print' @_('expression pr1 comma_thing') def res_write(self, p): return 'res_write' @_('COMMA res_write', 'empty') def comma_thing(self, p): return 'comma_thing' # Make quadruples if the stack of operators is not empty and Make print quadruple @_('') def pr1(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) if not lo['dir'] in range(4500, 6000) and self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if not ro['dir'] in range(4500, 6000) and self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if made_quad: last_quad = self.quadruples[-1] self.quadruples.append( Quadruple(-1, -1, 'print', last_quad.res)) self.quad_counter += 1 else: var = self.stack_of_stacks[-2].pop() if not var: return if self.check_var_is_array(var): var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = var['dir'] self.quadruples.append( Quadruple(-1, -1, 'print', var_dir)) self.quad_counter += 1 @_('TRUE', 'FALSE') def cte_bool(self, p): return p[0] @_('IF LPAREN expression dec1 RPAREN THEN LCURL statements RCURL else_stm') def decision_statement(self, p): return 'decision_statement' # Make quadruples if the stack of operators is not empty and goto_f quadruple @_('') def dec1(self, p): while len(self.stack_of_stacks[-1]): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + 3000 r_type = sm.checkOperation(lo['type'], ro['type'], op) self.stack_of_stacks[-2].append( {'value': 't' + str(self.temp_counter), 'type': r_type, 'dir': t_dir}) self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) if self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 lo = self.stack_of_stacks[-2].pop() if lo['type'] != 'bool': raise SyntaxError( 'Expression to evaluate in if statement is not boolean') else: self.quadruples.append(Quadruple(-1, lo['dir'], 'goto_f', -1)) self.jumps.append(self.quad_counter) self.quad_counter += 1 @_('dec2 ELSE LCURL statements RCURL dec3', 'empty dec4') def else_stm(self, p): return 'else_stm' # Makes goto quadruples @_('') def dec2(self, p): falso = self.jumps.pop() self.quadruples.append(Quadruple(-1, -1, 'goto', -1)) self.jumps.append(self.quad_counter) self.quad_counter += 1 self.quadruples[falso - 1].res = self.quad_counter # Actualizes goto quadruple @_('') def dec3(self, p): jump = self.jumps.pop() self.quadruples[jump - 1].res = self.quad_counter # Actualizes goto_f @_('') def dec4(self, p): jump = self.jumps.pop() self.quadruples[jump - 1].res = self.quad_counter @_('conditional', 'non_conditional') def repetition_statement(self, p): return 'repetition_statement' @_('WHILE LPAREN con0 expression con1 RPAREN DO LCURL statements RCURL con2') def conditional(self, p): return 'conditional' # Add quadruple counter to jumps stack @_('') def con0(self, p): self.jumps.append(self.quad_counter) # Make quadruples if the stack of operators is not empty and make goto_f quadruple @_('') def con1(self, p): while len(self.stack_of_stacks[-1]): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) r_type = sm.checkOperation(lo['type'], ro['type'], op) self.stack_of_stacks[-2].append( {'value': 't' + str(self.temp_counter), 'type': r_type, 'dir': t_dir}) if self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 if self.last_type != 'bool': raise SyntaxError( 'Expression to evaluate in if statement is not boolean') else: last_quad = self.quadruples[-1].res self.quadruples.append(Quadruple(-1, last_quad, 'goto_f', -1)) self.jumps.append(self.quad_counter) self.quad_counter += 1 # Make goto quadruple and actualize goto_f @_('') def con2(self, p): falso = self.jumps.pop() ret = self.jumps.pop() self.quadruples.append(Quadruple(-1, -1, 'goto', ret)) self.quadruples[falso - 1].res = self.quad_counter + 1 if len(self.break_stack): bq = self.break_stack.pop() self.quadruples[bq - 1].res = self.quad_counter + 1 self.quad_counter += 1 @_('FOR variable ass1 EQUALS expression ass2 nc0 UNTIL expression nc1 DO nc2 LCURL statements RCURL nc3') def non_conditional(self, p): return 'non_conditional' # Append the result of the last quadruple to the for_var_dir stack @_('') def nc0(self, p): self.for_var_dir.append(self.quadruples[-1].res) # Make quadruples if the stack of operators is not empty and do quadruple with <= as its operator @_('') def nc1(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) if not lo['dir'] in range(4500, 6000) and self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if not ro['dir'] in range(4500, 6000) and self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if made_quad: last_quad = self.quadruples[-1].res if (last_quad % 1500) // 300 != 0 and (last_quad % 1500) // 300 != 1: raise TypeError('Type mismatch') num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = 3 * 300 + \ int(num_temps.split('\u001f')[3]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 3) self.quadruples.append( Quadruple(self.for_var_dir[-1], last_quad, '<=', t_dir)) self.jumps.append(self.quad_counter) self.quad_counter += 1 self.temp_counter += 1 else: var = self.stack_of_stacks[-2].pop() if (var['dir'] % 1500) // 300 != 0 and (var['dir'] % 1500) // 300 != 1: raise TypeError('Type mismatch') num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = 3 * 300 + \ int(num_temps.split('\u001f')[3]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 3) if self.check_var_is_array(var): var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = var['dir'] self.quadruples.append( Quadruple(self.for_var_dir[-1], var_dir, '<=', t_dir)) self.jumps.append(self.quad_counter) self.quad_counter += 1 self.temp_counter += 1 # Make goto_f quadruple @_('') def nc2(self, p): last_quad = self.quadruples[-1].res self.quadruples.append(Quadruple(-1, last_quad, 'goto_f', -1)) self.jumps.append(self.quad_counter) self.quad_counter += 1 # Make goto quadruple and actualize goto_f @_('') def nc3(self, p): falso = self.jumps.pop() cond = self.jumps.pop() if not 1 in self.constant_table['int']: self.constant_table['int'].append(1) one_dir = self.constant_table['int'].index(1) + 4500 self.quadruples.append( Quadruple(self.for_var_dir[-1], one_dir, '+', self.for_var_dir[-1])) self.quad_counter += 1 self.quadruples.append(Quadruple(-1, -1, 'goto', cond)) self.quad_counter += 1 self.quadruples[falso - 1].res = self.quad_counter if len(self.break_stack): bq = self.break_stack.pop() self.quadruples[bq - 1].res = self.quad_counter + 1 self.for_var_dir.pop() @_('RETURN LPAREN expression fr0 RPAREN SEMI fr1') def function_returns(self, p): return 'function_returns' # Make return quadruple @_('') def fr0(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): self.make_and_push_quad() made_quad = True if made_quad: last_quad = self.quadruples[-1] self.quadruples.append( Quadruple(last_quad.res, -1, 'return', self.function_table[self.program_name]['vars'][self.curr_scope]['real_dir'])) self.quad_counter += 1 self.stack_of_stacks[-2].pop() else: self.quadruples.append( Quadruple(self.stack_of_stacks[-2].pop()['dir'], -1, 'return', self.function_table[self.program_name]['vars'][self.curr_scope]['real_dir'])) self.quad_counter += 1 # Actualize has_returned to true @_('') def fr1(self, p): self.has_returned = True @ _('function_or_method vf0 LPAREN func_params RPAREN fp2 fp3 SEMI') def call_to_void_function(self, p): return 'call_to_void_function' # Make gosub quadruple @ _('') def fp2(self, p): self.quadruples.append( Quadruple(self.called_func, -1, 'gosub', -1)) self.quad_counter += 1 # Equalize parameter counter to 0 @ _('') def fp3(self, p): self.param_counter = 0 # make era quadruple @ _('') def vf0(self, p): self.called_func, quads = p[-1] if self.current_class != None: self.called_func = self.current_class + '.' + self.called_func self.quadruples.append(Quadruple(self.called_func, -1, 'era', -1)) self.quad_counter += 1 if quads: for q in quads: self.quadruples.append(q) self.quad_counter += 1 @ _('expression fp1 param_list', 'empty') def func_params(self, p): return 'func_params' # Make quadruples if the stack of operators is not empty and do param quadruples @ _('') def fp1(self, p): if not self.called_func in self.function_table: for i in range(1, len(self.symstack)): if hasattr(self.symstack[i * -1], 'lineno'): lineno = self.symstack[i * -1].lineno break self.found_errors = True print('ERROR: No function\033[1m', self.called_func, '\033[0mwas found.') print(' Missing reference found on line', lineno) return made_quad = False while(len(self.stack_of_stacks[-1])): offset = 800 * len(self.types) * 2 ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) t_dir = idx * 300 + 3000 self.quadruples.append( Quadruple(lo['dir'], ro['dir'], op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if made_quad: last_quad = self.quadruples[-1] if self.param_counter == len(self.function_table[self.called_func]['params']): self.found_errors = True print( 'ERROR: Too many parameters passed in call to function on line', self.symstack[-2].lineno) return try: sm.checkAssignment(self.types[int(self.function_table[self.called_func] ['params'][self.param_counter])], self.types[(last_quad.res % 1500) // 300], '=') except TypeError: self.found_errors = True print( 'ERROR: Type mismatch on line', self.symstack[-2].lineno) print( ' Expected value of type', self.types[int(self.function_table[self.called_func]['params'][self.param_counter])], 'got value of type', self.types[(last_quad.res % 1500) // 300], 'instead') return self.quadruples.append( Quadruple(last_quad.res, -1, 'param', self.param_counter)) self.quad_counter += 1 self.param_counter += 1 else: val = self.stack_of_stacks[-2].pop() if self.param_counter == len(self.function_table[self.called_func]['params']): self.found_errors = True print( 'ERROR: Too many parameters passed in call to function on line', self.symstack[-2].lineno) return if not val: return try: sm.checkAssignment(self.types[int(self.function_table[self.called_func] ['params'][self.param_counter])], self.types[(val['dir'] % 1500) // 300], '=') except TypeError: self.found_errors = True print( 'ERROR: Type mismatch on line', self.symstack[-2].lineno) print( ' Expected value of type', self.types[int(self.function_table[self.called_func]['params'][self.param_counter])], 'got value of type', self.types[(val['dir'] % 1500) // 300], 'instead') return self.quadruples.append( Quadruple(val['dir'], -1, 'param', self.param_counter)) self.quad_counter += 1 self.param_counter += 1 @ _('MAIN m1_add_to_func_table LPAREN RPAREN LCURL main0 var_declaration statements RCURL main2') def main(self, p): return 'main' # Actualize the jump of the first goto made int he list of quadruples @ _('') def main0(self, p): self.quadruples[0].res = self.quad_counter # Do end quadruple and delete function and class tables @ _('') def main2(self, p): self.quadruples.append(Quadruple(-1, -1, 'end', -1)) del self.function_table[self.program_name]['vars'] del self.function_table['main']['vars'] for class_name in self.class_table: del self.class_table[class_name]['vars'] pass # Add main to function table @ _('') def m1_add_to_func_table(self, p): self.curr_scope = 'main' self.add_to_func_table('main', None) @ _('') def empty(self, p): pass def error(self, p): if not p: return print('ERROR: Syntax error found on line', p.lineno) if p.value == 'var': print( ' All variable declarations must be done before any other statement') elif p.value == '(': print( ' Parentheses are not allowed in this position.') elif p.value == '{': print( ' Curly brackets are not allowed in this position.') elif p.value == '[': print( ' Brackets are not allowed in this position.') elif p.value == ')': print( ' Closing parenthesis found without matching opening one.') elif p.value == '}': print( ' Closing curly bracket without an opening one.') elif p.value == ']': print( ' Closing bracket without an opening one.') elif p.value == ';': print( ' Must only be used at the end of statements') elif p.value == '=': print( ' Assignment is not allowed here. Perhaps you meant to use ==?') else: print( ' Keyword or id misplaced') if not self.found_errors: print( ' It\'s possible that all other syntax errors may be fixed by solving this one.') self.errok() self.found_errors = True while True: tok = next(self.tokens, None) if tok == None: raise EOFError() if tok.type == 'SEMI': tok = next(self.tokens, None) return tok
43.773619
213
0.565392
from sly import Parser from sly.yacc import _decorator as _ from .domas_lexer import DomasLexer from .domas_quadruples import Quadruple from .domas_errors import * from . import domas_semantic_cube as sm import json import os import copy os.system('color') class DomasParser(Parser): tokens = DomasLexer.tokens start = 'programa' function_table = {} class_table = {} constant_table = {'int': [], 'float': [], 'string': [], 'bool': []} stack_of_stacks = [[], []] stack_vars = [] last_arr_t = [] displacements = [] for_var_dir = [] break_stack = [] quadruples = [] jumps = [] quad_counter = 1 param_counter = 0 temp_counter = 0 attr_counter = 1 current_class = None last_arr_id = None last_type = None last_func_added = None has_returned = False found_errors = False types = ['int', 'float', 'string', 'bool', 'void'] operators = ['+', '-', '*', '/', '<', '>', '<=', '>=', '==', '<>', '&', '|'] def add_to_func_table(self, id, return_type): self.function_table[id] = { 'return_type': return_type, 'vars': {}, 'num_types': '0\u001f' * len(self.types), 'params': '', 'num_temps': '0\u001f' * len(self.types) } def check_variable_exists(self, var): if self.current_class != None: return var in self.function_table[self.curr_scope]['vars'] or var in self.class_table[self.current_class]['vars'] return var in self.function_table[self.curr_scope]['vars'] or var in self.function_table[self.program_name]['vars'] def get_var_type(self, var, tok): if not self.check_variable_exists(var): self.found_errors = True print('ERROR: No variable\033[1m', var, '\033[0mwas found.') print(' Missing reference found on line', tok.lineno) return None if self.current_class != None: if var in self.function_table[self.curr_scope]['vars']: return self.function_table[self.curr_scope]['vars'][var]['type'] return self.class_table[self.current_class]['vars'][var]['type'] if var in self.function_table[self.curr_scope]['vars']: return self.function_table[self.curr_scope]['vars'][var]['type'] return self.function_table[self.program_name]['vars'][var]['type'] def update_num_temps(self, func_num_temps, type_idx, quantity=1): lst = func_num_temps.split('\u001f') lst[type_idx] = str(int(lst[type_idx]) + quantity) return '\u001f'.join(lst) def check_var_is_array(self, var): if not var: return if var['dir'] >= 4500 and var['dir'] < 6000: return False if not self.check_variable_exists(var['value']): return False if self.current_class != None: if var['value'] in self.function_table[self.curr_scope]['vars']: return 'd1' in self.function_table[self.curr_scope]['vars'][var['value']] else: return 'd1' in self.class_table[self.current_class]['vars'][var['value']] elif var['value'] in self.function_table[self.curr_scope]['vars']: return 'd1' in self.function_table[self.curr_scope]['vars'][var['value']] else: return 'd1' in self.function_table[self.program_name]['vars'][var['value']] def make_and_push_quad(self): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() if not ro or not lo: raise SystemError("Reached unsolvable state") r_type = sm.checkOperation(lo['type'], ro['type'], op) self.last_type = r_type idx = self.types.index(r_type) num_temps = self.function_table[self.curr_scope]['num_temps'] self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.stack_of_stacks[-2].append( {'value': 't' + str(self.temp_counter), 'type': r_type, 'dir': t_dir}) if self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 @_('PROGRAM ID pro1 SEMI pro0 declarations') def programa(self, p): if self.found_errors: raise CompilationError() return (self.program_name, self.function_table, self.class_table, self.constant_table, self.quadruples) @_('') def pro0(self, p): self.quadruples.append(Quadruple(-1, -1, 'goto', -1)) self.quad_counter += 1 @_('') def pro1(self, p): self.program_name = p[-1] self.curr_scope = p[-1] self.function_table[p[-1]] = { 'return_type': None, 'vars': {}, 'num_types': '0\u001f0\u001f0\u001f0\u001f0\u001f' } @ _('class_declaration out_class var_declaration function_definition main') def declarations(self, p): return 'declarations' @ _('''CLASS ID cd1 inherits LCURL ATTRIBUTES attribute_declaration METHODS method_definition RCURL class_declaration''', 'empty') def class_declaration(self, p): return 'class_declaration' @ _('') def cd1(self, p): if p[-1] in self.class_table: self.found_errors = True print( 'ERROR: A class with the name\033[1m', p[-1], '\033[0mhas already been defined') print(' Redefinition found on line', self.symstack[-1].lineno) else: self.class_table[p[-1]] = { 'vars': {}, 'num_types': '0\u001f0\u001f0\u001f0\u001f0\u001f' } self.current_class = p[-1] @ _('INHERITS ID cd3', 'empty') def inherits(self, p): return 'inherits' @ _('') def cd3(self, p): if not p[-1] in self.class_table: self.found_errors = True print('ERROR: Id\033[1m', p[-1], '\033[0mis not defined as a class') print(' Missing reference found on line', self.symstack[-1].lineno) else: self.class_table[self.current_class] = copy.deepcopy( self.class_table[p[-1]]) @ _('VAR ID ad1 attr_vector', 'VAR ID ad1 attr_simple_var', 'empty') def attribute_declaration(self, p): return 'attribute_declaration' @ _('') def ad1(self, p): if p[-1] in self.class_table[self.current_class]['vars']: self.found_errors = True print('ERROR: An attribute\033[1m', p[-1], '\033[0mhas already been defined') print(' Redefinition found on line', self.symstack[-1].lineno) else: self.stack_vars.append(p[-1]) @ _('''LBRACKET CTE_I ad2 attr_multidim RBRACKET COLON simple_type ad4 SEMI attribute_declaration''') def attr_vector(self, p): return 'vector' # Adds the first dimension of an array to the information of the variable @ _('') def ad2(self, p): self.latest_var = self.stack_vars.pop() if self.class_table[self.current_class]['vars']: self.class_table[self.current_class]['vars'][self.latest_var] = { 'd1': p[-1]} else: self.class_table[self.current_class]['vars'] = { self.latest_var: {'d1': p[-1]}} @ _('COMMA CTE_I ad3', 'empty') def attr_multidim(self, p): return 'attr_multidim' # Adds the second dimension of an array to the information of the variable. @ _('') def ad3(self, p): self.class_table[self.current_class]['vars'][self.latest_var]['d2'] = p[-1] # Adds the type and direction to the information of the variable declared and updated the amount of types used in the class. @ _('') def ad4(self, p): idx = self.types.index(p[-1]) num_types = self.class_table[self.current_class]['num_types'] if 'd1' in self.class_table[self.current_class]['vars'][self.latest_var]: q = self.class_table[self.current_class]['vars'][self.latest_var]['d1'] if 'd2' in self.class_table[self.current_class]['vars'][self.latest_var]: q *= self.class_table[self.current_class]['vars'][self.latest_var]['d2'] self.class_table[self.current_class]['vars'][self.latest_var]['dir'] = 6000 + idx * \ 300 + int(num_types.split('\u001f')[idx]) self.class_table[self.current_class]['vars'][self.latest_var]['type'] = p[-1] self.class_table[self.current_class]['num_types'] = self.update_num_temps( num_types, idx, q) @ _('attr_var_list COLON simple_type ad5 SEMI attribute_declaration') def attr_simple_var(self, p): return 'attr_simple_var' @ _('COMMA ID ad1 attr_var_list', 'empty') def attr_var_list(self, p): return 'attr_var_list' # Pops the stack of variables and giving each their corresponding type. @ _('') def ad5(self, p): while len(self.stack_vars) > 0: curr_var = self.stack_vars.pop() if curr_var in self.class_table[self.current_class]['vars']: self.found_errors = True print('ERROR: An attribute\033[1m', curr_var, '\033[0mhas already been defined in class', self.current_class) print(' Redefinition found on line', self.symstack[-2].lineno) idx = self.types.index(p[-1]) num_types = self.class_table[self.current_class]['num_types'] self.class_table[self.current_class]['num_types'] = self.update_num_temps( num_types, idx) self.class_table[self.current_class]['vars'][curr_var] = { 'type': p[-1], 'dir': 6000 + idx * 300 + int(num_types.split('\u001f')[idx])} @ _('''def_type fd1 FUNCTION ID md3 LPAREN m_parameters RPAREN LCURL fd4 var_declaration statements RCURL fd5 fd6 method_definition''', 'empty') def method_definition(self, p): return 'method_definition' # Adds the id with the name of the class prefixed to the function table. @ _('') def md3(self, p): if p[-1] in self.class_table[self.current_class]['vars']: self.found_errors = True print('ERROR: An attribute or method\033[1m', p[-1], '\033[0mhas already been defined in class', self.current_class) print(' Redefinition found on line', self.symstack[-1].lineno) else: self.add_to_func_table( self.current_class + '.' + p[-1], self.curr_func_type) self.last_func_added = self.current_class + '.' + p[-1] self.curr_scope = self.last_func_added idx = self.types.index(self.curr_func_type) num_types = self.function_table[self.program_name]['num_types'] self.function_table[self.program_name]['num_types'] = self.update_num_temps( num_types, idx) self.function_table[self.program_name]['vars'][self.current_class + '.' + p[-1]] = { 'type': self.curr_func_type, 'dir': 0, 'real_dir': idx * 300 + int(num_types.split('\u001f')[idx])} @_('ID p1 COLON simple_type m2 m_param_choose', 'empty') def m_parameters(self, p): return 'parameters' # Adds the id and type of the parameter to the table of variables of the current method. @ _('') def m2(self, p): if self.latest_var in self.function_table[self.curr_scope]['vars']: self.found_errors = True print('ERROR: A parameter\033[1m', self.latest_var, '\033[0mhas already been declared for method', self.last_func_added.split('.')[1], 'in class', self.current_class) print(' Redefinition found on line', self.symstack[-2].lineno) else: idx = self.types.index(p[-1]) self.function_table[self.curr_scope]['params'] += str( idx) num_types = self.function_table[self.curr_scope]['num_types'] self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx) self.function_table[self.curr_scope]['vars'][self.latest_var] = { 'type': p[-1], 'dir': 1500 + idx * 300 + int(num_types.split('\u001f')[idx]) } @ _('COMMA m_parameters', 'empty') def m_param_choose(self, p): return 'm_param_choose' @ _('') def out_class(self, p): self.current_class = None self.curr_scope = self.program_name @ _('VAR ID gvd1 vector', 'VAR ID gvd1 simple_var', 'empty') def var_declaration(self, p): return p[0] # Appends the current var to the stack of variables of it doesn't exist. @ _('') def gvd1(self, p): if p[-1] in self.function_table: self.found_errors = True print('ERROR: A function with ID\033[1m', p[-1], '\033[0mhas already been declared. Variables may not share name with functions') print(' Redefinition found on line', self.symstack[-1].lineno) else: self.stack_vars.append(p[-1]) @ _('LBRACKET CTE_I gvd2 multidim RBRACKET COLON simple_type gvd4 SEMI var_declaration') def vector(self, p): return 'vector' @ _('') def gvd2(self, p): self.latest_var = self.stack_vars.pop() if self.function_table[self.curr_scope]['vars']: self.function_table[self.curr_scope]['vars'][self.latest_var] = { 'd1': p[-1] } else: self.function_table[self.curr_scope]['vars'] = { self.latest_var: {'d1': p[-1]} } @ _('COMMA CTE_I gvd3', 'empty') def multidim(self, p): return 'multidim' @ _('var_list COLON composite_type gvd6 SEMI var_declaration', 'var_list COLON simple_type gvd5 SEMI var_declaration') def simple_var(self, p): return 'simple_var' @ _('COMMA ID gvd1 var_list', 'empty') def var_list(self, p): return 'var_list' @ _('') def gvd3(self, p): self.function_table[self.curr_scope]['vars'][self.latest_var]['d2'] = p[-1] @ _('') def gvd4(self, p): idx = self.types.index(p[-1]) num_types = self.function_table[self.curr_scope]['num_types'] offset = 1500 if self.curr_scope != self.program_name else 0 if 'd1' in self.function_table[self.curr_scope]['vars'][self.latest_var]: q = self.function_table[self.curr_scope]['vars'][self.latest_var]['d1'] if 'd2' in self.function_table[self.curr_scope]['vars'][self.latest_var]: q *= self.function_table[self.curr_scope]['vars'][self.latest_var]['d2'] self.function_table[self.curr_scope]['vars'][self.latest_var]['dir'] = idx * \ 300 + int(num_types.split('\u001f')[idx]) + offset self.function_table[self.curr_scope]['vars'][self.latest_var]['type'] = p[-1] self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx, q) @ _('') def gvd5(self, p): while len(self.stack_vars) > 0: curr_var = self.stack_vars.pop() if curr_var in self.function_table[self.curr_scope]['vars']: self.found_errors = True print('ERROR: A variable\033[1m', curr_var, '\033[0mhas already been declared.') print(' Redefinition found on line', self.symstack[-5].lineno) idx = self.types.index(p[-1]) num_types = self.function_table[self.curr_scope]['num_types'] offset = 1500 if self.curr_scope != self.program_name else 0 self.function_table[self.curr_scope]['vars'][curr_var] = { 'type': p[-1], 'dir': idx * 300 + int(num_types.split('\u001f')[idx]) + offset} self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx) @ _('') def gvd6(self, p): while len(self.stack_vars) > 0: var_id = self.stack_vars.pop() if var_id in self.function_table[self.curr_scope]['vars']: self.found_errors = True print('ERROR: A variable\033[1m', var_id, '\033[0mhas already been declared.') print(' Redefinition found on line', self.symstack[-5].lineno) offset = 1500 if self.curr_scope != self.program_name else 0 num_types = self.function_table[self.curr_scope]['num_types'] base_addrs = [int(n) for n in num_types.split('\u001f')[:-1]] if not p[-1] in self.class_table: for i in range(1, len(self.symstack)): if hasattr(self.symstack[i * -1], 'lineno'): lineno = self.symstack[i * -1].lineno break print('ERROR: No class\033[1m', p[-1], '\033[0mwas found.') print(' Missing reference found on line', lineno) return for attr in self.class_table[p[-1]]['vars']: attr_type = self.class_table[p[-1]]['vars'][attr]['type'] idx = self.types.index(attr_type) num_types = self.function_table[self.curr_scope]['num_types'] q = 1 self.function_table[self.curr_scope]['vars'][var_id + '.' + attr] = { 'type': attr_type, 'dir': base_addrs[idx] + self.class_table[p[-1]]['vars'][attr]['dir'] - 6000 + offset } if 'd1' in self.class_table[p[-1]]['vars'][attr]: q = self.class_table[p[-1]]['vars'][attr]['d1'] self.function_table[self.curr_scope]['vars'][var_id + '.' + attr]['d1'] = q if 'd2' in self.class_table[p[-1]]['vars'][attr]: d2 = self.class_table[p[-1]]['vars'][attr]['d2'] q *= d2 self.function_table[self.curr_scope]['vars'][var_id + '.' + attr]['d2'] = d2 self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx, q) self.function_table[self.curr_scope]['vars'][var_id] = { 'type': p[-1]} @ _('def_type fd1 FUNCTION ID fd3 LPAREN parameters RPAREN LCURL fd4 var_declaration statements RCURL fd5 fd6 function_definition', 'empty') def function_definition(self, p): return 'function_definition' @ _('') def fd1(self, p): self.curr_func_type = p[-1] @ _('') def fd3(self, p): if p[-1] in self.function_table: self.found_errors = True print('ERROR: A function\033[1m', p[-1], '\033[0mhas already been defined.') print(' Redefinition found on line', self.symstack[-1].lineno) elif p[-1] in self.function_table[self.program_name]['vars']: self.found_errors = True print('ERROR: A global variable\033[1m', p[-1], '\033[0mhas been declared. Functions may not share names with global variables') print(' Redefinition found on line', self.symstack[-1].lineno) else: self.add_to_func_table(p[-1], self.curr_func_type) self.last_func_added = p[-1] self.curr_scope = self.last_func_added idx = self.types.index(self.curr_func_type) num_types = self.function_table[self.program_name]['num_types'] self.function_table[self.program_name]['num_types'] = self.update_num_temps( num_types, idx) self.function_table[self.program_name]['vars'][p[-1]] = { 'type': self.curr_func_type, 'dir': 0, 'real_dir': idx * 300 + int(num_types.split('\u001f')[idx])} @ _('') def fd4(self, p): if not self.last_func_added: return self.function_table[self.last_func_added]['start'] = self.quad_counter @ _('') def fd5(self, p): if not self.last_func_added: return del self.function_table[self.last_func_added]['vars'] @ _('') def fd6(self, p): if self.curr_func_type != 'void' and self.has_returned == False: self.found_errors = True print('ERROR: Function\033[1m', self.curr_scope, '\033[0mis missing a return statement') print(' Non-void functions must have a return statement.') self.quadruples.append(Quadruple(-1, -1, 'end_func', -1)) self.quad_counter += 1 self.temp_counter = 1 self.has_returned = False @ _('statement statements', 'empty') def statements(self, p): return 'statements' @ _('simple_type', 'VOID') def def_type(self, p): return p[0] @ _('INT', 'FLOAT', 'STRING', 'BOOL') def simple_type(self, p): return p[0] @ _('ID') def composite_type(self, p): return p[0] @ _('ID p1 COLON simple_type p2 param_choose', 'empty') def parameters(self, p): return 'parameters' @ _('COMMA parameters', 'empty') def param_choose(self, p): return 'param_choose' @ _('') def p1(self, p): self.latest_var = p[-1] @ _('') def p2(self, p): if self.latest_var in self.function_table[self.curr_scope]['vars']: self.found_errors = True print('ERROR: A parameter\033[1m', self.latest_var, '\033[0mhas already been declared for function', self.last_func_added) print(' Redefinition found on line', self.symstack[-2].lineno) idx = self.types.index(p[-1]) self.function_table[self.curr_scope]['params'] += str(idx) num_types = self.function_table[self.curr_scope]['num_types'] offset = 1500 if self.curr_scope != self.program_name else 0 self.function_table[self.curr_scope]['num_types'] = self.update_num_temps( num_types, idx) self.function_table[self.curr_scope]['vars'][self.latest_var] = { 'type': p[-1], 'dir': idx * 300 + int(num_types.split('\u001f')[idx]) + offset} @_('assignment', 'call_to_void_function', 'function_returns', 'read', 'print', 'decision_statement', 'repetition_statement', 'BREAK br0 SEMI') def statement(self, p): return 'statement' @_('') def br0(self, p): if len(self.jumps) == 0: self.found_errors = True print('ERROR: break statement on line', self.symstack[-1].lineno, 'used outside a loop') self.quadruples.append(Quadruple(-1, -1, 'goto', None)) self.break_stack.append(self.quad_counter) self.quad_counter += 1 @_('variable ass1 EQUALS expression ass2 SEMI') def assignment(self, p): return 'assignment' @_('') def ass1(self, p): self.latest_var = p[-1] @_('') def ass2(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): self.make_and_push_quad() made_quad = True lo = self.stack_of_stacks[-2].pop() if not lo: return v_type = self.get_var_type(self.latest_var, self.symstack[-2]) if not v_type: return self.last_type = sm.checkOperation(v_type, lo['type'], '=') if not made_quad and self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if self.current_class != None: if self.latest_var in self.class_table[self.current_class]['vars']: if 'd1' in self.class_table[self.current_class]['vars'][self.latest_var]: if not self.last_arr_t: return var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = self.class_table[self.current_class]['vars'][self.latest_var]['dir'] else: if 'd1' in self.function_table[self.curr_scope]['vars'][self.latest_var]: if not self.last_arr_t: return var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = self.function_table[self.curr_scope]['vars'][self.latest_var]['dir'] elif self.latest_var in self.function_table[self.curr_scope]['vars']: if 'd1' in self.function_table[self.curr_scope]['vars'][self.latest_var]: if not self.last_arr_t: return var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = self.function_table[self.curr_scope]['vars'][self.latest_var]['dir'] else: if 'd1' in self.function_table[self.program_name]['vars'][self.latest_var]: if not self.last_arr_t: return var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = self.function_table[self.program_name]['vars'][self.latest_var]['dir'] q = Quadruple(lo_dir, -1, '=', var_dir) self.quadruples.append(q) self.quad_counter += 1 @_('id_or_attribute', 'id_or_attribute v0 LBRACKET expression v1 RBRACKET', 'id_or_attribute v0 LBRACKET expression v2 COMMA v4 expression v3 RBRACKET') def variable(self, p): return p[0] @_('') def v0(self, p): self.check_variable_exists(p[-1]) if self.current_class != None: if not 'd1' in self.class_table[self.current_class]['vars'][p[-1]]: self.found_errors = True print('ERROR: Variable\033[1m', p[-1], '\033[0mis not an array or matrix.') self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) return elif p[-1] in self.function_table[self.curr_scope]['vars']: if not 'd1' in self.function_table[self.curr_scope]['vars'][p[-1]]: self.found_errors = True print('ERROR: Variable\033[1m', p[-1], '\033[0mis not an array or matrix.') self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) return elif not 'd1' in self.function_table[self.program_name]['vars'][p[-1]]: self.found_errors = True print('ERROR: Variable\033[1m', p[-1], '\033[0mis not an array or matrix.') self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) return self.last_arr_id = p[-1] self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) @_('') def v4(self, p): self.check_variable_exists(self.last_arr_id) if self.current_class != None: if not 'd2' in self.class_table[self.current_class]['vars'][self.last_arr_id]: self.found_errors = True print('ERROR: Variable\033[1m', self.last_arr_id, '\033[0mis not a matrix.') elif self.last_arr_id in self.function_table[self.curr_scope]['vars']: if not 'd2' in self.function_table[self.curr_scope]['vars'][self.last_arr_id]: self.found_errors = True print('ERROR: Variable\033[1m', self.last_arr_id, '\033[0mis not a matrix.') else: if not 'd2' in self.function_table[self.program_name]['vars'][self.last_arr_id]: self.found_errors = True print('ERROR: Variable\033[1m', self.last_arr_id, '\033[0mis not a matrix.') self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) @_('') def v1(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) self.quadruples.append( Quadruple(lo['dir'], ro['dir'], op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if self.current_class != None and self.last_arr_id in self.class_table[self.current_class]['vars']: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.class_table[self.current_class]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.class_table[self.current_class]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '+', t_dir)) self.quad_counter += 1 self.last_arr_t.append(t_dir) elif self.last_arr_id in self.function_table[self.curr_scope]['vars']: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '+', t_dir)) self.quad_counter += 1 self.last_arr_t.append(t_dir) else: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.program_name]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.function_table[self.program_name]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '+', t_dir)) self.quad_counter += 1 self.last_arr_t.append(t_dir) self.stack_of_stacks.pop() self.stack_of_stacks.pop() @_('') def v2(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) self.quadruples.append( Quadruple(lo['dir'], ro['dir'], op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if self.current_class != None: pass elif self.last_arr_id in self.function_table[self.curr_scope]['vars']: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 d2 = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['d2'] if not d2 in self.constant_table['int']: self.constant_table['int'].append(d2) cons_dir = self.constant_table['int'].index(d2) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '*', t_dir)) self.quad_counter += 1 self.displacements.append(t_dir) else: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.program_name]['vars'][self.last_arr_id]['d1'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 d2 = self.function_table[self.program_name]['vars'][self.last_arr_id]['d2'] if not d2 in self.constant_table['int']: self.constant_table['int'].append(d2) cons_dir = self.constant_table['int'].index(d2) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0) self.quadruples.append(Quadruple(cons_dir, t_addr, '*', t_dir)) self.quad_counter += 1 self.displacements.append(t_dir) self.stack_of_stacks.pop() self.stack_of_stacks.pop() @_('') def v3(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) self.quadruples.append( Quadruple(lo['dir'], ro['dir'], op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if self.current_class != None: pass elif self.last_arr_id in self.function_table[self.curr_scope]['vars']: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['d2'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.function_table[self.curr_scope]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0, 2) self.quadruples.append( Quadruple(self.displacements.pop(), t_addr, '+', t_dir)) self.quadruples.append(Quadruple(cons_dir, t_dir, '+', t_dir + 1)) self.quad_counter += 2 self.last_arr_t.append(t_dir + 1) else: t_addr = self.quadruples[-1].res if made_quad else self.stack_of_stacks[-2].pop()[ 'dir'] if (t_addr % 1500) // 300 != 0 and (t_addr % 1500) // 300 != 1: raise TypeError('Type mismatch') lms = self.function_table[self.program_name]['vars'][self.last_arr_id]['d2'] self.quadruples.append(Quadruple(0, lms, 'verify', t_addr)) self.quad_counter += 1 dir_b = self.function_table[self.program_name]['vars'][self.last_arr_id]['dir'] if not dir_b in self.constant_table['int']: self.constant_table['int'].append(dir_b) cons_dir = self.constant_table['int'].index(dir_b) + 4500 num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = int(num_temps.split('\u001f')[0]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 0, 2) self.quadruples.append( Quadruple(self.displacements.pop(), t_addr, '+', t_dir)) self.quadruples.append(Quadruple(cons_dir, t_dir, '+', t_dir + 1)) self.quad_counter += 2 self.last_arr_t.append(t_dir + 1) self.stack_of_stacks.pop() self.stack_of_stacks.pop() @_('ID', 'ID DOT ID') def id_or_attribute(self, p): if len(p) > 1: return p[0] + p[1] + p[2] return p[0] @_('variable', 'CTE_I', 'CTE_F', 'CTE_STRING', 'cte_bool', 'call_to_function') def var_cte(self, p): offset = 4500 if hasattr(p, 'CTE_I'): cte_type = 'int' if not p[0] in self.constant_table['int']: self.constant_table['int'].append(p[0]) cons_dir = self.constant_table['int'].index(p[0]) + offset elif hasattr(p, 'CTE_F'): cte_type = 'float' if not p[0] in self.constant_table['float']: self.constant_table['float'].append(p[0]) cons_dir = self.constant_table['float'].index(p[0]) + offset + 300 elif hasattr(p, 'CTE_STRING'): cte_type = 'string' if not p[0] in self.constant_table['string']: self.constant_table['string'].append(p[0]) cons_dir = self.constant_table['string'].index(p[0]) + offset + 600 elif hasattr(p, 'cte_bool'): cte_type = 'bool' if not p[0] in self.constant_table['bool']: self.constant_table['bool'].append(p[0]) cons_dir = self.constant_table['bool'].index(p[0]) + offset + 900 elif hasattr(p, 'call_to_function'): return p[0] else: if not self.check_variable_exists(p[0]): for i in range(1, len(self.symstack)): if hasattr(self.symstack[i * -1], 'lineno'): lineno = self.symstack[i * -1].lineno break self.found_errors = True print('ERROR: No variable\033[1m', p[0], '\033[0mwas found.') print(' Missing reference found on line', lineno) return if self.current_class != None and p[0] in self.class_table[self.current_class]['vars']: cte_type = self.class_table[self.current_class]['vars'][p[0]]['type'] cons_dir = self.class_table[self.current_class]['vars'][p[0]]['dir'] else: cte_type = self.get_var_type(p[0], self.symstack[-2]) if p[0] in self.function_table[self.curr_scope]['vars']: cons_dir = self.function_table[self.curr_scope]['vars'][p[0]]['dir'] else: cons_dir = self.function_table[self.program_name]['vars'][p[0]]['dir'] return {'value': p[0], 'type': cte_type, 'dir': cons_dir} @_('constant e2 operator e3 expression', 'constant e2', 'LPAREN e1 expression RPAREN e4', 'LPAREN e1 expression RPAREN e4 operator e3 expression') def expression(self, p): if hasattr(p, 'LPAREN'): return p[2] return p[0] @_('') def e1(self, p): self.stack_of_stacks[-1].append('(') @_('') def e2(self, p): self.stack_of_stacks[-2].append(p[-1]) @_('') def e3(self, p): if len(self.stack_of_stacks[-1]) == 0 or self.stack_of_stacks[-1][-1] == '(': self.stack_of_stacks[-1].append(p[-1]) elif self.stack_of_stacks[-1][-1] == '*' or self.stack_of_stacks[-1][-1] == '/': self.make_and_push_quad() if (self.stack_of_stacks[-1] and (self.stack_of_stacks[-1][-1] == '+' or self.stack_of_stacks[-1][-1] == '-')) and (p[-1] == '+' or p[-1] == '-'): self.make_and_push_quad() self.stack_of_stacks[-1].append(p[-1]) elif p[-1] == '*' or p[-1] == '/': self.stack_of_stacks[-1].append(p[-1]) elif self.stack_of_stacks[-1][-1] == '+' or self.stack_of_stacks[-1][-1] == '-': self.make_and_push_quad() self.stack_of_stacks[-1].append(p[-1]) elif p[-1] == '+' or p[-1] == '-': self.stack_of_stacks[-1].append(p[-1]) elif self.stack_of_stacks[-1][-1] in sm.comparison_ops or self.stack_of_stacks[-1][-1] in sm.equality_ops: self.make_and_push_quad() self.stack_of_stacks[-1].append(p[-1]) elif p[-1] in sm.comparison_ops or p[-1] in sm.equality_ops: self.stack_of_stacks[-1].append(p[-1]) elif self.stack_of_stacks[-1][-1] in sm.logic_ops: self.make_and_push_quad() self.stack_of_stacks[-1].append(p[-1]) elif p[-1] in sm.logic_ops: self.stack_of_stacks[-1].append(p[-1]) @_('') def e4(self, p): while(self.stack_of_stacks[-1][-1] != '('): self.make_and_push_quad() self.stack_of_stacks[-1].pop() @_('AND', 'OR') def logical_operator(self, p): return p[-1] @_('LT', 'GT', 'SAME', 'GEQ', 'LEQ', 'NEQ') def relational_operator(self, p): return p[0] @_('PLUS', 'MINUS', 'MULTIPLY', 'DIVIDE') def arithmetic_operator(self, p): return p[0] @_('logical_operator', 'relational_operator', 'arithmetic_operator') def operator(self, p): return p[0] @_('PLUS var_cte', 'MINUS var_cte', 'var_cte') def constant(self, p): if len(p) > 1 and p[1] == '-': return -p.var_cte else: return p.var_cte @_('READ LPAREN read_h') def read(self, p): return 'read' @_('variable r1 COMMA read_h', 'variable r1 RPAREN SEMI') def read_h(self, p): return 'read_h' @_('') def r1(self, p): if self.current_class != None and p[-1] in self.class_table[self.current_class]['vars']: if 'd1' in self.class_table[self.current_class]['vars'][p[-1]]: var_addr = '$' + str(self.last_arr_t.pop()) else: var_addr = self.class_table[self.current_class]['vars'][p[-1]]['dir'] elif p[-1] in self.function_table[self.curr_scope]['vars']: if 'd1' in self.function_table[self.curr_scope]['vars'][p[-1]]: var_addr = '$' + str(self.last_arr_t.pop()) else: var_addr = self.function_table[self.curr_scope]['vars'][p[-1]]['dir'] elif p[-1] in self.function_table[self.program_name]['vars']: if 'd1' in self.function_table[self.program_name]['vars'][p[-1]]: var_addr = '$' + str(self.last_arr_t.pop()) else: var_addr = self.function_table[self.program_name]['vars'][p[-1]]['dir'] else: raise UndeclaredIdError(p[-1]) self.quadruples.append(Quadruple(-1, -1, 'read', var_addr)) self.quad_counter += 1 @_('function_or_method vf0 ctf2 LPAREN func_params RPAREN fp2 fp3 ctf0 ctf3') def call_to_function(self, p): if not self.check_variable_exists(self.called_func): return func_dir = self.function_table[self.program_name]['vars'][self.called_func]['dir'] func_type = self.function_table[self.called_func]['return_type'] return {'value': 't' + str(self.temp_counter - 1), 'type': func_type, 'dir': func_dir} @_('') def ctf2(self, p): self.stack_of_stacks.append([]) self.stack_of_stacks.append([]) @_('') def ctf3(self, p): self.stack_of_stacks.pop() self.stack_of_stacks.pop() @_('') def ctf0(self, p): if not self.check_variable_exists(self.called_func): self.found_errors = True print('ERROR: No function\033[1m', self.called_func, '\033[0mwas found.') print(' Missing reference found on line', self.symstack[-3].lineno) return func_dir = self.function_table[self.program_name]['vars'][self.called_func]['real_dir'] func_type = self.function_table[self.program_name]['vars'][self.called_func]['type'] idx = self.types.index(func_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) self.quadruples.append( Quadruple(func_dir, -1, '=', t_dir)) self.function_table[self.program_name]['vars'][self.called_func]['dir'] = t_dir self.quad_counter += 1 self.temp_counter += 1 @_('ID ctf1', 'ID DOT ID') def function_or_method(self, p): if(len(p) == 2): return (p[0], None) else: var_type = self.get_var_type(p[0], self.symstack[-1]) quads = [] for attr in self.class_table[var_type]['vars']: var_dir = self.function_table[self.curr_scope]['vars'][p[0]+'.'+attr]['dir'] attr_dir = self.class_table[var_type]['vars'][attr]['dir'] quads.append(Quadruple(var_dir, -2, '=', attr_dir)) return (var_type + p[1] + p[2], quads) @_('') def ctf1(self, p): if self.current_class != None: if not self.current_class + '.' + p[-1] in self.function_table: self.found_errors = True print('ERROR: No function\033[1m', p[-1], '\033[0mwas found.') print(' Missing reference found on line', self.symstack[-1].lineno) elif not p[-1] in self.function_table: self.found_errors = True print('ERROR: No function\033[1m', p[-1], '\033[0mwas found.') print(' Missing reference found on line', self.symstack[-1].lineno) @_('COMMA expression fp1 param_list', 'empty') def param_list(self, p): return 'param_list' @_('PRINT LPAREN res_write RPAREN SEMI') def print(self, p): return 'print' @_('expression pr1 comma_thing') def res_write(self, p): return 'res_write' @_('COMMA res_write', 'empty') def comma_thing(self, p): return 'comma_thing' @_('') def pr1(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) if not lo['dir'] in range(4500, 6000) and self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if not ro['dir'] in range(4500, 6000) and self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if made_quad: last_quad = self.quadruples[-1] self.quadruples.append( Quadruple(-1, -1, 'print', last_quad.res)) self.quad_counter += 1 else: var = self.stack_of_stacks[-2].pop() if not var: return if self.check_var_is_array(var): var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = var['dir'] self.quadruples.append( Quadruple(-1, -1, 'print', var_dir)) self.quad_counter += 1 @_('TRUE', 'FALSE') def cte_bool(self, p): return p[0] @_('IF LPAREN expression dec1 RPAREN THEN LCURL statements RCURL else_stm') def decision_statement(self, p): return 'decision_statement' @_('') def dec1(self, p): while len(self.stack_of_stacks[-1]): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + 3000 r_type = sm.checkOperation(lo['type'], ro['type'], op) self.stack_of_stacks[-2].append( {'value': 't' + str(self.temp_counter), 'type': r_type, 'dir': t_dir}) self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) if self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 lo = self.stack_of_stacks[-2].pop() if lo['type'] != 'bool': raise SyntaxError( 'Expression to evaluate in if statement is not boolean') else: self.quadruples.append(Quadruple(-1, lo['dir'], 'goto_f', -1)) self.jumps.append(self.quad_counter) self.quad_counter += 1 @_('dec2 ELSE LCURL statements RCURL dec3', 'empty dec4') def else_stm(self, p): return 'else_stm' @_('') def dec2(self, p): falso = self.jumps.pop() self.quadruples.append(Quadruple(-1, -1, 'goto', -1)) self.jumps.append(self.quad_counter) self.quad_counter += 1 self.quadruples[falso - 1].res = self.quad_counter @_('') def dec3(self, p): jump = self.jumps.pop() self.quadruples[jump - 1].res = self.quad_counter @_('') def dec4(self, p): jump = self.jumps.pop() self.quadruples[jump - 1].res = self.quad_counter @_('conditional', 'non_conditional') def repetition_statement(self, p): return 'repetition_statement' @_('WHILE LPAREN con0 expression con1 RPAREN DO LCURL statements RCURL con2') def conditional(self, p): return 'conditional' @_('') def con0(self, p): self.jumps.append(self.quad_counter) @_('') def con1(self, p): while len(self.stack_of_stacks[-1]): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) r_type = sm.checkOperation(lo['type'], ro['type'], op) self.stack_of_stacks[-2].append( {'value': 't' + str(self.temp_counter), 'type': r_type, 'dir': t_dir}) if self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 if self.last_type != 'bool': raise SyntaxError( 'Expression to evaluate in if statement is not boolean') else: last_quad = self.quadruples[-1].res self.quadruples.append(Quadruple(-1, last_quad, 'goto_f', -1)) self.jumps.append(self.quad_counter) self.quad_counter += 1 @_('') def con2(self, p): falso = self.jumps.pop() ret = self.jumps.pop() self.quadruples.append(Quadruple(-1, -1, 'goto', ret)) self.quadruples[falso - 1].res = self.quad_counter + 1 if len(self.break_stack): bq = self.break_stack.pop() self.quadruples[bq - 1].res = self.quad_counter + 1 self.quad_counter += 1 @_('FOR variable ass1 EQUALS expression ass2 nc0 UNTIL expression nc1 DO nc2 LCURL statements RCURL nc3') def non_conditional(self, p): return 'non_conditional' @_('') def nc0(self, p): self.for_var_dir.append(self.quadruples[-1].res) @_('') def nc1(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = idx * 300 + \ int(num_temps.split('\u001f')[idx]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) if not lo['dir'] in range(4500, 6000) and self.check_var_is_array(lo): lo_dir = '$' + str(self.last_arr_t.pop()) else: lo_dir = lo['dir'] if not ro['dir'] in range(4500, 6000) and self.check_var_is_array(ro): ro_dir = '$' + str(self.last_arr_t.pop()) else: ro_dir = ro['dir'] self.quadruples.append( Quadruple(lo_dir, ro_dir, op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if made_quad: last_quad = self.quadruples[-1].res if (last_quad % 1500) // 300 != 0 and (last_quad % 1500) // 300 != 1: raise TypeError('Type mismatch') num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = 3 * 300 + \ int(num_temps.split('\u001f')[3]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 3) self.quadruples.append( Quadruple(self.for_var_dir[-1], last_quad, '<=', t_dir)) self.jumps.append(self.quad_counter) self.quad_counter += 1 self.temp_counter += 1 else: var = self.stack_of_stacks[-2].pop() if (var['dir'] % 1500) // 300 != 0 and (var['dir'] % 1500) // 300 != 1: raise TypeError('Type mismatch') num_temps = self.function_table[self.curr_scope]['num_temps'] t_dir = 3 * 300 + \ int(num_temps.split('\u001f')[3]) + 3000 self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, 3) if self.check_var_is_array(var): var_dir = '$' + str(self.last_arr_t.pop()) else: var_dir = var['dir'] self.quadruples.append( Quadruple(self.for_var_dir[-1], var_dir, '<=', t_dir)) self.jumps.append(self.quad_counter) self.quad_counter += 1 self.temp_counter += 1 @_('') def nc2(self, p): last_quad = self.quadruples[-1].res self.quadruples.append(Quadruple(-1, last_quad, 'goto_f', -1)) self.jumps.append(self.quad_counter) self.quad_counter += 1 @_('') def nc3(self, p): falso = self.jumps.pop() cond = self.jumps.pop() if not 1 in self.constant_table['int']: self.constant_table['int'].append(1) one_dir = self.constant_table['int'].index(1) + 4500 self.quadruples.append( Quadruple(self.for_var_dir[-1], one_dir, '+', self.for_var_dir[-1])) self.quad_counter += 1 self.quadruples.append(Quadruple(-1, -1, 'goto', cond)) self.quad_counter += 1 self.quadruples[falso - 1].res = self.quad_counter if len(self.break_stack): bq = self.break_stack.pop() self.quadruples[bq - 1].res = self.quad_counter + 1 self.for_var_dir.pop() @_('RETURN LPAREN expression fr0 RPAREN SEMI fr1') def function_returns(self, p): return 'function_returns' @_('') def fr0(self, p): made_quad = False while(len(self.stack_of_stacks[-1])): self.make_and_push_quad() made_quad = True if made_quad: last_quad = self.quadruples[-1] self.quadruples.append( Quadruple(last_quad.res, -1, 'return', self.function_table[self.program_name]['vars'][self.curr_scope]['real_dir'])) self.quad_counter += 1 self.stack_of_stacks[-2].pop() else: self.quadruples.append( Quadruple(self.stack_of_stacks[-2].pop()['dir'], -1, 'return', self.function_table[self.program_name]['vars'][self.curr_scope]['real_dir'])) self.quad_counter += 1 @_('') def fr1(self, p): self.has_returned = True @ _('function_or_method vf0 LPAREN func_params RPAREN fp2 fp3 SEMI') def call_to_void_function(self, p): return 'call_to_void_function' @ _('') def fp2(self, p): self.quadruples.append( Quadruple(self.called_func, -1, 'gosub', -1)) self.quad_counter += 1 @ _('') def fp3(self, p): self.param_counter = 0 @ _('') def vf0(self, p): self.called_func, quads = p[-1] if self.current_class != None: self.called_func = self.current_class + '.' + self.called_func self.quadruples.append(Quadruple(self.called_func, -1, 'era', -1)) self.quad_counter += 1 if quads: for q in quads: self.quadruples.append(q) self.quad_counter += 1 @ _('expression fp1 param_list', 'empty') def func_params(self, p): return 'func_params' @ _('') def fp1(self, p): if not self.called_func in self.function_table: for i in range(1, len(self.symstack)): if hasattr(self.symstack[i * -1], 'lineno'): lineno = self.symstack[i * -1].lineno break self.found_errors = True print('ERROR: No function\033[1m', self.called_func, '\033[0mwas found.') print(' Missing reference found on line', lineno) return made_quad = False while(len(self.stack_of_stacks[-1])): offset = 800 * len(self.types) * 2 ro = self.stack_of_stacks[-2].pop() lo = self.stack_of_stacks[-2].pop() op = self.stack_of_stacks[-1].pop() self.last_type = sm.checkOperation(lo['type'], ro['type'], op) idx = self.types.index(self.last_type) num_temps = self.function_table[self.curr_scope]['num_temps'] self.function_table[self.curr_scope]['num_temps'] = self.update_num_temps( num_temps, idx) t_dir = idx * 300 + 3000 self.quadruples.append( Quadruple(lo['dir'], ro['dir'], op, t_dir)) self.temp_counter += 1 self.quad_counter += 1 made_quad = True if made_quad: last_quad = self.quadruples[-1] if self.param_counter == len(self.function_table[self.called_func]['params']): self.found_errors = True print( 'ERROR: Too many parameters passed in call to function on line', self.symstack[-2].lineno) return try: sm.checkAssignment(self.types[int(self.function_table[self.called_func] ['params'][self.param_counter])], self.types[(last_quad.res % 1500) // 300], '=') except TypeError: self.found_errors = True print( 'ERROR: Type mismatch on line', self.symstack[-2].lineno) print( ' Expected value of type', self.types[int(self.function_table[self.called_func]['params'][self.param_counter])], 'got value of type', self.types[(last_quad.res % 1500) // 300], 'instead') return self.quadruples.append( Quadruple(last_quad.res, -1, 'param', self.param_counter)) self.quad_counter += 1 self.param_counter += 1 else: val = self.stack_of_stacks[-2].pop() if self.param_counter == len(self.function_table[self.called_func]['params']): self.found_errors = True print( 'ERROR: Too many parameters passed in call to function on line', self.symstack[-2].lineno) return if not val: return try: sm.checkAssignment(self.types[int(self.function_table[self.called_func] ['params'][self.param_counter])], self.types[(val['dir'] % 1500) // 300], '=') except TypeError: self.found_errors = True print( 'ERROR: Type mismatch on line', self.symstack[-2].lineno) print( ' Expected value of type', self.types[int(self.function_table[self.called_func]['params'][self.param_counter])], 'got value of type', self.types[(val['dir'] % 1500) // 300], 'instead') return self.quadruples.append( Quadruple(val['dir'], -1, 'param', self.param_counter)) self.quad_counter += 1 self.param_counter += 1 @ _('MAIN m1_add_to_func_table LPAREN RPAREN LCURL main0 var_declaration statements RCURL main2') def main(self, p): return 'main' @ _('') def main0(self, p): self.quadruples[0].res = self.quad_counter @ _('') def main2(self, p): self.quadruples.append(Quadruple(-1, -1, 'end', -1)) del self.function_table[self.program_name]['vars'] del self.function_table['main']['vars'] for class_name in self.class_table: del self.class_table[class_name]['vars'] pass @ _('') def m1_add_to_func_table(self, p): self.curr_scope = 'main' self.add_to_func_table('main', None) @ _('') def empty(self, p): pass def error(self, p): if not p: return print('ERROR: Syntax error found on line', p.lineno) if p.value == 'var': print( ' All variable declarations must be done before any other statement') elif p.value == '(': print( ' Parentheses are not allowed in this position.') elif p.value == '{': print( ' Curly brackets are not allowed in this position.') elif p.value == '[': print( ' Brackets are not allowed in this position.') elif p.value == ')': print( ' Closing parenthesis found without matching opening one.') elif p.value == '}': print( ' Closing curly bracket without an opening one.') elif p.value == ']': print( ' Closing bracket without an opening one.') elif p.value == ';': print( ' Must only be used at the end of statements') elif p.value == '=': print( ' Assignment is not allowed here. Perhaps you meant to use ==?') else: print( ' Keyword or id misplaced') if not self.found_errors: print( ' It\'s possible that all other syntax errors may be fixed by solving this one.') self.errok() self.found_errors = True while True: tok = next(self.tokens, None) if tok == None: raise EOFError() if tok.type == 'SEMI': tok = next(self.tokens, None) return tok
true
true
f70eb2c0d33d0d03e11a8847b577c694fe9dd459
1,322
py
Python
tracdap-runtime/python/test/tracdap_examples/test_chaining.py
martin-traverse/tracdap
2df2f08bee352f4f5188953efe5a33aa1ae51f2d
[ "Apache-2.0" ]
null
null
null
tracdap-runtime/python/test/tracdap_examples/test_chaining.py
martin-traverse/tracdap
2df2f08bee352f4f5188953efe5a33aa1ae51f2d
[ "Apache-2.0" ]
null
null
null
tracdap-runtime/python/test/tracdap_examples/test_chaining.py
martin-traverse/tracdap
2df2f08bee352f4f5188953efe5a33aa1ae51f2d
[ "Apache-2.0" ]
null
null
null
# Copyright 2022 Accenture Global Solutions Limited # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import unittest import pathlib import sys import tracdap.rt.launch as launch _ROOT_DIR = pathlib.Path(__file__).parent \ .joinpath("../../../..") \ .resolve() _EXAMPLES_DIR = _ROOT_DIR.joinpath("examples/models/python") class ChainingExample(unittest.TestCase): def test_chaining(self): job_config = _EXAMPLES_DIR.joinpath("chaining/chaining.yaml") sys_config = _EXAMPLES_DIR.joinpath("sys_config.yaml") test_dir = str(_EXAMPLES_DIR.joinpath("chaining")) try: sys.path.append(test_dir) launch.launch_job(job_config, sys_config, dev_mode=True) self.assertTrue(True) finally: sys.path.remove(test_dir)
28.12766
75
0.706505
import unittest import pathlib import sys import tracdap.rt.launch as launch _ROOT_DIR = pathlib.Path(__file__).parent \ .joinpath("../../../..") \ .resolve() _EXAMPLES_DIR = _ROOT_DIR.joinpath("examples/models/python") class ChainingExample(unittest.TestCase): def test_chaining(self): job_config = _EXAMPLES_DIR.joinpath("chaining/chaining.yaml") sys_config = _EXAMPLES_DIR.joinpath("sys_config.yaml") test_dir = str(_EXAMPLES_DIR.joinpath("chaining")) try: sys.path.append(test_dir) launch.launch_job(job_config, sys_config, dev_mode=True) self.assertTrue(True) finally: sys.path.remove(test_dir)
true
true
f70eb2d8117df4c61263da019a776ba2536c2d96
4,243
py
Python
day08/run.py
kung-foo/aoc2021
ce3111d50b7b3354331b8dc546b13c7fe8882ba9
[ "Apache-2.0" ]
null
null
null
day08/run.py
kung-foo/aoc2021
ce3111d50b7b3354331b8dc546b13c7fe8882ba9
[ "Apache-2.0" ]
null
null
null
day08/run.py
kung-foo/aoc2021
ce3111d50b7b3354331b8dc546b13c7fe8882ba9
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/env python3 import os import sys import random import numpy as np src = open("input.txt", "r").readlines() example = """ be cfbegad cbdgef fgaecd cgeb fdcge agebfd fecdb fabcd edb | fdgacbe cefdb cefbgd gcbe edbfga begcd cbg gc gcadebf fbgde acbgfd abcde gfcbed gfec | fcgedb cgb dgebacf gc fgaebd cg bdaec gdafb agbcfd gdcbef bgcad gfac gcb cdgabef | cg cg fdcagb cbg fbegcd cbd adcefb dageb afcb bc aefdc ecdab fgdeca fcdbega | efabcd cedba gadfec cb aecbfdg fbg gf bafeg dbefa fcge gcbea fcaegb dgceab fcbdga | gecf egdcabf bgf bfgea fgeab ca afcebg bdacfeg cfaedg gcfdb baec bfadeg bafgc acf | gebdcfa ecba ca fadegcb dbcfg fgd bdegcaf fgec aegbdf ecdfab fbedc dacgb gdcebf gf | cefg dcbef fcge gbcadfe bdfegc cbegaf gecbf dfcage bdacg ed bedf ced adcbefg gebcd | ed bcgafe cdgba cbgef egadfb cdbfeg cegd fecab cgb gbdefca cg fgcdab egfdb bfceg | gbdfcae bgc cg cgb gcafb gcf dcaebfg ecagb gf abcdeg gaef cafbge fdbac fegbdc | fgae cfgab fg bagce """.splitlines() # example = "acedgfb cdfbe gcdfa fbcad dab cefabd cdfgeb eafb cagedb ab | cdfeb fcadb cdfeb cdbaf".splitlines() """ 0: 1: 2: 3: 4: aaaa .... aaaa aaaa .... b c . c . c . c b c b c . c . c . c b c .... .... dddd dddd dddd e f . f e . . f . f e f . f e . . f . f gggg .... gggg gggg .... 5: 6: 7: 8: 9: aaaa aaaa aaaa aaaa aaaa b . b . . c b c b c b . b . . c b c b c dddd dddd .... dddd dddd . f e f . f e f . f . f e f . f e f . f gggg gggg .... gggg gggg """ # src = example src = [r.strip() for r in src if r.strip()] dc = [6, 2, 5, 5, 4, 5, 6, 3, 7, 6] # 0 is a subset of 8 # *1 is a subset of 0, 3, 4, 7, 8, 9 ^[2, 5, 6] # 2 is a subset of 8 # 3 is a subset of 8, 9 # *4 is a subset of 8, 9 # 5 is a subset of 8, 9 # 6 is a subset of 8 # *7 is a subset of 0, 3, 8, 9 ^[1, 2, 4, 5, 6] # *8 is a subset of # 9 is a subset of 8 part1 = 0 def collapse(digits): # clean up where we only have a "new" single guess for v in digits.values(): if len(v) == 1: for i, j in digits.items(): if len(j) == 1: continue j.difference_update(v) # cleanup where a digit has multiple guesses, but one of the guesses only appears once guesses = [0 for _ in range(10)] for d in digits.values(): for v in d: guesses[v] += 1 for gi, c in enumerate(guesses): if c > 1: continue for i, j in digits.items(): if gi in j: j.difference_update(j.difference({gi})) return digits def get_choices(digits, idx): choices = [] for k, v in digits.items(): if idx in v: choices.append(k) return choices total = 0 for line in src: scram, outp = line.split(" | ") scram = [frozenset(x) for x in scram.split()] outp = [frozenset(x) for x in outp.split()] for d in outp: if len(d) in (2, 4, 3, 7): part1 += 1 digits = {} for d in scram: if len(d) == 2: digits[d] = {1} one = d elif len(d) == 4: digits[d] = {4} four = d elif len(d) == 3: digits[d] = {7} elif len(d) == 7: digits[d] = {8} elif len(d) == 6: digits[d] = {0, 6, 9} elif len(d) == 5: digits[d] = {2, 3, 5} else: assert "wut" # reduce based on if it is a subset of 1 for d in scram: if one.issubset(d): digits[d].difference_update({2, 5, 6}) # four must be a subset of 9 for c in get_choices(digits, 9): if four.issubset(c): digits[c] = {9} nine = c # five must be a subset of nine for c in get_choices(digits, 5): if c.issubset(nine): digits[c] = {5} digits = collapse(digits) c = "" for d in outp: c += str(list(digits[d])[0]) total += int(c) print("part1:", part1) print("part1:", total)
26.51875
111
0.525572
import os import sys import random import numpy as np src = open("input.txt", "r").readlines() example = """ be cfbegad cbdgef fgaecd cgeb fdcge agebfd fecdb fabcd edb | fdgacbe cefdb cefbgd gcbe edbfga begcd cbg gc gcadebf fbgde acbgfd abcde gfcbed gfec | fcgedb cgb dgebacf gc fgaebd cg bdaec gdafb agbcfd gdcbef bgcad gfac gcb cdgabef | cg cg fdcagb cbg fbegcd cbd adcefb dageb afcb bc aefdc ecdab fgdeca fcdbega | efabcd cedba gadfec cb aecbfdg fbg gf bafeg dbefa fcge gcbea fcaegb dgceab fcbdga | gecf egdcabf bgf bfgea fgeab ca afcebg bdacfeg cfaedg gcfdb baec bfadeg bafgc acf | gebdcfa ecba ca fadegcb dbcfg fgd bdegcaf fgec aegbdf ecdfab fbedc dacgb gdcebf gf | cefg dcbef fcge gbcadfe bdfegc cbegaf gecbf dfcage bdacg ed bedf ced adcbefg gebcd | ed bcgafe cdgba cbgef egadfb cdbfeg cegd fecab cgb gbdefca cg fgcdab egfdb bfceg | gbdfcae bgc cg cgb gcafb gcf dcaebfg ecagb gf abcdeg gaef cafbge fdbac fegbdc | fgae cfgab fg bagce """.splitlines() src = [r.strip() for r in src if r.strip()] dc = [6, 2, 5, 5, 4, 5, 6, 3, 7, 6] part1 = 0 def collapse(digits): for v in digits.values(): if len(v) == 1: for i, j in digits.items(): if len(j) == 1: continue j.difference_update(v) guesses = [0 for _ in range(10)] for d in digits.values(): for v in d: guesses[v] += 1 for gi, c in enumerate(guesses): if c > 1: continue for i, j in digits.items(): if gi in j: j.difference_update(j.difference({gi})) return digits def get_choices(digits, idx): choices = [] for k, v in digits.items(): if idx in v: choices.append(k) return choices total = 0 for line in src: scram, outp = line.split(" | ") scram = [frozenset(x) for x in scram.split()] outp = [frozenset(x) for x in outp.split()] for d in outp: if len(d) in (2, 4, 3, 7): part1 += 1 digits = {} for d in scram: if len(d) == 2: digits[d] = {1} one = d elif len(d) == 4: digits[d] = {4} four = d elif len(d) == 3: digits[d] = {7} elif len(d) == 7: digits[d] = {8} elif len(d) == 6: digits[d] = {0, 6, 9} elif len(d) == 5: digits[d] = {2, 3, 5} else: assert "wut" for d in scram: if one.issubset(d): digits[d].difference_update({2, 5, 6}) for c in get_choices(digits, 9): if four.issubset(c): digits[c] = {9} nine = c for c in get_choices(digits, 5): if c.issubset(nine): digits[c] = {5} digits = collapse(digits) c = "" for d in outp: c += str(list(digits[d])[0]) total += int(c) print("part1:", part1) print("part1:", total)
true
true
f70eb2e3a9de09a2e1bdb89d0cfb67ccb56d4e41
7,146
py
Python
Tests/GUI/DMachineSetup/test_DMachineSetup.py
Superomeg4/pyleecan
2b695b5f39e77475a07aa0ea89489fb0a9659337
[ "Apache-2.0" ]
null
null
null
Tests/GUI/DMachineSetup/test_DMachineSetup.py
Superomeg4/pyleecan
2b695b5f39e77475a07aa0ea89489fb0a9659337
[ "Apache-2.0" ]
null
null
null
Tests/GUI/DMachineSetup/test_DMachineSetup.py
Superomeg4/pyleecan
2b695b5f39e77475a07aa0ea89489fb0a9659337
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- """ @date Created on Thu May 18 14:35:34 2017 @copyright (C) 2015-2016 EOMYS ENGINEERING. @author: pierre_b """ from os.path import join, isfile from os import remove import sys from unittest import TestCase from ddt import ddt, data import mock # for unittest of raw_input from PyQt5 import QtWidgets from pyleecan.Classes.MachineSyRM import MachineSyRM from pyleecan.Classes.MachineIPMSM import MachineIPMSM from pyleecan.Classes.MachineDFIM import MachineDFIM from pyleecan.Classes.MachineSCIM import MachineSCIM from pyleecan.Classes.MachineSIPMSM import MachineSIPMSM from pyleecan.Classes.MachineWRSM import MachineWRSM from pyleecan.Classes.MachineSRM import MachineSRM from pyleecan.GUI.Dialog.DMachineSetup.DMachineSetup import DMachineSetup from pyleecan.Tests import save_gui_path as save_path from pyleecan.GUI.Dialog.DMachineSetup.SMachineType.SMachineType import SMachineType from pyleecan.GUI.Dialog.DMachineSetup.SMagnet.SMagnet import SMagnet from pyleecan.GUI.Dialog.DMachineSetup.SWindParam.SWindParam import SWindParam from pyleecan.GUI.Dialog.DMachineSetup.SWindCond.SWindCond import SWindCond from pyleecan.GUI.Dialog.DMachineSetup.SBar.SBar import SBar from pyleecan.GUI.Dialog.DMachineSetup.SWSlot.SWSlot import SWSlot from pyleecan.GUI.Dialog.DMachineSetup.SMHoleMag.SMHoleMag import SMHoleMag import matplotlib.pyplot as plt from pyleecan.Tests import DATA_DIR load_test = list() load_test.append( # 1 {"type": "SCIM", "index": 0, "name": "SCIM_001", "p": 1, "count": 10} ) load_test.append( # 2 {"type": "DFIM", "index": 1, "name": "DFIM_001", "p": 2, "count": 12} ) load_test.append( # 3 {"type": "SyRM", "index": 2, "name": "SynRM_001", "p": 2, "count": 9} ) load_test.append( # 4 {"type": "SPMSM", "index": 3, "name": "SPMSM_001", "p": 4, "count": 9} ) load_test.append( # 5 {"type": "SIPMSM", "index": 4, "name": "SIPMSM_008", "p": 4, "count": 9} ) load_test.append( # 6 {"type": "IPMSM", "index": 5, "name": "machine_IPMSM_A", "p": 5, "count": 9} ) load_test.append( # 7 {"type": "WRSM", "index": 6, "name": "WRSM_001", "p": 6, "count": 12} ) load_test.append( # 8 {"type": "SRM", "index": 7, "name": "SRM_test_load", "p": 10, "count": 9} ) from PyQt5.QtCore import Qt ENABLE_ITEM = Qt.ItemIsSelectable | Qt.ItemIsEnabled @ddt class test_DMachineSetup(TestCase): """Test that the widget DMachineSetup behave like it should""" def setUp(self): """Run at the begining of every test to setup the gui""" self.widget = DMachineSetup(matlib_path="./MaterialData") @classmethod def setUpClass(cls): """Start the app for the test""" print("\nStart Test DMachineSetup") cls.app = QtWidgets.QApplication(sys.argv) @classmethod def tearDownClass(cls): """Exit the app after the test""" cls.app.quit() @data(*load_test) def test_load(self, test_dict): """Check that you can load a machine """ return_value = ( join(join(DATA_DIR, "Load_GUI"), test_dict["name"] + ".json"), "Json (*.json)", ) with mock.patch( "PyQt5.QtWidgets.QFileDialog.getOpenFileName", return_value=return_value ): # To trigger the slot self.widget.b_load.clicked.emit(True) # To remember to update when adding a new machine type self.assertEqual(self.widget.w_step.c_type.count(), 8) # Check load MachineType self.assertEqual(type(self.widget.w_step), SMachineType) self.assertEqual(self.widget.w_step.c_type.currentIndex(), test_dict["index"]) self.assertEqual(self.widget.w_step.c_type.currentText(), test_dict["type"]) self.assertEqual(self.widget.w_step.si_p.value(), test_dict["p"]) self.assertEqual(self.widget.w_step.le_name.text(), test_dict["name"]) # Check that the nav_step is correct self.assertEqual(self.widget.nav_step.count(), test_dict["count"]) def test_set_save_machine_type(self): """Check that the Widget allow to change the machine type and save """ # Check that all the machine type are available self.assertEqual(self.widget.w_step.c_type.count(), 8) # DFIM self.widget.w_step.c_type.setCurrentIndex(1) self.assertEqual(self.widget.w_step.c_type.currentText(), "DFIM") self.assertEqual(type(self.widget.machine), MachineDFIM) save_function(self, self.widget, "test_dfim_save") # SyRM self.widget.w_step.c_type.setCurrentIndex(2) self.assertEqual(self.widget.w_step.c_type.currentText(), "SyRM") self.assertEqual(type(self.widget.machine), MachineSyRM) save_function(self, self.widget, "test_syrm_save") # SPMSM self.widget.w_step.c_type.setCurrentIndex(3) self.assertEqual(self.widget.w_step.c_type.currentText(), "SPMSM") self.assertEqual(type(self.widget.machine), MachineSIPMSM) save_function(self, self.widget, "test_spmsm_save") # SIPMSM self.widget.w_step.c_type.setCurrentIndex(4) self.assertEqual(self.widget.w_step.c_type.currentText(), "SIPMSM") self.assertEqual(type(self.widget.machine), MachineSIPMSM) save_function(self, self.widget, "test_sipmsm_save") # IPMSM self.widget.w_step.c_type.setCurrentIndex(5) self.assertEqual(self.widget.w_step.c_type.currentText(), "IPMSM") self.assertEqual(type(self.widget.machine), MachineIPMSM) save_function(self, self.widget, "test_ipmsm_save") # WRSM self.widget.w_step.c_type.setCurrentIndex(6) self.assertEqual(self.widget.w_step.c_type.currentText(), "WRSM") self.assertEqual(type(self.widget.machine), MachineWRSM) save_function(self, self.widget, "test_wrsm_save") # SRM self.widget.w_step.c_type.setCurrentIndex(7) self.assertEqual(self.widget.w_step.c_type.currentText(), "SRM") self.assertEqual(type(self.widget.machine), MachineSRM) save_function(self, self.widget, "test_srm_save") # SCIM self.widget.w_step.c_type.setCurrentIndex(0) self.assertEqual(self.widget.w_step.c_type.currentText(), "SCIM") self.assertEqual(type(self.widget.machine), MachineSCIM) def save_function(self, widget, file_name): """Function to save a machine from the GUI """ file_path = join(save_path, file_name + ".json") # Check that the file didn't already exist if isfile(file_path): remove(file_path) self.assertFalse(isfile(file_path)) return_value = (file_path, "Json (*.json)") with mock.patch( "PyQt5.QtWidgets.QFileDialog.getSaveFileName", return_value=return_value ): # To trigger the slot widget.b_save.clicked.emit(True) # Check that the file now exist => delete for next test self.assertTrue(isfile(file_path)) remove(file_path) # Check that the GUI have been updated self.assertEqual(type(widget.w_step), SMachineType) self.assertEqual(widget.w_step.le_name.text(), file_name)
39.480663
86
0.687378
from os.path import join, isfile from os import remove import sys from unittest import TestCase from ddt import ddt, data import mock from PyQt5 import QtWidgets from pyleecan.Classes.MachineSyRM import MachineSyRM from pyleecan.Classes.MachineIPMSM import MachineIPMSM from pyleecan.Classes.MachineDFIM import MachineDFIM from pyleecan.Classes.MachineSCIM import MachineSCIM from pyleecan.Classes.MachineSIPMSM import MachineSIPMSM from pyleecan.Classes.MachineWRSM import MachineWRSM from pyleecan.Classes.MachineSRM import MachineSRM from pyleecan.GUI.Dialog.DMachineSetup.DMachineSetup import DMachineSetup from pyleecan.Tests import save_gui_path as save_path from pyleecan.GUI.Dialog.DMachineSetup.SMachineType.SMachineType import SMachineType from pyleecan.GUI.Dialog.DMachineSetup.SMagnet.SMagnet import SMagnet from pyleecan.GUI.Dialog.DMachineSetup.SWindParam.SWindParam import SWindParam from pyleecan.GUI.Dialog.DMachineSetup.SWindCond.SWindCond import SWindCond from pyleecan.GUI.Dialog.DMachineSetup.SBar.SBar import SBar from pyleecan.GUI.Dialog.DMachineSetup.SWSlot.SWSlot import SWSlot from pyleecan.GUI.Dialog.DMachineSetup.SMHoleMag.SMHoleMag import SMHoleMag import matplotlib.pyplot as plt from pyleecan.Tests import DATA_DIR load_test = list() load_test.append( {"type": "SCIM", "index": 0, "name": "SCIM_001", "p": 1, "count": 10} ) load_test.append( {"type": "DFIM", "index": 1, "name": "DFIM_001", "p": 2, "count": 12} ) load_test.append( {"type": "SyRM", "index": 2, "name": "SynRM_001", "p": 2, "count": 9} ) load_test.append( {"type": "SPMSM", "index": 3, "name": "SPMSM_001", "p": 4, "count": 9} ) load_test.append( {"type": "SIPMSM", "index": 4, "name": "SIPMSM_008", "p": 4, "count": 9} ) load_test.append( {"type": "IPMSM", "index": 5, "name": "machine_IPMSM_A", "p": 5, "count": 9} ) load_test.append( {"type": "WRSM", "index": 6, "name": "WRSM_001", "p": 6, "count": 12} ) load_test.append( {"type": "SRM", "index": 7, "name": "SRM_test_load", "p": 10, "count": 9} ) from PyQt5.QtCore import Qt ENABLE_ITEM = Qt.ItemIsSelectable | Qt.ItemIsEnabled @ddt class test_DMachineSetup(TestCase): def setUp(self): self.widget = DMachineSetup(matlib_path="./MaterialData") @classmethod def setUpClass(cls): print("\nStart Test DMachineSetup") cls.app = QtWidgets.QApplication(sys.argv) @classmethod def tearDownClass(cls): cls.app.quit() @data(*load_test) def test_load(self, test_dict): return_value = ( join(join(DATA_DIR, "Load_GUI"), test_dict["name"] + ".json"), "Json (*.json)", ) with mock.patch( "PyQt5.QtWidgets.QFileDialog.getOpenFileName", return_value=return_value ): self.widget.b_load.clicked.emit(True) self.assertEqual(self.widget.w_step.c_type.count(), 8) self.assertEqual(type(self.widget.w_step), SMachineType) self.assertEqual(self.widget.w_step.c_type.currentIndex(), test_dict["index"]) self.assertEqual(self.widget.w_step.c_type.currentText(), test_dict["type"]) self.assertEqual(self.widget.w_step.si_p.value(), test_dict["p"]) self.assertEqual(self.widget.w_step.le_name.text(), test_dict["name"]) self.assertEqual(self.widget.nav_step.count(), test_dict["count"]) def test_set_save_machine_type(self): self.assertEqual(self.widget.w_step.c_type.count(), 8) self.widget.w_step.c_type.setCurrentIndex(1) self.assertEqual(self.widget.w_step.c_type.currentText(), "DFIM") self.assertEqual(type(self.widget.machine), MachineDFIM) save_function(self, self.widget, "test_dfim_save") self.widget.w_step.c_type.setCurrentIndex(2) self.assertEqual(self.widget.w_step.c_type.currentText(), "SyRM") self.assertEqual(type(self.widget.machine), MachineSyRM) save_function(self, self.widget, "test_syrm_save") self.widget.w_step.c_type.setCurrentIndex(3) self.assertEqual(self.widget.w_step.c_type.currentText(), "SPMSM") self.assertEqual(type(self.widget.machine), MachineSIPMSM) save_function(self, self.widget, "test_spmsm_save") self.widget.w_step.c_type.setCurrentIndex(4) self.assertEqual(self.widget.w_step.c_type.currentText(), "SIPMSM") self.assertEqual(type(self.widget.machine), MachineSIPMSM) save_function(self, self.widget, "test_sipmsm_save") self.widget.w_step.c_type.setCurrentIndex(5) self.assertEqual(self.widget.w_step.c_type.currentText(), "IPMSM") self.assertEqual(type(self.widget.machine), MachineIPMSM) save_function(self, self.widget, "test_ipmsm_save") self.widget.w_step.c_type.setCurrentIndex(6) self.assertEqual(self.widget.w_step.c_type.currentText(), "WRSM") self.assertEqual(type(self.widget.machine), MachineWRSM) save_function(self, self.widget, "test_wrsm_save") self.widget.w_step.c_type.setCurrentIndex(7) self.assertEqual(self.widget.w_step.c_type.currentText(), "SRM") self.assertEqual(type(self.widget.machine), MachineSRM) save_function(self, self.widget, "test_srm_save") self.widget.w_step.c_type.setCurrentIndex(0) self.assertEqual(self.widget.w_step.c_type.currentText(), "SCIM") self.assertEqual(type(self.widget.machine), MachineSCIM) def save_function(self, widget, file_name): file_path = join(save_path, file_name + ".json") if isfile(file_path): remove(file_path) self.assertFalse(isfile(file_path)) return_value = (file_path, "Json (*.json)") with mock.patch( "PyQt5.QtWidgets.QFileDialog.getSaveFileName", return_value=return_value ): # To trigger the slot widget.b_save.clicked.emit(True) # Check that the file now exist => delete for next test self.assertTrue(isfile(file_path)) remove(file_path) # Check that the GUI have been updated self.assertEqual(type(widget.w_step), SMachineType) self.assertEqual(widget.w_step.le_name.text(), file_name)
true
true
f70eb2f3401ca84927cd6ad2318eeb8439e46c72
2,814
py
Python
core/queue/views.py
lottspot/prevention-point
e4d5eaa437c3e979e8585bdada4efd33e995e39e
[ "MIT" ]
35
2019-03-12T23:59:10.000Z
2021-04-05T15:07:38.000Z
core/queue/views.py
lottspot/prevention-point
e4d5eaa437c3e979e8585bdada4efd33e995e39e
[ "MIT" ]
365
2019-03-12T23:40:39.000Z
2022-02-10T11:07:26.000Z
core/queue/views.py
lottspot/prevention-point
e4d5eaa437c3e979e8585bdada4efd33e995e39e
[ "MIT" ]
20
2019-03-12T23:36:25.000Z
2021-12-30T00:05:42.000Z
import datetime from rest_framework import viewsets, status from rest_framework.response import Response from rest_framework.permissions import IsAuthenticated from core.permissions import DjangoModelPermissions from core.visits.serializer import PopulatedVisitSerializer from core.models import Visit, FrontDeskEvent, FrontDeskEventType from core.front_desk_events.serializer import FrontDeskEventForQueueSerializer from django.contrib.auth.models import User class QueueViewSet(viewsets.ViewSet): """ API endpoint that displays the queue uses regular ViewSet to be able to display adjacent model responses in one view, hence the permission classes being repeated here instead of using viewsets.py prototype """ # DjangoModelPermissions requires a queryset to function, # the next line is what the docs suggest as a 'sentinel queryset' queryset= FrontDeskEvent.objects.none() permission_classes = [DjangoModelPermissions, IsAuthenticated] def retrieve(self, request, program_id=None): """ retrieve most recent front desk event for each visit that is happening today, filtered by program """ # filter by visits that are happening today in a certain program visits_queryset = ( Visit.objects.select_related("participant", "program") .filter( program=program_id, created_at__date=datetime.date.today(), ) .order_by("urgency", "-created_at") ) todays_visit_data = PopulatedVisitSerializer( visits_queryset, many=True, context={"request": request} ).data active_visits_queue = [] front_desk_events = FrontDeskEvent.objects.select_related("visit").filter( visit__in=[dict(x)["id"] for x in todays_visit_data] ).order_by("-created_at").values("id", "visit", "event_type", "created_at") # for each visit, get the most recent front desk event, to glean current visit status for visit in todays_visit_data: events = list( filter(lambda x: x.get("visit") is visit.get("id"), front_desk_events) ) if events: event = events[0] event_type = event.get("event_type") if event_type in [ FrontDeskEventType.ARRIVED.name, FrontDeskEventType.STEPPED_OUT.name, FrontDeskEventType.CAME_BACK.name, ]: # if most recent front desk event is an 'active' status add it to visit object visit["status"] = event # then add it to the 'active visits queue' active_visits_queue.append(visit) return Response(active_visits_queue)
40.2
98
0.657783
import datetime from rest_framework import viewsets, status from rest_framework.response import Response from rest_framework.permissions import IsAuthenticated from core.permissions import DjangoModelPermissions from core.visits.serializer import PopulatedVisitSerializer from core.models import Visit, FrontDeskEvent, FrontDeskEventType from core.front_desk_events.serializer import FrontDeskEventForQueueSerializer from django.contrib.auth.models import User class QueueViewSet(viewsets.ViewSet): queryset= FrontDeskEvent.objects.none() permission_classes = [DjangoModelPermissions, IsAuthenticated] def retrieve(self, request, program_id=None): visits_queryset = ( Visit.objects.select_related("participant", "program") .filter( program=program_id, created_at__date=datetime.date.today(), ) .order_by("urgency", "-created_at") ) todays_visit_data = PopulatedVisitSerializer( visits_queryset, many=True, context={"request": request} ).data active_visits_queue = [] front_desk_events = FrontDeskEvent.objects.select_related("visit").filter( visit__in=[dict(x)["id"] for x in todays_visit_data] ).order_by("-created_at").values("id", "visit", "event_type", "created_at") for visit in todays_visit_data: events = list( filter(lambda x: x.get("visit") is visit.get("id"), front_desk_events) ) if events: event = events[0] event_type = event.get("event_type") if event_type in [ FrontDeskEventType.ARRIVED.name, FrontDeskEventType.STEPPED_OUT.name, FrontDeskEventType.CAME_BACK.name, ]: visit["status"] = event active_visits_queue.append(visit) return Response(active_visits_queue)
true
true
f70eb3a29c6a510eca3947f79787ed0abd7f6655
732
py
Python
day01_Sonar_Sweep/day01.py
anolivei/advent_of_code_2021
1eac988d37bf754bbee68fad2e927914351a5a2b
[ "MIT" ]
null
null
null
day01_Sonar_Sweep/day01.py
anolivei/advent_of_code_2021
1eac988d37bf754bbee68fad2e927914351a5a2b
[ "MIT" ]
null
null
null
day01_Sonar_Sweep/day01.py
anolivei/advent_of_code_2021
1eac988d37bf754bbee68fad2e927914351a5a2b
[ "MIT" ]
null
null
null
def open_input(): with open("input.txt") as fd: array = fd.read().splitlines() array = list(map(int, array)) return array def part_one(array): lenght = len(array) increased = 0 for i in range(0, lenght - 1): if array[i] < array[i + 1]: increased += 1 print("part one:", increased) def part_two(array): lenght = len(array) increased = 0 for i in range(0, lenght - 3): sum1 = array[i] + array[i + 1] + array[i + 2] sum2 = array[i + 1] + array[i + 2] + array[i + 3] if sum1 < sum2: increased += 1 print("part two:", increased) if (__name__ == "__main__"): array = open_input() part_one(array) part_two(array)
22.875
57
0.543716
def open_input(): with open("input.txt") as fd: array = fd.read().splitlines() array = list(map(int, array)) return array def part_one(array): lenght = len(array) increased = 0 for i in range(0, lenght - 1): if array[i] < array[i + 1]: increased += 1 print("part one:", increased) def part_two(array): lenght = len(array) increased = 0 for i in range(0, lenght - 3): sum1 = array[i] + array[i + 1] + array[i + 2] sum2 = array[i + 1] + array[i + 2] + array[i + 3] if sum1 < sum2: increased += 1 print("part two:", increased) if (__name__ == "__main__"): array = open_input() part_one(array) part_two(array)
true
true
f70eb3de404bce1c3ea4c27a5b70915529f3a386
6,153
py
Python
sdk/python/pulumi_azure_native/customproviders/v20180901preview/get_custom_resource_provider.py
sebtelko/pulumi-azure-native
711ec021b5c73da05611c56c8a35adb0ce3244e4
[ "Apache-2.0" ]
null
null
null
sdk/python/pulumi_azure_native/customproviders/v20180901preview/get_custom_resource_provider.py
sebtelko/pulumi-azure-native
711ec021b5c73da05611c56c8a35adb0ce3244e4
[ "Apache-2.0" ]
null
null
null
sdk/python/pulumi_azure_native/customproviders/v20180901preview/get_custom_resource_provider.py
sebtelko/pulumi-azure-native
711ec021b5c73da05611c56c8a35adb0ce3244e4
[ "Apache-2.0" ]
null
null
null
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from ... import _utilities from . import outputs __all__ = [ 'GetCustomResourceProviderResult', 'AwaitableGetCustomResourceProviderResult', 'get_custom_resource_provider', ] @pulumi.output_type class GetCustomResourceProviderResult: """ A manifest file that defines the custom resource provider resources. """ def __init__(__self__, actions=None, id=None, location=None, name=None, provisioning_state=None, resource_types=None, tags=None, type=None, validations=None): if actions and not isinstance(actions, list): raise TypeError("Expected argument 'actions' to be a list") pulumi.set(__self__, "actions", actions) if id and not isinstance(id, str): raise TypeError("Expected argument 'id' to be a str") pulumi.set(__self__, "id", id) if location and not isinstance(location, str): raise TypeError("Expected argument 'location' to be a str") pulumi.set(__self__, "location", location) if name and not isinstance(name, str): raise TypeError("Expected argument 'name' to be a str") pulumi.set(__self__, "name", name) if provisioning_state and not isinstance(provisioning_state, str): raise TypeError("Expected argument 'provisioning_state' to be a str") pulumi.set(__self__, "provisioning_state", provisioning_state) if resource_types and not isinstance(resource_types, list): raise TypeError("Expected argument 'resource_types' to be a list") pulumi.set(__self__, "resource_types", resource_types) if tags and not isinstance(tags, dict): raise TypeError("Expected argument 'tags' to be a dict") pulumi.set(__self__, "tags", tags) if type and not isinstance(type, str): raise TypeError("Expected argument 'type' to be a str") pulumi.set(__self__, "type", type) if validations and not isinstance(validations, list): raise TypeError("Expected argument 'validations' to be a list") pulumi.set(__self__, "validations", validations) @property @pulumi.getter def actions(self) -> Optional[Sequence['outputs.CustomRPActionRouteDefinitionResponse']]: """ A list of actions that the custom resource provider implements. """ return pulumi.get(self, "actions") @property @pulumi.getter def id(self) -> str: """ Resource Id """ return pulumi.get(self, "id") @property @pulumi.getter def location(self) -> str: """ Resource location """ return pulumi.get(self, "location") @property @pulumi.getter def name(self) -> str: """ Resource name """ return pulumi.get(self, "name") @property @pulumi.getter(name="provisioningState") def provisioning_state(self) -> str: """ The provisioning state of the resource provider. """ return pulumi.get(self, "provisioning_state") @property @pulumi.getter(name="resourceTypes") def resource_types(self) -> Optional[Sequence['outputs.CustomRPResourceTypeRouteDefinitionResponse']]: """ A list of resource types that the custom resource provider implements. """ return pulumi.get(self, "resource_types") @property @pulumi.getter def tags(self) -> Optional[Mapping[str, str]]: """ Resource tags """ return pulumi.get(self, "tags") @property @pulumi.getter def type(self) -> str: """ Resource type """ return pulumi.get(self, "type") @property @pulumi.getter def validations(self) -> Optional[Sequence['outputs.CustomRPValidationsResponse']]: """ A list of validations to run on the custom resource provider's requests. """ return pulumi.get(self, "validations") class AwaitableGetCustomResourceProviderResult(GetCustomResourceProviderResult): # pylint: disable=using-constant-test def __await__(self): if False: yield self return GetCustomResourceProviderResult( actions=self.actions, id=self.id, location=self.location, name=self.name, provisioning_state=self.provisioning_state, resource_types=self.resource_types, tags=self.tags, type=self.type, validations=self.validations) def get_custom_resource_provider(resource_group_name: Optional[str] = None, resource_provider_name: Optional[str] = None, opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetCustomResourceProviderResult: """ A manifest file that defines the custom resource provider resources. :param str resource_group_name: The name of the resource group. :param str resource_provider_name: The name of the resource provider. """ __args__ = dict() __args__['resourceGroupName'] = resource_group_name __args__['resourceProviderName'] = resource_provider_name if opts is None: opts = pulumi.InvokeOptions() if opts.version is None: opts.version = _utilities.get_version() __ret__ = pulumi.runtime.invoke('azure-native:customproviders/v20180901preview:getCustomResourceProvider', __args__, opts=opts, typ=GetCustomResourceProviderResult).value return AwaitableGetCustomResourceProviderResult( actions=__ret__.actions, id=__ret__.id, location=__ret__.location, name=__ret__.name, provisioning_state=__ret__.provisioning_state, resource_types=__ret__.resource_types, tags=__ret__.tags, type=__ret__.type, validations=__ret__.validations)
35.982456
174
0.653827
import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from ... import _utilities from . import outputs __all__ = [ 'GetCustomResourceProviderResult', 'AwaitableGetCustomResourceProviderResult', 'get_custom_resource_provider', ] @pulumi.output_type class GetCustomResourceProviderResult: def __init__(__self__, actions=None, id=None, location=None, name=None, provisioning_state=None, resource_types=None, tags=None, type=None, validations=None): if actions and not isinstance(actions, list): raise TypeError("Expected argument 'actions' to be a list") pulumi.set(__self__, "actions", actions) if id and not isinstance(id, str): raise TypeError("Expected argument 'id' to be a str") pulumi.set(__self__, "id", id) if location and not isinstance(location, str): raise TypeError("Expected argument 'location' to be a str") pulumi.set(__self__, "location", location) if name and not isinstance(name, str): raise TypeError("Expected argument 'name' to be a str") pulumi.set(__self__, "name", name) if provisioning_state and not isinstance(provisioning_state, str): raise TypeError("Expected argument 'provisioning_state' to be a str") pulumi.set(__self__, "provisioning_state", provisioning_state) if resource_types and not isinstance(resource_types, list): raise TypeError("Expected argument 'resource_types' to be a list") pulumi.set(__self__, "resource_types", resource_types) if tags and not isinstance(tags, dict): raise TypeError("Expected argument 'tags' to be a dict") pulumi.set(__self__, "tags", tags) if type and not isinstance(type, str): raise TypeError("Expected argument 'type' to be a str") pulumi.set(__self__, "type", type) if validations and not isinstance(validations, list): raise TypeError("Expected argument 'validations' to be a list") pulumi.set(__self__, "validations", validations) @property @pulumi.getter def actions(self) -> Optional[Sequence['outputs.CustomRPActionRouteDefinitionResponse']]: return pulumi.get(self, "actions") @property @pulumi.getter def id(self) -> str: return pulumi.get(self, "id") @property @pulumi.getter def location(self) -> str: return pulumi.get(self, "location") @property @pulumi.getter def name(self) -> str: return pulumi.get(self, "name") @property @pulumi.getter(name="provisioningState") def provisioning_state(self) -> str: return pulumi.get(self, "provisioning_state") @property @pulumi.getter(name="resourceTypes") def resource_types(self) -> Optional[Sequence['outputs.CustomRPResourceTypeRouteDefinitionResponse']]: return pulumi.get(self, "resource_types") @property @pulumi.getter def tags(self) -> Optional[Mapping[str, str]]: return pulumi.get(self, "tags") @property @pulumi.getter def type(self) -> str: return pulumi.get(self, "type") @property @pulumi.getter def validations(self) -> Optional[Sequence['outputs.CustomRPValidationsResponse']]: return pulumi.get(self, "validations") class AwaitableGetCustomResourceProviderResult(GetCustomResourceProviderResult): # pylint: disable=using-constant-test def __await__(self): if False: yield self return GetCustomResourceProviderResult( actions=self.actions, id=self.id, location=self.location, name=self.name, provisioning_state=self.provisioning_state, resource_types=self.resource_types, tags=self.tags, type=self.type, validations=self.validations) def get_custom_resource_provider(resource_group_name: Optional[str] = None, resource_provider_name: Optional[str] = None, opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetCustomResourceProviderResult: __args__ = dict() __args__['resourceGroupName'] = resource_group_name __args__['resourceProviderName'] = resource_provider_name if opts is None: opts = pulumi.InvokeOptions() if opts.version is None: opts.version = _utilities.get_version() __ret__ = pulumi.runtime.invoke('azure-native:customproviders/v20180901preview:getCustomResourceProvider', __args__, opts=opts, typ=GetCustomResourceProviderResult).value return AwaitableGetCustomResourceProviderResult( actions=__ret__.actions, id=__ret__.id, location=__ret__.location, name=__ret__.name, provisioning_state=__ret__.provisioning_state, resource_types=__ret__.resource_types, tags=__ret__.tags, type=__ret__.type, validations=__ret__.validations)
true
true
f70eb40817dabb9bff5e70b4866289c7586fcf23
1,189
py
Python
plugins/action/gitlab_modwrap.py
sma-de/ansible-collections-gitlab
5da99b04722fc016d3e8589635fcbb3579dcfda2
[ "BSD-3-Clause" ]
null
null
null
plugins/action/gitlab_modwrap.py
sma-de/ansible-collections-gitlab
5da99b04722fc016d3e8589635fcbb3579dcfda2
[ "BSD-3-Clause" ]
null
null
null
plugins/action/gitlab_modwrap.py
sma-de/ansible-collections-gitlab
5da99b04722fc016d3e8589635fcbb3579dcfda2
[ "BSD-3-Clause" ]
null
null
null
from __future__ import (absolute_import, division, print_function) __metaclass__ = type import collections ##from ansible.errors import AnsibleOptionsError, AnsibleModuleError##, AnsibleError ####from ansible.module_utils._text import to_native from ansible.module_utils.six import iteritems, string_types from ansible_collections.smabot.git.plugins.module_utils.plugins.gitlab_action import GitlabBase from ansible_collections.smabot.base.plugins.module_utils.utils.utils import ansible_assert class ActionModule(GitlabBase): def __init__(self, *args, **kwargs): super(ActionModule, self).__init__(*args, **kwargs) self._supports_check_mode = False self._supports_async = False @property def argspec(self): tmp = super(ActionModule, self).argspec tmp.update({ 'modname': (list(string_types)), 'modargs': ([collections.abc.Mapping], {}), }) return tmp def run_specific(self, result): cmdret = self.exec_gitlab_module( self.get_taskparam('modname'), modargs=self.get_taskparam('modargs') ) result.update(cmdret) return result
25.847826
96
0.699748
from __future__ import (absolute_import, division, print_function) __metaclass__ = type import collections ionModule(GitlabBase): def __init__(self, *args, **kwargs): super(ActionModule, self).__init__(*args, **kwargs) self._supports_check_mode = False self._supports_async = False @property def argspec(self): tmp = super(ActionModule, self).argspec tmp.update({ 'modname': (list(string_types)), 'modargs': ([collections.abc.Mapping], {}), }) return tmp def run_specific(self, result): cmdret = self.exec_gitlab_module( self.get_taskparam('modname'), modargs=self.get_taskparam('modargs') ) result.update(cmdret) return result
true
true
f70eb5b9f0829e7189fca574a45e2c0d95713fd9
2,964
py
Python
python/tests/test_oberon.py
tmarkovski/okapi
1ce36d2a82bb0e409a5183cd116d3a9eb474fc9f
[ "Apache-2.0" ]
null
null
null
python/tests/test_oberon.py
tmarkovski/okapi
1ce36d2a82bb0e409a5183cd116d3a9eb474fc9f
[ "Apache-2.0" ]
null
null
null
python/tests/test_oberon.py
tmarkovski/okapi
1ce36d2a82bb0e409a5183cd116d3a9eb474fc9f
[ "Apache-2.0" ]
null
null
null
import unittest from okapi.proto.okapi.security.v1 import CreateOberonKeyRequest, CreateOberonTokenRequest, CreateOberonProofRequest, \ VerifyOberonProofRequest, UnBlindOberonTokenRequest, BlindOberonTokenRequest from okapi.wrapper import Oberon class KeyTests(unittest.TestCase): def test_oberon_demo(self): key = Oberon.create_key(CreateOberonKeyRequest()) data = bytes("alice", "utf8") nonce = bytes("1234", "utf8") token = Oberon.create_token(CreateOberonTokenRequest(data=data, sk=key.sk)) proof = Oberon.create_proof(CreateOberonProofRequest(data=data, nonce=nonce, token=token.token)) result = Oberon.verify_proof(VerifyOberonProofRequest(data=data, nonce=nonce, pk=key.pk, proof=proof.proof)) self.assertTrue(result.valid, "Proof should verify") def test_demo_with_binding(self): key = Oberon.create_key(CreateOberonKeyRequest()) data = bytes("alice", "utf8") nonce = bytes("1234", "utf8") issuer_2fa = bytes("issuer code", "utf8") token_request = CreateOberonTokenRequest(data=data, sk=key.sk) token_request.blinding.append(issuer_2fa) blinded_token = Oberon.create_token(token_request) # Holder unblinds the token unblind_request = UnBlindOberonTokenRequest(token=blinded_token.token) unblind_request.blinding.append(issuer_2fa) token = Oberon.unblind_token(unblind_request) # Holder prepares a proof without blinding proof = Oberon.create_proof(CreateOberonProofRequest(data=data, nonce=nonce, token=token.token)) # Verifier verifies the proof result = Oberon.verify_proof(VerifyOberonProofRequest(data=data, nonce=nonce, pk=key.pk, proof=proof.proof)) self.assertTrue(result.valid) # Holder blinds the token with a personal pin user_pin = bytes("0042", "utf8") blind_request = BlindOberonTokenRequest(token=token.token) blind_request.blinding.append(user_pin) user_blinded_token = Oberon.blind_token(blind_request) proof_request = CreateOberonProofRequest(data=data, nonce=nonce, token=user_blinded_token.token) proof_request.blinding.append(user_pin) proof = Oberon.create_proof(proof_request) # Verifier verifies the proof result = Oberon.verify_proof(VerifyOberonProofRequest(data=data, nonce=nonce, pk=key.pk, proof=proof.proof)) self.assertTrue(result.valid) # Bad actor creates a proof with incorrect blinding pin proof_request = CreateOberonProofRequest(data=data, nonce=nonce, token=user_blinded_token.token) proof_request.blinding.append(bytes("invalid pin", "utf8")) proof = Oberon.create_proof(proof_request) # Verifies tries to verify proof, fails result = Oberon.verify_proof(VerifyOberonProofRequest(data=data, nonce=nonce, pk=key.pk, proof=proof.proof)) self.assertFalse(result.valid)
48.590164
119
0.721997
import unittest from okapi.proto.okapi.security.v1 import CreateOberonKeyRequest, CreateOberonTokenRequest, CreateOberonProofRequest, \ VerifyOberonProofRequest, UnBlindOberonTokenRequest, BlindOberonTokenRequest from okapi.wrapper import Oberon class KeyTests(unittest.TestCase): def test_oberon_demo(self): key = Oberon.create_key(CreateOberonKeyRequest()) data = bytes("alice", "utf8") nonce = bytes("1234", "utf8") token = Oberon.create_token(CreateOberonTokenRequest(data=data, sk=key.sk)) proof = Oberon.create_proof(CreateOberonProofRequest(data=data, nonce=nonce, token=token.token)) result = Oberon.verify_proof(VerifyOberonProofRequest(data=data, nonce=nonce, pk=key.pk, proof=proof.proof)) self.assertTrue(result.valid, "Proof should verify") def test_demo_with_binding(self): key = Oberon.create_key(CreateOberonKeyRequest()) data = bytes("alice", "utf8") nonce = bytes("1234", "utf8") issuer_2fa = bytes("issuer code", "utf8") token_request = CreateOberonTokenRequest(data=data, sk=key.sk) token_request.blinding.append(issuer_2fa) blinded_token = Oberon.create_token(token_request) unblind_request = UnBlindOberonTokenRequest(token=blinded_token.token) unblind_request.blinding.append(issuer_2fa) token = Oberon.unblind_token(unblind_request) proof = Oberon.create_proof(CreateOberonProofRequest(data=data, nonce=nonce, token=token.token)) result = Oberon.verify_proof(VerifyOberonProofRequest(data=data, nonce=nonce, pk=key.pk, proof=proof.proof)) self.assertTrue(result.valid) user_pin = bytes("0042", "utf8") blind_request = BlindOberonTokenRequest(token=token.token) blind_request.blinding.append(user_pin) user_blinded_token = Oberon.blind_token(blind_request) proof_request = CreateOberonProofRequest(data=data, nonce=nonce, token=user_blinded_token.token) proof_request.blinding.append(user_pin) proof = Oberon.create_proof(proof_request) result = Oberon.verify_proof(VerifyOberonProofRequest(data=data, nonce=nonce, pk=key.pk, proof=proof.proof)) self.assertTrue(result.valid) proof_request = CreateOberonProofRequest(data=data, nonce=nonce, token=user_blinded_token.token) proof_request.blinding.append(bytes("invalid pin", "utf8")) proof = Oberon.create_proof(proof_request) result = Oberon.verify_proof(VerifyOberonProofRequest(data=data, nonce=nonce, pk=key.pk, proof=proof.proof)) self.assertFalse(result.valid)
true
true
f70eb60e7035a3d1c3d3275fedadb650fd20fe9f
3,030
py
Python
searchmethods/modularGA.py
esnet/hps-rl
8426652e622394a955a44c42201e2204f6bfa0f2
[ "BSD-3-Clause-LBNL" ]
null
null
null
searchmethods/modularGA.py
esnet/hps-rl
8426652e622394a955a44c42201e2204f6bfa0f2
[ "BSD-3-Clause-LBNL" ]
null
null
null
searchmethods/modularGA.py
esnet/hps-rl
8426652e622394a955a44c42201e2204f6bfa0f2
[ "BSD-3-Clause-LBNL" ]
null
null
null
import numpy, random class Individual: def __init__(self,genome, llimits =[], ulimits=[], type=[], LEN = 1,fitness_func = None): if genome is None: self.genome = numpy.zeros(LEN,dtype=float) for gene in range(LEN): if type[gene] == "integer": self.genome[gene] = numpy.random.randint(llimits[gene], ulimits[gene]) else: self.genome[gene] = numpy.random.uniform(llimits[gene], ulimits[gene]) else: self.genome = genome self.fitness = fitness_func(self.genome) def __str__(self): return "".join(str(int(i)) for i in self.genome) def crossover(a, b, fitness): g, h = a.genome.copy(), b.genome.copy() for pt in range(len(g)): if numpy.random.random() < 0.5: g[pt], h[pt] = h[pt], g[pt] return (Individual(genome=g,fitness_func=fitness), Individual(genome=h,fitness_func=fitness)) def mutate(a, mut_prob,fitness): g = a.genome.copy() for pt in range(len(g)): if numpy.random.random() < mut_prob: g[pt] = not g[pt] return Individual(g,fitness_func=fitness) def stats(pop, gen,threshold): best = max(pop, key=lambda x: x.fitness) print("{0} {1:.2f} {2} {3}".format(gen, numpy.mean([i.fitness for i in pop]), best.fitness, str(best))) return (best.fitness >= threshold) def roulette(items, n): total = float(sum(w.fitness for w in items)) i = 0 w, v = items[0].fitness, items[0] while n: x = total * (1 - numpy.random.random() ** (1.0 / n)) total -= x while x > w: x -= w i += 1 w, v = items[i].fitness, items[i] w -= x yield v n -= 1 def tournament(items, n, tsize=5): for i in range(n): candidates = random.sample(items, tsize) yield max(candidates, key=lambda x: x.fitness) def step(pop,cross_prob,mut_prob,fitness): newpop = [] parents = roulette(pop, len(pop) + 1) # one extra for final xover while len(newpop) < len(pop): if numpy.random.random() < cross_prob: newpop.extend(map(mutate, crossover(next(parents), next(parents),fitness=fitness),[mut_prob,mut_prob],[fitness,fitness])) else: newpop.append(mutate(next(parents),mut_prob=mut_prob,fitness=fitness)) return newpop def run(llimit, ulimit, type, GENERATIONS, CROSSOVER_PROB, POPSIZE, LEN, MUTATION_PROB,FITNESS,THRESHOLD): numpy.random.seed(100) pop = [Individual(None,llimit,ulimit,type,LEN,FITNESS) for i in range(POPSIZE)] print(pop) stats(pop, 0, THRESHOLD) for gen in range(1, GENERATIONS): pop = step(pop,CROSSOVER_PROB,MUTATION_PROB,FITNESS) if stats(pop, gen, THRESHOLD): print("Success") llimit = [0.5,1e-6,1e-6,0] ulimit = [1.5,0.1,0.1,3] type = ['real','real','real','integer'] LEN = 4 FITNESS, SUCCESS_THRESHOLD = (numpy.sum, LEN) run(llimit,ulimit,type,100,1,100,4,0.9,FITNESS,10)
34.044944
133
0.59703
import numpy, random class Individual: def __init__(self,genome, llimits =[], ulimits=[], type=[], LEN = 1,fitness_func = None): if genome is None: self.genome = numpy.zeros(LEN,dtype=float) for gene in range(LEN): if type[gene] == "integer": self.genome[gene] = numpy.random.randint(llimits[gene], ulimits[gene]) else: self.genome[gene] = numpy.random.uniform(llimits[gene], ulimits[gene]) else: self.genome = genome self.fitness = fitness_func(self.genome) def __str__(self): return "".join(str(int(i)) for i in self.genome) def crossover(a, b, fitness): g, h = a.genome.copy(), b.genome.copy() for pt in range(len(g)): if numpy.random.random() < 0.5: g[pt], h[pt] = h[pt], g[pt] return (Individual(genome=g,fitness_func=fitness), Individual(genome=h,fitness_func=fitness)) def mutate(a, mut_prob,fitness): g = a.genome.copy() for pt in range(len(g)): if numpy.random.random() < mut_prob: g[pt] = not g[pt] return Individual(g,fitness_func=fitness) def stats(pop, gen,threshold): best = max(pop, key=lambda x: x.fitness) print("{0} {1:.2f} {2} {3}".format(gen, numpy.mean([i.fitness for i in pop]), best.fitness, str(best))) return (best.fitness >= threshold) def roulette(items, n): total = float(sum(w.fitness for w in items)) i = 0 w, v = items[0].fitness, items[0] while n: x = total * (1 - numpy.random.random() ** (1.0 / n)) total -= x while x > w: x -= w i += 1 w, v = items[i].fitness, items[i] w -= x yield v n -= 1 def tournament(items, n, tsize=5): for i in range(n): candidates = random.sample(items, tsize) yield max(candidates, key=lambda x: x.fitness) def step(pop,cross_prob,mut_prob,fitness): newpop = [] parents = roulette(pop, len(pop) + 1) while len(newpop) < len(pop): if numpy.random.random() < cross_prob: newpop.extend(map(mutate, crossover(next(parents), next(parents),fitness=fitness),[mut_prob,mut_prob],[fitness,fitness])) else: newpop.append(mutate(next(parents),mut_prob=mut_prob,fitness=fitness)) return newpop def run(llimit, ulimit, type, GENERATIONS, CROSSOVER_PROB, POPSIZE, LEN, MUTATION_PROB,FITNESS,THRESHOLD): numpy.random.seed(100) pop = [Individual(None,llimit,ulimit,type,LEN,FITNESS) for i in range(POPSIZE)] print(pop) stats(pop, 0, THRESHOLD) for gen in range(1, GENERATIONS): pop = step(pop,CROSSOVER_PROB,MUTATION_PROB,FITNESS) if stats(pop, gen, THRESHOLD): print("Success") llimit = [0.5,1e-6,1e-6,0] ulimit = [1.5,0.1,0.1,3] type = ['real','real','real','integer'] LEN = 4 FITNESS, SUCCESS_THRESHOLD = (numpy.sum, LEN) run(llimit,ulimit,type,100,1,100,4,0.9,FITNESS,10)
true
true
f70eb629524c29581304d4058b938e52261dca03
10,701
py
Python
kubernetes_asyncio/client/models/v1_topology_spread_constraint.py
lsst-sqre/kubernetes_asyncio
f028cc793e3a2c519be6a52a49fb77ff0b014c9b
[ "Apache-2.0" ]
null
null
null
kubernetes_asyncio/client/models/v1_topology_spread_constraint.py
lsst-sqre/kubernetes_asyncio
f028cc793e3a2c519be6a52a49fb77ff0b014c9b
[ "Apache-2.0" ]
null
null
null
kubernetes_asyncio/client/models/v1_topology_spread_constraint.py
lsst-sqre/kubernetes_asyncio
f028cc793e3a2c519be6a52a49fb77ff0b014c9b
[ "Apache-2.0" ]
null
null
null
# coding: utf-8 """ Kubernetes No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator) # noqa: E501 The version of the OpenAPI document: v1.19.15 Generated by: https://openapi-generator.tech """ import pprint import re # noqa: F401 import six from kubernetes_asyncio.client.configuration import Configuration class V1TopologySpreadConstraint(object): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech Do not edit the class manually. """ """ Attributes: openapi_types (dict): The key is attribute name and the value is attribute type. attribute_map (dict): The key is attribute name and the value is json key in definition. """ openapi_types = { 'label_selector': 'V1LabelSelector', 'max_skew': 'int', 'topology_key': 'str', 'when_unsatisfiable': 'str' } attribute_map = { 'label_selector': 'labelSelector', 'max_skew': 'maxSkew', 'topology_key': 'topologyKey', 'when_unsatisfiable': 'whenUnsatisfiable' } def __init__(self, label_selector=None, max_skew=None, topology_key=None, when_unsatisfiable=None, local_vars_configuration=None): # noqa: E501 """V1TopologySpreadConstraint - a model defined in OpenAPI""" # noqa: E501 if local_vars_configuration is None: local_vars_configuration = Configuration() self.local_vars_configuration = local_vars_configuration self._label_selector = None self._max_skew = None self._topology_key = None self._when_unsatisfiable = None self.discriminator = None if label_selector is not None: self.label_selector = label_selector self.max_skew = max_skew self.topology_key = topology_key self.when_unsatisfiable = when_unsatisfiable @property def label_selector(self): """Gets the label_selector of this V1TopologySpreadConstraint. # noqa: E501 :return: The label_selector of this V1TopologySpreadConstraint. # noqa: E501 :rtype: V1LabelSelector """ return self._label_selector @label_selector.setter def label_selector(self, label_selector): """Sets the label_selector of this V1TopologySpreadConstraint. :param label_selector: The label_selector of this V1TopologySpreadConstraint. # noqa: E501 :type: V1LabelSelector """ self._label_selector = label_selector @property def max_skew(self): """Gets the max_skew of this V1TopologySpreadConstraint. # noqa: E501 MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 1/1/0: | zone1 | zone2 | zone3 | | P | P | | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 1/1/1; scheduling it onto zone1(zone2) would make the ActualSkew(2-0) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed. # noqa: E501 :return: The max_skew of this V1TopologySpreadConstraint. # noqa: E501 :rtype: int """ return self._max_skew @max_skew.setter def max_skew(self, max_skew): """Sets the max_skew of this V1TopologySpreadConstraint. MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 1/1/0: | zone1 | zone2 | zone3 | | P | P | | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 1/1/1; scheduling it onto zone1(zone2) would make the ActualSkew(2-0) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed. # noqa: E501 :param max_skew: The max_skew of this V1TopologySpreadConstraint. # noqa: E501 :type: int """ if self.local_vars_configuration.client_side_validation and max_skew is None: # noqa: E501 raise ValueError("Invalid value for `max_skew`, must not be `None`") # noqa: E501 self._max_skew = max_skew @property def topology_key(self): """Gets the topology_key of this V1TopologySpreadConstraint. # noqa: E501 TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each <key, value> as a \"bucket\", and try to put balanced number of pods into each bucket. It's a required field. # noqa: E501 :return: The topology_key of this V1TopologySpreadConstraint. # noqa: E501 :rtype: str """ return self._topology_key @topology_key.setter def topology_key(self, topology_key): """Sets the topology_key of this V1TopologySpreadConstraint. TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each <key, value> as a \"bucket\", and try to put balanced number of pods into each bucket. It's a required field. # noqa: E501 :param topology_key: The topology_key of this V1TopologySpreadConstraint. # noqa: E501 :type: str """ if self.local_vars_configuration.client_side_validation and topology_key is None: # noqa: E501 raise ValueError("Invalid value for `topology_key`, must not be `None`") # noqa: E501 self._topology_key = topology_key @property def when_unsatisfiable(self): """Gets the when_unsatisfiable of this V1TopologySpreadConstraint. # noqa: E501 WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location, but giving higher precedence to topologies that would help reduce the skew. A constraint is considered \"Unsatisfiable\" for an incoming pod if and only if every possible node assigment for that pod would violate \"MaxSkew\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field. # noqa: E501 :return: The when_unsatisfiable of this V1TopologySpreadConstraint. # noqa: E501 :rtype: str """ return self._when_unsatisfiable @when_unsatisfiable.setter def when_unsatisfiable(self, when_unsatisfiable): """Sets the when_unsatisfiable of this V1TopologySpreadConstraint. WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location, but giving higher precedence to topologies that would help reduce the skew. A constraint is considered \"Unsatisfiable\" for an incoming pod if and only if every possible node assigment for that pod would violate \"MaxSkew\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field. # noqa: E501 :param when_unsatisfiable: The when_unsatisfiable of this V1TopologySpreadConstraint. # noqa: E501 :type: str """ if self.local_vars_configuration.client_side_validation and when_unsatisfiable is None: # noqa: E501 raise ValueError("Invalid value for `when_unsatisfiable`, must not be `None`") # noqa: E501 self._when_unsatisfiable = when_unsatisfiable def to_dict(self): """Returns the model properties as a dict""" result = {} for attr, _ in six.iteritems(self.openapi_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value return result def to_str(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict()) def __repr__(self): """For `print` and `pprint`""" return self.to_str() def __eq__(self, other): """Returns true if both objects are equal""" if not isinstance(other, V1TopologySpreadConstraint): return False return self.to_dict() == other.to_dict() def __ne__(self, other): """Returns true if both objects are not equal""" if not isinstance(other, V1TopologySpreadConstraint): return True return self.to_dict() != other.to_dict()
51.447115
965
0.680404
import pprint import re import six from kubernetes_asyncio.client.configuration import Configuration class V1TopologySpreadConstraint(object): openapi_types = { 'label_selector': 'V1LabelSelector', 'max_skew': 'int', 'topology_key': 'str', 'when_unsatisfiable': 'str' } attribute_map = { 'label_selector': 'labelSelector', 'max_skew': 'maxSkew', 'topology_key': 'topologyKey', 'when_unsatisfiable': 'whenUnsatisfiable' } def __init__(self, label_selector=None, max_skew=None, topology_key=None, when_unsatisfiable=None, local_vars_configuration=None): if local_vars_configuration is None: local_vars_configuration = Configuration() self.local_vars_configuration = local_vars_configuration self._label_selector = None self._max_skew = None self._topology_key = None self._when_unsatisfiable = None self.discriminator = None if label_selector is not None: self.label_selector = label_selector self.max_skew = max_skew self.topology_key = topology_key self.when_unsatisfiable = when_unsatisfiable @property def label_selector(self): return self._label_selector @label_selector.setter def label_selector(self, label_selector): self._label_selector = label_selector @property def max_skew(self): return self._max_skew @max_skew.setter def max_skew(self, max_skew): if self.local_vars_configuration.client_side_validation and max_skew is None: raise ValueError("Invalid value for `max_skew`, must not be `None`") self._max_skew = max_skew @property def topology_key(self): return self._topology_key @topology_key.setter def topology_key(self, topology_key): if self.local_vars_configuration.client_side_validation and topology_key is None: raise ValueError("Invalid value for `topology_key`, must not be `None`") self._topology_key = topology_key @property def when_unsatisfiable(self): return self._when_unsatisfiable @when_unsatisfiable.setter def when_unsatisfiable(self, when_unsatisfiable): if self.local_vars_configuration.client_side_validation and when_unsatisfiable is None: raise ValueError("Invalid value for `when_unsatisfiable`, must not be `None`") self._when_unsatisfiable = when_unsatisfiable def to_dict(self): result = {} for attr, _ in six.iteritems(self.openapi_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value return result def to_str(self): return pprint.pformat(self.to_dict()) def __repr__(self): return self.to_str() def __eq__(self, other): if not isinstance(other, V1TopologySpreadConstraint): return False return self.to_dict() == other.to_dict() def __ne__(self, other): if not isinstance(other, V1TopologySpreadConstraint): return True return self.to_dict() != other.to_dict()
true
true
f70eb689c0446ccc43bdd2bc306542d157e6219f
589
py
Python
splango/urls.py
shimon/Splango
a89408b8b62d631763d42e049dd1b7d6ef72b02f
[ "MIT" ]
3
2016-06-02T10:18:57.000Z
2022-03-16T06:10:21.000Z
splango/urls.py
shimon/Splango
a89408b8b62d631763d42e049dd1b7d6ef72b02f
[ "MIT" ]
null
null
null
splango/urls.py
shimon/Splango
a89408b8b62d631763d42e049dd1b7d6ef72b02f
[ "MIT" ]
1
2022-03-16T06:10:33.000Z
2022-03-16T06:10:33.000Z
from django.conf.urls.defaults import * urlpatterns = patterns( 'splango.views', url(r'^confirm_human/$', 'confirm_human', name="splango-confirm-human"), url(r'^admin/$', 'experiments_overview', name="splango-admin"), url(r'^admin/exp/(?P<expname>[^/]+)/$', 'experiment_detail', name="splango-experiment-detail"), url(r'^admin/exp/(?P<expname>[^/]+)/(?P<report_id>\d+)/$', 'experiment_report', name="splango-experiment-report"), url(r'^admin/exp/(?P<expname>[^/]+)/(?P<variant>[^/]+)/(?P<goal>[^/]+)/$', 'experiment_log', name="splango-experiment-log"), )
39.266667
128
0.629881
from django.conf.urls.defaults import * urlpatterns = patterns( 'splango.views', url(r'^confirm_human/$', 'confirm_human', name="splango-confirm-human"), url(r'^admin/$', 'experiments_overview', name="splango-admin"), url(r'^admin/exp/(?P<expname>[^/]+)/$', 'experiment_detail', name="splango-experiment-detail"), url(r'^admin/exp/(?P<expname>[^/]+)/(?P<report_id>\d+)/$', 'experiment_report', name="splango-experiment-report"), url(r'^admin/exp/(?P<expname>[^/]+)/(?P<variant>[^/]+)/(?P<goal>[^/]+)/$', 'experiment_log', name="splango-experiment-log"), )
true
true
f70eb6f89adb3426524735b14874f388558e4b0c
5,289
py
Python
data_creation.py
Bruce-zxy/deep-study-lenet5
bba6531c9234c077107f79ff852f141cfed58229
[ "MIT" ]
null
null
null
data_creation.py
Bruce-zxy/deep-study-lenet5
bba6531c9234c077107f79ff852f141cfed58229
[ "MIT" ]
null
null
null
data_creation.py
Bruce-zxy/deep-study-lenet5
bba6531c9234c077107f79ff852f141cfed58229
[ "MIT" ]
null
null
null
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import os import sys import random import numpy as np import pandas as pd import h5py import matplotlib.pyplot as plt from math import cos, sin, atan2, sqrt, pi, radians, degrees, ceil, isnan from skimage import io, transform BASE_DIR = os.path.dirname(os.path.abspath(__file__)) sys.path.append(BASE_DIR) TRAIN_CSV_PATH = './pointdata4/traindata/' TEST_CSV_PATH = './pointdata4/testdata/' data_path = './h5/' train_file_path = data_path + 'initial_train_data.h5' test_file_path = data_path + 'initial_test_data.h5' # 按旋转角度分类的子级目录 label_dirs = [[16, 19], [43,71,129, 260], [95,128,129, 274]] # 按道路分类的父级目录 label_set = [0, 1, 2] # 获取二维点集的中心点坐标 def get_centroid(point_set): c_x, c_y = zip(*point_set) centroid_x = sum(c_x)/len(c_x) centroid_y = sum(c_y)/len(c_y) return centroid_x, centroid_y # 逆时针旋转坐标点 def n_rotate(angle, valuex, valuey, centerx, centery): valuex = np.array(valuex) valuey = np.array(valuey) nRotatex = (valuex-centerx)*cos(angle) - \ (valuey-centery)*sin(angle) + centerx nRotatey = (valuex-centerx)*sin(angle) + \ (valuey-centery)*cos(angle) + centery return nRotatex, nRotatey # 获取csv文件的列表 def get_csv_list(path): csv_file_list = [] file_list = os.listdir(path) for file_name in file_list: if file_name.endswith('csv'): csv_file_list.append(path + "/" + file_name) return csv_file_list # 获取csv文件中的点集数据 def get_csv_data(path_list): # 创建空的定维数组 sum_data = np.empty([0, 1024, 2], dtype=np.float32) # 遍历每个csv文件 for path in path_list: # 将每个csv文件读取为Numpy的数据 data = np.genfromtxt(path, delimiter=',', dtype=np.float32)[:, :2] data_len = len(data) empty_len = 1024 - data_len # 完整的1024个元数据=csv文件数据+在csv文件中随机指定下标数据 count = 0 while count < empty_len: data = np.append( data, [data[random.randint(0, data_len-1)]], axis=0) count += 1 sum_data = np.append(sum_data, [data], axis=0) print(sum_data.shape) return sum_data # 随机打乱点集数据 def exchange_data_index(sum_data, label_data): cursor_index = 0 max_range = len(sum_data) while cursor_index < max_range: random_index = random.randint(0, max_range-1) temp_sum_data = sum_data[0] temp_label_data = label_data[0] sum_data = np.delete(sum_data, 0, axis=0) label_data = np.delete(label_data, 0, axis=0) sum_data = np.insert(sum_data, random_index, temp_sum_data, axis=0) label_data = np.insert(label_data, random_index, temp_label_data, axis=0) cursor_index += 1 return sum_data, label_data def get_label_and_data(root_path, label_dirs): sum_data = np.empty([0, 1024, 2], dtype=np.float32) typical_data = np.empty([0], dtype=np.int32) for data_type, label_dir_set in enumerate(label_dirs): print(">> 现在进入【第%d类】数据" % (data_type+1)) for rotate_angle in label_dir_set: print("-- 需要旋转%d度的数据集:" % (rotate_angle)) # 获取csv文件列表 csv_list = get_csv_list( root_path + str(data_type) + '/' + str(rotate_angle)) # 获取csv文件点集数据 csv_data = get_csv_data(csv_list) # 遍历样本数据 for i, sample_data in enumerate(csv_data): # 求出点集的中心坐标点 centroid_x, centroid_y = get_centroid(sample_data) # 根据中心坐标点旋转点集中的点 for index, coordinate in enumerate(sample_data): x, y = coordinate n_x, n_y = n_rotate( radians(rotate_angle), x, y, centroid_x, centroid_y) # 旋转后的点集坐标中心化 sample_data[index] = [n_x-centroid_x, n_y-centroid_y] # 旋转后的点集回归原列表 csv_data[i] = sample_data # 归集点集标签 typical_data = np.append(typical_data, [data_type], axis=0) # 将每个不同数量的样本合并到主列表中(n,1024,2)=>(m,n,1024,2) sum_data = np.append(sum_data, csv_data, axis=0) return sum_data, typical_data if __name__ == "__main__": sum_train_data, train_typical_data = get_label_and_data( TRAIN_CSV_PATH, label_dirs) sum_test_data, test_typical_data = get_label_and_data( TEST_CSV_PATH, label_dirs) # 随机打乱点集数据 rand_sum_train_data, rand_train_typical_data = exchange_data_index( sum_train_data, train_typical_data) rand_sum_test_data, rand_test_typical_data = exchange_data_index( sum_test_data, test_typical_data) if os.access(data_path, os.F_OK) == False: os.mkdir(data_path) if os.access(train_file_path, os.F_OK) == True: os.remove(train_file_path) open(train_file_path, 'w') with h5py.File(train_file_path, 'r+') as f: f.create_dataset('data', data=rand_sum_train_data) f.create_dataset('label', data=rand_train_typical_data) if os.access(test_file_path, os.F_OK) == True: os.remove(test_file_path) open(test_file_path, 'w') with h5py.File(test_file_path, 'r+') as f: f.create_dataset('data', data=rand_sum_test_data) f.create_dataset('label', data=rand_test_typical_data)
31.861446
76
0.636226
import os import sys import random import numpy as np import pandas as pd import h5py import matplotlib.pyplot as plt from math import cos, sin, atan2, sqrt, pi, radians, degrees, ceil, isnan from skimage import io, transform BASE_DIR = os.path.dirname(os.path.abspath(__file__)) sys.path.append(BASE_DIR) TRAIN_CSV_PATH = './pointdata4/traindata/' TEST_CSV_PATH = './pointdata4/testdata/' data_path = './h5/' train_file_path = data_path + 'initial_train_data.h5' test_file_path = data_path + 'initial_test_data.h5' label_dirs = [[16, 19], [43,71,129, 260], [95,128,129, 274]] label_set = [0, 1, 2] def get_centroid(point_set): c_x, c_y = zip(*point_set) centroid_x = sum(c_x)/len(c_x) centroid_y = sum(c_y)/len(c_y) return centroid_x, centroid_y def n_rotate(angle, valuex, valuey, centerx, centery): valuex = np.array(valuex) valuey = np.array(valuey) nRotatex = (valuex-centerx)*cos(angle) - \ (valuey-centery)*sin(angle) + centerx nRotatey = (valuex-centerx)*sin(angle) + \ (valuey-centery)*cos(angle) + centery return nRotatex, nRotatey def get_csv_list(path): csv_file_list = [] file_list = os.listdir(path) for file_name in file_list: if file_name.endswith('csv'): csv_file_list.append(path + "/" + file_name) return csv_file_list def get_csv_data(path_list): sum_data = np.empty([0, 1024, 2], dtype=np.float32) for path in path_list: data = np.genfromtxt(path, delimiter=',', dtype=np.float32)[:, :2] data_len = len(data) empty_len = 1024 - data_len count = 0 while count < empty_len: data = np.append( data, [data[random.randint(0, data_len-1)]], axis=0) count += 1 sum_data = np.append(sum_data, [data], axis=0) print(sum_data.shape) return sum_data def exchange_data_index(sum_data, label_data): cursor_index = 0 max_range = len(sum_data) while cursor_index < max_range: random_index = random.randint(0, max_range-1) temp_sum_data = sum_data[0] temp_label_data = label_data[0] sum_data = np.delete(sum_data, 0, axis=0) label_data = np.delete(label_data, 0, axis=0) sum_data = np.insert(sum_data, random_index, temp_sum_data, axis=0) label_data = np.insert(label_data, random_index, temp_label_data, axis=0) cursor_index += 1 return sum_data, label_data def get_label_and_data(root_path, label_dirs): sum_data = np.empty([0, 1024, 2], dtype=np.float32) typical_data = np.empty([0], dtype=np.int32) for data_type, label_dir_set in enumerate(label_dirs): print(">> 现在进入【第%d类】数据" % (data_type+1)) for rotate_angle in label_dir_set: print("-- 需要旋转%d度的数据集:" % (rotate_angle)) csv_list = get_csv_list( root_path + str(data_type) + '/' + str(rotate_angle)) csv_data = get_csv_data(csv_list) for i, sample_data in enumerate(csv_data): centroid_x, centroid_y = get_centroid(sample_data) for index, coordinate in enumerate(sample_data): x, y = coordinate n_x, n_y = n_rotate( radians(rotate_angle), x, y, centroid_x, centroid_y) sample_data[index] = [n_x-centroid_x, n_y-centroid_y] csv_data[i] = sample_data typical_data = np.append(typical_data, [data_type], axis=0) sum_data = np.append(sum_data, csv_data, axis=0) return sum_data, typical_data if __name__ == "__main__": sum_train_data, train_typical_data = get_label_and_data( TRAIN_CSV_PATH, label_dirs) sum_test_data, test_typical_data = get_label_and_data( TEST_CSV_PATH, label_dirs) rand_sum_train_data, rand_train_typical_data = exchange_data_index( sum_train_data, train_typical_data) rand_sum_test_data, rand_test_typical_data = exchange_data_index( sum_test_data, test_typical_data) if os.access(data_path, os.F_OK) == False: os.mkdir(data_path) if os.access(train_file_path, os.F_OK) == True: os.remove(train_file_path) open(train_file_path, 'w') with h5py.File(train_file_path, 'r+') as f: f.create_dataset('data', data=rand_sum_train_data) f.create_dataset('label', data=rand_train_typical_data) if os.access(test_file_path, os.F_OK) == True: os.remove(test_file_path) open(test_file_path, 'w') with h5py.File(test_file_path, 'r+') as f: f.create_dataset('data', data=rand_sum_test_data) f.create_dataset('label', data=rand_test_typical_data)
true
true
f70eb7f417149b9a8081bb318f4066b6168a6304
1,289
py
Python
362 Design Hit Counter.py
ChiFire/legend_LeetCode
93fe97fef7e929fdbdc25fbb53955d44e14ecff8
[ "MIT" ]
872
2015-06-15T12:02:41.000Z
2022-03-30T08:44:35.000Z
362 Design Hit Counter.py
ChiFire/legend_LeetCode
93fe97fef7e929fdbdc25fbb53955d44e14ecff8
[ "MIT" ]
8
2015-06-21T15:11:59.000Z
2022-02-01T11:22:34.000Z
362 Design Hit Counter.py
ChiFire/legend_LeetCode
93fe97fef7e929fdbdc25fbb53955d44e14ecff8
[ "MIT" ]
328
2015-06-28T03:10:35.000Z
2022-03-29T11:05:28.000Z
""" Premium Question """ from collections import deque __author__ = 'Daniel' class HitCounter(object): def __init__(self): """ Initialize your data structure here. calls are being made to the system in chronological order. It is possible that several hits arrive roughly at the same time. What if the number of hits per second could be very large? Does your design scale? # use counter """ self.q = deque() def hit(self, timestamp): """ Record a hit. @param timestamp - The current timestamp (in seconds granularity). :type timestamp: int :rtype: void """ self.pop(timestamp) self.q.append(timestamp) def getHits(self, timestamp): """ Return the number of hits in the past 5 minutes. @param timestamp - The current timestamp (in seconds granularity). :type timestamp: int :rtype: int """ self.pop(timestamp) return len(self.q) def pop(self, timestamp): while self.q and timestamp - self.q[0] >= 300: self.q.popleft() # Your HitCounter object will be instantiated and called as such: # obj = HitCounter() # obj.hit(timestamp) # param_2 = obj.getHits(timestamp)
26.854167
105
0.612878
from collections import deque __author__ = 'Daniel' class HitCounter(object): def __init__(self): self.q = deque() def hit(self, timestamp): self.pop(timestamp) self.q.append(timestamp) def getHits(self, timestamp): self.pop(timestamp) return len(self.q) def pop(self, timestamp): while self.q and timestamp - self.q[0] >= 300: self.q.popleft()
true
true
f70eb992a38c3e76ec1dd03ed607975174757fd1
5,136
py
Python
pyinfra/operations/yum.py
vamshi091211/pyinfra
6e14b039422e00ebc68110eabbc6a3a543c96279
[ "MIT" ]
1
2022-03-24T05:44:45.000Z
2022-03-24T05:44:45.000Z
pyinfra/operations/yum.py
marinakravchenko21/pyinfra
6e14b039422e00ebc68110eabbc6a3a543c96279
[ "MIT" ]
null
null
null
pyinfra/operations/yum.py
marinakravchenko21/pyinfra
6e14b039422e00ebc68110eabbc6a3a543c96279
[ "MIT" ]
1
2021-11-12T18:36:01.000Z
2021-11-12T18:36:01.000Z
''' Manage yum packages and repositories. Note that yum package names are case-sensitive. ''' from __future__ import unicode_literals from pyinfra.api import operation from . import files from .util.packaging import ensure_packages, ensure_rpm, ensure_yum_repo @operation def key(state, host, key): ''' Add yum gpg keys with ``rpm``. + key: filename or URL Note: always returns one command, not state checking Example: .. code:: python linux_id = host.fact.linux_distribution['release_meta'].get('ID') yum.key( {'Add the Docker CentOS gpg key'}, 'https://download.docker.com/linux/{}/gpg'.format(linux_id), ) ''' yield 'rpm --import {0}'.format(key) @operation def repo( state, host, name, baseurl=None, present=True, description=None, enabled=True, gpgcheck=True, gpgkey=None, ): # NOTE: if updating this docstring also update `dnf.repo` # COMPAT: on v1 rearrange baseurl/present kwargs ''' Add/remove/update yum repositories. + name: URL or name for the ``.repo`` file + baseurl: the baseurl of the repo (if ``name`` is not a URL) + present: whether the ``.repo`` file should be present + description: optional verbose description + enabled: whether this repo is enabled + gpgcheck: whether set ``gpgcheck=1`` + gpgkey: the URL to the gpg key for this repo ``Baseurl``/``description``/``gpgcheck``/``gpgkey``: These are only valid when ``name`` is a filename (ie not a URL). This is for manual construction of repository files. Use a URL to download and install remote repository files. Examples: .. code:: python # Download a repository file yum.repo( {'Install Docker-CE repo via URL'}, 'https://download.docker.com/linux/centos/docker-ce.repo', ) # Create the repository file from baseurl/etc yum.repo( {'Add the Docker CentOS repo'}, name='DockerCE', baseurl='https://download.docker.com/linux/centos/7/$basearch/stable', ) ''' yield ensure_yum_repo( state, host, files, name, baseurl, present, description, enabled, gpgcheck, gpgkey, 'yum-config-manager', ) @operation def rpm(state, host, source, present=True): # NOTE: if updating this docstring also update `dnf.rpm` ''' Add/remove ``.rpm`` file packages. + source: filename or URL of the ``.rpm`` package + present: whether ore not the package should exist on the system URL sources with ``present=False``: If the ``.rpm`` file isn't downloaded, pyinfra can't remove any existing package as the file won't exist until mid-deploy. Example: .. code:: python yum.rpm( {'Install EPEL rpm to enable EPEL repo'}, 'https://dl.fedoraproject.org/pub/epel/epel-release-latest-' '{{ host.fact.linux_distribution.major }}.noarch.rpm', ) ''' yield ensure_rpm(state, host, files, source, present, 'yum') @operation def update(state, host): ''' Updates all yum packages. ''' yield 'yum update -y' _update = update # noqa: E305 (for use below where update is a kwarg) @operation def packages( state, host, packages=None, present=True, latest=False, update=False, clean=False, nobest=False, extra_install_args='', extra_uninstall_args='', ): ''' Install/remove/update yum packages & updates. + packages: list of packages to ensure + present: whether the packages should be installed + latest: whether to upgrade packages without a specified version + update: run yum update + clean: run yum clean + nobest: add the no best option to install + extra_install_args: additional arguments to the yum install command + extra_uninstall_args: additional arguments to the yum uninstall command Versions: Package versions can be pinned like yum: ``<pkg>-<version>`` Examples: .. code:: python # Update package list and install packages yum.packages( {'Install Vim and Vim enhanced'}, ['vim-enhanced', 'vim'], update=True, ) # Install the latest versions of packages (always check) yum.packages( {'Install latest Vim'}, ['vim'], latest=True, ) ''' if clean: yield 'yum clean all' if update: yield _update(state, host) nobest_option = '' if nobest: nobest_option = ' --nobest' if extra_install_args != '': extra_install_args = ' ' + extra_install_args if extra_uninstall_args != '': extra_uninstall_args = ' ' + extra_uninstall_args yield ensure_packages( packages, host.fact.rpm_packages, present, install_command='yum install -y' + nobest_option + extra_install_args, uninstall_command='yum remove -y' + extra_uninstall_args, upgrade_command='yum update -y', version_join='-', latest=latest, )
27.319149
85
0.626558
from __future__ import unicode_literals from pyinfra.api import operation from . import files from .util.packaging import ensure_packages, ensure_rpm, ensure_yum_repo @operation def key(state, host, key): yield 'rpm --import {0}'.format(key) @operation def repo( state, host, name, baseurl=None, present=True, description=None, enabled=True, gpgcheck=True, gpgkey=None, ): yield ensure_yum_repo( state, host, files, name, baseurl, present, description, enabled, gpgcheck, gpgkey, 'yum-config-manager', ) @operation def rpm(state, host, source, present=True): yield ensure_rpm(state, host, files, source, present, 'yum') @operation def update(state, host): yield 'yum update -y' _update = update @operation def packages( state, host, packages=None, present=True, latest=False, update=False, clean=False, nobest=False, extra_install_args='', extra_uninstall_args='', ): if clean: yield 'yum clean all' if update: yield _update(state, host) nobest_option = '' if nobest: nobest_option = ' --nobest' if extra_install_args != '': extra_install_args = ' ' + extra_install_args if extra_uninstall_args != '': extra_uninstall_args = ' ' + extra_uninstall_args yield ensure_packages( packages, host.fact.rpm_packages, present, install_command='yum install -y' + nobest_option + extra_install_args, uninstall_command='yum remove -y' + extra_uninstall_args, upgrade_command='yum update -y', version_join='-', latest=latest, )
true
true
f70eba1aa1744645a51c91f874da542a669aeca5
168
py
Python
setup.py
jovanzac/Server
90932560b61f378122355bb8df18309b245858d5
[ "MIT" ]
null
null
null
setup.py
jovanzac/Server
90932560b61f378122355bb8df18309b245858d5
[ "MIT" ]
null
null
null
setup.py
jovanzac/Server
90932560b61f378122355bb8df18309b245858d5
[ "MIT" ]
null
null
null
from cx_Freeze import setup, Executable setup(name = "Server" , version = "1.0" , description = "" , executables = [Executable("server.py")])
21
46
0.583333
from cx_Freeze import setup, Executable setup(name = "Server" , version = "1.0" , description = "" , executables = [Executable("server.py")])
true
true
f70eba2ea25aab68aa58ed217078ed6007e84dc0
23,807
py
Python
conans/client/build/cmake.py
sigiesec/conan
f966d516452380918437888811bc833c804dac39
[ "MIT" ]
null
null
null
conans/client/build/cmake.py
sigiesec/conan
f966d516452380918437888811bc833c804dac39
[ "MIT" ]
null
null
null
conans/client/build/cmake.py
sigiesec/conan
f966d516452380918437888811bc833c804dac39
[ "MIT" ]
null
null
null
import os import platform from collections import OrderedDict from itertools import chain from conans.client import defs_to_string, join_arguments from conans.client.build.cppstd_flags import cppstd_flag from conans.client.tools import cross_building from conans.client.tools.oss import get_cross_building_settings from conans.errors import ConanException from conans.model.conan_file import ConanFile from conans.model.version import Version from conans.util.env_reader import get_env from conans.util.files import mkdir, get_abs_path from conans.tools import cpu_count, args_to_string from conans import tools from conans.util.log import logger from conans.util.config_parser import get_bool_from_text from conans.client.build.compiler_flags import architecture_flag def _get_env_cmake_system_name(): env_system_name = get_env("CONAN_CMAKE_SYSTEM_NAME", "") return {"False": False, "True": True, "": None}.get(env_system_name, env_system_name) class CMake(object): def __init__(self, conanfile, generator=None, cmake_system_name=True, parallel=True, build_type=None, toolset=None, make_program=None, set_cmake_flags=False): """ :param settings_or_conanfile: Conanfile instance (or settings for retro compatibility) :param generator: Generator name to use or none to autodetect :param cmake_system_name: False to not use CMAKE_SYSTEM_NAME variable, True for auto-detect or directly a string with the system name :param parallel: Try to build with multiple cores if available :param build_type: Overrides default build type comming from settings :param toolset: Toolset name to use (such as llvm-vs2014) or none for default one, applies only to certain generators (e.g. Visual Studio) :param set_cmake_flags: whether or not to set CMake flags like CMAKE_CXX_FLAGS, CMAKE_C_FLAGS, etc. it's vital to set for certain projects (e.g. using CMAKE_SIZEOF_VOID_P or CMAKE_LIBRARY_ARCHITECTURE) """ if not isinstance(conanfile, ConanFile): raise ConanException("First argument of CMake() has to be ConanFile. Use CMake(self)") self._settings = conanfile.settings self._conanfile = conanfile self._os = self._settings.get_safe("os") self._os_build, _, self._os_host, _ = get_cross_building_settings(self._settings) self._compiler = self._settings.get_safe("compiler") self._compiler_version = self._settings.get_safe("compiler.version") self._arch = self._settings.get_safe("arch") os_ver_str = "os.api_level" if self._os == "Android" else "os.version" self._op_system_version = self._settings.get_safe(os_ver_str) self._libcxx = self._settings.get_safe("compiler.libcxx") self._runtime = self._settings.get_safe("compiler.runtime") self._build_type = self._settings.get_safe("build_type") self._cppstd = self._settings.get_safe("cppstd") self.generator = generator or self._generator() self.toolset = self._toolset(toolset) self.build_dir = None self._cmake_system_name = _get_env_cmake_system_name() if self._cmake_system_name is None: # Not overwritten using environment self._cmake_system_name = cmake_system_name self.parallel = parallel self._set_cmake_flags = set_cmake_flags self.definitions = self._get_cmake_definitions() if build_type and build_type != self._build_type: # Call the setter to warn and update the definitions if needed self.build_type = build_type make_program = os.getenv("CONAN_MAKE_PROGRAM") or make_program if make_program: if not tools.which(make_program): self._conanfile.output.warn("The specified make program '%s' cannot be found" "and will be ignored" % make_program) else: self._conanfile.output.info("Using '%s' as CMAKE_MAKE_PROGRAM" % make_program) self.definitions["CMAKE_MAKE_PROGRAM"] = make_program @property def build_folder(self): return self.build_dir @build_folder.setter def build_folder(self, value): self.build_dir = value @property def build_type(self): return self._build_type @build_type.setter def build_type(self, build_type): settings_build_type = self._settings.get_safe("build_type") if build_type != settings_build_type: self._conanfile.output.warn( 'Set CMake build type "%s" is different than the settings build_type "%s"' % (build_type, settings_build_type)) self._build_type = build_type self.definitions.update(self._build_type_definition()) @property def flags(self): return defs_to_string(self.definitions) def _generator(self): if "CONAN_CMAKE_GENERATOR" in os.environ: return os.environ["CONAN_CMAKE_GENERATOR"] if not self._compiler or not self._compiler_version or not self._arch: if self._os_build == "Windows": # Not enough settings to set a generator in Windows return None return "Unix Makefiles" if self._compiler == "Visual Studio": _visuals = {'8': '8 2005', '9': '9 2008', '10': '10 2010', '11': '11 2012', '12': '12 2013', '14': '14 2015', '15': '15 2017'} base = "Visual Studio %s" % _visuals.get(self._compiler_version, "UnknownVersion %s" % self._compiler_version) if self._arch == "x86_64": return base + " Win64" elif "arm" in self._arch: return base + " ARM" else: return base # The generator depends on the build machine, not the target if self._os_build == "Windows": return "MinGW Makefiles" # it is valid only under Windows return "Unix Makefiles" def _toolset(self, toolset=None): if toolset: return toolset elif self._settings.get_safe("compiler") == "Visual Studio": subs_toolset = self._settings.get_safe("compiler.toolset") if subs_toolset: return subs_toolset return None def _cmake_compiler_options(self): cmake_definitions = OrderedDict() if str(self._os).lower() == "macos": if self._arch == "x86": cmake_definitions["CMAKE_OSX_ARCHITECTURES"] = "i386" return cmake_definitions def _cmake_cross_build_defines(self): ret = OrderedDict() os_ver = get_env("CONAN_CMAKE_SYSTEM_VERSION", self._op_system_version) toolchain_file = get_env("CONAN_CMAKE_TOOLCHAIN_FILE", "") if toolchain_file != "": logger.info("Setting Cross build toolchain file: %s" % toolchain_file) ret["CMAKE_TOOLCHAIN_FILE"] = toolchain_file return ret if self._cmake_system_name is False: return ret # System name and system version if self._cmake_system_name is not True: # String not empty ret["CMAKE_SYSTEM_NAME"] = self._cmake_system_name ret["CMAKE_SYSTEM_VERSION"] = os_ver else: # detect if we are cross building and the system name and version if cross_building(self._conanfile.settings): # We are cross building if self._os != self._os_build: if self._os: # the_os is the host (regular setting) ret["CMAKE_SYSTEM_NAME"] = "Darwin" if self._os in ["iOS", "tvOS", "watchOS"] else self._os if os_ver: ret["CMAKE_SYSTEM_VERSION"] = os_ver else: ret["CMAKE_SYSTEM_NAME"] = "Generic" # system processor cmake_system_processor = os.getenv("CONAN_CMAKE_SYSTEM_PROCESSOR", None) if cmake_system_processor: ret["CMAKE_SYSTEM_PROCESSOR"] = cmake_system_processor if ret: # If enabled cross compile for env_var in ["CONAN_CMAKE_FIND_ROOT_PATH", "CONAN_CMAKE_FIND_ROOT_PATH_MODE_PROGRAM", "CONAN_CMAKE_FIND_ROOT_PATH_MODE_LIBRARY", "CONAN_CMAKE_FIND_ROOT_PATH_MODE_INCLUDE"]: value = os.getenv(env_var, None) if value: ret[env_var] = value if self._conanfile and self._conanfile.deps_cpp_info.sysroot: sysroot_path = self._conanfile.deps_cpp_info.sysroot else: sysroot_path = os.getenv("CONAN_CMAKE_FIND_ROOT_PATH", None) if sysroot_path: # Needs to be set here, can't be managed in the cmake generator, CMake needs # to know about the sysroot before any other thing ret["CMAKE_SYSROOT"] = sysroot_path.replace("\\", "/") # Adjust Android stuff if self._os == "Android": arch_abi_settings = {"armv8": "arm64-v8a", "armv7": "armeabi-v7a", "armv7hf": "armeabi-v7a", "armv6": "armeabi-v6", "armv5": "armeabi" }.get(self._arch, self._arch) if arch_abi_settings: ret["CMAKE_ANDROID_ARCH_ABI"] = arch_abi_settings logger.info("Setting Cross build flags: %s" % ", ".join(["%s=%s" % (k, v) for k, v in ret.items()])) return ret @property def is_multi_configuration(self): """ some IDEs are multi-configuration, as Visual. Makefiles or Ninja are single-conf """ if "Visual" in self.generator or "Xcode" in self.generator: return True # TODO: complete logic return False @property def command_line(self): args = ['-G "%s"' % self.generator] if self.generator else [] args.append(self.flags) args.append('-Wno-dev') if self.toolset: args.append('-T "%s"' % self.toolset) return join_arguments(args) def _build_type_definition(self): if self._build_type and not self.is_multi_configuration: return {'CMAKE_BUILD_TYPE': self._build_type} return {} @property def runtime(self): return defs_to_string(self._runtime_definition()) def _runtime_definition(self): if self._runtime: return {"CONAN_LINK_RUNTIME": "/%s" % self._runtime} return {} @property def build_config(self): """ cmake --build tool have a --config option for Multi-configuration IDEs """ if self._build_type and self.is_multi_configuration: return "--config %s" % self._build_type return "" def _get_cmake_definitions(self): def add_cmake_flag(cmake_flags, name, flag): """ appends compiler linker flags (if already present), or just sets """ if flag: if name not in cmake_flags: cmake_flags[name] = flag else: cmake_flags[name] = ' ' + flag return cmake_flags ret = OrderedDict() ret.update(self._build_type_definition()) ret.update(self._runtime_definition()) ret.update(self._cmake_compiler_options()) ret.update(self._cmake_cross_build_defines()) ret.update(self._get_cpp_standard_vars()) ret["CONAN_EXPORTED"] = "1" if self._compiler: ret["CONAN_COMPILER"] = self._compiler if self._compiler_version: ret["CONAN_COMPILER_VERSION"] = str(self._compiler_version) # Force compiler flags -- TODO: give as environment/setting parameter? arch_flag = architecture_flag(compiler=self._compiler, arch=self._arch) ret = add_cmake_flag(ret, 'CONAN_CXX_FLAGS', arch_flag) ret = add_cmake_flag(ret, 'CONAN_SHARED_LINKER_FLAGS', arch_flag) ret = add_cmake_flag(ret, 'CONAN_C_FLAGS', arch_flag) if self._set_cmake_flags: ret = add_cmake_flag(ret, 'CMAKE_CXX_FLAGS', arch_flag) ret = add_cmake_flag(ret, 'CMAKE_SHARED_LINKER_FLAGS', arch_flag) ret = add_cmake_flag(ret, 'CMAKE_C_FLAGS', arch_flag) if self._libcxx: ret["CONAN_LIBCXX"] = self._libcxx # Shared library try: ret["BUILD_SHARED_LIBS"] = "ON" if self._conanfile.options.shared else "OFF" except ConanException: pass # Install to package folder try: if self._conanfile.package_folder: ret["CMAKE_INSTALL_PREFIX"] = self._conanfile.package_folder except AttributeError: pass if str(self._os) in ["Windows", "WindowsStore"] and self._compiler == "Visual Studio": if self.parallel: cpus = tools.cpu_count() ret["CONAN_CXX_FLAGS"] = "/MP%s" % cpus ret["CONAN_C_FLAGS"] = "/MP%s" % cpus # fpic if str(self._os) not in ["Windows", "WindowsStore"]: fpic = self._conanfile.options.get_safe("fPIC") if fpic is not None: shared = self._conanfile.options.get_safe("shared") ret["CONAN_CMAKE_POSITION_INDEPENDENT_CODE"] = "ON" if (fpic or shared) else "OFF" # Adjust automatically the module path in case the conanfile is using the cmake_find_package if "cmake_find_package" in self._conanfile.generators: ret["CMAKE_MODULE_PATH"] = self._conanfile.install_folder.replace("\\", "/") # Disable CMake export registry #3070 (CMake installing modules in user home's) ret["CMAKE_EXPORT_NO_PACKAGE_REGISTRY"] = "ON" return ret def _get_dirs(self, source_folder, build_folder, source_dir, build_dir, cache_build_folder): if (source_folder or build_folder) and (source_dir or build_dir): raise ConanException("Use 'build_folder'/'source_folder' arguments") def get_dir(folder, origin): if folder: if os.path.isabs(folder): return folder return os.path.join(origin, folder) return origin if source_dir or build_dir: # OLD MODE build_ret = build_dir or self.build_dir or self._conanfile.build_folder source_ret = source_dir or self._conanfile.source_folder else: build_ret = get_dir(build_folder, self._conanfile.build_folder) source_ret = get_dir(source_folder, self._conanfile.source_folder) if self._conanfile.in_local_cache and cache_build_folder: build_ret = get_dir(cache_build_folder, self._conanfile.build_folder) return source_ret, build_ret def _run(self, command): if self._compiler == 'Visual Studio' and self.generator in ['Ninja', 'NMake Makefiles', 'NMake Makefiles JOM']: with tools.vcvars(self._settings, force=True, filter_known_paths=False): self._conanfile.run(command) else: self._conanfile.run(command) def configure(self, args=None, defs=None, source_dir=None, build_dir=None, source_folder=None, build_folder=None, cache_build_folder=None, pkg_config_paths=None): # TODO: Deprecate source_dir and build_dir in favor of xxx_folder if not self._conanfile.should_configure: return args = args or [] defs = defs or {} source_dir, self.build_dir = self._get_dirs(source_folder, build_folder, source_dir, build_dir, cache_build_folder) mkdir(self.build_dir) arg_list = join_arguments([ self.command_line, args_to_string(args), defs_to_string(defs), args_to_string([source_dir]) ]) if pkg_config_paths: pkg_env = {"PKG_CONFIG_PATH": os.pathsep.join(get_abs_path(f, self._conanfile.install_folder) for f in pkg_config_paths)} else: # If we are using pkg_config generator automate the pcs location, otherwise it could # read wrong files set_env = "pkg_config" in self._conanfile.generators \ and "PKG_CONFIG_PATH" not in os.environ pkg_env = {"PKG_CONFIG_PATH": self._conanfile.install_folder} if set_env else {} with tools.environment_append(pkg_env): command = "cd %s && cmake %s" % (args_to_string([self.build_dir]), arg_list) if platform.system() == "Windows" and self.generator == "MinGW Makefiles": with tools.remove_from_path("sh"): self._conanfile.run(command) else: self._conanfile.run(command) def build(self, args=None, build_dir=None, target=None): if not self._conanfile.should_build: return args = args or [] build_dir = build_dir or self.build_dir or self._conanfile.build_folder if target is not None: args = ["--target", target] + args if self.generator and self.parallel: if "Makefiles" in self.generator and "NMake" not in self.generator: if "--" not in args: args.append("--") args.append("-j%i" % cpu_count()) elif "Visual Studio" in self.generator and \ self._compiler_version and Version(self._compiler_version) >= "10": if "--" not in args: args.append("--") args.append("/m:%i" % cpu_count()) arg_list = join_arguments([ args_to_string([build_dir]), self.build_config, args_to_string(args) ]) command = "cmake --build %s" % arg_list self._run(command) def install(self, args=None, build_dir=None): if not self._conanfile.should_install: return mkdir(self._conanfile.package_folder) if not self.definitions.get("CMAKE_INSTALL_PREFIX"): raise ConanException("CMAKE_INSTALL_PREFIX not defined for 'cmake.install()'\n" "Make sure 'package_folder' is defined") self.build(args=args, build_dir=build_dir, target="install") def test(self, args=None, build_dir=None, target=None): if not self._conanfile.should_test: return if not target: target = "RUN_TESTS" if self.is_multi_configuration else "test" self.build(args=args, build_dir=build_dir, target=target) @property def verbose(self): try: verbose = self.definitions["CMAKE_VERBOSE_MAKEFILE"] return get_bool_from_text(str(verbose)) except KeyError: return False @verbose.setter def verbose(self, value): self.definitions["CMAKE_VERBOSE_MAKEFILE"] = "ON" if value else "OFF" def patch_config_paths(self): """ changes references to the absolute path of the installed package and its dependencies in exported cmake config files to the appropriate conan variable. This makes most (sensible) cmake config files portable. For example, if a package foo installs a file called "fooConfig.cmake" to be used by cmake's find_package method, normally this file will contain absolute paths to the installed package folder, for example it will contain a line such as: SET(Foo_INSTALL_DIR /home/developer/.conan/data/Foo/1.0.0/...) This will cause cmake find_package() method to fail when someone else installs the package via conan. This function will replace such mentions to SET(Foo_INSTALL_DIR ${CONAN_FOO_ROOT}) which is a variable that is set by conanbuildinfo.cmake, so that find_package() now correctly works on this conan package. For dependent packages, if a package foo installs a file called "fooConfig.cmake" to be used by cmake's find_package method and if it depends to a package bar, normally this file will contain absolute paths to the bar package folder, for example it will contain a line such as: SET_TARGET_PROPERTIES(foo PROPERTIES INTERFACE_INCLUDE_DIRECTORIES "/home/developer/.conan/data/Bar/1.0.0/user/channel/id/include") This function will replace such mentions to SET_TARGET_PROPERTIES(foo PROPERTIES INTERFACE_INCLUDE_DIRECTORIES "${CONAN_BAR_ROOT}/include") If the install() method of the CMake object in the conan file is used, this function should be called _after_ that invocation. For example: def build(self): cmake = CMake(self) cmake.configure() cmake.build() cmake.install() cmake.patch_config_paths() """ if not self._conanfile.should_install: return if not self._conanfile.name: raise ConanException("cmake.patch_config_paths() can't work without package name. " "Define name in your recipe") pf = self.definitions.get("CMAKE_INSTALL_PREFIX") replstr = "${CONAN_%s_ROOT}" % self._conanfile.name.upper() allwalk = chain(os.walk(self._conanfile.build_folder), os.walk(self._conanfile.package_folder)) for root, _, files in allwalk: for f in files: if f.endswith(".cmake"): path = os.path.join(root, f) tools.replace_in_file(path, pf, replstr, strict=False) # patch paths of dependent packages that are found in any cmake files of the current package path_content = tools.load(path) for dep in self._conanfile.deps_cpp_info.deps: from_str = self._conanfile.deps_cpp_info[dep].rootpath # try to replace only if from str is found if path_content.find(from_str) != -1: dep_str = "${CONAN_%s_ROOT}" % dep.upper() self._conanfile.output.info("Patching paths for %s: %s to %s" % (dep, from_str, dep_str)) tools.replace_in_file(path, from_str, dep_str, strict=False) def _get_cpp_standard_vars(self): if not self._cppstd: return {} ret = {} if self._cppstd.startswith("gnu"): ret["CONAN_CMAKE_CXX_STANDARD"] = self._cppstd[3:] ret["CONAN_CMAKE_CXX_EXTENSIONS"] = "ON" else: ret["CONAN_CMAKE_CXX_STANDARD"] = self._cppstd ret["CONAN_CMAKE_CXX_EXTENSIONS"] = "OFF" ret["CONAN_STD_CXX_FLAG"] = cppstd_flag(self._compiler, self._compiler_version, self._cppstd) return ret
42.588551
119
0.604192
import os import platform from collections import OrderedDict from itertools import chain from conans.client import defs_to_string, join_arguments from conans.client.build.cppstd_flags import cppstd_flag from conans.client.tools import cross_building from conans.client.tools.oss import get_cross_building_settings from conans.errors import ConanException from conans.model.conan_file import ConanFile from conans.model.version import Version from conans.util.env_reader import get_env from conans.util.files import mkdir, get_abs_path from conans.tools import cpu_count, args_to_string from conans import tools from conans.util.log import logger from conans.util.config_parser import get_bool_from_text from conans.client.build.compiler_flags import architecture_flag def _get_env_cmake_system_name(): env_system_name = get_env("CONAN_CMAKE_SYSTEM_NAME", "") return {"False": False, "True": True, "": None}.get(env_system_name, env_system_name) class CMake(object): def __init__(self, conanfile, generator=None, cmake_system_name=True, parallel=True, build_type=None, toolset=None, make_program=None, set_cmake_flags=False): if not isinstance(conanfile, ConanFile): raise ConanException("First argument of CMake() has to be ConanFile. Use CMake(self)") self._settings = conanfile.settings self._conanfile = conanfile self._os = self._settings.get_safe("os") self._os_build, _, self._os_host, _ = get_cross_building_settings(self._settings) self._compiler = self._settings.get_safe("compiler") self._compiler_version = self._settings.get_safe("compiler.version") self._arch = self._settings.get_safe("arch") os_ver_str = "os.api_level" if self._os == "Android" else "os.version" self._op_system_version = self._settings.get_safe(os_ver_str) self._libcxx = self._settings.get_safe("compiler.libcxx") self._runtime = self._settings.get_safe("compiler.runtime") self._build_type = self._settings.get_safe("build_type") self._cppstd = self._settings.get_safe("cppstd") self.generator = generator or self._generator() self.toolset = self._toolset(toolset) self.build_dir = None self._cmake_system_name = _get_env_cmake_system_name() if self._cmake_system_name is None: self._cmake_system_name = cmake_system_name self.parallel = parallel self._set_cmake_flags = set_cmake_flags self.definitions = self._get_cmake_definitions() if build_type and build_type != self._build_type: self.build_type = build_type make_program = os.getenv("CONAN_MAKE_PROGRAM") or make_program if make_program: if not tools.which(make_program): self._conanfile.output.warn("The specified make program '%s' cannot be found" "and will be ignored" % make_program) else: self._conanfile.output.info("Using '%s' as CMAKE_MAKE_PROGRAM" % make_program) self.definitions["CMAKE_MAKE_PROGRAM"] = make_program @property def build_folder(self): return self.build_dir @build_folder.setter def build_folder(self, value): self.build_dir = value @property def build_type(self): return self._build_type @build_type.setter def build_type(self, build_type): settings_build_type = self._settings.get_safe("build_type") if build_type != settings_build_type: self._conanfile.output.warn( 'Set CMake build type "%s" is different than the settings build_type "%s"' % (build_type, settings_build_type)) self._build_type = build_type self.definitions.update(self._build_type_definition()) @property def flags(self): return defs_to_string(self.definitions) def _generator(self): if "CONAN_CMAKE_GENERATOR" in os.environ: return os.environ["CONAN_CMAKE_GENERATOR"] if not self._compiler or not self._compiler_version or not self._arch: if self._os_build == "Windows": return None return "Unix Makefiles" if self._compiler == "Visual Studio": _visuals = {'8': '8 2005', '9': '9 2008', '10': '10 2010', '11': '11 2012', '12': '12 2013', '14': '14 2015', '15': '15 2017'} base = "Visual Studio %s" % _visuals.get(self._compiler_version, "UnknownVersion %s" % self._compiler_version) if self._arch == "x86_64": return base + " Win64" elif "arm" in self._arch: return base + " ARM" else: return base if self._os_build == "Windows": return "MinGW Makefiles" return "Unix Makefiles" def _toolset(self, toolset=None): if toolset: return toolset elif self._settings.get_safe("compiler") == "Visual Studio": subs_toolset = self._settings.get_safe("compiler.toolset") if subs_toolset: return subs_toolset return None def _cmake_compiler_options(self): cmake_definitions = OrderedDict() if str(self._os).lower() == "macos": if self._arch == "x86": cmake_definitions["CMAKE_OSX_ARCHITECTURES"] = "i386" return cmake_definitions def _cmake_cross_build_defines(self): ret = OrderedDict() os_ver = get_env("CONAN_CMAKE_SYSTEM_VERSION", self._op_system_version) toolchain_file = get_env("CONAN_CMAKE_TOOLCHAIN_FILE", "") if toolchain_file != "": logger.info("Setting Cross build toolchain file: %s" % toolchain_file) ret["CMAKE_TOOLCHAIN_FILE"] = toolchain_file return ret if self._cmake_system_name is False: return ret if self._cmake_system_name is not True: ret["CMAKE_SYSTEM_NAME"] = self._cmake_system_name ret["CMAKE_SYSTEM_VERSION"] = os_ver else: if cross_building(self._conanfile.settings): if self._os != self._os_build: if self._os: ret["CMAKE_SYSTEM_NAME"] = "Darwin" if self._os in ["iOS", "tvOS", "watchOS"] else self._os if os_ver: ret["CMAKE_SYSTEM_VERSION"] = os_ver else: ret["CMAKE_SYSTEM_NAME"] = "Generic" cmake_system_processor = os.getenv("CONAN_CMAKE_SYSTEM_PROCESSOR", None) if cmake_system_processor: ret["CMAKE_SYSTEM_PROCESSOR"] = cmake_system_processor if ret: for env_var in ["CONAN_CMAKE_FIND_ROOT_PATH", "CONAN_CMAKE_FIND_ROOT_PATH_MODE_PROGRAM", "CONAN_CMAKE_FIND_ROOT_PATH_MODE_LIBRARY", "CONAN_CMAKE_FIND_ROOT_PATH_MODE_INCLUDE"]: value = os.getenv(env_var, None) if value: ret[env_var] = value if self._conanfile and self._conanfile.deps_cpp_info.sysroot: sysroot_path = self._conanfile.deps_cpp_info.sysroot else: sysroot_path = os.getenv("CONAN_CMAKE_FIND_ROOT_PATH", None) if sysroot_path: # to know about the sysroot before any other thing ret["CMAKE_SYSROOT"] = sysroot_path.replace("\\", "/") # Adjust Android stuff if self._os == "Android": arch_abi_settings = {"armv8": "arm64-v8a", "armv7": "armeabi-v7a", "armv7hf": "armeabi-v7a", "armv6": "armeabi-v6", "armv5": "armeabi" }.get(self._arch, self._arch) if arch_abi_settings: ret["CMAKE_ANDROID_ARCH_ABI"] = arch_abi_settings logger.info("Setting Cross build flags: %s" % ", ".join(["%s=%s" % (k, v) for k, v in ret.items()])) return ret @property def is_multi_configuration(self): if "Visual" in self.generator or "Xcode" in self.generator: return True # TODO: complete logic return False @property def command_line(self): args = ['-G "%s"' % self.generator] if self.generator else [] args.append(self.flags) args.append('-Wno-dev') if self.toolset: args.append('-T "%s"' % self.toolset) return join_arguments(args) def _build_type_definition(self): if self._build_type and not self.is_multi_configuration: return {'CMAKE_BUILD_TYPE': self._build_type} return {} @property def runtime(self): return defs_to_string(self._runtime_definition()) def _runtime_definition(self): if self._runtime: return {"CONAN_LINK_RUNTIME": "/%s" % self._runtime} return {} @property def build_config(self): if self._build_type and self.is_multi_configuration: return "--config %s" % self._build_type return "" def _get_cmake_definitions(self): def add_cmake_flag(cmake_flags, name, flag): if flag: if name not in cmake_flags: cmake_flags[name] = flag else: cmake_flags[name] = ' ' + flag return cmake_flags ret = OrderedDict() ret.update(self._build_type_definition()) ret.update(self._runtime_definition()) ret.update(self._cmake_compiler_options()) ret.update(self._cmake_cross_build_defines()) ret.update(self._get_cpp_standard_vars()) ret["CONAN_EXPORTED"] = "1" if self._compiler: ret["CONAN_COMPILER"] = self._compiler if self._compiler_version: ret["CONAN_COMPILER_VERSION"] = str(self._compiler_version) # Force compiler flags -- TODO: give as environment/setting parameter? arch_flag = architecture_flag(compiler=self._compiler, arch=self._arch) ret = add_cmake_flag(ret, 'CONAN_CXX_FLAGS', arch_flag) ret = add_cmake_flag(ret, 'CONAN_SHARED_LINKER_FLAGS', arch_flag) ret = add_cmake_flag(ret, 'CONAN_C_FLAGS', arch_flag) if self._set_cmake_flags: ret = add_cmake_flag(ret, 'CMAKE_CXX_FLAGS', arch_flag) ret = add_cmake_flag(ret, 'CMAKE_SHARED_LINKER_FLAGS', arch_flag) ret = add_cmake_flag(ret, 'CMAKE_C_FLAGS', arch_flag) if self._libcxx: ret["CONAN_LIBCXX"] = self._libcxx # Shared library try: ret["BUILD_SHARED_LIBS"] = "ON" if self._conanfile.options.shared else "OFF" except ConanException: pass # Install to package folder try: if self._conanfile.package_folder: ret["CMAKE_INSTALL_PREFIX"] = self._conanfile.package_folder except AttributeError: pass if str(self._os) in ["Windows", "WindowsStore"] and self._compiler == "Visual Studio": if self.parallel: cpus = tools.cpu_count() ret["CONAN_CXX_FLAGS"] = "/MP%s" % cpus ret["CONAN_C_FLAGS"] = "/MP%s" % cpus # fpic if str(self._os) not in ["Windows", "WindowsStore"]: fpic = self._conanfile.options.get_safe("fPIC") if fpic is not None: shared = self._conanfile.options.get_safe("shared") ret["CONAN_CMAKE_POSITION_INDEPENDENT_CODE"] = "ON" if (fpic or shared) else "OFF" # Adjust automatically the module path in case the conanfile is using the cmake_find_package if "cmake_find_package" in self._conanfile.generators: ret["CMAKE_MODULE_PATH"] = self._conanfile.install_folder.replace("\\", "/") # Disable CMake export registry #3070 (CMake installing modules in user home's) ret["CMAKE_EXPORT_NO_PACKAGE_REGISTRY"] = "ON" return ret def _get_dirs(self, source_folder, build_folder, source_dir, build_dir, cache_build_folder): if (source_folder or build_folder) and (source_dir or build_dir): raise ConanException("Use 'build_folder'/'source_folder' arguments") def get_dir(folder, origin): if folder: if os.path.isabs(folder): return folder return os.path.join(origin, folder) return origin if source_dir or build_dir: build_ret = build_dir or self.build_dir or self._conanfile.build_folder source_ret = source_dir or self._conanfile.source_folder else: build_ret = get_dir(build_folder, self._conanfile.build_folder) source_ret = get_dir(source_folder, self._conanfile.source_folder) if self._conanfile.in_local_cache and cache_build_folder: build_ret = get_dir(cache_build_folder, self._conanfile.build_folder) return source_ret, build_ret def _run(self, command): if self._compiler == 'Visual Studio' and self.generator in ['Ninja', 'NMake Makefiles', 'NMake Makefiles JOM']: with tools.vcvars(self._settings, force=True, filter_known_paths=False): self._conanfile.run(command) else: self._conanfile.run(command) def configure(self, args=None, defs=None, source_dir=None, build_dir=None, source_folder=None, build_folder=None, cache_build_folder=None, pkg_config_paths=None): if not self._conanfile.should_configure: return args = args or [] defs = defs or {} source_dir, self.build_dir = self._get_dirs(source_folder, build_folder, source_dir, build_dir, cache_build_folder) mkdir(self.build_dir) arg_list = join_arguments([ self.command_line, args_to_string(args), defs_to_string(defs), args_to_string([source_dir]) ]) if pkg_config_paths: pkg_env = {"PKG_CONFIG_PATH": os.pathsep.join(get_abs_path(f, self._conanfile.install_folder) for f in pkg_config_paths)} else: set_env = "pkg_config" in self._conanfile.generators \ and "PKG_CONFIG_PATH" not in os.environ pkg_env = {"PKG_CONFIG_PATH": self._conanfile.install_folder} if set_env else {} with tools.environment_append(pkg_env): command = "cd %s && cmake %s" % (args_to_string([self.build_dir]), arg_list) if platform.system() == "Windows" and self.generator == "MinGW Makefiles": with tools.remove_from_path("sh"): self._conanfile.run(command) else: self._conanfile.run(command) def build(self, args=None, build_dir=None, target=None): if not self._conanfile.should_build: return args = args or [] build_dir = build_dir or self.build_dir or self._conanfile.build_folder if target is not None: args = ["--target", target] + args if self.generator and self.parallel: if "Makefiles" in self.generator and "NMake" not in self.generator: if "--" not in args: args.append("--") args.append("-j%i" % cpu_count()) elif "Visual Studio" in self.generator and \ self._compiler_version and Version(self._compiler_version) >= "10": if "--" not in args: args.append("--") args.append("/m:%i" % cpu_count()) arg_list = join_arguments([ args_to_string([build_dir]), self.build_config, args_to_string(args) ]) command = "cmake --build %s" % arg_list self._run(command) def install(self, args=None, build_dir=None): if not self._conanfile.should_install: return mkdir(self._conanfile.package_folder) if not self.definitions.get("CMAKE_INSTALL_PREFIX"): raise ConanException("CMAKE_INSTALL_PREFIX not defined for 'cmake.install()'\n" "Make sure 'package_folder' is defined") self.build(args=args, build_dir=build_dir, target="install") def test(self, args=None, build_dir=None, target=None): if not self._conanfile.should_test: return if not target: target = "RUN_TESTS" if self.is_multi_configuration else "test" self.build(args=args, build_dir=build_dir, target=target) @property def verbose(self): try: verbose = self.definitions["CMAKE_VERBOSE_MAKEFILE"] return get_bool_from_text(str(verbose)) except KeyError: return False @verbose.setter def verbose(self, value): self.definitions["CMAKE_VERBOSE_MAKEFILE"] = "ON" if value else "OFF" def patch_config_paths(self): if not self._conanfile.should_install: return if not self._conanfile.name: raise ConanException("cmake.patch_config_paths() can't work without package name. " "Define name in your recipe") pf = self.definitions.get("CMAKE_INSTALL_PREFIX") replstr = "${CONAN_%s_ROOT}" % self._conanfile.name.upper() allwalk = chain(os.walk(self._conanfile.build_folder), os.walk(self._conanfile.package_folder)) for root, _, files in allwalk: for f in files: if f.endswith(".cmake"): path = os.path.join(root, f) tools.replace_in_file(path, pf, replstr, strict=False) # patch paths of dependent packages that are found in any cmake files of the current package path_content = tools.load(path) for dep in self._conanfile.deps_cpp_info.deps: from_str = self._conanfile.deps_cpp_info[dep].rootpath # try to replace only if from str is found if path_content.find(from_str) != -1: dep_str = "${CONAN_%s_ROOT}" % dep.upper() self._conanfile.output.info("Patching paths for %s: %s to %s" % (dep, from_str, dep_str)) tools.replace_in_file(path, from_str, dep_str, strict=False) def _get_cpp_standard_vars(self): if not self._cppstd: return {} ret = {} if self._cppstd.startswith("gnu"): ret["CONAN_CMAKE_CXX_STANDARD"] = self._cppstd[3:] ret["CONAN_CMAKE_CXX_EXTENSIONS"] = "ON" else: ret["CONAN_CMAKE_CXX_STANDARD"] = self._cppstd ret["CONAN_CMAKE_CXX_EXTENSIONS"] = "OFF" ret["CONAN_STD_CXX_FLAG"] = cppstd_flag(self._compiler, self._compiler_version, self._cppstd) return ret
true
true
f70eba7ac72db4241c482950c5d46e65d867d233
2,161
py
Python
apps/erms/api.py
remocrevo/celus
682b13168eb475d7f970502113e756e40a899877
[ "MIT" ]
7
2020-02-20T13:24:40.000Z
2022-01-28T19:36:04.000Z
apps/erms/api.py
remocrevo/celus
682b13168eb475d7f970502113e756e40a899877
[ "MIT" ]
15
2020-04-28T13:09:02.000Z
2021-11-03T15:21:24.000Z
apps/erms/api.py
remocrevo/celus
682b13168eb475d7f970502113e756e40a899877
[ "MIT" ]
4
2020-02-20T13:48:30.000Z
2021-03-19T00:33:34.000Z
import urllib.parse import requests class ERMSError(Exception): pass class ERMS(object): """ Possible queries: /object?id=eq.574 /object?id=in.(574,575) """ # endpoints EP_OBJECT = 'object' EP_IDENTITY = 'identity' EP_CONSORTIUM = 'consortium' EP_CONSORTIUM_MEMBER = 'consortium_member' EP_ACQUISITION = 'acquisition' EP_PROCUREMENT = 'procurement' EP_OFFER = 'offer' EP_OFFER_SPLIT = 'offer_split' # object classes CLS_PERSON = 'Person' CLS_ORGANIZATION = 'Organization' CLS_PLATFORM = 'Platform' def __init__(self, base_url="https://erms.czechelib.cz/api/"): self.base_url = base_url.rstrip('/') self.session = requests.Session() @classmethod def _construct_query_string(cls, value): if type(value) in (list, tuple, set): return 'in.({})'.format(','.join(str(_id) for _id in value)) return f'eq.{value}' def construct_object_url(self, cls=None, object_id=None): params = {} if cls: params['class'] = self._construct_query_string(cls) if object_id: params['id'] = self._construct_query_string(object_id) else: params['order'] = 'id' query = urllib.parse.urlencode(params) return f'{self.base_url}/{self.EP_OBJECT}?{query}' def fetch_url(self, url): response = self.session.get(url) if response.status_code == 200: return response.json() raise ERMSError(response) def fetch_objects(self, cls=None, object_id=None): url = self.construct_object_url(cls=cls, object_id=object_id) data = self.fetch_url(url) return data def fetch_endpoint(self, endpoint, object_id=None, **kwargs): url = f'{self.base_url}/{endpoint}' params = {} if object_id: params['id'] = self._construct_query_string(object_id) for key, value in kwargs.items(): params[key] = self._construct_query_string(value) if params: url += '?{}'.format(urllib.parse.urlencode(params)) return self.fetch_url(url)
28.064935
72
0.61777
import urllib.parse import requests class ERMSError(Exception): pass class ERMS(object): EP_OBJECT = 'object' EP_IDENTITY = 'identity' EP_CONSORTIUM = 'consortium' EP_CONSORTIUM_MEMBER = 'consortium_member' EP_ACQUISITION = 'acquisition' EP_PROCUREMENT = 'procurement' EP_OFFER = 'offer' EP_OFFER_SPLIT = 'offer_split' CLS_PERSON = 'Person' CLS_ORGANIZATION = 'Organization' CLS_PLATFORM = 'Platform' def __init__(self, base_url="https://erms.czechelib.cz/api/"): self.base_url = base_url.rstrip('/') self.session = requests.Session() @classmethod def _construct_query_string(cls, value): if type(value) in (list, tuple, set): return 'in.({})'.format(','.join(str(_id) for _id in value)) return f'eq.{value}' def construct_object_url(self, cls=None, object_id=None): params = {} if cls: params['class'] = self._construct_query_string(cls) if object_id: params['id'] = self._construct_query_string(object_id) else: params['order'] = 'id' query = urllib.parse.urlencode(params) return f'{self.base_url}/{self.EP_OBJECT}?{query}' def fetch_url(self, url): response = self.session.get(url) if response.status_code == 200: return response.json() raise ERMSError(response) def fetch_objects(self, cls=None, object_id=None): url = self.construct_object_url(cls=cls, object_id=object_id) data = self.fetch_url(url) return data def fetch_endpoint(self, endpoint, object_id=None, **kwargs): url = f'{self.base_url}/{endpoint}' params = {} if object_id: params['id'] = self._construct_query_string(object_id) for key, value in kwargs.items(): params[key] = self._construct_query_string(value) if params: url += '?{}'.format(urllib.parse.urlencode(params)) return self.fetch_url(url)
true
true
f70ebad00452b2d47ad5070a859d615e6a790ba0
5,906
py
Python
dash/html/Code.py
sthagen/plotly-dash
4a9412efe5229809016eee9a5f1c50f882720c4c
[ "MIT" ]
null
null
null
dash/html/Code.py
sthagen/plotly-dash
4a9412efe5229809016eee9a5f1c50f882720c4c
[ "MIT" ]
null
null
null
dash/html/Code.py
sthagen/plotly-dash
4a9412efe5229809016eee9a5f1c50f882720c4c
[ "MIT" ]
null
null
null
# AUTO GENERATED FILE - DO NOT EDIT from dash.development.base_component import Component, _explicitize_args class Code(Component): """A Code component. Code is a wrapper for the <code> HTML5 element. For detailed attribute info see: https://developer.mozilla.org/en-US/docs/Web/HTML/Element/code Keyword arguments: - children (a list of or a singular dash component, string or number; optional): The children of this component. - id (string; optional): The ID of this component, used to identify dash components in callbacks. The ID needs to be unique across all of the components in an app. - accessKey (string; optional): Keyboard shortcut to activate or add focus to the element. - aria-* (string; optional): A wildcard aria attribute. - className (string; optional): Often used with CSS to style elements with common properties. - contentEditable (string; optional): Indicates whether the element's content is editable. - contextMenu (string; optional): Defines the ID of a <menu> element which will serve as the element's context menu. - data-* (string; optional): A wildcard data attribute. - dir (string; optional): Defines the text direction. Allowed values are ltr (Left-To-Right) or rtl (Right-To-Left). - draggable (string; optional): Defines whether the element can be dragged. - hidden (a value equal to: 'hidden', 'HIDDEN' | boolean; optional): Prevents rendering of given element, while keeping child elements, e.g. script elements, active. - key (string; optional): A unique identifier for the component, used to improve performance by React.js while rendering components See https://reactjs.org/docs/lists-and-keys.html for more info. - lang (string; optional): Defines the language used in the element. - loading_state (dict; optional): Object that holds the loading state object coming from dash-renderer. `loading_state` is a dict with keys: - component_name (string; optional): Holds the name of the component that is loading. - is_loading (boolean; optional): Determines if the component is loading or not. - prop_name (string; optional): Holds which property is loading. - n_clicks (number; default 0): An integer that represents the number of times that this element has been clicked on. - n_clicks_timestamp (number; default -1): An integer that represents the time (in ms since 1970) at which n_clicks changed. This can be used to tell which button was changed most recently. - role (string; optional): Defines an explicit role for an element for use by assistive technologies. - spellCheck (string; optional): Indicates whether spell checking is allowed for the element. - style (dict; optional): Defines CSS styles which will override styles previously set. - tabIndex (string; optional): Overrides the browser's default tab order and follows the one specified instead. - title (string; optional): Text to be displayed in a tooltip when hovering over the element.""" _children_props = [] _base_nodes = ["children"] _namespace = "dash_html_components" _type = "Code" @_explicitize_args def __init__( self, children=None, id=Component.UNDEFINED, n_clicks=Component.UNDEFINED, n_clicks_timestamp=Component.UNDEFINED, key=Component.UNDEFINED, accessKey=Component.UNDEFINED, className=Component.UNDEFINED, contentEditable=Component.UNDEFINED, contextMenu=Component.UNDEFINED, dir=Component.UNDEFINED, draggable=Component.UNDEFINED, hidden=Component.UNDEFINED, lang=Component.UNDEFINED, role=Component.UNDEFINED, spellCheck=Component.UNDEFINED, style=Component.UNDEFINED, tabIndex=Component.UNDEFINED, title=Component.UNDEFINED, loading_state=Component.UNDEFINED, **kwargs ): self._prop_names = [ "children", "id", "accessKey", "aria-*", "className", "contentEditable", "contextMenu", "data-*", "dir", "draggable", "hidden", "key", "lang", "loading_state", "n_clicks", "n_clicks_timestamp", "role", "spellCheck", "style", "tabIndex", "title", ] self._valid_wildcard_attributes = ["data-", "aria-"] self.available_properties = [ "children", "id", "accessKey", "aria-*", "className", "contentEditable", "contextMenu", "data-*", "dir", "draggable", "hidden", "key", "lang", "loading_state", "n_clicks", "n_clicks_timestamp", "role", "spellCheck", "style", "tabIndex", "title", ] self.available_wildcard_properties = ["data-", "aria-"] _explicit_args = kwargs.pop("_explicit_args") _locals = locals() _locals.update(kwargs) # For wildcard attrs and excess named props args = {k: _locals[k] for k in _explicit_args if k != "children"} for k in []: if k not in args: raise TypeError("Required argument `" + k + "` was not specified.") super(Code, self).__init__(children=children, **args)
31.752688
84
0.596173
from dash.development.base_component import Component, _explicitize_args class Code(Component): _children_props = [] _base_nodes = ["children"] _namespace = "dash_html_components" _type = "Code" @_explicitize_args def __init__( self, children=None, id=Component.UNDEFINED, n_clicks=Component.UNDEFINED, n_clicks_timestamp=Component.UNDEFINED, key=Component.UNDEFINED, accessKey=Component.UNDEFINED, className=Component.UNDEFINED, contentEditable=Component.UNDEFINED, contextMenu=Component.UNDEFINED, dir=Component.UNDEFINED, draggable=Component.UNDEFINED, hidden=Component.UNDEFINED, lang=Component.UNDEFINED, role=Component.UNDEFINED, spellCheck=Component.UNDEFINED, style=Component.UNDEFINED, tabIndex=Component.UNDEFINED, title=Component.UNDEFINED, loading_state=Component.UNDEFINED, **kwargs ): self._prop_names = [ "children", "id", "accessKey", "aria-*", "className", "contentEditable", "contextMenu", "data-*", "dir", "draggable", "hidden", "key", "lang", "loading_state", "n_clicks", "n_clicks_timestamp", "role", "spellCheck", "style", "tabIndex", "title", ] self._valid_wildcard_attributes = ["data-", "aria-"] self.available_properties = [ "children", "id", "accessKey", "aria-*", "className", "contentEditable", "contextMenu", "data-*", "dir", "draggable", "hidden", "key", "lang", "loading_state", "n_clicks", "n_clicks_timestamp", "role", "spellCheck", "style", "tabIndex", "title", ] self.available_wildcard_properties = ["data-", "aria-"] _explicit_args = kwargs.pop("_explicit_args") _locals = locals() _locals.update(kwargs) args = {k: _locals[k] for k in _explicit_args if k != "children"} for k in []: if k not in args: raise TypeError("Required argument `" + k + "` was not specified.") super(Code, self).__init__(children=children, **args)
true
true
f70ebb69800c57b924e20d307a2ea6cd4f9f431c
11,439
py
Python
python/train_model.py
skinnider/low-data-generative-models
6e743b6d1ba3265f58fcbd33f2c60e633cf25999
[ "MIT" ]
10
2021-02-01T06:21:52.000Z
2022-03-02T23:59:30.000Z
python/train_model.py
skinnider/low-data-generative-models
6e743b6d1ba3265f58fcbd33f2c60e633cf25999
[ "MIT" ]
1
2021-09-21T02:38:07.000Z
2021-09-21T02:38:07.000Z
python/train_model.py
skinnider/low-data-generative-models
6e743b6d1ba3265f58fcbd33f2c60e633cf25999
[ "MIT" ]
5
2021-04-16T18:11:48.000Z
2021-12-08T07:21:09.000Z
""" Train a language model to generate SMILES. """ import argparse import os import numpy as np import pandas as pd import random import sys import torch import torch.nn as nn import torch.optim as optim from torch.utils.data import DataLoader from tqdm import tqdm # suppress Chem.MolFromSmiles error output from rdkit import rdBase rdBase.DisableLog('rdApp.error') # set working directory git_dir = os.path.expanduser("~/git/low-data-generative-models") python_dir = git_dir + "/python" os.chdir(python_dir) # import classes from models import RNN, OneHotRNN, EarlyStopping from datasets import SmilesDataset, SelfiesDataset, SmilesCollate from functions import decrease_learning_rate, print_update, track_loss, \ sample_smiles, write_smiles ### CLI parser = argparse.ArgumentParser( description='Chemical structure language model interface') # input file parser.add_argument('--smiles_file', type=str, help='location of the SMILES file to train on') parser.add_argument('--selfies', dest='selfies', action='store_true') parser.set_defaults(selfies=False) # output files parser.add_argument('--output_dir', type=str, help='directory to save trained models to') # RNN parameters parser.add_argument('--rnn_type', type=str, choices=['RNN', 'LSTM', 'GRU'], default='GRU', help='type of language model to train') parser.add_argument('--embedding_size', type=int, default=128, help='size of vocabulary embedding') parser.add_argument('--hidden_size', type=int, default=512, help='size of language model hidden layers') parser.add_argument('--n_layers', type=int, default=3, help='number of layers in language model') parser.add_argument('--dropout', type=float, default=0, help='amount of dropout (0-1) to apply to model') parser.add_argument('--bidirectional', type=bool, default=False, help='for LSTMs only, train a bidirectional model') parser.add_argument('--nonlinearity', type=str, choices=['tanh', 'relu'], default='tanh', help='for RNNs only, nonlinearity to use') parser.add_argument('--tie_weights', dest='tie_weights', help='require embedding/dense linear layers use the ' +\ 'same weights', action='store_true') parser.set_defaults(tie_weights=False) # optimization parameters parser.add_argument('--learning_rate', type=float, default=0.001, help='initial learning rate') parser.add_argument('--learning_rate_decay', default=None, # type=float, help='amount (0-1) to decrease learning rate by every ' +\ 'fixed number of steps') parser.add_argument('--learning_rate_decay_steps', default=10000, type=int, help='# of steps between learning rate decrements') parser.add_argument('--gradient_clip', default=None, # type=float, help='amount to which to clip the gradients') # training schedule parser.add_argument('--seed', type=int, default=0, help='seed for random number generator') parser.add_argument('--batch_size', type=int, default=128, help='batch size') parser.add_argument('--max_epochs', type=int, default=1000, help='maximum number of epochs to train for') parser.add_argument('--patience', type=int, default=100, help='patience for early stopping') # sampling from trained models parser.add_argument('--sample_idx', type=int, default=0, help='index of the model being trained (zero-indexed)') parser.add_argument('--sample_every_epochs', type=int, help='if set, sample SMILES from the trained model' + 'every n epochs') parser.add_argument('--sample_every_steps', type=int, help='if set, sample SMILES from the trained model' + 'every n steps') parser.add_argument('--log_every_epochs', type=int, help='log training/validation losses every n epochs') parser.add_argument('--log_every_steps', type=int, help='log training/validation losses every n steps') parser.add_argument('--sample_size', type=int, default=100000, help='size of each sample from the trained model') # start with pretrained model parser.add_argument('--pretrain_model', type=str, default=None, help='load parameters from a pretrained model') # enforce a larger vocabulary parser.add_argument('--vocab_file', type=str, default=None, help='file containing all tokens in vocabulary') # for use in grid parser.add_argument('--stop_if_exists', dest='stop_if_exists', action='store_true') parser.set_defaults(stop_if_exists=False) # parse arguments args = parser.parse_args() # manually deal with gradient clipping try: args.gradient_clip = float(args.gradient_clip) except (ValueError, TypeError): args.gradient_clip = None # manually deal with learning rate decay try: args.learning_rate_decay = float(args.learning_rate_decay) except (ValueError, TypeError): args.learning_rate_decay = None # log args (make searching through logging directory easier) for arg in vars(args): print(arg, ": ", getattr(args, arg), "(", type(getattr(args, arg)), ")") # optionally stop if output file already exists if args.selfies: smiles_filename = "sample-" + str(args.sample_idx + 1) + "-SELFIES.smi" else: smiles_filename = "sample-" + str(args.sample_idx + 1) + "-SMILES.smi" smiles_file = os.path.join(args.output_dir, smiles_filename) if os.path.isfile(smiles_file) and args.stop_if_exists: print("output file " + smiles_file + " exists: stopping early") sys.exit() # make output directories if not os.path.isdir(args.output_dir): try: os.makedirs(args.output_dir) except FileExistsError: pass ## seed all RNGs torch.manual_seed(args.seed) random.seed(args.seed) np.random.seed(args.seed) if torch.cuda.is_available(): print("using cuda") torch.cuda.manual_seed_all(args.seed) # set up dataset if args.selfies: dataset = SelfiesDataset(selfies_file=args.smiles_file) else: dataset = SmilesDataset(smiles_file=args.smiles_file, vocab_file=args.vocab_file) # set up batching loader = DataLoader(dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=SmilesCollate(dataset.vocabulary)) # set up model if args.embedding_size > 0: model = RNN(vocabulary=dataset.vocabulary, rnn_type=args.rnn_type, embedding_size=args.embedding_size, hidden_size=args.hidden_size, n_layers=args.n_layers, dropout=args.dropout, bidirectional=args.bidirectional, tie_weights=args.tie_weights, nonlinearity=args.nonlinearity) else: # no embedding layer (one-hot encoding) model = OneHotRNN(vocabulary=dataset.vocabulary, rnn_type=args.rnn_type, hidden_size=args.hidden_size, n_layers=args.n_layers, dropout=args.dropout, bidirectional=args.bidirectional, nonlinearity=args.nonlinearity) # optionally, load model parameters from file if args.pretrain_model is not None: model.load_state_dict(torch.load(args.pretrain_model)) # set up optimizer optimizer = optim.Adam(model.parameters(), betas=(0.9, 0.999), ## default eps=1e-08, ## default lr=args.learning_rate) # set up early stopping early_stop = EarlyStopping(patience=args.patience) # set up training schedule file sched_filename = "training_schedule-" + str(args.sample_idx + 1) + ".csv" sched_file = os.path.join(args.output_dir, sched_filename) # iterate over epochs counter = 0 for epoch in range(args.max_epochs): # iterate over batches for batch_idx, batch in tqdm(enumerate(loader), total=len(loader)): batch, lengths = batch # increment counter counter += 1 # calculate loss log_p = model.loss(batch, lengths) loss = log_p.mean() # zero gradients, calculate new gradients, and take a step optimizer.zero_grad() loss.backward() # clip gradient if args.gradient_clip is not None: nn.utils.clip_grad_norm_(model.parameters(), args.gradient_clip) optimizer.step() # check learning rate decay if args.learning_rate_decay is not None and \ counter % args.learning_rate_decay_steps == 0: decrease_learning_rate(optimizer, multiplier=args.learning_rate_decay) # print update and write training schedule? if args.log_every_steps is not None: if counter % args.log_every_steps == 0: print_update(model, dataset, epoch, batch_idx + 1, loss.item(), args.batch_size, selfies=args.selfies) track_loss(sched_file, model, dataset, epoch, counter, loss.item(), args.batch_size) # save SMILES? if args.sample_every_steps is not None: if counter % args.sample_every_steps == 0: sample_smiles(args.output_dir, args.sample_idx, model, args.sample_size, epoch, counter) # calculate validation loss validation, lengths = dataset.get_validation(args.batch_size) validation_loss = model.loss(validation, lengths).mean().detach() # check early stopping model_filename = "model-" + str(args.sample_idx + 1) + ".pt" model_file = os.path.join(args.output_dir, model_filename) early_stop(validation_loss.item(), model, model_file, counter) if early_stop.stop: break # print update and write training schedule? if args.log_every_epochs is not None: print_update(model, dataset, epoch, 'NA', loss.item(), args.batch_size) track_loss(sched_file, model, dataset, epoch, counter, loss.item(), args.batch_size) # save SMILES? if args.sample_every_epochs is not None: sample_smiles(args.output_dir, args.sample_idx, model, args.sample_size, epoch, counter) if early_stop.stop: break # append information about final training step if args.log_every_epochs is not None or args.log_every_steps is not None: sched = pd.DataFrame({'epoch': [None], 'step': [early_stop.step_at_best], 'outcome': ['training loss'], 'value': [early_stop.best_loss]}) sched.to_csv(sched_file, index=False, mode='a', header=False) # load the best model model.load_state_dict(torch.load(model_file)) model.eval() ## enable evaluation modes # sample a set of SMILES from the final, trained model sampled_smiles = [] while len(sampled_smiles) < args.sample_size: sampled_smiles.extend(model.sample(args.batch_size, return_smiles=True)) # write sampled SMILES write_smiles(sampled_smiles, smiles_file)
39.309278
79
0.652417
import argparse import os import numpy as np import pandas as pd import random import sys import torch import torch.nn as nn import torch.optim as optim from torch.utils.data import DataLoader from tqdm import tqdm from rdkit import rdBase rdBase.DisableLog('rdApp.error') git_dir = os.path.expanduser("~/git/low-data-generative-models") python_dir = git_dir + "/python" os.chdir(python_dir) from models import RNN, OneHotRNN, EarlyStopping from datasets import SmilesDataset, SelfiesDataset, SmilesCollate from functions import decrease_learning_rate, print_update, track_loss, \ sample_smiles, write_smiles rgparse.ArgumentParser( description='Chemical structure language model interface') parser.add_argument('--smiles_file', type=str, help='location of the SMILES file to train on') parser.add_argument('--selfies', dest='selfies', action='store_true') parser.set_defaults(selfies=False) parser.add_argument('--output_dir', type=str, help='directory to save trained models to') parser.add_argument('--rnn_type', type=str, choices=['RNN', 'LSTM', 'GRU'], default='GRU', help='type of language model to train') parser.add_argument('--embedding_size', type=int, default=128, help='size of vocabulary embedding') parser.add_argument('--hidden_size', type=int, default=512, help='size of language model hidden layers') parser.add_argument('--n_layers', type=int, default=3, help='number of layers in language model') parser.add_argument('--dropout', type=float, default=0, help='amount of dropout (0-1) to apply to model') parser.add_argument('--bidirectional', type=bool, default=False, help='for LSTMs only, train a bidirectional model') parser.add_argument('--nonlinearity', type=str, choices=['tanh', 'relu'], default='tanh', help='for RNNs only, nonlinearity to use') parser.add_argument('--tie_weights', dest='tie_weights', help='require embedding/dense linear layers use the ' +\ 'same weights', action='store_true') parser.set_defaults(tie_weights=False) parser.add_argument('--learning_rate', type=float, default=0.001, help='initial learning rate') parser.add_argument('--learning_rate_decay', default=None, help='amount (0-1) to decrease learning rate by every ' +\ 'fixed number of steps') parser.add_argument('--learning_rate_decay_steps', default=10000, type=int, help='# of steps between learning rate decrements') parser.add_argument('--gradient_clip', default=None, help='amount to which to clip the gradients') parser.add_argument('--seed', type=int, default=0, help='seed for random number generator') parser.add_argument('--batch_size', type=int, default=128, help='batch size') parser.add_argument('--max_epochs', type=int, default=1000, help='maximum number of epochs to train for') parser.add_argument('--patience', type=int, default=100, help='patience for early stopping') parser.add_argument('--sample_idx', type=int, default=0, help='index of the model being trained (zero-indexed)') parser.add_argument('--sample_every_epochs', type=int, help='if set, sample SMILES from the trained model' + 'every n epochs') parser.add_argument('--sample_every_steps', type=int, help='if set, sample SMILES from the trained model' + 'every n steps') parser.add_argument('--log_every_epochs', type=int, help='log training/validation losses every n epochs') parser.add_argument('--log_every_steps', type=int, help='log training/validation losses every n steps') parser.add_argument('--sample_size', type=int, default=100000, help='size of each sample from the trained model') parser.add_argument('--pretrain_model', type=str, default=None, help='load parameters from a pretrained model') parser.add_argument('--vocab_file', type=str, default=None, help='file containing all tokens in vocabulary') parser.add_argument('--stop_if_exists', dest='stop_if_exists', action='store_true') parser.set_defaults(stop_if_exists=False) args = parser.parse_args() try: args.gradient_clip = float(args.gradient_clip) except (ValueError, TypeError): args.gradient_clip = None try: args.learning_rate_decay = float(args.learning_rate_decay) except (ValueError, TypeError): args.learning_rate_decay = None for arg in vars(args): print(arg, ": ", getattr(args, arg), "(", type(getattr(args, arg)), ")") if args.selfies: smiles_filename = "sample-" + str(args.sample_idx + 1) + "-SELFIES.smi" else: smiles_filename = "sample-" + str(args.sample_idx + 1) + "-SMILES.smi" smiles_file = os.path.join(args.output_dir, smiles_filename) if os.path.isfile(smiles_file) and args.stop_if_exists: print("output file " + smiles_file + " exists: stopping early") sys.exit() if not os.path.isdir(args.output_dir): try: os.makedirs(args.output_dir) except FileExistsError: pass eed(args.seed) random.seed(args.seed) np.random.seed(args.seed) if torch.cuda.is_available(): print("using cuda") torch.cuda.manual_seed_all(args.seed) if args.selfies: dataset = SelfiesDataset(selfies_file=args.smiles_file) else: dataset = SmilesDataset(smiles_file=args.smiles_file, vocab_file=args.vocab_file) loader = DataLoader(dataset, batch_size=args.batch_size, shuffle=True, drop_last=True, collate_fn=SmilesCollate(dataset.vocabulary)) if args.embedding_size > 0: model = RNN(vocabulary=dataset.vocabulary, rnn_type=args.rnn_type, embedding_size=args.embedding_size, hidden_size=args.hidden_size, n_layers=args.n_layers, dropout=args.dropout, bidirectional=args.bidirectional, tie_weights=args.tie_weights, nonlinearity=args.nonlinearity) else: model = OneHotRNN(vocabulary=dataset.vocabulary, rnn_type=args.rnn_type, hidden_size=args.hidden_size, n_layers=args.n_layers, dropout=args.dropout, bidirectional=args.bidirectional, nonlinearity=args.nonlinearity) if args.pretrain_model is not None: model.load_state_dict(torch.load(args.pretrain_model)) optimizer = optim.Adam(model.parameters(), betas=(0.9, 0.999), eps=1e-08, lr=args.learning_rate) early_stop = EarlyStopping(patience=args.patience) sched_filename = "training_schedule-" + str(args.sample_idx + 1) + ".csv" sched_file = os.path.join(args.output_dir, sched_filename) counter = 0 for epoch in range(args.max_epochs): for batch_idx, batch in tqdm(enumerate(loader), total=len(loader)): batch, lengths = batch counter += 1 log_p = model.loss(batch, lengths) loss = log_p.mean() optimizer.zero_grad() loss.backward() if args.gradient_clip is not None: nn.utils.clip_grad_norm_(model.parameters(), args.gradient_clip) optimizer.step() if args.learning_rate_decay is not None and \ counter % args.learning_rate_decay_steps == 0: decrease_learning_rate(optimizer, multiplier=args.learning_rate_decay) if args.log_every_steps is not None: if counter % args.log_every_steps == 0: print_update(model, dataset, epoch, batch_idx + 1, loss.item(), args.batch_size, selfies=args.selfies) track_loss(sched_file, model, dataset, epoch, counter, loss.item(), args.batch_size) if args.sample_every_steps is not None: if counter % args.sample_every_steps == 0: sample_smiles(args.output_dir, args.sample_idx, model, args.sample_size, epoch, counter) validation, lengths = dataset.get_validation(args.batch_size) validation_loss = model.loss(validation, lengths).mean().detach() model_filename = "model-" + str(args.sample_idx + 1) + ".pt" model_file = os.path.join(args.output_dir, model_filename) early_stop(validation_loss.item(), model, model_file, counter) if early_stop.stop: break if args.log_every_epochs is not None: print_update(model, dataset, epoch, 'NA', loss.item(), args.batch_size) track_loss(sched_file, model, dataset, epoch, counter, loss.item(), args.batch_size) if args.sample_every_epochs is not None: sample_smiles(args.output_dir, args.sample_idx, model, args.sample_size, epoch, counter) if early_stop.stop: break if args.log_every_epochs is not None or args.log_every_steps is not None: sched = pd.DataFrame({'epoch': [None], 'step': [early_stop.step_at_best], 'outcome': ['training loss'], 'value': [early_stop.best_loss]}) sched.to_csv(sched_file, index=False, mode='a', header=False) model.load_state_dict(torch.load(model_file)) model.eval() ile len(sampled_smiles) < args.sample_size: sampled_smiles.extend(model.sample(args.batch_size, return_smiles=True)) write_smiles(sampled_smiles, smiles_file)
true
true
f70ebc9c8b640280b26f6fd66deed90179dd2995
31,646
py
Python
pytype/analyze.py
Hirni-Meshram4/pytype
24b04237c15822b26f51c545646e4e5fff1ab709
[ "Apache-2.0" ]
null
null
null
pytype/analyze.py
Hirni-Meshram4/pytype
24b04237c15822b26f51c545646e4e5fff1ab709
[ "Apache-2.0" ]
null
null
null
pytype/analyze.py
Hirni-Meshram4/pytype
24b04237c15822b26f51c545646e4e5fff1ab709
[ "Apache-2.0" ]
null
null
null
"""Code for checking and inferring types.""" import collections import logging import re import subprocess from typing import Any, Dict, Union from pytype import abstract from pytype import abstract_utils from pytype import convert_structural from pytype import debug from pytype import function from pytype import metrics from pytype import special_builtins from pytype import state as frame_state from pytype import vm from pytype.overlays import typing_overlay from pytype.pytd import builtins from pytype.pytd import escape from pytype.pytd import optimize from pytype.pytd import pytd from pytype.pytd import pytd_utils from pytype.pytd import visitors from pytype.typegraph import cfg log = logging.getLogger(__name__) # Most interpreter functions (including lambdas) need to be analyzed as # stand-alone functions. The exceptions are comprehensions and generators, which # have names like "<listcomp>" and "<genexpr>". _SKIP_FUNCTION_RE = re.compile("<(?!lambda).+>$") CallRecord = collections.namedtuple( "CallRecord", ["node", "function", "signatures", "positional_arguments", "keyword_arguments", "return_value"]) # How deep to follow call chains: INIT_MAXIMUM_DEPTH = 4 # during module loading MAXIMUM_DEPTH = 3 # during non-quick analysis QUICK_CHECK_MAXIMUM_DEPTH = 2 # during quick checking QUICK_INFER_MAXIMUM_DEPTH = 1 # during quick inference class _Initializing: pass class CallTracer(vm.VirtualMachine): """Virtual machine that records all function calls. Attributes: exitpoint: A CFG node representing the program exit. Needs to be set before analyze_types. """ _CONSTRUCTORS = ("__new__", "__init__") def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self._unknowns = {} self._calls = set() self._method_calls = set() # Used by init_class. self._instance_cache: Dict[Any, Union[_Initializing, cfg.Variable]] = {} # Used by call_init. Can differ from _instance_cache because we also call # __init__ on classes not initialized via init_class. self._initialized_instances = set() self._interpreter_functions = [] self._interpreter_classes = [] self._analyzed_functions = set() self._analyzed_classes = set() self._generated_classes = {} self.exitpoint = None def create_varargs(self, node): value = abstract.Instance(self.convert.tuple_type, self) value.merge_instance_type_parameter( node, abstract_utils.T, self.convert.create_new_unknown(node)) return value.to_variable(node) def create_kwargs(self, node): key_type = self.convert.primitive_class_instances[str].to_variable(node) value_type = self.convert.create_new_unknown(node) kwargs = abstract.Instance(self.convert.dict_type, self) kwargs.merge_instance_type_parameter(node, abstract_utils.K, key_type) kwargs.merge_instance_type_parameter(node, abstract_utils.V, value_type) return kwargs.to_variable(node) def create_method_arguments(self, node, method, use_defaults=False): """Create arguments for the given method. Creates Unknown objects as arguments for the given method. Note that we don't need to take parameter annotations into account as InterpreterFunction.call() will take care of that. Args: node: The current node. method: An abstract.InterpreterFunction. use_defaults: Whether to use parameter defaults for arguments. When True, unknown arguments are created with force=False, as it is fine to use Unsolvable rather than Unknown objects for type-checking defaults. Returns: A tuple of a node and a function.Args object. """ args = [] num_posargs = method.argcount(node) num_posargs_no_default = num_posargs - len(method.defaults) for i in range(num_posargs): default_idx = i - num_posargs_no_default if use_defaults and default_idx >= 0: arg = method.defaults[default_idx] else: arg = self.convert.create_new_unknown(node, force=not use_defaults) args.append(arg) kws = {} for key in method.signature.kwonly_params: if use_defaults and key in method.kw_defaults: kws[key] = method.kw_defaults[key] else: kws[key] = self.convert.create_new_unknown(node, force=not use_defaults) starargs = self.create_varargs(node) if method.has_varargs() else None starstarargs = self.create_kwargs(node) if method.has_kwargs() else None return node, function.Args(posargs=tuple(args), namedargs=kws, starargs=starargs, starstarargs=starstarargs) def call_function_with_args(self, node, val, args): """Call a function. Args: node: The given node. val: A cfg.Binding containing the function. args: A function.Args object. Returns: A tuple of (1) a node and (2) a cfg.Variable of the return value. """ fvar = val.AssignToNewVariable(node) with val.data.record_calls(): new_node, ret = self.call_function_in_frame(node, fvar, *args) return new_node, ret def call_function_in_frame(self, node, var, args, kwargs, starargs, starstarargs): frame = frame_state.SimpleFrame(node=node) self.push_frame(frame) log.info("Analyzing %r", [v.name for v in var.data]) state = frame_state.FrameState.init(node, self) state, ret = self.call_function_with_state( state, var, args, kwargs, starargs, starstarargs) self.pop_frame(frame) return state.node, ret def _maybe_fix_classmethod_cls_arg(self, node, cls, func, args): sig = func.signature if (args.posargs and sig.param_names and (sig.param_names[0] not in sig.annotations)): # fix "cls" parameter return args._replace( posargs=(cls.AssignToNewVariable(node),) + args.posargs[1:]) else: return args def maybe_analyze_method(self, node, val, cls=None): method = val.data fname = val.data.name if isinstance(method, abstract.INTERPRETER_FUNCTION_TYPES): self._analyzed_functions.add(method.get_first_opcode()) if (not self.options.analyze_annotated and (method.signature.has_return_annotation or method.has_overloads) and fname.rsplit(".", 1)[-1] not in self._CONSTRUCTORS): log.info("%r has annotations, not analyzing further.", fname) else: for f in method.iter_signature_functions(): node, args = self.create_method_arguments(node, f) if f.is_classmethod and cls: args = self._maybe_fix_classmethod_cls_arg(node, cls, f, args) node, _ = self.call_function_with_args(node, val, args) return node def _call_with_fake_args(self, node0, funcv): """Attempt to call the given function with made-up arguments.""" # TODO(tsudol): If expand this beyond __init__, need to handle # DictKeyMissing nodes = [] rets = [] for funcb in funcv.bindings: func = funcb.data log.info("Trying %s with fake arguments", func) if isinstance(func, abstract.INTERPRETER_FUNCTION_TYPES): node1, args = self.create_method_arguments(node0, func) # Once the args are generated, try calling the function. # call_function will check fallback_to_unsolvable if a DictKeyMissing or # FailedFunctionCall error is raised when the target function is called. # DictKeyMissing doesn't trigger call_with_fake_args, so that shouldn't # be raised again, and generating fake arguments should avoid any # FailedFunctionCall errors. To prevent an infinite recursion loop, set # fallback_to_unsolvable to False just in case. # This means any additional errors that may be raised will be passed to # the call_function that called this method in the first place. node2, ret = self.call_function(node1, funcb.AssignToNewVariable(), args, fallback_to_unsolvable=False) nodes.append(node2) rets.append(ret) if nodes: ret = self.join_variables(node0, rets) node = self.join_cfg_nodes(nodes) if ret.bindings: return node, ret else: node = node0 log.info("Unable to generate fake arguments for %s", funcv) return node, self.new_unsolvable(node) def analyze_method_var(self, node0, name, var, cls=None): log.info("Analyzing %s", name) node1 = node0.ConnectNew(name) for val in var.bindings: node2 = self.maybe_analyze_method(node1, val, cls) node2.ConnectTo(node0) return node0 def bind_method(self, node, name, methodvar, instance_var): bound = self.program.NewVariable() for m in methodvar.Data(node): if isinstance(m, special_builtins.ClassMethodInstance): m = m.func.data[0] is_cls = True else: is_cls = (m.isinstance_InterpreterFunction() and m.is_classmethod) bound.AddBinding(m.property_get(instance_var, is_cls), [], node) return bound def _instantiate_binding(self, node0, cls, container): """Instantiate a class binding.""" node1, new = cls.data.get_own_new(node0, cls) if not new or ( any(not isinstance(f, abstract.InterpreterFunction) for f in new.data)): # This assumes that any inherited __new__ method defined in a pyi file # returns an instance of the current class. return node0, cls.data.instantiate(node0, container=container) instance = self.program.NewVariable() nodes = [] for b in new.bindings: self._analyzed_functions.add(b.data.get_first_opcode()) node2, args = self.create_method_arguments(node1, b.data) args = self._maybe_fix_classmethod_cls_arg(node0, cls, b.data, args) node3 = node2.ConnectNew() node4, ret = self.call_function_with_args(node3, b, args) instance.PasteVariable(ret) nodes.append(node4) return self.join_cfg_nodes(nodes), instance def _instantiate_var(self, node, clsv, container): """Build an (dummy) instance from a class, for analyzing it.""" n = self.program.NewVariable() for cls in clsv.Bindings(node, strict=False): node, var = self._instantiate_binding(node, cls, container) n.PasteVariable(var) return node, n def _mark_maybe_missing_members(self, values): """Set maybe_missing_members to True on these values and their type params. Args: values: A list of BaseValue objects. On every instance among the values, recursively set maybe_missing_members to True on the instance and its type parameters. """ values = list(values) seen = set() while values: v = values.pop(0) if v not in seen: seen.add(v) if isinstance(v, abstract.SimpleValue): v.maybe_missing_members = True for child in v.instance_type_parameters.values(): values.extend(child.data) def init_class(self, node, cls, container=None, extra_key=None): """Instantiate a class, and also call __init__. Calling __init__ can be expensive, so this method caches its created instances. If you don't need __init__ called, use cls.instantiate instead. Args: node: The current node. cls: The class to instantiate. container: Optionally, a container to pass to the class's instantiate() method, so that type parameters in the container's template are instantiated to TypeParameterInstance. extra_key: Optionally, extra information about the location at which the instantion occurs. By default, this method keys on the current opcode and the class, which sometimes isn't enough to disambiguate callers that shouldn't get back the same cached instance. Returns: A tuple of node and instance variable. """ key = (self.frame and self.frame.current_opcode, extra_key, cls) instance = self._instance_cache.get(key) if not instance or isinstance(instance, _Initializing): clsvar = cls.to_variable(node) node, instance = self._instantiate_var(node, clsvar, container) if key in self._instance_cache: # We've encountered a recursive pattern such as # class A: # def __init__(self, x: "A"): ... # Calling __init__ again would lead to an infinite loop, so # we instead create an incomplete instance that will be # overwritten later. Note that we have to create a new # instance rather than using the one that we're already in # the process of initializing - otherwise, setting # maybe_missing_members to True would cause pytype to ignore # all attribute errors on self in __init__. self._mark_maybe_missing_members(instance.data) else: self._instance_cache[key] = _Initializing() node = self.call_init(node, instance) self._instance_cache[key] = instance return node, instance def _call_method(self, node, binding, method_name): node, method = self.attribute_handler.get_attribute( node, binding.data.get_class(), method_name, binding) if method: bound_method = self.bind_method( node, method_name, method, binding.AssignToNewVariable()) node = self.analyze_method_var(node, method_name, bound_method) return node def _call_init_on_binding(self, node, b): if isinstance(b.data, abstract.SimpleValue): for param in b.data.instance_type_parameters.values(): node = self.call_init(node, param) node = self._call_method(node, b, "__init__") cls = b.data.get_class() if isinstance(cls, abstract.InterpreterClass): # Call any additional initalizers the class has registered. for method in cls.additional_init_methods: node = self._call_method(node, b, method) return node def call_init(self, node, instance): # Call __init__ on each binding. for b in instance.bindings: if b.data in self._initialized_instances: continue self._initialized_instances.add(b.data) node = self._call_init_on_binding(node, b) return node def reinitialize_if_initialized(self, node, instance): if instance in self._initialized_instances: self._call_init_on_binding(node, instance.to_binding(node)) def analyze_class(self, node, val): self._analyzed_classes.add(val.data) node, instance = self.init_class(node, val.data) good_instances = [b for b in instance.bindings if val.data == b.data.cls] if not good_instances: # __new__ returned something that's not an instance of our class. instance = val.data.instantiate(node) node = self.call_init(node, instance) elif len(good_instances) != len(instance.bindings): # __new__ returned some extra possibilities we don't need. instance = self.join_bindings(node, good_instances) for instance_value in instance.data: val.data.register_canonical_instance(instance_value) methods = sorted(val.data.members.items()) while methods: name, methodvar = methods.pop(0) if name in self._CONSTRUCTORS: continue # We already called this method during initialization. for v in methodvar.data: if (self.options.bind_properties and isinstance(v, special_builtins.PropertyInstance)): for m in (v.fget, v.fset, v.fdel): if m: methods.insert(0, (name, m)) b = self.bind_method(node, name, methodvar, instance) node = self.analyze_method_var(node, name, b, val) return node def analyze_function(self, node0, val): if val.data.is_attribute_of_class: # We'll analyze this function as part of a class. log.info("Analyze functions: Skipping class method %s", val.data.name) else: node1 = node0.ConnectNew(val.data.name) node2 = self.maybe_analyze_method(node1, val) node2.ConnectTo(node0) return node0 def _should_analyze_as_interpreter_function(self, data): # We record analyzed functions by opcode rather than function object. The # two ways of recording are equivalent except for closures, which are # re-generated when the variables they close over change, but we don't want # to re-analyze them. return (isinstance(data, abstract.InterpreterFunction) and not data.is_overload and not data.is_class_builder and data.get_first_opcode() not in self._analyzed_functions and not _SKIP_FUNCTION_RE.search(data.name)) def analyze_toplevel(self, node, defs): for name, var in sorted(defs.items()): # sort, for determinicity if not self._is_typing_member(name, var): for value in var.bindings: if isinstance(value.data, abstract.InterpreterClass): new_node = self.analyze_class(node, value) elif (isinstance(value.data, abstract.INTERPRETER_FUNCTION_TYPES) and not value.data.is_overload): new_node = self.analyze_function(node, value) else: continue if new_node is not node: new_node.ConnectTo(node) # Now go through all functions and classes we haven't analyzed yet. # These are typically hidden under a decorator. # Go through classes first so that the `is_attribute_of_class` will # be set for all functions in class. for c in self._interpreter_classes: for value in c.bindings: if (isinstance(value.data, abstract.InterpreterClass) and value.data not in self._analyzed_classes): node = self.analyze_class(node, value) for f in self._interpreter_functions: for value in f.bindings: if self._should_analyze_as_interpreter_function(value.data): node = self.analyze_function(node, value) return node def analyze(self, node, defs, maximum_depth): assert not self.frame self.maximum_depth = maximum_depth self._analyzing = True node = node.ConnectNew(name="Analyze") return self.analyze_toplevel(node, defs) def trace_unknown(self, name, unknown_binding): self._unknowns[name] = unknown_binding def trace_call(self, node, func, sigs, posargs, namedargs, result): """Add an entry into the call trace. Args: node: The CFG node right after this function call. func: A cfg.Binding of a function that was called. sigs: The signatures that the function might have been called with. posargs: The positional arguments, an iterable over cfg.Value. namedargs: The keyword arguments, a dict mapping str to cfg.Value. result: A Variable of the possible result values. """ log.debug("Logging call to %r with %d args, return %r", func, len(posargs), result) args = tuple(posargs) kwargs = tuple((namedargs or {}).items()) record = CallRecord(node, func, sigs, args, kwargs, result) if isinstance(func.data, abstract.BoundPyTDFunction): self._method_calls.add(record) elif isinstance(func.data, abstract.PyTDFunction): self._calls.add(record) def trace_functiondef(self, f): self._interpreter_functions.append(f) def trace_classdef(self, c): self._interpreter_classes.append(c) def trace_namedtuple(self, nt): # All namedtuple instances with the same name are equal, so it's fine to # overwrite previous instances. self._generated_classes[nt.name] = nt def pytd_classes_for_unknowns(self): classes = [] for name, val in self._unknowns.items(): if val in val.variable.Filter(self.exitpoint, strict=False): classes.append(val.data.to_structural_def(self.exitpoint, name)) return classes def pytd_for_types(self, defs): # If a variable is annotated, we'll always output that type. annotated_names = set() data = [] pytd_convert = self.convert.pytd_convert annots = abstract_utils.get_annotations_dict(defs) for name, t in pytd_convert.annotations_to_instance_types( self.exitpoint, annots): annotated_names.add(name) data.append(pytd.Constant(name, t)) for name, var in defs.items(): if (name in abstract_utils.TOP_LEVEL_IGNORE or name in annotated_names or self._is_typing_member(name, var)): continue options = var.FilteredData(self.exitpoint, strict=False) if (len(options) > 1 and not all(isinstance(o, abstract.FUNCTION_TYPES) for o in options)): if all(isinstance(o, (abstract.ParameterizedClass, abstract.TypeParameter, abstract.Union)) for o in options ) and self.options.preserve_union_macros: # type alias data.append(pytd_utils.JoinTypes(t.to_pytd_def(self.exitpoint, name) for t in options)) else: # It's ambiguous whether this is a type, a function or something # else, so encode it as a constant. combined_types = pytd_utils.JoinTypes(t.to_type(self.exitpoint) for t in options) data.append(pytd.Constant(name, combined_types)) elif options: for option in options: try: d = option.to_pytd_def(self.exitpoint, name) # Deep definition except NotImplementedError: d = option.to_type(self.exitpoint) # Type only if isinstance(d, pytd.NothingType): if isinstance(option, abstract.Empty): d = pytd.AnythingType() else: assert isinstance(option, typing_overlay.NoReturn) if isinstance(d, pytd.Type) and not isinstance(d, pytd.TypeParameter): data.append(pytd.Constant(name, d)) else: data.append(d) else: log.error("No visible options for %s", name) data.append(pytd.Constant(name, pytd.AnythingType())) return pytd_utils.WrapTypeDeclUnit("inferred", data) @staticmethod def _call_traces_to_function(call_traces, name_transform=lambda x: x): funcs = collections.defaultdict(pytd_utils.OrderedSet) for node, func, sigs, args, kws, retvar in call_traces: # The lengths may be different in the presence of optional and kw args. arg_names = max((sig.get_positional_names() for sig in sigs), key=len) for i in range(len(arg_names)): if not isinstance(func.data, abstract.BoundFunction) or i > 0: arg_names[i] = function.argname(i) arg_types = (a.data.to_type(node) for a in args) ret = pytd_utils.JoinTypes(t.to_type(node) for t in retvar.data) starargs = None starstarargs = None funcs[func.data.name].add(pytd.Signature( tuple(pytd.Parameter(n, t, False, False, None) for n, t in zip(arg_names, arg_types)) + tuple(pytd.Parameter(name, a.data.to_type(node), False, False, None) for name, a in kws), starargs, starstarargs, ret, exceptions=(), template=())) functions = [] for name, signatures in funcs.items(): functions.append(pytd.Function(name_transform(name), tuple(signatures), pytd.MethodTypes.METHOD)) return functions def _is_typing_member(self, name, var): for module_name in ("typing", "typing_extensions"): if module_name not in self.loaded_overlays: continue module = self.loaded_overlays[module_name].get_module(name) if name in module.members and module.members[name].data == var.data: return True return False def pytd_functions_for_call_traces(self): return self._call_traces_to_function(self._calls, escape.pack_partial) def pytd_classes_for_call_traces(self): class_to_records = collections.defaultdict(list) for call_record in self._method_calls: args = call_record.positional_arguments if not any(isinstance(a.data, abstract.Unknown) for a in args): # We don't need to record call signatures that don't involve # unknowns - there's nothing to solve for. continue cls = args[0].data.get_class() if isinstance(cls, abstract.PyTDClass): class_to_records[cls].append(call_record) classes = [] for cls, call_records in class_to_records.items(): full_name = cls.module + "." + cls.name if cls.module else cls.name classes.append(pytd.Class( name=escape.pack_partial(full_name), metaclass=None, parents=(pytd.NamedType("builtins.object"),), # not used in solver methods=tuple(self._call_traces_to_function(call_records)), constants=(), classes=(), decorators=(), slots=None, template=(), )) return classes def pytd_classes_for_namedtuple_instances(self): return tuple(v.generate_ast() for v in self._generated_classes.values()) def compute_types(self, defs): classes = (tuple(self.pytd_classes_for_unknowns()) + tuple(self.pytd_classes_for_call_traces()) + self.pytd_classes_for_namedtuple_instances()) functions = tuple(self.pytd_functions_for_call_traces()) aliases = () # aliases are instead recorded as constants ty = pytd_utils.Concat( self.pytd_for_types(defs), pytd_utils.CreateModule("unknowns", classes=classes, functions=functions, aliases=aliases)) ty = ty.Visit(optimize.CombineReturnsAndExceptions()) ty = ty.Visit(optimize.PullInMethodClasses()) ty = ty.Visit(visitors.DefaceUnresolved( [ty, self.loader.concat_all()], escape.UNKNOWN)) return ty.Visit(visitors.AdjustTypeParameters()) def _check_return(self, node, actual, formal): if not self.options.report_errors: return True views = abstract_utils.get_views([actual], node) # Check for typevars in the return value first, since bad_matches # expects not to get any. bad = [view for view in views if actual in view and view[actual].data.formal] if not bad: bad = self.matcher(node).bad_matches(actual, formal) if bad: self.errorlog.bad_return_type( self.frames, node, formal, actual, bad) return not bad def check_types(src, filename, errorlog, options, loader, deep=True, init_maximum_depth=INIT_MAXIMUM_DEPTH, maximum_depth=None, **kwargs): """Verify the Python code.""" tracer = CallTracer(errorlog=errorlog, options=options, generate_unknowns=False, loader=loader, **kwargs) loc, defs = tracer.run_program(src, filename, init_maximum_depth) snapshotter = metrics.get_metric("memory", metrics.Snapshot) snapshotter.take_snapshot("analyze:check_types:tracer") if deep: if maximum_depth is None: maximum_depth = ( QUICK_CHECK_MAXIMUM_DEPTH if options.quick else MAXIMUM_DEPTH) tracer.analyze(loc, defs, maximum_depth=maximum_depth) snapshotter.take_snapshot("analyze:check_types:post") _maybe_output_debug(options, tracer.program) def infer_types(src, errorlog, options, loader, filename=None, deep=True, init_maximum_depth=INIT_MAXIMUM_DEPTH, show_library_calls=False, maximum_depth=None, tracer_vm=None, **kwargs): """Given Python source return its types. Args: src: A string containing Python source code. errorlog: Where error messages go. Instance of errors.ErrorLog. options: config.Options object loader: A load_pytd.Loader instance to load PYI information. filename: Filename of the program we're parsing. deep: If True, analyze all functions, even the ones not called by the main execution flow. init_maximum_depth: Depth of analysis during module loading. show_library_calls: If True, call traces are kept in the output. maximum_depth: Depth of the analysis. Default: unlimited. tracer_vm: An instance of CallTracer, in case the caller wants to instantiate and retain the vm used for type inference. **kwargs: Additional parameters to pass to vm.VirtualMachine Returns: A tuple of (ast: TypeDeclUnit, builtins: TypeDeclUnit) Raises: AssertionError: In case of a bad parameter combination. """ # If the caller has passed in a vm, use that. if tracer_vm: assert isinstance(tracer_vm, CallTracer) tracer = tracer_vm else: tracer = CallTracer(errorlog=errorlog, options=options, generate_unknowns=options.protocols, store_all_calls=not deep, loader=loader, **kwargs) loc, defs = tracer.run_program(src, filename, init_maximum_depth) log.info("===Done running definitions and module-level code===") snapshotter = metrics.get_metric("memory", metrics.Snapshot) snapshotter.take_snapshot("analyze:infer_types:tracer") if deep: if maximum_depth is None: if not options.quick: maximum_depth = MAXIMUM_DEPTH elif options.analyze_annotated: # Since there's no point in analyzing annotated functions for inference, # the presence of this option means that the user wants checking, too. maximum_depth = QUICK_CHECK_MAXIMUM_DEPTH else: maximum_depth = QUICK_INFER_MAXIMUM_DEPTH tracer.exitpoint = tracer.analyze(loc, defs, maximum_depth) else: tracer.exitpoint = loc snapshotter.take_snapshot("analyze:infer_types:post") ast = tracer.compute_types(defs) ast = tracer.loader.resolve_ast(ast) if tracer.has_unknown_wildcard_imports or any( a in defs for a in abstract_utils.DYNAMIC_ATTRIBUTE_MARKERS): if "__getattr__" not in ast: ast = pytd_utils.Concat( ast, builtins.GetDefaultAst(options.python_version)) # If merged with other if statement, triggers a ValueError: Unresolved class # when attempts to load from the protocols file if options.protocols: protocols_pytd = tracer.loader.import_name("protocols") else: protocols_pytd = None builtins_pytd = tracer.loader.concat_all() # Insert type parameters, where appropriate ast = ast.Visit(visitors.CreateTypeParametersForSignatures()) if options.protocols: log.info("=========== PyTD to solve =============\n%s", pytd_utils.Print(ast)) ast = convert_structural.convert_pytd(ast, builtins_pytd, protocols_pytd) elif not show_library_calls: log.info("Solving is turned off. Discarding call traces.") # Rename remaining "~unknown" to "?" ast = ast.Visit(visitors.RemoveUnknownClasses()) # Remove "~list" etc.: ast = convert_structural.extract_local(ast) _maybe_output_debug(options, tracer.program) return ast, builtins_pytd def _maybe_output_debug(options, program): """Maybe emit debugging output.""" if options.output_cfg or options.output_typegraph: dot = debug.program_to_dot(program, set([]), bool(options.output_cfg)) svg_file = options.output_cfg or options.output_typegraph with subprocess.Popen( ["/usr/bin/dot", "-T", "svg", "-o", svg_file], stdin=subprocess.PIPE, universal_newlines=True) as proc: (_, stderr) = proc.communicate(dot) if stderr: log.info("Failed to create %s: %s", svg_file, stderr) if options.output_debug: text = debug.program_to_text(program) if options.output_debug == "-": log.info("=========== Program Dump =============\n%s", text) else: with options.open_function(options.output_debug, "w") as fi: fi.write(text)
41.694335
80
0.684573
import collections import logging import re import subprocess from typing import Any, Dict, Union from pytype import abstract from pytype import abstract_utils from pytype import convert_structural from pytype import debug from pytype import function from pytype import metrics from pytype import special_builtins from pytype import state as frame_state from pytype import vm from pytype.overlays import typing_overlay from pytype.pytd import builtins from pytype.pytd import escape from pytype.pytd import optimize from pytype.pytd import pytd from pytype.pytd import pytd_utils from pytype.pytd import visitors from pytype.typegraph import cfg log = logging.getLogger(__name__) _SKIP_FUNCTION_RE = re.compile("<(?!lambda).+>$") CallRecord = collections.namedtuple( "CallRecord", ["node", "function", "signatures", "positional_arguments", "keyword_arguments", "return_value"]) INIT_MAXIMUM_DEPTH = 4 MAXIMUM_DEPTH = 3 QUICK_CHECK_MAXIMUM_DEPTH = 2 QUICK_INFER_MAXIMUM_DEPTH = 1 class _Initializing: pass class CallTracer(vm.VirtualMachine): _CONSTRUCTORS = ("__new__", "__init__") def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self._unknowns = {} self._calls = set() self._method_calls = set() self._instance_cache: Dict[Any, Union[_Initializing, cfg.Variable]] = {} self._initialized_instances = set() self._interpreter_functions = [] self._interpreter_classes = [] self._analyzed_functions = set() self._analyzed_classes = set() self._generated_classes = {} self.exitpoint = None def create_varargs(self, node): value = abstract.Instance(self.convert.tuple_type, self) value.merge_instance_type_parameter( node, abstract_utils.T, self.convert.create_new_unknown(node)) return value.to_variable(node) def create_kwargs(self, node): key_type = self.convert.primitive_class_instances[str].to_variable(node) value_type = self.convert.create_new_unknown(node) kwargs = abstract.Instance(self.convert.dict_type, self) kwargs.merge_instance_type_parameter(node, abstract_utils.K, key_type) kwargs.merge_instance_type_parameter(node, abstract_utils.V, value_type) return kwargs.to_variable(node) def create_method_arguments(self, node, method, use_defaults=False): args = [] num_posargs = method.argcount(node) num_posargs_no_default = num_posargs - len(method.defaults) for i in range(num_posargs): default_idx = i - num_posargs_no_default if use_defaults and default_idx >= 0: arg = method.defaults[default_idx] else: arg = self.convert.create_new_unknown(node, force=not use_defaults) args.append(arg) kws = {} for key in method.signature.kwonly_params: if use_defaults and key in method.kw_defaults: kws[key] = method.kw_defaults[key] else: kws[key] = self.convert.create_new_unknown(node, force=not use_defaults) starargs = self.create_varargs(node) if method.has_varargs() else None starstarargs = self.create_kwargs(node) if method.has_kwargs() else None return node, function.Args(posargs=tuple(args), namedargs=kws, starargs=starargs, starstarargs=starstarargs) def call_function_with_args(self, node, val, args): fvar = val.AssignToNewVariable(node) with val.data.record_calls(): new_node, ret = self.call_function_in_frame(node, fvar, *args) return new_node, ret def call_function_in_frame(self, node, var, args, kwargs, starargs, starstarargs): frame = frame_state.SimpleFrame(node=node) self.push_frame(frame) log.info("Analyzing %r", [v.name for v in var.data]) state = frame_state.FrameState.init(node, self) state, ret = self.call_function_with_state( state, var, args, kwargs, starargs, starstarargs) self.pop_frame(frame) return state.node, ret def _maybe_fix_classmethod_cls_arg(self, node, cls, func, args): sig = func.signature if (args.posargs and sig.param_names and (sig.param_names[0] not in sig.annotations)): return args._replace( posargs=(cls.AssignToNewVariable(node),) + args.posargs[1:]) else: return args def maybe_analyze_method(self, node, val, cls=None): method = val.data fname = val.data.name if isinstance(method, abstract.INTERPRETER_FUNCTION_TYPES): self._analyzed_functions.add(method.get_first_opcode()) if (not self.options.analyze_annotated and (method.signature.has_return_annotation or method.has_overloads) and fname.rsplit(".", 1)[-1] not in self._CONSTRUCTORS): log.info("%r has annotations, not analyzing further.", fname) else: for f in method.iter_signature_functions(): node, args = self.create_method_arguments(node, f) if f.is_classmethod and cls: args = self._maybe_fix_classmethod_cls_arg(node, cls, f, args) node, _ = self.call_function_with_args(node, val, args) return node def _call_with_fake_args(self, node0, funcv): nodes = [] rets = [] for funcb in funcv.bindings: func = funcb.data log.info("Trying %s with fake arguments", func) if isinstance(func, abstract.INTERPRETER_FUNCTION_TYPES): node1, args = self.create_method_arguments(node0, func) node2, ret = self.call_function(node1, funcb.AssignToNewVariable(), args, fallback_to_unsolvable=False) nodes.append(node2) rets.append(ret) if nodes: ret = self.join_variables(node0, rets) node = self.join_cfg_nodes(nodes) if ret.bindings: return node, ret else: node = node0 log.info("Unable to generate fake arguments for %s", funcv) return node, self.new_unsolvable(node) def analyze_method_var(self, node0, name, var, cls=None): log.info("Analyzing %s", name) node1 = node0.ConnectNew(name) for val in var.bindings: node2 = self.maybe_analyze_method(node1, val, cls) node2.ConnectTo(node0) return node0 def bind_method(self, node, name, methodvar, instance_var): bound = self.program.NewVariable() for m in methodvar.Data(node): if isinstance(m, special_builtins.ClassMethodInstance): m = m.func.data[0] is_cls = True else: is_cls = (m.isinstance_InterpreterFunction() and m.is_classmethod) bound.AddBinding(m.property_get(instance_var, is_cls), [], node) return bound def _instantiate_binding(self, node0, cls, container): node1, new = cls.data.get_own_new(node0, cls) if not new or ( any(not isinstance(f, abstract.InterpreterFunction) for f in new.data)): return node0, cls.data.instantiate(node0, container=container) instance = self.program.NewVariable() nodes = [] for b in new.bindings: self._analyzed_functions.add(b.data.get_first_opcode()) node2, args = self.create_method_arguments(node1, b.data) args = self._maybe_fix_classmethod_cls_arg(node0, cls, b.data, args) node3 = node2.ConnectNew() node4, ret = self.call_function_with_args(node3, b, args) instance.PasteVariable(ret) nodes.append(node4) return self.join_cfg_nodes(nodes), instance def _instantiate_var(self, node, clsv, container): n = self.program.NewVariable() for cls in clsv.Bindings(node, strict=False): node, var = self._instantiate_binding(node, cls, container) n.PasteVariable(var) return node, n def _mark_maybe_missing_members(self, values): values = list(values) seen = set() while values: v = values.pop(0) if v not in seen: seen.add(v) if isinstance(v, abstract.SimpleValue): v.maybe_missing_members = True for child in v.instance_type_parameters.values(): values.extend(child.data) def init_class(self, node, cls, container=None, extra_key=None): key = (self.frame and self.frame.current_opcode, extra_key, cls) instance = self._instance_cache.get(key) if not instance or isinstance(instance, _Initializing): clsvar = cls.to_variable(node) node, instance = self._instantiate_var(node, clsvar, container) if key in self._instance_cache: # class A: # def __init__(self, x: "A"): ... # Calling __init__ again would lead to an infinite loop, so # we instead create an incomplete instance that will be # overwritten later. Note that we have to create a new # instance rather than using the one that we're already in self._mark_maybe_missing_members(instance.data) else: self._instance_cache[key] = _Initializing() node = self.call_init(node, instance) self._instance_cache[key] = instance return node, instance def _call_method(self, node, binding, method_name): node, method = self.attribute_handler.get_attribute( node, binding.data.get_class(), method_name, binding) if method: bound_method = self.bind_method( node, method_name, method, binding.AssignToNewVariable()) node = self.analyze_method_var(node, method_name, bound_method) return node def _call_init_on_binding(self, node, b): if isinstance(b.data, abstract.SimpleValue): for param in b.data.instance_type_parameters.values(): node = self.call_init(node, param) node = self._call_method(node, b, "__init__") cls = b.data.get_class() if isinstance(cls, abstract.InterpreterClass): for method in cls.additional_init_methods: node = self._call_method(node, b, method) return node def call_init(self, node, instance): for b in instance.bindings: if b.data in self._initialized_instances: continue self._initialized_instances.add(b.data) node = self._call_init_on_binding(node, b) return node def reinitialize_if_initialized(self, node, instance): if instance in self._initialized_instances: self._call_init_on_binding(node, instance.to_binding(node)) def analyze_class(self, node, val): self._analyzed_classes.add(val.data) node, instance = self.init_class(node, val.data) good_instances = [b for b in instance.bindings if val.data == b.data.cls] if not good_instances: instance = val.data.instantiate(node) node = self.call_init(node, instance) elif len(good_instances) != len(instance.bindings): # __new__ returned some extra possibilities we don't need. instance = self.join_bindings(node, good_instances) for instance_value in instance.data: val.data.register_canonical_instance(instance_value) methods = sorted(val.data.members.items()) while methods: name, methodvar = methods.pop(0) if name in self._CONSTRUCTORS: continue for v in methodvar.data: if (self.options.bind_properties and isinstance(v, special_builtins.PropertyInstance)): for m in (v.fget, v.fset, v.fdel): if m: methods.insert(0, (name, m)) b = self.bind_method(node, name, methodvar, instance) node = self.analyze_method_var(node, name, b, val) return node def analyze_function(self, node0, val): if val.data.is_attribute_of_class: log.info("Analyze functions: Skipping class method %s", val.data.name) else: node1 = node0.ConnectNew(val.data.name) node2 = self.maybe_analyze_method(node1, val) node2.ConnectTo(node0) return node0 def _should_analyze_as_interpreter_function(self, data): # We record analyzed functions by opcode rather than function object. The # two ways of recording are equivalent except for closures, which are # re-generated when the variables they close over change, but we don't want return (isinstance(data, abstract.InterpreterFunction) and not data.is_overload and not data.is_class_builder and data.get_first_opcode() not in self._analyzed_functions and not _SKIP_FUNCTION_RE.search(data.name)) def analyze_toplevel(self, node, defs): for name, var in sorted(defs.items()): if not self._is_typing_member(name, var): for value in var.bindings: if isinstance(value.data, abstract.InterpreterClass): new_node = self.analyze_class(node, value) elif (isinstance(value.data, abstract.INTERPRETER_FUNCTION_TYPES) and not value.data.is_overload): new_node = self.analyze_function(node, value) else: continue if new_node is not node: new_node.ConnectTo(node) # These are typically hidden under a decorator. # Go through classes first so that the `is_attribute_of_class` will # be set for all functions in class. for c in self._interpreter_classes: for value in c.bindings: if (isinstance(value.data, abstract.InterpreterClass) and value.data not in self._analyzed_classes): node = self.analyze_class(node, value) for f in self._interpreter_functions: for value in f.bindings: if self._should_analyze_as_interpreter_function(value.data): node = self.analyze_function(node, value) return node def analyze(self, node, defs, maximum_depth): assert not self.frame self.maximum_depth = maximum_depth self._analyzing = True node = node.ConnectNew(name="Analyze") return self.analyze_toplevel(node, defs) def trace_unknown(self, name, unknown_binding): self._unknowns[name] = unknown_binding def trace_call(self, node, func, sigs, posargs, namedargs, result): log.debug("Logging call to %r with %d args, return %r", func, len(posargs), result) args = tuple(posargs) kwargs = tuple((namedargs or {}).items()) record = CallRecord(node, func, sigs, args, kwargs, result) if isinstance(func.data, abstract.BoundPyTDFunction): self._method_calls.add(record) elif isinstance(func.data, abstract.PyTDFunction): self._calls.add(record) def trace_functiondef(self, f): self._interpreter_functions.append(f) def trace_classdef(self, c): self._interpreter_classes.append(c) def trace_namedtuple(self, nt): # All namedtuple instances with the same name are equal, so it's fine to self._generated_classes[nt.name] = nt def pytd_classes_for_unknowns(self): classes = [] for name, val in self._unknowns.items(): if val in val.variable.Filter(self.exitpoint, strict=False): classes.append(val.data.to_structural_def(self.exitpoint, name)) return classes def pytd_for_types(self, defs): annotated_names = set() data = [] pytd_convert = self.convert.pytd_convert annots = abstract_utils.get_annotations_dict(defs) for name, t in pytd_convert.annotations_to_instance_types( self.exitpoint, annots): annotated_names.add(name) data.append(pytd.Constant(name, t)) for name, var in defs.items(): if (name in abstract_utils.TOP_LEVEL_IGNORE or name in annotated_names or self._is_typing_member(name, var)): continue options = var.FilteredData(self.exitpoint, strict=False) if (len(options) > 1 and not all(isinstance(o, abstract.FUNCTION_TYPES) for o in options)): if all(isinstance(o, (abstract.ParameterizedClass, abstract.TypeParameter, abstract.Union)) for o in options ) and self.options.preserve_union_macros: # type alias data.append(pytd_utils.JoinTypes(t.to_pytd_def(self.exitpoint, name) for t in options)) else: # It's ambiguous whether this is a type, a function or something combined_types = pytd_utils.JoinTypes(t.to_type(self.exitpoint) for t in options) data.append(pytd.Constant(name, combined_types)) elif options: for option in options: try: d = option.to_pytd_def(self.exitpoint, name) except NotImplementedError: d = option.to_type(self.exitpoint) if isinstance(d, pytd.NothingType): if isinstance(option, abstract.Empty): d = pytd.AnythingType() else: assert isinstance(option, typing_overlay.NoReturn) if isinstance(d, pytd.Type) and not isinstance(d, pytd.TypeParameter): data.append(pytd.Constant(name, d)) else: data.append(d) else: log.error("No visible options for %s", name) data.append(pytd.Constant(name, pytd.AnythingType())) return pytd_utils.WrapTypeDeclUnit("inferred", data) @staticmethod def _call_traces_to_function(call_traces, name_transform=lambda x: x): funcs = collections.defaultdict(pytd_utils.OrderedSet) for node, func, sigs, args, kws, retvar in call_traces: arg_names = max((sig.get_positional_names() for sig in sigs), key=len) for i in range(len(arg_names)): if not isinstance(func.data, abstract.BoundFunction) or i > 0: arg_names[i] = function.argname(i) arg_types = (a.data.to_type(node) for a in args) ret = pytd_utils.JoinTypes(t.to_type(node) for t in retvar.data) starargs = None starstarargs = None funcs[func.data.name].add(pytd.Signature( tuple(pytd.Parameter(n, t, False, False, None) for n, t in zip(arg_names, arg_types)) + tuple(pytd.Parameter(name, a.data.to_type(node), False, False, None) for name, a in kws), starargs, starstarargs, ret, exceptions=(), template=())) functions = [] for name, signatures in funcs.items(): functions.append(pytd.Function(name_transform(name), tuple(signatures), pytd.MethodTypes.METHOD)) return functions def _is_typing_member(self, name, var): for module_name in ("typing", "typing_extensions"): if module_name not in self.loaded_overlays: continue module = self.loaded_overlays[module_name].get_module(name) if name in module.members and module.members[name].data == var.data: return True return False def pytd_functions_for_call_traces(self): return self._call_traces_to_function(self._calls, escape.pack_partial) def pytd_classes_for_call_traces(self): class_to_records = collections.defaultdict(list) for call_record in self._method_calls: args = call_record.positional_arguments if not any(isinstance(a.data, abstract.Unknown) for a in args): continue cls = args[0].data.get_class() if isinstance(cls, abstract.PyTDClass): class_to_records[cls].append(call_record) classes = [] for cls, call_records in class_to_records.items(): full_name = cls.module + "." + cls.name if cls.module else cls.name classes.append(pytd.Class( name=escape.pack_partial(full_name), metaclass=None, parents=(pytd.NamedType("builtins.object"),), # not used in solver methods=tuple(self._call_traces_to_function(call_records)), constants=(), classes=(), decorators=(), slots=None, template=(), )) return classes def pytd_classes_for_namedtuple_instances(self): return tuple(v.generate_ast() for v in self._generated_classes.values()) def compute_types(self, defs): classes = (tuple(self.pytd_classes_for_unknowns()) + tuple(self.pytd_classes_for_call_traces()) + self.pytd_classes_for_namedtuple_instances()) functions = tuple(self.pytd_functions_for_call_traces()) aliases = () # aliases are instead recorded as constants ty = pytd_utils.Concat( self.pytd_for_types(defs), pytd_utils.CreateModule("unknowns", classes=classes, functions=functions, aliases=aliases)) ty = ty.Visit(optimize.CombineReturnsAndExceptions()) ty = ty.Visit(optimize.PullInMethodClasses()) ty = ty.Visit(visitors.DefaceUnresolved( [ty, self.loader.concat_all()], escape.UNKNOWN)) return ty.Visit(visitors.AdjustTypeParameters()) def _check_return(self, node, actual, formal): if not self.options.report_errors: return True views = abstract_utils.get_views([actual], node) # Check for typevars in the return value first, since bad_matches # expects not to get any. bad = [view for view in views if actual in view and view[actual].data.formal] if not bad: bad = self.matcher(node).bad_matches(actual, formal) if bad: self.errorlog.bad_return_type( self.frames, node, formal, actual, bad) return not bad def check_types(src, filename, errorlog, options, loader, deep=True, init_maximum_depth=INIT_MAXIMUM_DEPTH, maximum_depth=None, **kwargs): tracer = CallTracer(errorlog=errorlog, options=options, generate_unknowns=False, loader=loader, **kwargs) loc, defs = tracer.run_program(src, filename, init_maximum_depth) snapshotter = metrics.get_metric("memory", metrics.Snapshot) snapshotter.take_snapshot("analyze:check_types:tracer") if deep: if maximum_depth is None: maximum_depth = ( QUICK_CHECK_MAXIMUM_DEPTH if options.quick else MAXIMUM_DEPTH) tracer.analyze(loc, defs, maximum_depth=maximum_depth) snapshotter.take_snapshot("analyze:check_types:post") _maybe_output_debug(options, tracer.program) def infer_types(src, errorlog, options, loader, filename=None, deep=True, init_maximum_depth=INIT_MAXIMUM_DEPTH, show_library_calls=False, maximum_depth=None, tracer_vm=None, **kwargs): # If the caller has passed in a vm, use that. if tracer_vm: assert isinstance(tracer_vm, CallTracer) tracer = tracer_vm else: tracer = CallTracer(errorlog=errorlog, options=options, generate_unknowns=options.protocols, store_all_calls=not deep, loader=loader, **kwargs) loc, defs = tracer.run_program(src, filename, init_maximum_depth) log.info("===Done running definitions and module-level code===") snapshotter = metrics.get_metric("memory", metrics.Snapshot) snapshotter.take_snapshot("analyze:infer_types:tracer") if deep: if maximum_depth is None: if not options.quick: maximum_depth = MAXIMUM_DEPTH elif options.analyze_annotated: # Since there's no point in analyzing annotated functions for inference, maximum_depth = QUICK_CHECK_MAXIMUM_DEPTH else: maximum_depth = QUICK_INFER_MAXIMUM_DEPTH tracer.exitpoint = tracer.analyze(loc, defs, maximum_depth) else: tracer.exitpoint = loc snapshotter.take_snapshot("analyze:infer_types:post") ast = tracer.compute_types(defs) ast = tracer.loader.resolve_ast(ast) if tracer.has_unknown_wildcard_imports or any( a in defs for a in abstract_utils.DYNAMIC_ATTRIBUTE_MARKERS): if "__getattr__" not in ast: ast = pytd_utils.Concat( ast, builtins.GetDefaultAst(options.python_version)) if options.protocols: protocols_pytd = tracer.loader.import_name("protocols") else: protocols_pytd = None builtins_pytd = tracer.loader.concat_all() ast = ast.Visit(visitors.CreateTypeParametersForSignatures()) if options.protocols: log.info("=========== PyTD to solve =============\n%s", pytd_utils.Print(ast)) ast = convert_structural.convert_pytd(ast, builtins_pytd, protocols_pytd) elif not show_library_calls: log.info("Solving is turned off. Discarding call traces.") ast = ast.Visit(visitors.RemoveUnknownClasses()) ast = convert_structural.extract_local(ast) _maybe_output_debug(options, tracer.program) return ast, builtins_pytd def _maybe_output_debug(options, program): if options.output_cfg or options.output_typegraph: dot = debug.program_to_dot(program, set([]), bool(options.output_cfg)) svg_file = options.output_cfg or options.output_typegraph with subprocess.Popen( ["/usr/bin/dot", "-T", "svg", "-o", svg_file], stdin=subprocess.PIPE, universal_newlines=True) as proc: (_, stderr) = proc.communicate(dot) if stderr: log.info("Failed to create %s: %s", svg_file, stderr) if options.output_debug: text = debug.program_to_text(program) if options.output_debug == "-": log.info("=========== Program Dump =============\n%s", text) else: with options.open_function(options.output_debug, "w") as fi: fi.write(text)
true
true
f70ebcc7d351e260c5318ca161f5f74fcdefc39f
2,579
py
Python
item_44_pickle_copyreg.py
nickaigi/effective_python_tips
1a68b6eaed2e946b003c0cd0bdea03e79b8e8990
[ "Unlicense" ]
null
null
null
item_44_pickle_copyreg.py
nickaigi/effective_python_tips
1a68b6eaed2e946b003c0cd0bdea03e79b8e8990
[ "Unlicense" ]
null
null
null
item_44_pickle_copyreg.py
nickaigi/effective_python_tips
1a68b6eaed2e946b003c0cd0bdea03e79b8e8990
[ "Unlicense" ]
null
null
null
""" pickle can serialized python objects into a stream of bytes and deserialize bytes back into objects. Note: by design, pickle is unsafe! """ import pickle state_path = 'game_state.bin' class GameState(object): def __init__(self): self.level = 0 self.lives = 4 def save_game(state): with open(state_path, 'wb') as f: pickle.dump(state, f) def load_game(): state_after = {} with open(state_path, 'rb') as f: state_after = pickle.load(f) return state_after def example_one(): """ >>> {'level': 1, 'lives': 3} """ state = GameState() state.level += 1 state.lives -= 1 save_game(state) saved_state = load_game() print(saved_state.__dict__) class GameStateNew(object): def __init__(self): self.level = 0 self.lives = 4 self.points = 0 def example_two(): """ >>> {'level': 0, 'lives': 4, 'points': 0} """ state = GameStateNew() serialized = pickle.dumps(state) # dumps state_after = pickle.loads(serialized) # loads print(state_after.__dict__) def example_three(): """ - What happens when we try to access an older saved GameState but the defination of GameState has changed to GameStateNew ? """ state_after = load_game() try: assert isinstance(state_after, GameStateNew) except AssertionError: print('AssertionError: We knew') class GameStateDefaults(object): def __init__(self, level=0, lives=4, points=0): self.level = level self.lives = lives self.points = points def pickle_game_state(game_state): kwargs = game_state.__dict__ return unpickle_game_state, (kwargs, ) def unpickle_game_state(kwargs): return GameStateDefaults(**kwargs) class GameStateDefaultsMagic(object): def __init__(self, level=0, lives=4, points=0, magic=5): self.level = level self.lives = lives self.points = points self.magic = magic def example_four(): """ >>> {'level': 0, 'lives': 4, 'points': 1000} """ state = GameStateDefaults() state.points += 1000 serialized = pickle.dumps(state) state_after = pickle.loads(serialized) print(state_after.__dict__) def main(): example_four() # TODO things got unclear from this point onwards, # book describes removing lives form the game state, and versioning. # I have been putting off this book for 4 days, not making progress. # moving to item 45 if __name__ == '__main__': main()
20.468254
73
0.632803
import pickle state_path = 'game_state.bin' class GameState(object): def __init__(self): self.level = 0 self.lives = 4 def save_game(state): with open(state_path, 'wb') as f: pickle.dump(state, f) def load_game(): state_after = {} with open(state_path, 'rb') as f: state_after = pickle.load(f) return state_after def example_one(): state = GameState() state.level += 1 state.lives -= 1 save_game(state) saved_state = load_game() print(saved_state.__dict__) class GameStateNew(object): def __init__(self): self.level = 0 self.lives = 4 self.points = 0 def example_two(): state = GameStateNew() serialized = pickle.dumps(state) state_after = pickle.loads(serialized) print(state_after.__dict__) def example_three(): state_after = load_game() try: assert isinstance(state_after, GameStateNew) except AssertionError: print('AssertionError: We knew') class GameStateDefaults(object): def __init__(self, level=0, lives=4, points=0): self.level = level self.lives = lives self.points = points def pickle_game_state(game_state): kwargs = game_state.__dict__ return unpickle_game_state, (kwargs, ) def unpickle_game_state(kwargs): return GameStateDefaults(**kwargs) class GameStateDefaultsMagic(object): def __init__(self, level=0, lives=4, points=0, magic=5): self.level = level self.lives = lives self.points = points self.magic = magic def example_four(): state = GameStateDefaults() state.points += 1000 serialized = pickle.dumps(state) state_after = pickle.loads(serialized) print(state_after.__dict__) def main(): example_four() if __name__ == '__main__': main()
true
true
f70ebd3a3a9e6e7198ad4ce295f8d246601bb1e8
5,826
py
Python
homeassistant/components/push/camera.py
MrDelik/core
93a66cc357b226389967668441000498a10453bb
[ "Apache-2.0" ]
30,023
2016-04-13T10:17:53.000Z
2020-03-02T12:56:31.000Z
homeassistant/components/push/camera.py
MrDelik/core
93a66cc357b226389967668441000498a10453bb
[ "Apache-2.0" ]
24,710
2016-04-13T08:27:26.000Z
2020-03-02T12:59:13.000Z
homeassistant/components/push/camera.py
MrDelik/core
93a66cc357b226389967668441000498a10453bb
[ "Apache-2.0" ]
11,956
2016-04-13T18:42:31.000Z
2020-03-02T09:32:12.000Z
"""Camera platform that receives images through HTTP POST.""" from __future__ import annotations import asyncio from collections import deque from datetime import timedelta import logging import aiohttp import async_timeout import voluptuous as vol from homeassistant.components import webhook from homeassistant.components.camera import PLATFORM_SCHEMA, STATE_IDLE, Camera from homeassistant.components.camera.const import DOMAIN from homeassistant.const import CONF_NAME, CONF_TIMEOUT, CONF_WEBHOOK_ID from homeassistant.core import HomeAssistant, callback from homeassistant.helpers import config_validation as cv from homeassistant.helpers.entity_platform import AddEntitiesCallback from homeassistant.helpers.event import async_track_point_in_utc_time from homeassistant.helpers.typing import ConfigType, DiscoveryInfoType import homeassistant.util.dt as dt_util _LOGGER = logging.getLogger(__name__) CONF_BUFFER_SIZE = "buffer" CONF_IMAGE_FIELD = "field" DEFAULT_NAME = "Push Camera" ATTR_FILENAME = "filename" ATTR_LAST_TRIP = "last_trip" PUSH_CAMERA_DATA = "push_camera" PLATFORM_SCHEMA = PLATFORM_SCHEMA.extend( { vol.Optional(CONF_NAME, default=DEFAULT_NAME): cv.string, vol.Optional(CONF_BUFFER_SIZE, default=1): cv.positive_int, vol.Optional(CONF_TIMEOUT, default=timedelta(seconds=5)): vol.All( cv.time_period, cv.positive_timedelta ), vol.Optional(CONF_IMAGE_FIELD, default="image"): cv.string, vol.Required(CONF_WEBHOOK_ID): cv.string, } ) async def async_setup_platform( hass: HomeAssistant, config: ConfigType, async_add_entities: AddEntitiesCallback, discovery_info: DiscoveryInfoType | None = None, ) -> None: """Set up the Push Camera platform.""" if PUSH_CAMERA_DATA not in hass.data: hass.data[PUSH_CAMERA_DATA] = {} webhook_id = config.get(CONF_WEBHOOK_ID) cameras = [ PushCamera( hass, config[CONF_NAME], config[CONF_BUFFER_SIZE], config[CONF_TIMEOUT], config[CONF_IMAGE_FIELD], webhook_id, ) ] async_add_entities(cameras) async def handle_webhook(hass, webhook_id, request): """Handle incoming webhook POST with image files.""" try: async with async_timeout.timeout(5): data = dict(await request.post()) except (asyncio.TimeoutError, aiohttp.web.HTTPException) as error: _LOGGER.error("Could not get information from POST <%s>", error) return camera = hass.data[PUSH_CAMERA_DATA][webhook_id] if camera.image_field not in data: _LOGGER.warning("Webhook call without POST parameter <%s>", camera.image_field) return await camera.update_image( data[camera.image_field].file.read(), data[camera.image_field].filename ) class PushCamera(Camera): """The representation of a Push camera.""" def __init__(self, hass, name, buffer_size, timeout, image_field, webhook_id): """Initialize push camera component.""" super().__init__() self._name = name self._last_trip = None self._filename = None self._expired_listener = None self._timeout = timeout self.queue = deque([], buffer_size) self._current_image = None self._image_field = image_field self.webhook_id = webhook_id self.webhook_url = webhook.async_generate_url(hass, webhook_id) async def async_added_to_hass(self): """Call when entity is added to hass.""" self.hass.data[PUSH_CAMERA_DATA][self.webhook_id] = self try: webhook.async_register( self.hass, DOMAIN, self.name, self.webhook_id, handle_webhook ) except ValueError: _LOGGER.error( "In <%s>, webhook_id <%s> already used", self.name, self.webhook_id ) @property def image_field(self): """HTTP field containing the image file.""" return self._image_field async def update_image(self, image, filename): """Update the camera image.""" if self.state == STATE_IDLE: self._attr_is_recording = True self._last_trip = dt_util.utcnow() self.queue.clear() self._filename = filename self.queue.appendleft(image) @callback def reset_state(now): """Set state to idle after no new images for a period of time.""" self._attr_is_recording = False self._expired_listener = None _LOGGER.debug("Reset state") self.async_write_ha_state() if self._expired_listener: self._expired_listener() self._expired_listener = async_track_point_in_utc_time( self.hass, reset_state, dt_util.utcnow() + self._timeout ) self.async_write_ha_state() async def async_camera_image( self, width: int | None = None, height: int | None = None ) -> bytes | None: """Return a still image response.""" if self.queue: if self.state == STATE_IDLE: self.queue.rotate(1) self._current_image = self.queue[0] return self._current_image @property def name(self): """Return the name of this camera.""" return self._name @property def motion_detection_enabled(self): """Camera Motion Detection Status.""" return False @property def extra_state_attributes(self): """Return the state attributes.""" return { name: value for name, value in ( (ATTR_LAST_TRIP, self._last_trip), (ATTR_FILENAME, self._filename), ) if value is not None }
30.825397
87
0.656368
from __future__ import annotations import asyncio from collections import deque from datetime import timedelta import logging import aiohttp import async_timeout import voluptuous as vol from homeassistant.components import webhook from homeassistant.components.camera import PLATFORM_SCHEMA, STATE_IDLE, Camera from homeassistant.components.camera.const import DOMAIN from homeassistant.const import CONF_NAME, CONF_TIMEOUT, CONF_WEBHOOK_ID from homeassistant.core import HomeAssistant, callback from homeassistant.helpers import config_validation as cv from homeassistant.helpers.entity_platform import AddEntitiesCallback from homeassistant.helpers.event import async_track_point_in_utc_time from homeassistant.helpers.typing import ConfigType, DiscoveryInfoType import homeassistant.util.dt as dt_util _LOGGER = logging.getLogger(__name__) CONF_BUFFER_SIZE = "buffer" CONF_IMAGE_FIELD = "field" DEFAULT_NAME = "Push Camera" ATTR_FILENAME = "filename" ATTR_LAST_TRIP = "last_trip" PUSH_CAMERA_DATA = "push_camera" PLATFORM_SCHEMA = PLATFORM_SCHEMA.extend( { vol.Optional(CONF_NAME, default=DEFAULT_NAME): cv.string, vol.Optional(CONF_BUFFER_SIZE, default=1): cv.positive_int, vol.Optional(CONF_TIMEOUT, default=timedelta(seconds=5)): vol.All( cv.time_period, cv.positive_timedelta ), vol.Optional(CONF_IMAGE_FIELD, default="image"): cv.string, vol.Required(CONF_WEBHOOK_ID): cv.string, } ) async def async_setup_platform( hass: HomeAssistant, config: ConfigType, async_add_entities: AddEntitiesCallback, discovery_info: DiscoveryInfoType | None = None, ) -> None: if PUSH_CAMERA_DATA not in hass.data: hass.data[PUSH_CAMERA_DATA] = {} webhook_id = config.get(CONF_WEBHOOK_ID) cameras = [ PushCamera( hass, config[CONF_NAME], config[CONF_BUFFER_SIZE], config[CONF_TIMEOUT], config[CONF_IMAGE_FIELD], webhook_id, ) ] async_add_entities(cameras) async def handle_webhook(hass, webhook_id, request): try: async with async_timeout.timeout(5): data = dict(await request.post()) except (asyncio.TimeoutError, aiohttp.web.HTTPException) as error: _LOGGER.error("Could not get information from POST <%s>", error) return camera = hass.data[PUSH_CAMERA_DATA][webhook_id] if camera.image_field not in data: _LOGGER.warning("Webhook call without POST parameter <%s>", camera.image_field) return await camera.update_image( data[camera.image_field].file.read(), data[camera.image_field].filename ) class PushCamera(Camera): def __init__(self, hass, name, buffer_size, timeout, image_field, webhook_id): super().__init__() self._name = name self._last_trip = None self._filename = None self._expired_listener = None self._timeout = timeout self.queue = deque([], buffer_size) self._current_image = None self._image_field = image_field self.webhook_id = webhook_id self.webhook_url = webhook.async_generate_url(hass, webhook_id) async def async_added_to_hass(self): self.hass.data[PUSH_CAMERA_DATA][self.webhook_id] = self try: webhook.async_register( self.hass, DOMAIN, self.name, self.webhook_id, handle_webhook ) except ValueError: _LOGGER.error( "In <%s>, webhook_id <%s> already used", self.name, self.webhook_id ) @property def image_field(self): return self._image_field async def update_image(self, image, filename): if self.state == STATE_IDLE: self._attr_is_recording = True self._last_trip = dt_util.utcnow() self.queue.clear() self._filename = filename self.queue.appendleft(image) @callback def reset_state(now): self._attr_is_recording = False self._expired_listener = None _LOGGER.debug("Reset state") self.async_write_ha_state() if self._expired_listener: self._expired_listener() self._expired_listener = async_track_point_in_utc_time( self.hass, reset_state, dt_util.utcnow() + self._timeout ) self.async_write_ha_state() async def async_camera_image( self, width: int | None = None, height: int | None = None ) -> bytes | None: if self.queue: if self.state == STATE_IDLE: self.queue.rotate(1) self._current_image = self.queue[0] return self._current_image @property def name(self): return self._name @property def motion_detection_enabled(self): return False @property def extra_state_attributes(self): return { name: value for name, value in ( (ATTR_LAST_TRIP, self._last_trip), (ATTR_FILENAME, self._filename), ) if value is not None }
true
true
f70ebd9d986c52d7cd9e9a67cef2870a625bb652
3,080
py
Python
scripts/eval_bleu.py
nng555/fairseq
c9730a125825a85f33042e1b9fd1959b8ca829e5
[ "MIT" ]
2
2020-10-05T08:52:01.000Z
2021-03-03T15:26:35.000Z
scripts/eval_bleu.py
nng555/fairseq
c9730a125825a85f33042e1b9fd1959b8ca829e5
[ "MIT" ]
null
null
null
scripts/eval_bleu.py
nng555/fairseq
c9730a125825a85f33042e1b9fd1959b8ca829e5
[ "MIT" ]
null
null
null
import os import sys import subprocess import hydra from omegaconf import DictConfig from hydra import slurm_utils @hydra.main(config_path='/h/nng/conf/robust/config.yaml') def gen_neighborhood_labels(cfg: DictConfig): base_path = '/h/nng/data' model_data_path = os.path.join(base_path, cfg.data.task, cfg.eval.model.data) eval_data_path = os.path.join(base_path, cfg.data.task, cfg.eval.data) model_path = os.path.join('/h/nng/slurm', cfg.eval.model.date, slurm_utils.resolve_name(cfg.eval.model.name)) if not os.path.exists(os.path.join(model_path, 'checkpoint_best.pt')): for f in sorted(os.listdir(model_path))[::-1]: if os.path.exists(os.path.join(model_path, f, 'checkpoint_best.pt')): model_path = os.path.join(model_path, f) break model_path = os.path.join(model_path, 'checkpoint_best.pt') bin_path = os.path.join(model_data_path, cfg.data.fdset, cfg.data.bin, 'bin') t_path = os.path.join(eval_data_path, cfg.data.tdset, 'orig', cfg.eval.split + '.bpe.' + cfg.data.src) ref_path = os.path.join(eval_data_path, cfg.data.tdset, 'orig', cfg.eval.split + '.raw.' + cfg.data.tgt) bpe_path = '/h/nng/programs/subword-nmt/subword_nmt' if cfg.data.fdset == 'iwslt': fair_sh = ['fairseq-generate', bin_path, \ '--path', model_path, \ '--beam', '10', \ '--remove-bpe', \ '--batch-size', '128', \ '--quiet'] fair_p = subprocess.Popen(fair_sh, stdout=subprocess.PIPE) output, err = fair_p.communicate() print(output) else: cat_sh = ['cat', t_path] fair_sh = ['fairseq-interactive', bin_path, \ '--path', model_path, \ '-s', cfg.data.src, \ '-t', cfg.data.tgt, \ '--beam', '10', \ '--remove-bpe', \ '--buffer-size', '1024', \ '--max-tokens', '8000'] grep_sh = ['grep', '^H-'] cut_sh = ['cut', '-f', '3-'] detoken_sh = ['sacremoses', 'detokenize', '-l', cfg.data.tgt, '-q'] score_sh = ['sacrebleu', ref_path, '-l', cfg.data.src + '-' + cfg.data.tgt, '-w', '2'] cat_p = subprocess.Popen(cat_sh, stdout=subprocess.PIPE) fair_p = subprocess.Popen(fair_sh, stdin=cat_p.stdout, stdout=subprocess.PIPE) cat_p.stdout.close() grep_p = subprocess.Popen(grep_sh, stdin=fair_p.stdout, stdout=subprocess.PIPE) fair_p.stdout.close() cut_p = subprocess.Popen(cut_sh, stdin=grep_p.stdout, stdout=subprocess.PIPE) grep_p.stdout.close() detoken_p = subprocess.Popen(detoken_sh, stdin=cut_p.stdout, stdout=subprocess.PIPE) cut_p.stdout.close() score_p = subprocess.Popen(score_sh, stdin=detoken_p.stdout, stdout=subprocess.PIPE) detoken_p.stdout.close() output, err = score_p.communicate() print(output) if __name__ == "__main__": gen_neighborhood_labels()
40.526316
113
0.593831
import os import sys import subprocess import hydra from omegaconf import DictConfig from hydra import slurm_utils @hydra.main(config_path='/h/nng/conf/robust/config.yaml') def gen_neighborhood_labels(cfg: DictConfig): base_path = '/h/nng/data' model_data_path = os.path.join(base_path, cfg.data.task, cfg.eval.model.data) eval_data_path = os.path.join(base_path, cfg.data.task, cfg.eval.data) model_path = os.path.join('/h/nng/slurm', cfg.eval.model.date, slurm_utils.resolve_name(cfg.eval.model.name)) if not os.path.exists(os.path.join(model_path, 'checkpoint_best.pt')): for f in sorted(os.listdir(model_path))[::-1]: if os.path.exists(os.path.join(model_path, f, 'checkpoint_best.pt')): model_path = os.path.join(model_path, f) break model_path = os.path.join(model_path, 'checkpoint_best.pt') bin_path = os.path.join(model_data_path, cfg.data.fdset, cfg.data.bin, 'bin') t_path = os.path.join(eval_data_path, cfg.data.tdset, 'orig', cfg.eval.split + '.bpe.' + cfg.data.src) ref_path = os.path.join(eval_data_path, cfg.data.tdset, 'orig', cfg.eval.split + '.raw.' + cfg.data.tgt) bpe_path = '/h/nng/programs/subword-nmt/subword_nmt' if cfg.data.fdset == 'iwslt': fair_sh = ['fairseq-generate', bin_path, \ '--path', model_path, \ '--beam', '10', \ '--remove-bpe', \ '--batch-size', '128', \ '--quiet'] fair_p = subprocess.Popen(fair_sh, stdout=subprocess.PIPE) output, err = fair_p.communicate() print(output) else: cat_sh = ['cat', t_path] fair_sh = ['fairseq-interactive', bin_path, \ '--path', model_path, \ '-s', cfg.data.src, \ '-t', cfg.data.tgt, \ '--beam', '10', \ '--remove-bpe', \ '--buffer-size', '1024', \ '--max-tokens', '8000'] grep_sh = ['grep', '^H-'] cut_sh = ['cut', '-f', '3-'] detoken_sh = ['sacremoses', 'detokenize', '-l', cfg.data.tgt, '-q'] score_sh = ['sacrebleu', ref_path, '-l', cfg.data.src + '-' + cfg.data.tgt, '-w', '2'] cat_p = subprocess.Popen(cat_sh, stdout=subprocess.PIPE) fair_p = subprocess.Popen(fair_sh, stdin=cat_p.stdout, stdout=subprocess.PIPE) cat_p.stdout.close() grep_p = subprocess.Popen(grep_sh, stdin=fair_p.stdout, stdout=subprocess.PIPE) fair_p.stdout.close() cut_p = subprocess.Popen(cut_sh, stdin=grep_p.stdout, stdout=subprocess.PIPE) grep_p.stdout.close() detoken_p = subprocess.Popen(detoken_sh, stdin=cut_p.stdout, stdout=subprocess.PIPE) cut_p.stdout.close() score_p = subprocess.Popen(score_sh, stdin=detoken_p.stdout, stdout=subprocess.PIPE) detoken_p.stdout.close() output, err = score_p.communicate() print(output) if __name__ == "__main__": gen_neighborhood_labels()
true
true
f70ebea5539a99338b6dc9117844ec03bdd9efa1
79,789
py
Python
src/spring-cloud/azext_spring_cloud/vendored_sdks/appplatform/v2022_01_01_preview/operations/_deployments_operations.py
haroonf/azure-cli-extensions
61c044d34c224372f186934fa7c9313f1cd3a525
[ "MIT" ]
207
2017-11-29T06:59:41.000Z
2022-03-31T10:00:53.000Z
src/spring-cloud/azext_spring_cloud/vendored_sdks/appplatform/v2022_01_01_preview/operations/_deployments_operations.py
haroonf/azure-cli-extensions
61c044d34c224372f186934fa7c9313f1cd3a525
[ "MIT" ]
4,061
2017-10-27T23:19:56.000Z
2022-03-31T23:18:30.000Z
src/spring-cloud/azext_spring_cloud/vendored_sdks/appplatform/v2022_01_01_preview/operations/_deployments_operations.py
haroonf/azure-cli-extensions
61c044d34c224372f186934fa7c9313f1cd3a525
[ "MIT" ]
802
2017-10-11T17:36:26.000Z
2022-03-31T22:24:32.000Z
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- from typing import TYPE_CHECKING import warnings from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error from azure.core.paging import ItemPaged from azure.core.pipeline import PipelineResponse from azure.core.pipeline.transport import HttpRequest, HttpResponse from azure.core.polling import LROPoller, NoPolling, PollingMethod from azure.mgmt.core.exceptions import ARMErrorFormat from azure.mgmt.core.polling.arm_polling import ARMPolling from .. import models as _models if TYPE_CHECKING: # pylint: disable=unused-import,ungrouped-imports from typing import Any, Callable, Dict, Generic, Iterable, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] class DeploymentsOperations(object): """DeploymentsOperations operations. You should not instantiate this class directly. Instead, you should create a Client instance that instantiates it for you and attaches it as an attribute. :ivar models: Alias to model classes used in this operation group. :type models: ~azure.mgmt.appplatform.v2022_01_01_preview.models :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An object model deserializer. """ models = _models def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer self._config = config def get( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> "_models.DeploymentResource" """Get a Deployment and its properties. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: DeploymentResource, or the result of cls(response) :rtype: ~azure.mgmt.appplatform.v2022_01_01_preview.models.DeploymentResource :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DeploymentResource"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" # Construct URL url = self.get.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.get(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized get.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} # type: ignore def _create_or_update_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str deployment_resource, # type: "_models.DeploymentResource" **kwargs # type: Any ): # type: (...) -> "_models.DeploymentResource" cls = kwargs.pop('cls', None) # type: ClsType["_models.DeploymentResource"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL url = self._create_or_update_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] body_content = self._serialize.body(deployment_resource, 'DeploymentResource') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 201, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: deserialized = self._deserialize('DeploymentResource', pipeline_response) if response.status_code == 201: deserialized = self._deserialize('DeploymentResource', pipeline_response) if response.status_code == 202: deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized _create_or_update_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} # type: ignore def begin_create_or_update( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str deployment_resource, # type: "_models.DeploymentResource" **kwargs # type: Any ): # type: (...) -> LROPoller["_models.DeploymentResource"] """Create a new Deployment or update an exiting Deployment. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :param deployment_resource: Parameters for the create or update operation. :type deployment_resource: ~azure.mgmt.appplatform.v2022_01_01_preview.models.DeploymentResource :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either DeploymentResource or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.appplatform.v2022_01_01_preview.models.DeploymentResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType["_models.DeploymentResource"] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._create_or_update_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, deployment_resource=deployment_resource, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_create_or_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} # type: ignore def _delete_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> None cls = kwargs.pop('cls', None) # type: ClsType[None] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" # Construct URL url = self._delete_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.delete(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _delete_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} # type: ignore def begin_delete( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> LROPoller[None] """Operation to delete a Deployment. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either None or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[None] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType[None] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._delete_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} # type: ignore def _update_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str deployment_resource, # type: "_models.DeploymentResource" **kwargs # type: Any ): # type: (...) -> "_models.DeploymentResource" cls = kwargs.pop('cls', None) # type: ClsType["_models.DeploymentResource"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL url = self._update_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] body_content = self._serialize.body(deployment_resource, 'DeploymentResource') body_content_kwargs['content'] = body_content request = self._client.patch(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: deserialized = self._deserialize('DeploymentResource', pipeline_response) if response.status_code == 202: deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized _update_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} # type: ignore def begin_update( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str deployment_resource, # type: "_models.DeploymentResource" **kwargs # type: Any ): # type: (...) -> LROPoller["_models.DeploymentResource"] """Operation to update an exiting Deployment. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :param deployment_resource: Parameters for the update operation. :type deployment_resource: ~azure.mgmt.appplatform.v2022_01_01_preview.models.DeploymentResource :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either DeploymentResource or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[~azure.mgmt.appplatform.v2022_01_01_preview.models.DeploymentResource] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType["_models.DeploymentResource"] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._update_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, deployment_resource=deployment_resource, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} # type: ignore def list( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str version=None, # type: Optional[List[str]] **kwargs # type: Any ): # type: (...) -> Iterable["_models.DeploymentResourceCollection"] """Handles requests to list all resources in an App. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param version: Version of the deployments to be listed. :type version: list[str] :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either DeploymentResourceCollection or the result of cls(response) :rtype: ~azure.core.paging.ItemPaged[~azure.mgmt.appplatform.v2022_01_01_preview.models.DeploymentResourceCollection] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DeploymentResourceCollection"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" def prepare_request(next_link=None): # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') if not next_link: # Construct URL url = self.list.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') if version is not None: query_parameters['version'] = [self._serialize.query("version", q, 'str') if q is not None else '' for q in version] request = self._client.get(url, query_parameters, header_parameters) else: url = next_link query_parameters = {} # type: Dict[str, Any] request = self._client.get(url, query_parameters, header_parameters) return request def extract_data(pipeline_response): deserialized = self._deserialize('DeploymentResourceCollection', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) return deserialized.next_link or None, iter(list_of_elem) def get_next(next_link=None): request = prepare_request(next_link) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) return pipeline_response return ItemPaged( get_next, extract_data ) list.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments'} # type: ignore def list_for_cluster( self, resource_group_name, # type: str service_name, # type: str version=None, # type: Optional[List[str]] **kwargs # type: Any ): # type: (...) -> Iterable["_models.DeploymentResourceCollection"] """List deployments for a certain service. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param version: Version of the deployments to be listed. :type version: list[str] :keyword callable cls: A custom type or function that will be passed the direct response :return: An iterator like instance of either DeploymentResourceCollection or the result of cls(response) :rtype: ~azure.core.paging.ItemPaged[~azure.mgmt.appplatform.v2022_01_01_preview.models.DeploymentResourceCollection] :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType["_models.DeploymentResourceCollection"] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" def prepare_request(next_link=None): # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') if not next_link: # Construct URL url = self.list_for_cluster.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') if version is not None: query_parameters['version'] = [self._serialize.query("version", q, 'str') if q is not None else '' for q in version] request = self._client.get(url, query_parameters, header_parameters) else: url = next_link query_parameters = {} # type: Dict[str, Any] request = self._client.get(url, query_parameters, header_parameters) return request def extract_data(pipeline_response): deserialized = self._deserialize('DeploymentResourceCollection', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) return deserialized.next_link or None, iter(list_of_elem) def get_next(next_link=None): request = prepare_request(next_link) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) return pipeline_response return ItemPaged( get_next, extract_data ) list_for_cluster.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/deployments'} # type: ignore def _start_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> None cls = kwargs.pop('cls', None) # type: ClsType[None] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" # Construct URL url = self._start_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.post(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _start_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/start'} # type: ignore def begin_start( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> LROPoller[None] """Start the deployment. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either None or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[None] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType[None] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._start_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_start.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/start'} # type: ignore def _stop_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> None cls = kwargs.pop('cls', None) # type: ClsType[None] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" # Construct URL url = self._stop_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.post(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _stop_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/stop'} # type: ignore def begin_stop( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> LROPoller[None] """Stop the deployment. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either None or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[None] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType[None] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._stop_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_stop.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/stop'} # type: ignore def _restart_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> None cls = kwargs.pop('cls', None) # type: ClsType[None] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" # Construct URL url = self._restart_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.post(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _restart_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/restart'} # type: ignore def begin_restart( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> LROPoller[None] """Restart the deployment. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either None or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[None] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType[None] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._restart_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_restart.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/restart'} # type: ignore def get_log_file_url( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str **kwargs # type: Any ): # type: (...) -> Optional["_models.LogFileUrlResponse"] """Get deployment log file URL. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :keyword callable cls: A custom type or function that will be passed the direct response :return: LogFileUrlResponse, or the result of cls(response) :rtype: ~azure.mgmt.appplatform.v2022_01_01_preview.models.LogFileUrlResponse or None :raises: ~azure.core.exceptions.HttpResponseError """ cls = kwargs.pop('cls', None) # type: ClsType[Optional["_models.LogFileUrlResponse"]] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" # Construct URL url = self.get_log_file_url.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.post(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) deserialized = None if response.status_code == 200: deserialized = self._deserialize('LogFileUrlResponse', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized get_log_file_url.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/getLogFileUrl'} # type: ignore def _generate_heap_dump_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str diagnostic_parameters, # type: "_models.DiagnosticParameters" **kwargs # type: Any ): # type: (...) -> None cls = kwargs.pop('cls', None) # type: ClsType[None] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL url = self._generate_heap_dump_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] body_content = self._serialize.body(diagnostic_parameters, 'DiagnosticParameters') body_content_kwargs['content'] = body_content request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _generate_heap_dump_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/generateHeapDump'} # type: ignore def begin_generate_heap_dump( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str diagnostic_parameters, # type: "_models.DiagnosticParameters" **kwargs # type: Any ): # type: (...) -> LROPoller[None] """Generate Heap Dump. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :param diagnostic_parameters: Parameters for the diagnostic operation. :type diagnostic_parameters: ~azure.mgmt.appplatform.v2022_01_01_preview.models.DiagnosticParameters :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either None or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[None] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType[None] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._generate_heap_dump_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, diagnostic_parameters=diagnostic_parameters, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_generate_heap_dump.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/generateHeapDump'} # type: ignore def _generate_thread_dump_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str diagnostic_parameters, # type: "_models.DiagnosticParameters" **kwargs # type: Any ): # type: (...) -> None cls = kwargs.pop('cls', None) # type: ClsType[None] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL url = self._generate_thread_dump_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] body_content = self._serialize.body(diagnostic_parameters, 'DiagnosticParameters') body_content_kwargs['content'] = body_content request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _generate_thread_dump_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/generateThreadDump'} # type: ignore def begin_generate_thread_dump( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str diagnostic_parameters, # type: "_models.DiagnosticParameters" **kwargs # type: Any ): # type: (...) -> LROPoller[None] """Generate Thread Dump. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :param diagnostic_parameters: Parameters for the diagnostic operation. :type diagnostic_parameters: ~azure.mgmt.appplatform.v2022_01_01_preview.models.DiagnosticParameters :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either None or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[None] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType[None] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._generate_thread_dump_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, diagnostic_parameters=diagnostic_parameters, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_generate_thread_dump.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/generateThreadDump'} # type: ignore def _start_jfr_initial( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str diagnostic_parameters, # type: "_models.DiagnosticParameters" **kwargs # type: Any ): # type: (...) -> None cls = kwargs.pop('cls', None) # type: ClsType[None] error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" # Construct URL url = self._start_jfr_initial.metadata['url'] # type: ignore path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) # Construct parameters query_parameters = {} # type: Dict[str, Any] query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') # Construct headers header_parameters = {} # type: Dict[str, Any] header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} # type: Dict[str, Any] body_content = self._serialize.body(diagnostic_parameters, 'DiagnosticParameters') body_content_kwargs['content'] = body_content request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _start_jfr_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/startJFR'} # type: ignore def begin_start_jfr( self, resource_group_name, # type: str service_name, # type: str app_name, # type: str deployment_name, # type: str diagnostic_parameters, # type: "_models.DiagnosticParameters" **kwargs # type: Any ): # type: (...) -> LROPoller[None] """Start JFR. :param resource_group_name: The name of the resource group that contains the resource. You can obtain this value from the Azure Resource Manager API or the portal. :type resource_group_name: str :param service_name: The name of the Service resource. :type service_name: str :param app_name: The name of the App resource. :type app_name: str :param deployment_name: The name of the Deployment resource. :type deployment_name: str :param diagnostic_parameters: Parameters for the diagnostic operation. :type diagnostic_parameters: ~azure.mgmt.appplatform.v2022_01_01_preview.models.DiagnosticParameters :keyword callable cls: A custom type or function that will be passed the direct response :keyword str continuation_token: A continuation token to restart a poller from a saved state. :keyword polling: By default, your polling method will be ARMPolling. Pass in False for this operation to not poll, or pass in your own initialized polling object for a personal polling strategy. :paramtype polling: bool or ~azure.core.polling.PollingMethod :keyword int polling_interval: Default waiting time between two polls for LRO operations if no Retry-After header is present. :return: An instance of LROPoller that returns either None or the result of cls(response) :rtype: ~azure.core.polling.LROPoller[None] :raises ~azure.core.exceptions.HttpResponseError: """ polling = kwargs.pop('polling', True) # type: Union[bool, PollingMethod] cls = kwargs.pop('cls', None) # type: ClsType[None] lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) # type: Optional[str] if cont_token is None: raw_result = self._start_jfr_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, diagnostic_parameters=diagnostic_parameters, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_start_jfr.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/startJFR'} # type: ignore
51.443585
254
0.660492
from typing import TYPE_CHECKING import warnings from azure.core.exceptions import ClientAuthenticationError, HttpResponseError, ResourceExistsError, ResourceNotFoundError, map_error from azure.core.paging import ItemPaged from azure.core.pipeline import PipelineResponse from azure.core.pipeline.transport import HttpRequest, HttpResponse from azure.core.polling import LROPoller, NoPolling, PollingMethod from azure.mgmt.core.exceptions import ARMErrorFormat from azure.mgmt.core.polling.arm_polling import ARMPolling from .. import models as _models if TYPE_CHECKING: from typing import Any, Callable, Dict, Generic, Iterable, List, Optional, TypeVar, Union T = TypeVar('T') ClsType = Optional[Callable[[PipelineResponse[HttpRequest, HttpResponse], T, Dict[str, Any]], Any]] class DeploymentsOperations(object): models = _models def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer self._config = config def get( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" url = self.get.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.get(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized get.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} def _create_or_update_initial( self, resource_group_name, service_name, app_name, deployment_name, deployment_resource, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" url = self._create_or_update_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} body_content = self._serialize.body(deployment_resource, 'DeploymentResource') body_content_kwargs['content'] = body_content request = self._client.put(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 201, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: deserialized = self._deserialize('DeploymentResource', pipeline_response) if response.status_code == 201: deserialized = self._deserialize('DeploymentResource', pipeline_response) if response.status_code == 202: deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized _create_or_update_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} def begin_create_or_update( self, resource_group_name, service_name, app_name, deployment_name, deployment_resource, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._create_or_update_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, deployment_resource=deployment_resource, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_create_or_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} def _delete_initial( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" url = self._delete_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.delete(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _delete_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} def begin_delete( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._delete_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_delete.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} def _update_initial( self, resource_group_name, service_name, app_name, deployment_name, deployment_resource, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" url = self._update_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} body_content = self._serialize.body(deployment_resource, 'DeploymentResource') body_content_kwargs['content'] = body_content request = self._client.patch(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if response.status_code == 200: deserialized = self._deserialize('DeploymentResource', pipeline_response) if response.status_code == 202: deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized _update_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} def begin_update( self, resource_group_name, service_name, app_name, deployment_name, deployment_resource, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._update_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, deployment_resource=deployment_resource, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): deserialized = self._deserialize('DeploymentResource', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_update.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}'} def list( self, resource_group_name, service_name, app_name, version=None, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" def prepare_request(next_link=None): header_parameters = {} header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') if not next_link: url = self.list.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') if version is not None: query_parameters['version'] = [self._serialize.query("version", q, 'str') if q is not None else '' for q in version] request = self._client.get(url, query_parameters, header_parameters) else: url = next_link query_parameters = {} request = self._client.get(url, query_parameters, header_parameters) return request def extract_data(pipeline_response): deserialized = self._deserialize('DeploymentResourceCollection', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) return deserialized.next_link or None, iter(list_of_elem) def get_next(next_link=None): request = prepare_request(next_link) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) return pipeline_response return ItemPaged( get_next, extract_data ) list.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments'} def list_for_cluster( self, resource_group_name, service_name, version=None, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" def prepare_request(next_link=None): header_parameters = {} header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') if not next_link: url = self.list_for_cluster.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') if version is not None: query_parameters['version'] = [self._serialize.query("version", q, 'str') if q is not None else '' for q in version] request = self._client.get(url, query_parameters, header_parameters) else: url = next_link query_parameters = {} request = self._client.get(url, query_parameters, header_parameters) return request def extract_data(pipeline_response): deserialized = self._deserialize('DeploymentResourceCollection', pipeline_response) list_of_elem = deserialized.value if cls: list_of_elem = cls(list_of_elem) return deserialized.next_link or None, iter(list_of_elem) def get_next(next_link=None): request = prepare_request(next_link) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) return pipeline_response return ItemPaged( get_next, extract_data ) list_for_cluster.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/deployments'} def _start_initial( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" url = self._start_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.post(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _start_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/start'} def begin_start( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._start_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_start.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/start'} def _stop_initial( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" url = self._stop_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.post(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _stop_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/stop'} def begin_stop( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._stop_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_stop.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/stop'} def _restart_initial( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" url = self._restart_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.post(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _restart_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/restart'} def begin_restart( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._restart_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_restart.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/restart'} def get_log_file_url( self, resource_group_name, service_name, app_name, deployment_name, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" accept = "application/json" url = self.get_log_file_url.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') request = self._client.post(url, query_parameters, header_parameters) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 204]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) deserialized = None if response.status_code == 200: deserialized = self._deserialize('LogFileUrlResponse', pipeline_response) if cls: return cls(pipeline_response, deserialized, {}) return deserialized get_log_file_url.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/getLogFileUrl'} def _generate_heap_dump_initial( self, resource_group_name, service_name, app_name, deployment_name, diagnostic_parameters, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" url = self._generate_heap_dump_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} body_content = self._serialize.body(diagnostic_parameters, 'DiagnosticParameters') body_content_kwargs['content'] = body_content request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _generate_heap_dump_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/generateHeapDump'} def begin_generate_heap_dump( self, resource_group_name, service_name, app_name, deployment_name, diagnostic_parameters, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._generate_heap_dump_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, diagnostic_parameters=diagnostic_parameters, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_generate_heap_dump.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/generateHeapDump'} def _generate_thread_dump_initial( self, resource_group_name, service_name, app_name, deployment_name, diagnostic_parameters, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" url = self._generate_thread_dump_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} body_content = self._serialize.body(diagnostic_parameters, 'DiagnosticParameters') body_content_kwargs['content'] = body_content request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _generate_thread_dump_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/generateThreadDump'} def begin_generate_thread_dump( self, resource_group_name, service_name, app_name, deployment_name, diagnostic_parameters, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._generate_thread_dump_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, diagnostic_parameters=diagnostic_parameters, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_generate_thread_dump.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/generateThreadDump'} def _start_jfr_initial( self, resource_group_name, service_name, app_name, deployment_name, diagnostic_parameters, **kwargs ): cls = kwargs.pop('cls', None) error_map = { 401: ClientAuthenticationError, 404: ResourceNotFoundError, 409: ResourceExistsError } error_map.update(kwargs.pop('error_map', {})) api_version = "2022-01-01-preview" content_type = kwargs.pop("content_type", "application/json") accept = "application/json" url = self._start_jfr_initial.metadata['url'] path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } url = self._client.format_url(url, **path_format_arguments) query_parameters = {} query_parameters['api-version'] = self._serialize.query("api_version", api_version, 'str') header_parameters = {} header_parameters['Content-Type'] = self._serialize.header("content_type", content_type, 'str') header_parameters['Accept'] = self._serialize.header("accept", accept, 'str') body_content_kwargs = {} body_content = self._serialize.body(diagnostic_parameters, 'DiagnosticParameters') body_content_kwargs['content'] = body_content request = self._client.post(url, query_parameters, header_parameters, **body_content_kwargs) pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) response = pipeline_response.http_response if response.status_code not in [200, 202]: map_error(status_code=response.status_code, response=response, error_map=error_map) raise HttpResponseError(response=response, error_format=ARMErrorFormat) if cls: return cls(pipeline_response, None, {}) _start_jfr_initial.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/startJFR'} def begin_start_jfr( self, resource_group_name, service_name, app_name, deployment_name, diagnostic_parameters, **kwargs ): polling = kwargs.pop('polling', True) cls = kwargs.pop('cls', None) lro_delay = kwargs.pop( 'polling_interval', self._config.polling_interval ) cont_token = kwargs.pop('continuation_token', None) if cont_token is None: raw_result = self._start_jfr_initial( resource_group_name=resource_group_name, service_name=service_name, app_name=app_name, deployment_name=deployment_name, diagnostic_parameters=diagnostic_parameters, cls=lambda x,y,z: x, **kwargs ) kwargs.pop('error_map', None) kwargs.pop('content_type', None) def get_long_running_output(pipeline_response): if cls: return cls(pipeline_response, None, {}) path_format_arguments = { 'subscriptionId': self._serialize.url("self._config.subscription_id", self._config.subscription_id, 'str'), 'resourceGroupName': self._serialize.url("resource_group_name", resource_group_name, 'str'), 'serviceName': self._serialize.url("service_name", service_name, 'str'), 'appName': self._serialize.url("app_name", app_name, 'str'), 'deploymentName': self._serialize.url("deployment_name", deployment_name, 'str'), } if polling is True: polling_method = ARMPolling(lro_delay, lro_options={'final-state-via': 'azure-async-operation'}, path_format_arguments=path_format_arguments, **kwargs) elif polling is False: polling_method = NoPolling() else: polling_method = polling if cont_token: return LROPoller.from_continuation_token( polling_method=polling_method, continuation_token=cont_token, client=self._client, deserialization_callback=get_long_running_output ) else: return LROPoller(self._client, raw_result, get_long_running_output, polling_method) begin_start_jfr.metadata = {'url': '/subscriptions/{subscriptionId}/resourceGroups/{resourceGroupName}/providers/Microsoft.AppPlatform/Spring/{serviceName}/apps/{appName}/deployments/{deploymentName}/startJFR'}
true
true
f70ec1157e54153c23a33efa2de815ddfff14a48
4,440
py
Python
src/helper/trainer.py
rlaehgns5399/GoogLeNet-Inception-tf
eb9597634eec9a7b511e967ad8c7b2552563755f
[ "MIT" ]
3
2018-12-02T13:34:32.000Z
2019-03-18T03:23:37.000Z
src/helper/trainer.py
rlaehgns5399/GoogLeNet-Inception-tf
eb9597634eec9a7b511e967ad8c7b2552563755f
[ "MIT" ]
null
null
null
src/helper/trainer.py
rlaehgns5399/GoogLeNet-Inception-tf
eb9597634eec9a7b511e967ad8c7b2552563755f
[ "MIT" ]
null
null
null
#!/usr/bin/env python # -*- coding: utf-8 -*- # File: trainer.py # Author: Qian Ge <geqian1001@gmail.com> import os import numpy as np import tensorflow as tf def display(global_step, step, scaler_sum_list, name_list, collection, summary_val=None, summary_writer=None, ): print('[step: {}]'.format(global_step), end='') for val, name in zip(scaler_sum_list, name_list): print(' {}: {:.4f}'.format(name, val * 1. / step), end='') print('') if summary_writer is not None: s = tf.Summary() for val, name in zip(scaler_sum_list, name_list): s.value.add(tag='{}/{}'.format(collection, name), simple_value=val * 1. / step) summary_writer.add_summary(s, global_step) if summary_val is not None: summary_writer.add_summary(summary_val, global_step) class Trainer(object): def __init__(self, train_model, valid_model, train_data, init_lr=1e-3): self._t_model = train_model self._v_model = valid_model self._train_data = train_data self._init_lr = init_lr self._train_op = train_model.get_train_op() self._train_loss_op = train_model.get_loss() self._train_accuracy_op = train_model.get_accuracy() self._valid_loss_op = valid_model.get_loss() self._valid_accuracy_op = valid_model.get_accuracy() # self._train_summary_op = train_model.get_train_summary() # self._valid_summary_op = train_model.get_valid_summary() self.global_step = 0 self.epoch_id = 0 def train_epoch(self, sess, keep_prob=1., summary_writer=None): if self.epoch_id < 35: self._lr = self._init_lr elif self.epoch_id < 50: self._lr = self._init_lr / 10. else: self._lr = self._init_lr / 100. # self._t_model.set_is_training(True) display_name_list = ['loss', 'accuracy'] cur_summary = None cur_epoch = self._train_data.epochs_completed step = 0 loss_sum = 0 acc_sum = 0 self.epoch_id += 1 while cur_epoch == self._train_data.epochs_completed: self.global_step += 1 step += 1 batch_data = self._train_data.next_batch_dict() im = batch_data['image'] label = batch_data['label'] _, loss, acc = sess.run( [self._train_op, self._train_loss_op, self._train_accuracy_op], feed_dict={self._t_model.image: im, self._t_model.label: label, self._t_model.lr: self._lr, self._t_model.keep_prob: keep_prob}) loss_sum += loss acc_sum += acc if step % 100 == 0 or step == 1: display(self.global_step, step, [loss_sum, acc_sum], display_name_list, 'train', summary_val=cur_summary, summary_writer=summary_writer) print('==== epoch: {}, lr:{} ===='.format(cur_epoch, self._lr)) display(self.global_step, step, [loss_sum, acc_sum], display_name_list, 'train', summary_val=cur_summary, summary_writer=summary_writer) def valid_epoch(self, sess, dataflow, summary_writer=None): display_name_list = ['loss', 'accuracy'] cur_summary = None dataflow.reset_epoch() step = 0 loss_sum = 0 acc_sum = 0 while dataflow.epochs_completed < 1: step += 1 batch_data = dataflow.next_batch_dict() im = batch_data['image'] label = batch_data['label'] loss, acc = sess.run( [self._valid_loss_op, self._valid_accuracy_op], feed_dict={self._v_model.image: im, self._v_model.label: label}) loss_sum += loss acc_sum += acc print('[Valid]: ', end='') display(self.global_step, step, [loss_sum, acc_sum], display_name_list, 'valid', summary_val=cur_summary, summary_writer=summary_writer)
33.134328
80
0.547072
import os import numpy as np import tensorflow as tf def display(global_step, step, scaler_sum_list, name_list, collection, summary_val=None, summary_writer=None, ): print('[step: {}]'.format(global_step), end='') for val, name in zip(scaler_sum_list, name_list): print(' {}: {:.4f}'.format(name, val * 1. / step), end='') print('') if summary_writer is not None: s = tf.Summary() for val, name in zip(scaler_sum_list, name_list): s.value.add(tag='{}/{}'.format(collection, name), simple_value=val * 1. / step) summary_writer.add_summary(s, global_step) if summary_val is not None: summary_writer.add_summary(summary_val, global_step) class Trainer(object): def __init__(self, train_model, valid_model, train_data, init_lr=1e-3): self._t_model = train_model self._v_model = valid_model self._train_data = train_data self._init_lr = init_lr self._train_op = train_model.get_train_op() self._train_loss_op = train_model.get_loss() self._train_accuracy_op = train_model.get_accuracy() self._valid_loss_op = valid_model.get_loss() self._valid_accuracy_op = valid_model.get_accuracy() self.global_step = 0 self.epoch_id = 0 def train_epoch(self, sess, keep_prob=1., summary_writer=None): if self.epoch_id < 35: self._lr = self._init_lr elif self.epoch_id < 50: self._lr = self._init_lr / 10. else: self._lr = self._init_lr / 100. display_name_list = ['loss', 'accuracy'] cur_summary = None cur_epoch = self._train_data.epochs_completed step = 0 loss_sum = 0 acc_sum = 0 self.epoch_id += 1 while cur_epoch == self._train_data.epochs_completed: self.global_step += 1 step += 1 batch_data = self._train_data.next_batch_dict() im = batch_data['image'] label = batch_data['label'] _, loss, acc = sess.run( [self._train_op, self._train_loss_op, self._train_accuracy_op], feed_dict={self._t_model.image: im, self._t_model.label: label, self._t_model.lr: self._lr, self._t_model.keep_prob: keep_prob}) loss_sum += loss acc_sum += acc if step % 100 == 0 or step == 1: display(self.global_step, step, [loss_sum, acc_sum], display_name_list, 'train', summary_val=cur_summary, summary_writer=summary_writer) print('==== epoch: {}, lr:{} ===='.format(cur_epoch, self._lr)) display(self.global_step, step, [loss_sum, acc_sum], display_name_list, 'train', summary_val=cur_summary, summary_writer=summary_writer) def valid_epoch(self, sess, dataflow, summary_writer=None): display_name_list = ['loss', 'accuracy'] cur_summary = None dataflow.reset_epoch() step = 0 loss_sum = 0 acc_sum = 0 while dataflow.epochs_completed < 1: step += 1 batch_data = dataflow.next_batch_dict() im = batch_data['image'] label = batch_data['label'] loss, acc = sess.run( [self._valid_loss_op, self._valid_accuracy_op], feed_dict={self._v_model.image: im, self._v_model.label: label}) loss_sum += loss acc_sum += acc print('[Valid]: ', end='') display(self.global_step, step, [loss_sum, acc_sum], display_name_list, 'valid', summary_val=cur_summary, summary_writer=summary_writer)
true
true
f70ec119a082e7c27fa7682d87edbb9dac979d7d
601
py
Python
ndg/__init__.py
philipkershaw/ndg_xacml
f24ccbb4d4c053c585b0369dfe5a5949d49f9715
[ "BSD-3-Clause" ]
null
null
null
ndg/__init__.py
philipkershaw/ndg_xacml
f24ccbb4d4c053c585b0369dfe5a5949d49f9715
[ "BSD-3-Clause" ]
null
null
null
ndg/__init__.py
philipkershaw/ndg_xacml
f24ccbb4d4c053c585b0369dfe5a5949d49f9715
[ "BSD-3-Clause" ]
null
null
null
"""NDG XACML ndg namespace package NERC DataGrid This is a setuptools namespace_package. DO NOT place any other code in this file! There is no guarantee that it will be installed with easy_install. See: http://peak.telecommunity.com/DevCenter/setuptools#namespace-packages ... for details. """ __author__ = "P J Kershaw" __date__ = "19/02/10" __copyright__ = "(C) 2010 Science and Technology Facilities Council" __license__ = "BSD - see LICENSE file in top-level directory" __contact__ = "Philip.Kershaw@stfc.ac.uk" __revision__ = '$Id$' __import__('pkg_resources').declare_namespace(__name__)
30.05
69
0.768719
__author__ = "P J Kershaw" __date__ = "19/02/10" __copyright__ = "(C) 2010 Science and Technology Facilities Council" __license__ = "BSD - see LICENSE file in top-level directory" __contact__ = "Philip.Kershaw@stfc.ac.uk" __revision__ = '$Id$' __import__('pkg_resources').declare_namespace(__name__)
true
true
f70ec43dc91c15bd7888f7714dfbcacfc1d5d148
4,513
py
Python
fastai/torch_imports.py
CalebEverett/fastai-dl2
64d23592eddca6ca1f3647e73c319e97c8eb392b
[ "Apache-2.0" ]
4
2018-03-29T20:47:24.000Z
2019-05-06T12:31:11.000Z
fastai/torch_imports.py
CalebEverett/fastai-dl2
64d23592eddca6ca1f3647e73c319e97c8eb392b
[ "Apache-2.0" ]
null
null
null
fastai/torch_imports.py
CalebEverett/fastai-dl2
64d23592eddca6ca1f3647e73c319e97c8eb392b
[ "Apache-2.0" ]
null
null
null
import os import torch, torchvision, torchtext from torch import nn, cuda, backends, FloatTensor, LongTensor, optim import torch.nn.functional as F from torch.autograd import Variable from torch.utils.data import Dataset, TensorDataset from torch.nn.init import kaiming_uniform, kaiming_normal from torchvision.transforms import Compose from torchvision.models import resnet18, resnet34, resnet50, resnet101, resnet152 from torchvision.models import vgg16_bn, vgg19_bn from torchvision.models import densenet121, densenet161, densenet169, densenet201 from .models.resnext_50_32x4d import resnext_50_32x4d from .models.resnext_101_32x4d import resnext_101_32x4d from .models.resnext_101_64x4d import resnext_101_64x4d from .models.wrn_50_2f import wrn_50_2f from .models.inceptionresnetv2 import InceptionResnetV2 from .models.inceptionv4 import InceptionV4 from .models.nasnet import nasnetalarge from unet_models import unet11 import warnings warnings.filterwarnings('ignore', message='Implicit dimension choice', category=UserWarning) def children(m): return m if isinstance(m, (list, tuple)) else list(m.children()) def save_model(m, p): torch.save(m.state_dict(), p) def load_model(m, p): m.load_state_dict(torch.load(p, map_location=lambda storage, loc: storage)) def load_pre(pre, f, fn): m = f() path = os.path.dirname(__file__) if pre: load_model(m, f'{path}/weights/{fn}.pth') return m def _fastai_model(name, paper_title, paper_href): def add_docs_wrapper(f): f.__doc__ = f"""{name} model from `"{paper_title}" <{paper_href}>`_ Args: pre (bool): If True, returns a model pre-trained on ImageNet """ return f return add_docs_wrapper @_fastai_model('Inception 4', 'Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning', 'https://arxiv.org/pdf/1602.07261.pdf') def inception_4(pre): return children(inceptionv4(pretrained=pre))[0] @_fastai_model('Inception 4', 'Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning', 'https://arxiv.org/pdf/1602.07261.pdf') def inceptionresnet_2(pre): return load_pre(pre, InceptionResnetV2, 'inceptionresnetv2-d579a627') @_fastai_model('ResNeXt 50', 'Aggregated Residual Transformations for Deep Neural Networks', 'https://arxiv.org/abs/1611.05431') def resnext50(pre): return load_pre(pre, resnext_50_32x4d, 'resnext_50_32x4d') @_fastai_model('ResNeXt 101_32', 'Aggregated Residual Transformations for Deep Neural Networks', 'https://arxiv.org/abs/1611.05431') def resnext101(pre): return load_pre(pre, resnext_101_32x4d, 'resnext_101_32x4d') @_fastai_model('ResNeXt 101_64', 'Aggregated Residual Transformations for Deep Neural Networks', 'https://arxiv.org/abs/1611.05431') def resnext101_64(pre): return load_pre(pre, resnext_101_64x4d, 'resnext_101_64x4d') @_fastai_model('Wide Residual Networks', 'Wide Residual Networks', 'https://arxiv.org/pdf/1605.07146.pdf') def wrn(pre): return load_pre(pre, wrn_50_2f, 'wrn_50_2f') @_fastai_model('Densenet-121', 'Densely Connected Convolutional Networks', 'https://arxiv.org/pdf/1608.06993.pdf') def dn121(pre): return children(densenet121(pre))[0] @_fastai_model('Densenet-169', 'Densely Connected Convolutional Networks', 'https://arxiv.org/pdf/1608.06993.pdf') def dn161(pre): return children(densenet161(pre))[0] @_fastai_model('Densenet-161', 'Densely Connected Convolutional Networks', 'https://arxiv.org/pdf/1608.06993.pdf') def dn169(pre): return children(densenet169(pre))[0] @_fastai_model('Densenet-201', 'Densely Connected Convolutional Networks', 'https://arxiv.org/pdf/1608.06993.pdf') def dn201(pre): return children(densenet201(pre))[0] @_fastai_model('Vgg-16 with batch norm added', 'Very Deep Convolutional Networks for Large-Scale Image Recognition', 'https://arxiv.org/pdf/1409.1556.pdf') def vgg16(pre): return children(vgg16_bn(pre))[0] @_fastai_model('Vgg-19 with batch norm added', 'Very Deep Convolutional Networks for Large-Scale Image Recognition', 'https://arxiv.org/pdf/1409.1556.pdf') def vgg19(pre): return children(vgg19_bn(pre))[0] @_fastai_model('Vgg-11 with U-Net', 'TernausNet: U-Net with VGG11 Encoder Pre-Trained on ImageNet for Image Segmentation', 'https://arxiv.org/pdf/1801.05746.pdf') def ternausnet(pre): return children(unet11(pre))
46.525773
122
0.739198
import os import torch, torchvision, torchtext from torch import nn, cuda, backends, FloatTensor, LongTensor, optim import torch.nn.functional as F from torch.autograd import Variable from torch.utils.data import Dataset, TensorDataset from torch.nn.init import kaiming_uniform, kaiming_normal from torchvision.transforms import Compose from torchvision.models import resnet18, resnet34, resnet50, resnet101, resnet152 from torchvision.models import vgg16_bn, vgg19_bn from torchvision.models import densenet121, densenet161, densenet169, densenet201 from .models.resnext_50_32x4d import resnext_50_32x4d from .models.resnext_101_32x4d import resnext_101_32x4d from .models.resnext_101_64x4d import resnext_101_64x4d from .models.wrn_50_2f import wrn_50_2f from .models.inceptionresnetv2 import InceptionResnetV2 from .models.inceptionv4 import InceptionV4 from .models.nasnet import nasnetalarge from unet_models import unet11 import warnings warnings.filterwarnings('ignore', message='Implicit dimension choice', category=UserWarning) def children(m): return m if isinstance(m, (list, tuple)) else list(m.children()) def save_model(m, p): torch.save(m.state_dict(), p) def load_model(m, p): m.load_state_dict(torch.load(p, map_location=lambda storage, loc: storage)) def load_pre(pre, f, fn): m = f() path = os.path.dirname(__file__) if pre: load_model(m, f'{path}/weights/{fn}.pth') return m def _fastai_model(name, paper_title, paper_href): def add_docs_wrapper(f): f.__doc__ = f"""{name} model from `"{paper_title}" <{paper_href}>`_ Args: pre (bool): If True, returns a model pre-trained on ImageNet """ return f return add_docs_wrapper @_fastai_model('Inception 4', 'Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning', 'https://arxiv.org/pdf/1602.07261.pdf') def inception_4(pre): return children(inceptionv4(pretrained=pre))[0] @_fastai_model('Inception 4', 'Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning', 'https://arxiv.org/pdf/1602.07261.pdf') def inceptionresnet_2(pre): return load_pre(pre, InceptionResnetV2, 'inceptionresnetv2-d579a627') @_fastai_model('ResNeXt 50', 'Aggregated Residual Transformations for Deep Neural Networks', 'https://arxiv.org/abs/1611.05431') def resnext50(pre): return load_pre(pre, resnext_50_32x4d, 'resnext_50_32x4d') @_fastai_model('ResNeXt 101_32', 'Aggregated Residual Transformations for Deep Neural Networks', 'https://arxiv.org/abs/1611.05431') def resnext101(pre): return load_pre(pre, resnext_101_32x4d, 'resnext_101_32x4d') @_fastai_model('ResNeXt 101_64', 'Aggregated Residual Transformations for Deep Neural Networks', 'https://arxiv.org/abs/1611.05431') def resnext101_64(pre): return load_pre(pre, resnext_101_64x4d, 'resnext_101_64x4d') @_fastai_model('Wide Residual Networks', 'Wide Residual Networks', 'https://arxiv.org/pdf/1605.07146.pdf') def wrn(pre): return load_pre(pre, wrn_50_2f, 'wrn_50_2f') @_fastai_model('Densenet-121', 'Densely Connected Convolutional Networks', 'https://arxiv.org/pdf/1608.06993.pdf') def dn121(pre): return children(densenet121(pre))[0] @_fastai_model('Densenet-169', 'Densely Connected Convolutional Networks', 'https://arxiv.org/pdf/1608.06993.pdf') def dn161(pre): return children(densenet161(pre))[0] @_fastai_model('Densenet-161', 'Densely Connected Convolutional Networks', 'https://arxiv.org/pdf/1608.06993.pdf') def dn169(pre): return children(densenet169(pre))[0] @_fastai_model('Densenet-201', 'Densely Connected Convolutional Networks', 'https://arxiv.org/pdf/1608.06993.pdf') def dn201(pre): return children(densenet201(pre))[0] @_fastai_model('Vgg-16 with batch norm added', 'Very Deep Convolutional Networks for Large-Scale Image Recognition', 'https://arxiv.org/pdf/1409.1556.pdf') def vgg16(pre): return children(vgg16_bn(pre))[0] @_fastai_model('Vgg-19 with batch norm added', 'Very Deep Convolutional Networks for Large-Scale Image Recognition', 'https://arxiv.org/pdf/1409.1556.pdf') def vgg19(pre): return children(vgg19_bn(pre))[0] @_fastai_model('Vgg-11 with U-Net', 'TernausNet: U-Net with VGG11 Encoder Pre-Trained on ImageNet for Image Segmentation', 'https://arxiv.org/pdf/1801.05746.pdf') def ternausnet(pre): return children(unet11(pre))
true
true
f70ec5c93336e5b029a6279f82499cd519691765
6,514
py
Python
models/model.py
qq456cvb/CPPF
79366978854ae18b14c69ac850ea64b9dc286081
[ "MIT" ]
23
2022-03-06T12:01:00.000Z
2022-03-31T19:26:39.000Z
models/model.py
qq456cvb/CPPF
79366978854ae18b14c69ac850ea64b9dc286081
[ "MIT" ]
null
null
null
models/model.py
qq456cvb/CPPF
79366978854ae18b14c69ac850ea64b9dc286081
[ "MIT" ]
2
2022-03-23T07:14:13.000Z
2022-03-24T07:18:55.000Z
import torch import torch.nn as nn import torch.nn.functional as F from .sprin import GlobalInfoProp, SparseSO3Conv import numpy as np class ResLayer(torch.nn.Module): def __init__(self, dim_in, dim_out, bn=False) -> None: super().__init__() assert(bn is False) self.fc1 = torch.nn.Linear(dim_in, dim_out) if bn: self.bn1 = torch.nn.BatchNorm1d(dim_out) else: self.bn1 = lambda x: x self.fc2 = torch.nn.Linear(dim_out, dim_out) if bn: self.bn2 = torch.nn.BatchNorm1d(dim_out) else: self.bn2 = lambda x: x if dim_in != dim_out: self.fc0 = torch.nn.Linear(dim_in, dim_out) else: self.fc0 = None def forward(self, x): x_res = x if self.fc0 is None else self.fc0(x) x = F.relu(self.bn1(self.fc1(x))) x = self.bn2(self.fc2(x)) return x + x_res class PointEncoder(nn.Module): def __init__(self, k, spfcs, out_dim, num_layers=2, num_nbr_feats=2) -> None: super().__init__() self.k = k self.spconvs = nn.ModuleList() self.spconvs.append(SparseSO3Conv(32, num_nbr_feats, out_dim, *spfcs)) self.aggrs = nn.ModuleList() self.aggrs.append(GlobalInfoProp(out_dim, out_dim // 4)) for _ in range(num_layers - 1): self.spconvs.append(SparseSO3Conv(32, out_dim + out_dim // 4, out_dim, *spfcs)) self.aggrs.append(GlobalInfoProp(out_dim, out_dim // 4)) def forward(self, pc, pc_normal, dist): nbrs_idx = torch.topk(dist, self.k, largest=False, sorted=False)[1] #[..., N, K] pc_nbrs = torch.gather(pc.unsqueeze(-3).expand(*pc.shape[:-1], *pc.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, pc.shape[-1])) #[..., N, K, 3] pc_nbrs_centered = pc_nbrs - pc.unsqueeze(-2) #[..., N, K, 3] pc_nbrs_norm = torch.norm(pc_nbrs_centered, dim=-1, keepdim=True) pc_normal_nbrs = torch.gather(pc_normal.unsqueeze(-3).expand(*pc_normal.shape[:-1], *pc_normal.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, pc_normal.shape[-1])) #[..., N, K, 3] pc_normal_cos = torch.sum(pc_normal_nbrs * pc_normal.unsqueeze(-2), -1, keepdim=True) feat = self.aggrs[0](self.spconvs[0](pc_nbrs, torch.cat([pc_nbrs_norm, pc_normal_cos], -1), pc)) for i in range(len(self.spconvs) - 1): spconv = self.spconvs[i + 1] aggr = self.aggrs[i + 1] feat_nbrs = torch.gather(feat.unsqueeze(-3).expand(*feat.shape[:-1], *feat.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, feat.shape[-1])) feat = aggr(spconv(pc_nbrs, feat_nbrs, pc)) return feat def forward_nbrs(self, pc, pc_normal, nbrs_idx): pc_nbrs = torch.gather(pc.unsqueeze(-3).expand(*pc.shape[:-1], *pc.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, pc.shape[-1])) #[..., N, K, 3] pc_nbrs_centered = pc_nbrs - pc.unsqueeze(-2) #[..., N, K, 3] pc_nbrs_norm = torch.norm(pc_nbrs_centered, dim=-1, keepdim=True) pc_normal_nbrs = torch.gather(pc_normal.unsqueeze(-3).expand(*pc_normal.shape[:-1], *pc_normal.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, pc_normal.shape[-1])) #[..., N, K, 3] pc_normal_cos = torch.sum(pc_normal_nbrs * pc_normal.unsqueeze(-2), -1, keepdim=True) feat = self.aggrs[0](self.spconvs[0](pc_nbrs, torch.cat([pc_nbrs_norm, pc_normal_cos], -1), pc)) for i in range(len(self.spconvs) - 1): spconv = self.spconvs[i + 1] aggr = self.aggrs[i + 1] feat_nbrs = torch.gather(feat.unsqueeze(-3).expand(*feat.shape[:-1], *feat.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, feat.shape[-1])) feat = aggr(spconv(pc_nbrs, feat_nbrs, pc)) return feat class PPFEncoder(nn.Module): def __init__(self, ppffcs, out_dim) -> None: super().__init__() self.res_layers = nn.ModuleList() for i in range(len(ppffcs) - 1): dim_in, dim_out = ppffcs[i], ppffcs[i + 1] self.res_layers.append(ResLayer(dim_in, dim_out, bn=False)) self.final = nn.Linear(ppffcs[-1], out_dim) def forward(self, pc, pc_normal, feat, dist=None, idxs=None): if idxs is not None: return self.forward_with_idx(pc[0], pc_normal[0], feat[0], idxs)[None] xx = pc.unsqueeze(-2) - pc.unsqueeze(-3) xx_normed = xx / (dist[..., None] + 1e-7) outputs = [] for idx in torch.chunk(torch.arange(pc.shape[1]), 5): feat_chunk = feat[..., idx, :] target_shape = [*feat_chunk.shape[:-2], feat_chunk.shape[-2], feat.shape[-2], feat_chunk.shape[-1]] # B x NC x N x F xx_normed_chunk = xx_normed[..., idx, :, :] ppf = torch.cat([ torch.sum(pc_normal[..., idx, :].unsqueeze(-2) * xx_normed_chunk, -1, keepdim=True), torch.sum(pc_normal.unsqueeze(-3) * xx_normed_chunk, -1, keepdim=True), torch.sum(pc_normal[..., idx, :].unsqueeze(-2) * pc_normal.unsqueeze(-3), -1, keepdim=True), dist[..., idx, :, None], ], -1) # ppf.zero_() final_feat = torch.cat([feat_chunk[..., None, :].expand(*target_shape), feat[..., None, :, :].expand(*target_shape), ppf], -1) output = final_feat for res_layer in self.res_layers: output = res_layer(output) outputs.append(output) output = torch.cat(outputs, dim=-3) return self.final(output) def forward_with_idx(self, pc, pc_normal, feat, idxs): a_idxs = idxs[:, 0] b_idxs = idxs[:, 1] xy = pc[a_idxs] - pc[b_idxs] xy_norm = torch.norm(xy, dim=-1) xy_normed = xy / (xy_norm[..., None] + 1e-7) pnormal_cos = pc_normal[a_idxs] * pc_normal[b_idxs] ppf = torch.cat([ torch.sum(pc_normal[a_idxs] * xy_normed, -1, keepdim=True), torch.sum(pc_normal[b_idxs] * xy_normed, -1, keepdim=True), torch.sum(pnormal_cos, -1, keepdim=True), xy_norm[..., None], ], -1) # ppf.zero_() final_feat = torch.cat([feat[a_idxs], feat[b_idxs], ppf], -1) output = final_feat for res_layer in self.res_layers: output = res_layer(output) return self.final(output)
47.202899
202
0.57553
import torch import torch.nn as nn import torch.nn.functional as F from .sprin import GlobalInfoProp, SparseSO3Conv import numpy as np class ResLayer(torch.nn.Module): def __init__(self, dim_in, dim_out, bn=False) -> None: super().__init__() assert(bn is False) self.fc1 = torch.nn.Linear(dim_in, dim_out) if bn: self.bn1 = torch.nn.BatchNorm1d(dim_out) else: self.bn1 = lambda x: x self.fc2 = torch.nn.Linear(dim_out, dim_out) if bn: self.bn2 = torch.nn.BatchNorm1d(dim_out) else: self.bn2 = lambda x: x if dim_in != dim_out: self.fc0 = torch.nn.Linear(dim_in, dim_out) else: self.fc0 = None def forward(self, x): x_res = x if self.fc0 is None else self.fc0(x) x = F.relu(self.bn1(self.fc1(x))) x = self.bn2(self.fc2(x)) return x + x_res class PointEncoder(nn.Module): def __init__(self, k, spfcs, out_dim, num_layers=2, num_nbr_feats=2) -> None: super().__init__() self.k = k self.spconvs = nn.ModuleList() self.spconvs.append(SparseSO3Conv(32, num_nbr_feats, out_dim, *spfcs)) self.aggrs = nn.ModuleList() self.aggrs.append(GlobalInfoProp(out_dim, out_dim // 4)) for _ in range(num_layers - 1): self.spconvs.append(SparseSO3Conv(32, out_dim + out_dim // 4, out_dim, *spfcs)) self.aggrs.append(GlobalInfoProp(out_dim, out_dim // 4)) def forward(self, pc, pc_normal, dist): nbrs_idx = torch.topk(dist, self.k, largest=False, sorted=False)[1] pc_nbrs = torch.gather(pc.unsqueeze(-3).expand(*pc.shape[:-1], *pc.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, pc.shape[-1])) pc_nbrs_centered = pc_nbrs - pc.unsqueeze(-2) pc_nbrs_norm = torch.norm(pc_nbrs_centered, dim=-1, keepdim=True) pc_normal_nbrs = torch.gather(pc_normal.unsqueeze(-3).expand(*pc_normal.shape[:-1], *pc_normal.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, pc_normal.shape[-1])) pc_normal_cos = torch.sum(pc_normal_nbrs * pc_normal.unsqueeze(-2), -1, keepdim=True) feat = self.aggrs[0](self.spconvs[0](pc_nbrs, torch.cat([pc_nbrs_norm, pc_normal_cos], -1), pc)) for i in range(len(self.spconvs) - 1): spconv = self.spconvs[i + 1] aggr = self.aggrs[i + 1] feat_nbrs = torch.gather(feat.unsqueeze(-3).expand(*feat.shape[:-1], *feat.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, feat.shape[-1])) feat = aggr(spconv(pc_nbrs, feat_nbrs, pc)) return feat def forward_nbrs(self, pc, pc_normal, nbrs_idx): pc_nbrs = torch.gather(pc.unsqueeze(-3).expand(*pc.shape[:-1], *pc.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, pc.shape[-1])) pc_nbrs_centered = pc_nbrs - pc.unsqueeze(-2) pc_nbrs_norm = torch.norm(pc_nbrs_centered, dim=-1, keepdim=True) pc_normal_nbrs = torch.gather(pc_normal.unsqueeze(-3).expand(*pc_normal.shape[:-1], *pc_normal.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, pc_normal.shape[-1])) pc_normal_cos = torch.sum(pc_normal_nbrs * pc_normal.unsqueeze(-2), -1, keepdim=True) feat = self.aggrs[0](self.spconvs[0](pc_nbrs, torch.cat([pc_nbrs_norm, pc_normal_cos], -1), pc)) for i in range(len(self.spconvs) - 1): spconv = self.spconvs[i + 1] aggr = self.aggrs[i + 1] feat_nbrs = torch.gather(feat.unsqueeze(-3).expand(*feat.shape[:-1], *feat.shape[-2:]), -2, nbrs_idx[..., None].expand(*nbrs_idx.shape, feat.shape[-1])) feat = aggr(spconv(pc_nbrs, feat_nbrs, pc)) return feat class PPFEncoder(nn.Module): def __init__(self, ppffcs, out_dim) -> None: super().__init__() self.res_layers = nn.ModuleList() for i in range(len(ppffcs) - 1): dim_in, dim_out = ppffcs[i], ppffcs[i + 1] self.res_layers.append(ResLayer(dim_in, dim_out, bn=False)) self.final = nn.Linear(ppffcs[-1], out_dim) def forward(self, pc, pc_normal, feat, dist=None, idxs=None): if idxs is not None: return self.forward_with_idx(pc[0], pc_normal[0], feat[0], idxs)[None] xx = pc.unsqueeze(-2) - pc.unsqueeze(-3) xx_normed = xx / (dist[..., None] + 1e-7) outputs = [] for idx in torch.chunk(torch.arange(pc.shape[1]), 5): feat_chunk = feat[..., idx, :] target_shape = [*feat_chunk.shape[:-2], feat_chunk.shape[-2], feat.shape[-2], feat_chunk.shape[-1]] xx_normed_chunk = xx_normed[..., idx, :, :] ppf = torch.cat([ torch.sum(pc_normal[..., idx, :].unsqueeze(-2) * xx_normed_chunk, -1, keepdim=True), torch.sum(pc_normal.unsqueeze(-3) * xx_normed_chunk, -1, keepdim=True), torch.sum(pc_normal[..., idx, :].unsqueeze(-2) * pc_normal.unsqueeze(-3), -1, keepdim=True), dist[..., idx, :, None], ], -1) final_feat = torch.cat([feat_chunk[..., None, :].expand(*target_shape), feat[..., None, :, :].expand(*target_shape), ppf], -1) output = final_feat for res_layer in self.res_layers: output = res_layer(output) outputs.append(output) output = torch.cat(outputs, dim=-3) return self.final(output) def forward_with_idx(self, pc, pc_normal, feat, idxs): a_idxs = idxs[:, 0] b_idxs = idxs[:, 1] xy = pc[a_idxs] - pc[b_idxs] xy_norm = torch.norm(xy, dim=-1) xy_normed = xy / (xy_norm[..., None] + 1e-7) pnormal_cos = pc_normal[a_idxs] * pc_normal[b_idxs] ppf = torch.cat([ torch.sum(pc_normal[a_idxs] * xy_normed, -1, keepdim=True), torch.sum(pc_normal[b_idxs] * xy_normed, -1, keepdim=True), torch.sum(pnormal_cos, -1, keepdim=True), xy_norm[..., None], ], -1) final_feat = torch.cat([feat[a_idxs], feat[b_idxs], ppf], -1) output = final_feat for res_layer in self.res_layers: output = res_layer(output) return self.final(output)
true
true
f70ec5d7df1d46b90aac95d7005cb457c693a297
4,095
py
Python
ea_sim/visualization_adv_tasks/convert_history_to_archive.py
lis-epfl/Tensoft-G21
7a83c5dabc12906c0a6bd1da0a28a131e9d5e144
[ "Apache-2.0" ]
1
2021-08-03T10:52:20.000Z
2021-08-03T10:52:20.000Z
ea_sim/visualization_adv_tasks/convert_history_to_archive.py
lis-epfl/Tensoft-G21
7a83c5dabc12906c0a6bd1da0a28a131e9d5e144
[ "Apache-2.0" ]
null
null
null
ea_sim/visualization_adv_tasks/convert_history_to_archive.py
lis-epfl/Tensoft-G21
7a83c5dabc12906c0a6bd1da0a28a131e9d5e144
[ "Apache-2.0" ]
1
2021-09-18T07:23:35.000Z
2021-09-18T07:23:35.000Z
import argparse import json import numpy as np import os import subprocess import sys sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from params_conf import N_MODULES, MIN_NUM_MODULES, STIFF_TABLE from utils import parse_robot_string def convert_h_to_arch(h_file, morph_file, best_function, out_dir, seed): with open(morph_file) as morph_file: morphologies = json.load(morph_file) inverse_morph_dict = {v: k for k, v in morphologies.items()} archive = np.full(shape=(len(N_MODULES), len(STIFF_TABLE)), fill_value=None) last_gen = 0 with open(h_file) as hf: _ = hf.readline() # skip header for line in hf: last_gen, m_id, c_id, fit, _ = line.split(',') last_gen, m_id, c_id, fit = int(last_gen), int(m_id), int(c_id), float(fit) robot = parse_robot_string(inverse_morph_dict[m_id]) index_1 = len(robot) - MIN_NUM_MODULES index_2 = STIFF_TABLE.index(robot[0]['stiff']) value_1 = len(robot) value_2 = robot[0]['stiff'] if archive[index_1, index_2] is None or fit == best_function(archive[index_1, index_2][4], fit): archive[index_1, index_2] = (index_1, index_2, value_1, value_2, fit, m_id, c_id) out_filename = os.path.join(out_dir, 'entity_archive_{}_ngen_{}.csv'.format(seed, last_gen)) with open(out_filename, 'w') as out_file: out_file.write('1st_dim_indx,2nd_dim_indx,1st_dim:num_modules,2nd_dim:stiffness_value,fitness,e_id,nn_id\n') for entry in archive.flatten(): if entry is not None: index_1, index_2, value_1, value_2, fit, m_id, c_id = entry out_file.write('{},{},{},{},{:.4f},{},{}\n'.format( index_1, index_2, value_1, value_2, fit, m_id, c_id )) if __name__ == '__main__': parser = argparse.ArgumentParser(description='Script for converting an history file into a archive based on ' 'morphological features.') parser.add_argument('res_dirs', metavar='res_dirs', type=str, nargs='+', help='list of folders containing evolution results') parser.add_argument('--fit-func-best', metavar='fitness_function_best', type=str, nargs='?', default='min', help='function used to determine the best fitness, allowed values are: min, max') parser.add_argument('--owner', metavar='owner', type=str, action='store', nargs='+', default=None, help='User and group for chown') args = parser.parse_args() if args.fit_func_best not in ['min', 'max']: raise Exception('The function provided to determine the best fitness' + 'is not valid') else: fit_func_best = min if args.fit_func_best == 'min' else max if args.res_dirs is not None: for res_dir in args.res_dirs: settings_file = os.path.join(res_dir, 'settings.json') with open(settings_file) as sf: settings = json.load(sf) abs_res_dir_path = os.path.abspath(res_dir) evo_info_dir = os.path.join(abs_res_dir_path, 'evolution_info') out_dir = os.path.join(abs_res_dir_path, 'archives') os.makedirs(out_dir, exist_ok=True) for seed in settings['seeds']: history_file = os.path.join(evo_info_dir, 'history_{}.csv'.format(seed)) morphologies_file = os.path.join(abs_res_dir_path, 'morphologies', 'morphologies_{}.json'.format(seed)) convert_h_to_arch(history_file, morphologies_file, fit_func_best, out_dir, seed) if args.owner is not None and len(args.owner) == 2: try: exec_string = 'chown -R {}:{} {}'.format(args.owner[0], args.owner[1], out_dir) c_proc = subprocess.run(exec_string.split(' '), capture_output=True) except: raise Exception('An error occurred during the owner setting')
45.5
119
0.620757
import argparse import json import numpy as np import os import subprocess import sys sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) from params_conf import N_MODULES, MIN_NUM_MODULES, STIFF_TABLE from utils import parse_robot_string def convert_h_to_arch(h_file, morph_file, best_function, out_dir, seed): with open(morph_file) as morph_file: morphologies = json.load(morph_file) inverse_morph_dict = {v: k for k, v in morphologies.items()} archive = np.full(shape=(len(N_MODULES), len(STIFF_TABLE)), fill_value=None) last_gen = 0 with open(h_file) as hf: _ = hf.readline() for line in hf: last_gen, m_id, c_id, fit, _ = line.split(',') last_gen, m_id, c_id, fit = int(last_gen), int(m_id), int(c_id), float(fit) robot = parse_robot_string(inverse_morph_dict[m_id]) index_1 = len(robot) - MIN_NUM_MODULES index_2 = STIFF_TABLE.index(robot[0]['stiff']) value_1 = len(robot) value_2 = robot[0]['stiff'] if archive[index_1, index_2] is None or fit == best_function(archive[index_1, index_2][4], fit): archive[index_1, index_2] = (index_1, index_2, value_1, value_2, fit, m_id, c_id) out_filename = os.path.join(out_dir, 'entity_archive_{}_ngen_{}.csv'.format(seed, last_gen)) with open(out_filename, 'w') as out_file: out_file.write('1st_dim_indx,2nd_dim_indx,1st_dim:num_modules,2nd_dim:stiffness_value,fitness,e_id,nn_id\n') for entry in archive.flatten(): if entry is not None: index_1, index_2, value_1, value_2, fit, m_id, c_id = entry out_file.write('{},{},{},{},{:.4f},{},{}\n'.format( index_1, index_2, value_1, value_2, fit, m_id, c_id )) if __name__ == '__main__': parser = argparse.ArgumentParser(description='Script for converting an history file into a archive based on ' 'morphological features.') parser.add_argument('res_dirs', metavar='res_dirs', type=str, nargs='+', help='list of folders containing evolution results') parser.add_argument('--fit-func-best', metavar='fitness_function_best', type=str, nargs='?', default='min', help='function used to determine the best fitness, allowed values are: min, max') parser.add_argument('--owner', metavar='owner', type=str, action='store', nargs='+', default=None, help='User and group for chown') args = parser.parse_args() if args.fit_func_best not in ['min', 'max']: raise Exception('The function provided to determine the best fitness' + 'is not valid') else: fit_func_best = min if args.fit_func_best == 'min' else max if args.res_dirs is not None: for res_dir in args.res_dirs: settings_file = os.path.join(res_dir, 'settings.json') with open(settings_file) as sf: settings = json.load(sf) abs_res_dir_path = os.path.abspath(res_dir) evo_info_dir = os.path.join(abs_res_dir_path, 'evolution_info') out_dir = os.path.join(abs_res_dir_path, 'archives') os.makedirs(out_dir, exist_ok=True) for seed in settings['seeds']: history_file = os.path.join(evo_info_dir, 'history_{}.csv'.format(seed)) morphologies_file = os.path.join(abs_res_dir_path, 'morphologies', 'morphologies_{}.json'.format(seed)) convert_h_to_arch(history_file, morphologies_file, fit_func_best, out_dir, seed) if args.owner is not None and len(args.owner) == 2: try: exec_string = 'chown -R {}:{} {}'.format(args.owner[0], args.owner[1], out_dir) c_proc = subprocess.run(exec_string.split(' '), capture_output=True) except: raise Exception('An error occurred during the owner setting')
true
true
f70ec64b9e31daafd1fb2f1ca0a900fb5ba86171
3,473
py
Python
pyexcel_xls/xlsw.py
pyexcel/pyexcel-xls
995cfd273d5360947a528ff3a1ed3f9e52a429ad
[ "BSD-3-Clause" ]
40
2016-05-18T20:09:39.000Z
2022-02-09T06:39:41.000Z
pyexcel_xls/xlsw.py
wenxuefeng3930/pyexcel-xls
995cfd273d5360947a528ff3a1ed3f9e52a429ad
[ "BSD-3-Clause" ]
46
2016-02-01T22:12:31.000Z
2021-10-07T18:57:05.000Z
pyexcel_xls/xlsw.py
wenxuefeng3930/pyexcel-xls
995cfd273d5360947a528ff3a1ed3f9e52a429ad
[ "BSD-3-Clause" ]
24
2016-01-29T12:26:27.000Z
2021-10-31T15:37:15.000Z
""" pyexcel_xlsw ~~~~~~~~~~~~~~~~~~~ The lower level xls file format handler using xlwt :copyright: (c) 2016-2021 by Onni Software Ltd :license: New BSD License """ import datetime import xlrd from xlwt import XFStyle, Workbook from pyexcel_io import constants from pyexcel_io.plugin_api import IWriter, ISheetWriter DEFAULT_DATE_FORMAT = "DD/MM/YY" DEFAULT_TIME_FORMAT = "HH:MM:SS" DEFAULT_LONGTIME_FORMAT = "[HH]:MM:SS" DEFAULT_DATETIME_FORMAT = "%s %s" % (DEFAULT_DATE_FORMAT, DEFAULT_TIME_FORMAT) EMPTY_SHEET_NOT_ALLOWED = "xlwt does not support a book without any sheets" class XLSheetWriter(ISheetWriter): """ xls sheet writer """ def __init__(self, xls_book, xls_sheet, sheet_name): if sheet_name is None: sheet_name = constants.DEFAULT_SHEET_NAME self._xls_book = xls_book self._xls_sheet = xls_sheet self._xls_sheet = self._xls_book.add_sheet(sheet_name) self.current_row = 0 def write_row(self, array): """ write a row into the file """ for i, value in enumerate(array): style = None tmp_array = [] if isinstance(value, datetime.datetime): tmp_array = [ value.year, value.month, value.day, value.hour, value.minute, value.second, ] value = xlrd.xldate.xldate_from_datetime_tuple(tmp_array, 0) style = XFStyle() style.num_format_str = DEFAULT_DATETIME_FORMAT elif isinstance(value, datetime.timedelta): value = value.days + value.seconds / 86_400 style = XFStyle() style.num_format_str = DEFAULT_LONGTIME_FORMAT elif isinstance(value, datetime.date): tmp_array = [value.year, value.month, value.day] value = xlrd.xldate.xldate_from_date_tuple(tmp_array, 0) style = XFStyle() style.num_format_str = DEFAULT_DATE_FORMAT elif isinstance(value, datetime.time): tmp_array = [value.hour, value.minute, value.second] value = xlrd.xldate.xldate_from_time_tuple(tmp_array) style = XFStyle() style.num_format_str = DEFAULT_TIME_FORMAT if style: self._xls_sheet.write(self.current_row, i, value, style) else: self._xls_sheet.write(self.current_row, i, value) self.current_row += 1 def close(self): pass class XLSWriter(IWriter): """ xls writer """ def __init__( self, file_alike_object, _, # file_type not used encoding="ascii", style_compression=2, **keywords, ): self.file_alike_object = file_alike_object self.work_book = Workbook( style_compression=style_compression, encoding=encoding ) def create_sheet(self, name): return XLSheetWriter(self.work_book, None, name) def write(self, incoming_dict): if incoming_dict: IWriter.write(self, incoming_dict) else: raise NotImplementedError(EMPTY_SHEET_NOT_ALLOWED) def close(self): """ This call actually save the file """ self.work_book.save(self.file_alike_object)
31.008929
78
0.589692
import datetime import xlrd from xlwt import XFStyle, Workbook from pyexcel_io import constants from pyexcel_io.plugin_api import IWriter, ISheetWriter DEFAULT_DATE_FORMAT = "DD/MM/YY" DEFAULT_TIME_FORMAT = "HH:MM:SS" DEFAULT_LONGTIME_FORMAT = "[HH]:MM:SS" DEFAULT_DATETIME_FORMAT = "%s %s" % (DEFAULT_DATE_FORMAT, DEFAULT_TIME_FORMAT) EMPTY_SHEET_NOT_ALLOWED = "xlwt does not support a book without any sheets" class XLSheetWriter(ISheetWriter): def __init__(self, xls_book, xls_sheet, sheet_name): if sheet_name is None: sheet_name = constants.DEFAULT_SHEET_NAME self._xls_book = xls_book self._xls_sheet = xls_sheet self._xls_sheet = self._xls_book.add_sheet(sheet_name) self.current_row = 0 def write_row(self, array): for i, value in enumerate(array): style = None tmp_array = [] if isinstance(value, datetime.datetime): tmp_array = [ value.year, value.month, value.day, value.hour, value.minute, value.second, ] value = xlrd.xldate.xldate_from_datetime_tuple(tmp_array, 0) style = XFStyle() style.num_format_str = DEFAULT_DATETIME_FORMAT elif isinstance(value, datetime.timedelta): value = value.days + value.seconds / 86_400 style = XFStyle() style.num_format_str = DEFAULT_LONGTIME_FORMAT elif isinstance(value, datetime.date): tmp_array = [value.year, value.month, value.day] value = xlrd.xldate.xldate_from_date_tuple(tmp_array, 0) style = XFStyle() style.num_format_str = DEFAULT_DATE_FORMAT elif isinstance(value, datetime.time): tmp_array = [value.hour, value.minute, value.second] value = xlrd.xldate.xldate_from_time_tuple(tmp_array) style = XFStyle() style.num_format_str = DEFAULT_TIME_FORMAT if style: self._xls_sheet.write(self.current_row, i, value, style) else: self._xls_sheet.write(self.current_row, i, value) self.current_row += 1 def close(self): pass class XLSWriter(IWriter): def __init__( self, file_alike_object, _, encoding="ascii", style_compression=2, **keywords, ): self.file_alike_object = file_alike_object self.work_book = Workbook( style_compression=style_compression, encoding=encoding ) def create_sheet(self, name): return XLSheetWriter(self.work_book, None, name) def write(self, incoming_dict): if incoming_dict: IWriter.write(self, incoming_dict) else: raise NotImplementedError(EMPTY_SHEET_NOT_ALLOWED) def close(self): self.work_book.save(self.file_alike_object)
true
true
f70ec7201f8c108b954356e237165ff9018c482b
3,781
py
Python
ethgreen/wallet/wallet_interested_store.py
ethgreen/ethgreen-blockchain
8f1a450897ab7a82326aea7e57e18ac2c03a9e83
[ "Apache-2.0" ]
11
2021-11-10T19:30:12.000Z
2022-02-09T04:30:29.000Z
ethgreen/wallet/wallet_interested_store.py
ethgreen/ethgreen-blockchain
8f1a450897ab7a82326aea7e57e18ac2c03a9e83
[ "Apache-2.0" ]
6
2021-11-16T17:11:03.000Z
2021-12-28T17:11:20.000Z
ethgreen/wallet/wallet_interested_store.py
ethgreen/ethgreen-blockchain
8f1a450897ab7a82326aea7e57e18ac2c03a9e83
[ "Apache-2.0" ]
3
2021-11-21T02:27:10.000Z
2022-03-15T08:34:47.000Z
from typing import List, Tuple, Optional import aiosqlite from ethgreen.types.blockchain_format.sized_bytes import bytes32 from ethgreen.util.db_wrapper import DBWrapper class WalletInterestedStore: """ Stores coin ids that we are interested in receiving """ db_connection: aiosqlite.Connection db_wrapper: DBWrapper @classmethod async def create(cls, wrapper: DBWrapper): self = cls() self.db_connection = wrapper.db self.db_wrapper = wrapper await self.db_connection.execute("CREATE TABLE IF NOT EXISTS interested_coins(coin_name text PRIMARY KEY)") await self.db_connection.execute( "CREATE TABLE IF NOT EXISTS interested_puzzle_hashes(puzzle_hash text PRIMARY KEY, wallet_id integer)" ) await self.db_connection.commit() return self async def _clear_database(self): cursor = await self.db_connection.execute("DELETE FROM puzzle_hashes") await cursor.close() cursor = await self.db_connection.execute("DELETE FROM interested_coins") await cursor.close() await self.db_connection.commit() async def get_interested_coin_ids(self) -> List[bytes32]: cursor = await self.db_connection.execute("SELECT coin_name FROM interested_coins") rows_hex = await cursor.fetchall() return [bytes32(bytes.fromhex(row[0])) for row in rows_hex] async def add_interested_coin_id(self, coin_id: bytes32, in_transaction: bool = False) -> None: if not in_transaction: await self.db_wrapper.lock.acquire() try: cursor = await self.db_connection.execute( "INSERT OR REPLACE INTO interested_coins VALUES (?)", (coin_id.hex(),) ) await cursor.close() finally: if not in_transaction: await self.db_connection.commit() self.db_wrapper.lock.release() async def get_interested_puzzle_hashes(self) -> List[Tuple[bytes32, int]]: cursor = await self.db_connection.execute("SELECT puzzle_hash, wallet_id FROM interested_puzzle_hashes") rows_hex = await cursor.fetchall() return [(bytes32(bytes.fromhex(row[0])), row[1]) for row in rows_hex] async def get_interested_puzzle_hash_wallet_id(self, puzzle_hash: bytes32) -> Optional[int]: cursor = await self.db_connection.execute( "SELECT wallet_id FROM interested_puzzle_hashes WHERE puzzle_hash=?", (puzzle_hash.hex(),) ) row = await cursor.fetchone() if row is None: return None return row[0] async def add_interested_puzzle_hash( self, puzzle_hash: bytes32, wallet_id: int, in_transaction: bool = False ) -> None: if not in_transaction: await self.db_wrapper.lock.acquire() try: cursor = await self.db_connection.execute( "INSERT OR REPLACE INTO interested_puzzle_hashes VALUES (?, ?)", (puzzle_hash.hex(), wallet_id) ) await cursor.close() finally: if not in_transaction: await self.db_connection.commit() self.db_wrapper.lock.release() async def remove_interested_puzzle_hash(self, puzzle_hash: bytes32, in_transaction: bool = False) -> None: if not in_transaction: await self.db_wrapper.lock.acquire() try: cursor = await self.db_connection.execute( "DELETE FROM interested_puzzle_hashes WHERE puzzle_hash=?", (puzzle_hash.hex(),) ) await cursor.close() finally: if not in_transaction: await self.db_connection.commit() self.db_wrapper.lock.release()
37.81
115
0.649035
from typing import List, Tuple, Optional import aiosqlite from ethgreen.types.blockchain_format.sized_bytes import bytes32 from ethgreen.util.db_wrapper import DBWrapper class WalletInterestedStore: db_connection: aiosqlite.Connection db_wrapper: DBWrapper @classmethod async def create(cls, wrapper: DBWrapper): self = cls() self.db_connection = wrapper.db self.db_wrapper = wrapper await self.db_connection.execute("CREATE TABLE IF NOT EXISTS interested_coins(coin_name text PRIMARY KEY)") await self.db_connection.execute( "CREATE TABLE IF NOT EXISTS interested_puzzle_hashes(puzzle_hash text PRIMARY KEY, wallet_id integer)" ) await self.db_connection.commit() return self async def _clear_database(self): cursor = await self.db_connection.execute("DELETE FROM puzzle_hashes") await cursor.close() cursor = await self.db_connection.execute("DELETE FROM interested_coins") await cursor.close() await self.db_connection.commit() async def get_interested_coin_ids(self) -> List[bytes32]: cursor = await self.db_connection.execute("SELECT coin_name FROM interested_coins") rows_hex = await cursor.fetchall() return [bytes32(bytes.fromhex(row[0])) for row in rows_hex] async def add_interested_coin_id(self, coin_id: bytes32, in_transaction: bool = False) -> None: if not in_transaction: await self.db_wrapper.lock.acquire() try: cursor = await self.db_connection.execute( "INSERT OR REPLACE INTO interested_coins VALUES (?)", (coin_id.hex(),) ) await cursor.close() finally: if not in_transaction: await self.db_connection.commit() self.db_wrapper.lock.release() async def get_interested_puzzle_hashes(self) -> List[Tuple[bytes32, int]]: cursor = await self.db_connection.execute("SELECT puzzle_hash, wallet_id FROM interested_puzzle_hashes") rows_hex = await cursor.fetchall() return [(bytes32(bytes.fromhex(row[0])), row[1]) for row in rows_hex] async def get_interested_puzzle_hash_wallet_id(self, puzzle_hash: bytes32) -> Optional[int]: cursor = await self.db_connection.execute( "SELECT wallet_id FROM interested_puzzle_hashes WHERE puzzle_hash=?", (puzzle_hash.hex(),) ) row = await cursor.fetchone() if row is None: return None return row[0] async def add_interested_puzzle_hash( self, puzzle_hash: bytes32, wallet_id: int, in_transaction: bool = False ) -> None: if not in_transaction: await self.db_wrapper.lock.acquire() try: cursor = await self.db_connection.execute( "INSERT OR REPLACE INTO interested_puzzle_hashes VALUES (?, ?)", (puzzle_hash.hex(), wallet_id) ) await cursor.close() finally: if not in_transaction: await self.db_connection.commit() self.db_wrapper.lock.release() async def remove_interested_puzzle_hash(self, puzzle_hash: bytes32, in_transaction: bool = False) -> None: if not in_transaction: await self.db_wrapper.lock.acquire() try: cursor = await self.db_connection.execute( "DELETE FROM interested_puzzle_hashes WHERE puzzle_hash=?", (puzzle_hash.hex(),) ) await cursor.close() finally: if not in_transaction: await self.db_connection.commit() self.db_wrapper.lock.release()
true
true
f70ec721475523ddf5e94b147c1e716254668fe7
574
py
Python
ui/maintenance_protocols/decompress.py
liyao001/BioQueue
2b2c9f023b988fd926a037eb4755f639632b2991
[ "Apache-2.0" ]
33
2017-03-12T16:26:45.000Z
2021-04-30T05:37:35.000Z
ui/maintenance_protocols/decompress.py
liyao001/BioQueue
2b2c9f023b988fd926a037eb4755f639632b2991
[ "Apache-2.0" ]
6
2017-04-21T08:44:47.000Z
2018-11-11T16:20:22.000Z
ui/maintenance_protocols/decompress.py
liyao001/BioQueue
2b2c9f023b988fd926a037eb4755f639632b2991
[ "Apache-2.0" ]
13
2017-03-12T16:26:56.000Z
2020-04-20T05:35:00.000Z
#!/usr/bin/env python # -*- coding: utf-8 -*- # @Time : 14/01/2018 01:04 AM # @Project : BioQueue # @Author : Li Yao # @File : gunzip.py def get_sub_protocol(db_obj, protocol_parent, step_order_start=1): steps = list() steps.append(db_obj(software='gunzip', parameter='{{InputFile}}', parent=protocol_parent, user_id=0, hash='541df26aff8e4d054a57c7e3717e91ca', step_order=step_order_start)) return step_order_start+len(steps), steps
31.888889
66
0.560976
def get_sub_protocol(db_obj, protocol_parent, step_order_start=1): steps = list() steps.append(db_obj(software='gunzip', parameter='{{InputFile}}', parent=protocol_parent, user_id=0, hash='541df26aff8e4d054a57c7e3717e91ca', step_order=step_order_start)) return step_order_start+len(steps), steps
true
true
f70ec72a7dd5c36e1338a62ba92bead14d1e8e76
846
py
Python
altair_transform/transform/joinaggregate.py
jakevdp/altair-transform
63094097e05891a8d12005b411ad8585b94e81fd
[ "MIT" ]
38
2019-03-23T20:12:06.000Z
2021-10-07T21:03:24.000Z
altair_transform/transform/joinaggregate.py
altair-viz/altair-transform
b65bf854de1e80f931e063d8fb2ec938773826fb
[ "MIT" ]
13
2019-07-19T03:33:07.000Z
2021-06-29T15:34:19.000Z
altair_transform/transform/joinaggregate.py
altair-viz/altair-transform
b65bf854de1e80f931e063d8fb2ec938773826fb
[ "MIT" ]
11
2019-07-19T02:48:35.000Z
2021-11-01T00:07:41.000Z
import altair as alt import pandas as pd from .visitor import visit from .aggregate import AGG_REPLACEMENTS @visit.register(alt.JoinAggregateTransform) def visit_joinaggregate( transform: alt.JoinAggregateTransform, df: pd.DataFrame ) -> pd.DataFrame: transform = transform.to_dict() groupby = transform.get("groupby") for aggregate in transform["joinaggregate"]: op = aggregate["op"] field = aggregate["field"] col = aggregate["as"] op = AGG_REPLACEMENTS.get(op, op) if field == "*" and field not in df.columns: field = df.columns[0] if groupby is None: df[col] = df[field].aggregate(op) else: result = df.groupby(groupby)[field].aggregate(op) result.name = col df = df.join(result, on=groupby) return df
29.172414
61
0.631206
import altair as alt import pandas as pd from .visitor import visit from .aggregate import AGG_REPLACEMENTS @visit.register(alt.JoinAggregateTransform) def visit_joinaggregate( transform: alt.JoinAggregateTransform, df: pd.DataFrame ) -> pd.DataFrame: transform = transform.to_dict() groupby = transform.get("groupby") for aggregate in transform["joinaggregate"]: op = aggregate["op"] field = aggregate["field"] col = aggregate["as"] op = AGG_REPLACEMENTS.get(op, op) if field == "*" and field not in df.columns: field = df.columns[0] if groupby is None: df[col] = df[field].aggregate(op) else: result = df.groupby(groupby)[field].aggregate(op) result.name = col df = df.join(result, on=groupby) return df
true
true
f70ec8116b154a5c5324c8498dcdda97090753ab
9,785
py
Python
habitat/tasks/nav/object_nav_task.py
Ram81/habitat-imitation-baselines
c6e11c8ebadbf1260e1bed58a5b8dfb7faf6a505
[ "MIT" ]
null
null
null
habitat/tasks/nav/object_nav_task.py
Ram81/habitat-imitation-baselines
c6e11c8ebadbf1260e1bed58a5b8dfb7faf6a505
[ "MIT" ]
null
null
null
habitat/tasks/nav/object_nav_task.py
Ram81/habitat-imitation-baselines
c6e11c8ebadbf1260e1bed58a5b8dfb7faf6a505
[ "MIT" ]
null
null
null
# Copyright (c) Facebook, Inc. and its affiliates. # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import os from typing import Any, List, Optional import attr from cv2 import log import numpy as np from gym import spaces from habitat.config import Config from habitat.core.dataset import SceneState from habitat.core.logging import logger from habitat.core.registry import registry from habitat.core.simulator import AgentState, Sensor, SensorTypes from habitat.core.utils import not_none_validator from habitat.tasks.nav.nav import ( NavigationEpisode, NavigationGoal, NavigationTask ) try: from habitat.datasets.object_nav.object_nav_dataset import ( ObjectNavDatasetV1, ) except ImportError: pass task_cat2mpcat40 = [ 3, # ('chair', 2, 0) 5, # ('table', 4, 1) 6, # ('picture', 5, 2) 7, # ('cabinet', 6, 3) 8, # ('cushion', 7, 4) 10, # ('sofa', 9, 5), 11, # ('bed', 10, 6) 13, # ('chest_of_drawers', 12, 7), 14, # ('plant', 13, 8) 15, # ('sink', 14, 9) 18, # ('toilet', 17, 10), 19, # ('stool', 18, 11), 20, # ('towel', 19, 12) 22, # ('tv_monitor', 21, 13) 23, # ('shower', 22, 14) 25, # ('bathtub', 24, 15) 26, # ('counter', 25, 16), 27, # ('fireplace', 26, 17), 33, # ('gym_equipment', 32, 18), 34, # ('seating', 33, 19), 38, # ('clothes', 37, 20), 43, # ('foodstuff', 42, 21), 44, # ('stationery', 43, 22), 45, # ('fruit', 44, 23), 46, # ('plaything', 45, 24), 47, # ('hand_tool', 46, 25), 48, # ('game_equipment', 47, 26), 49, # ('kitchenware', 48, 27) ] mapping_mpcat40_to_goal21 = { 3: 1, 5: 2, 6: 3, 7: 4, 8: 5, 10: 6, 11: 7, 13: 8, 14: 9, 15: 10, 18: 11, 19: 12, 20: 13, 22: 14, 23: 15, 25: 16, 26: 17, 27: 18, 33: 19, 34: 20, 38: 21, 43: 22, # ('foodstuff', 42, task_cat: 21) 44: 28, # ('stationery', 43, task_cat: 22) 45: 26, # ('fruit', 44, task_cat: 23) 46: 25, # ('plaything', 45, task_cat: 24) 47: 24, # ('hand_tool', 46, task_cat: 25) 48: 23, # ('game_equipment', 47, task_cat: 26) 49: 27, # ('kitchenware', 48, task_cat: 27) } @attr.s(auto_attribs=True, kw_only=True) class AgentStateSpec: r"""Agent data specifications that capture states of agent and sensor in replay state. """ position: Optional[List[float]] = attr.ib(default=None) rotation: Optional[List[float]] = attr.ib(default=None) sensor_data: Optional[dict] = attr.ib(default=None) @attr.s(auto_attribs=True, kw_only=True) class ReplayActionSpec: r"""Replay specifications that capture metadata associated with action. """ action: str = attr.ib(default=None, validator=not_none_validator) agent_state: Optional[AgentStateSpec] = attr.ib(default=None) @attr.s(auto_attribs=True, kw_only=True) class ObjectGoalNavEpisode(NavigationEpisode): r"""ObjectGoal Navigation Episode :param object_category: Category of the obect """ object_category: Optional[str] = None reference_replay: Optional[List[ReplayActionSpec]] = None scene_state: Optional[List[SceneState]] = None is_thda: Optional[bool] = False scene_dataset: Optional[str] = "mp3d" @property def goals_key(self) -> str: r"""The key to retrieve the goals""" return f"{os.path.basename(self.scene_id)}_{self.object_category}" @attr.s(auto_attribs=True) class ObjectViewLocation: r"""ObjectViewLocation provides information about a position around an object goal usually that is navigable and the object is visible with specific agent configuration that episode's dataset was created. that is target for navigation. That can be specify object_id, position and object category. An important part for metrics calculation are view points that describe success area for the navigation. Args: agent_state: navigable AgentState with a position and a rotation where the object is visible. iou: an intersection of a union of the object and a rectangle in the center of view. This metric is used to evaluate how good is the object view form current position. Higher iou means better view, iou equals 1.0 if whole object is inside of the rectangle and no pixel inside the rectangle belongs to anything except the object. """ agent_state: AgentState iou: Optional[float] @attr.s(auto_attribs=True, kw_only=True) class ObjectGoal(NavigationGoal): r"""Object goal provides information about an object that is target for navigation. That can be specify object_id, position and object category. An important part for metrics calculation are view points that describe success area for the navigation. Args: object_id: id that can be used to retrieve object from the semantic scene annotation object_name: name of the object object_category: object category name usually similar to scene semantic categories room_id: id of a room where object is located, can be used to retrieve room from the semantic scene annotation room_name: name of the room, where object is located view_points: navigable positions around the object with specified proximity of the object surface used for navigation metrics calculation. The object is visible from these positions. """ object_id: str = attr.ib(default=None, validator=not_none_validator) object_name: Optional[str] = None object_name_id: Optional[int] = None object_category: Optional[str] = None room_id: Optional[str] = None room_name: Optional[str] = None view_points: Optional[List[ObjectViewLocation]] = None @registry.register_sensor class ObjectGoalSensor(Sensor): r"""A sensor for Object Goal specification as observations which is used in ObjectGoal Navigation. The goal is expected to be specified by object_id or semantic category id. For the agent in simulator the forward direction is along negative-z. In polar coordinate format the angle returned is azimuth to the goal. Args: sim: a reference to the simulator for calculating task observations. config: a config for the ObjectGoalSensor sensor. Can contain field GOAL_SPEC that specifies which id use for goal specification, GOAL_SPEC_MAX_VAL the maximum object_id possible used for observation space definition. dataset: a Object Goal navigation dataset that contains dictionaries of categories id to text mapping. """ cls_uuid: str = "objectgoal" def __init__( self, sim, config: Config, dataset: "ObjectNavDatasetV1", *args: Any, **kwargs: Any, ): self._sim = sim self._dataset = dataset super().__init__(config=config) def _get_uuid(self, *args: Any, **kwargs: Any) -> str: return self.cls_uuid def _get_sensor_type(self, *args: Any, **kwargs: Any): return SensorTypes.SEMANTIC def _get_observation_space(self, *args: Any, **kwargs: Any): sensor_shape = (1,) max_value = self.config.GOAL_SPEC_MAX_VAL - 1 if self.config.GOAL_SPEC == "TASK_CATEGORY_ID": max_value = max( self._dataset.category_to_task_category_id.values() ) logger.info("max object cat: {}".format(max_value)) logger.info("cats: {}".format(self._dataset.category_to_task_category_id.values())) return spaces.Box( low=0, high=max_value, shape=sensor_shape, dtype=np.int64 ) def get_observation( self, observations, *args: Any, episode: ObjectGoalNavEpisode, **kwargs: Any, ) -> Optional[int]: if len(episode.goals) == 0: logger.error( f"No goal specified for episode {episode.episode_id}." ) return None if not isinstance(episode.goals[0], ObjectGoal): logger.error( f"First goal should be ObjectGoal, episode {episode.episode_id}." ) return None category_name = episode.object_category if self.config.GOAL_SPEC == "TASK_CATEGORY_ID": return np.array( [self._dataset.category_to_task_category_id[category_name]], dtype=np.int64, ) elif self.config.GOAL_SPEC == "OBJECT_ID": obj_goal = episode.goals[0] assert isinstance(obj_goal, ObjectGoal) # for type checking return np.array([obj_goal.object_name_id], dtype=np.int64) else: raise RuntimeError( "Wrong GOAL_SPEC specified for ObjectGoalSensor." ) @registry.register_task(name="ObjectNav-v1") class ObjectNavigationTask(NavigationTask): r"""An Object Navigation Task class for a task specific methods. Used to explicitly state a type of the task in config. """ _is_episode_active: bool _prev_action: int def __init__(self, **kwargs) -> None: super().__init__(**kwargs) self._is_episode_active = False def overwrite_sim_config(self, sim_config, episode): super().overwrite_sim_config(sim_config, episode) sim_config.defrost() sim_config.scene_state = episode.scene_state sim_config.freeze() return sim_config def _check_episode_is_active(self, action, *args: Any, **kwargs: Any) -> bool: return not getattr(self, "is_stop_called", False)
33.62543
95
0.6465
import os from typing import Any, List, Optional import attr from cv2 import log import numpy as np from gym import spaces from habitat.config import Config from habitat.core.dataset import SceneState from habitat.core.logging import logger from habitat.core.registry import registry from habitat.core.simulator import AgentState, Sensor, SensorTypes from habitat.core.utils import not_none_validator from habitat.tasks.nav.nav import ( NavigationEpisode, NavigationGoal, NavigationTask ) try: from habitat.datasets.object_nav.object_nav_dataset import ( ObjectNavDatasetV1, ) except ImportError: pass task_cat2mpcat40 = [ 3, 5, 6, 7, 8, 10, 11, 13, 14, 15, 18, 19, 20, 22, 23, 25, 26, 27, 33, 34, 38, 43, 44, 45, 46, 47, 48, 49, ] mapping_mpcat40_to_goal21 = { 3: 1, 5: 2, 6: 3, 7: 4, 8: 5, 10: 6, 11: 7, 13: 8, 14: 9, 15: 10, 18: 11, 19: 12, 20: 13, 22: 14, 23: 15, 25: 16, 26: 17, 27: 18, 33: 19, 34: 20, 38: 21, 43: 22, 44: 28, 45: 26, 46: 25, 47: 24, 48: 23, 49: 27, } @attr.s(auto_attribs=True, kw_only=True) class AgentStateSpec: position: Optional[List[float]] = attr.ib(default=None) rotation: Optional[List[float]] = attr.ib(default=None) sensor_data: Optional[dict] = attr.ib(default=None) @attr.s(auto_attribs=True, kw_only=True) class ReplayActionSpec: action: str = attr.ib(default=None, validator=not_none_validator) agent_state: Optional[AgentStateSpec] = attr.ib(default=None) @attr.s(auto_attribs=True, kw_only=True) class ObjectGoalNavEpisode(NavigationEpisode): object_category: Optional[str] = None reference_replay: Optional[List[ReplayActionSpec]] = None scene_state: Optional[List[SceneState]] = None is_thda: Optional[bool] = False scene_dataset: Optional[str] = "mp3d" @property def goals_key(self) -> str: return f"{os.path.basename(self.scene_id)}_{self.object_category}" @attr.s(auto_attribs=True) class ObjectViewLocation: agent_state: AgentState iou: Optional[float] @attr.s(auto_attribs=True, kw_only=True) class ObjectGoal(NavigationGoal): object_id: str = attr.ib(default=None, validator=not_none_validator) object_name: Optional[str] = None object_name_id: Optional[int] = None object_category: Optional[str] = None room_id: Optional[str] = None room_name: Optional[str] = None view_points: Optional[List[ObjectViewLocation]] = None @registry.register_sensor class ObjectGoalSensor(Sensor): cls_uuid: str = "objectgoal" def __init__( self, sim, config: Config, dataset: "ObjectNavDatasetV1", *args: Any, **kwargs: Any, ): self._sim = sim self._dataset = dataset super().__init__(config=config) def _get_uuid(self, *args: Any, **kwargs: Any) -> str: return self.cls_uuid def _get_sensor_type(self, *args: Any, **kwargs: Any): return SensorTypes.SEMANTIC def _get_observation_space(self, *args: Any, **kwargs: Any): sensor_shape = (1,) max_value = self.config.GOAL_SPEC_MAX_VAL - 1 if self.config.GOAL_SPEC == "TASK_CATEGORY_ID": max_value = max( self._dataset.category_to_task_category_id.values() ) logger.info("max object cat: {}".format(max_value)) logger.info("cats: {}".format(self._dataset.category_to_task_category_id.values())) return spaces.Box( low=0, high=max_value, shape=sensor_shape, dtype=np.int64 ) def get_observation( self, observations, *args: Any, episode: ObjectGoalNavEpisode, **kwargs: Any, ) -> Optional[int]: if len(episode.goals) == 0: logger.error( f"No goal specified for episode {episode.episode_id}." ) return None if not isinstance(episode.goals[0], ObjectGoal): logger.error( f"First goal should be ObjectGoal, episode {episode.episode_id}." ) return None category_name = episode.object_category if self.config.GOAL_SPEC == "TASK_CATEGORY_ID": return np.array( [self._dataset.category_to_task_category_id[category_name]], dtype=np.int64, ) elif self.config.GOAL_SPEC == "OBJECT_ID": obj_goal = episode.goals[0] assert isinstance(obj_goal, ObjectGoal) return np.array([obj_goal.object_name_id], dtype=np.int64) else: raise RuntimeError( "Wrong GOAL_SPEC specified for ObjectGoalSensor." ) @registry.register_task(name="ObjectNav-v1") class ObjectNavigationTask(NavigationTask): _is_episode_active: bool _prev_action: int def __init__(self, **kwargs) -> None: super().__init__(**kwargs) self._is_episode_active = False def overwrite_sim_config(self, sim_config, episode): super().overwrite_sim_config(sim_config, episode) sim_config.defrost() sim_config.scene_state = episode.scene_state sim_config.freeze() return sim_config def _check_episode_is_active(self, action, *args: Any, **kwargs: Any) -> bool: return not getattr(self, "is_stop_called", False)
true
true
f70ec8b62d9552a0f4c485bf78d176b374cc936b
8,312
py
Python
kubernetes_asyncio/client/models/v2beta1_metric_spec.py
tomplus/kubernetes-asyncio
11c3eb4d50ae822545572aa7b8c15f7153f65a1c
[ "Apache-2.0" ]
null
null
null
kubernetes_asyncio/client/models/v2beta1_metric_spec.py
tomplus/kubernetes-asyncio
11c3eb4d50ae822545572aa7b8c15f7153f65a1c
[ "Apache-2.0" ]
null
null
null
kubernetes_asyncio/client/models/v2beta1_metric_spec.py
tomplus/kubernetes-asyncio
11c3eb4d50ae822545572aa7b8c15f7153f65a1c
[ "Apache-2.0" ]
null
null
null
# coding: utf-8 """ Kubernetes No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator) # noqa: E501 The version of the OpenAPI document: v1.23.6 Generated by: https://openapi-generator.tech """ try: from inspect import getfullargspec except ImportError: from inspect import getargspec as getfullargspec import pprint import re # noqa: F401 import six from kubernetes_asyncio.client.configuration import Configuration class V2beta1MetricSpec(object): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech Do not edit the class manually. """ """ Attributes: openapi_types (dict): The key is attribute name and the value is attribute type. attribute_map (dict): The key is attribute name and the value is json key in definition. """ openapi_types = { 'container_resource': 'V2beta1ContainerResourceMetricSource', 'external': 'V2beta1ExternalMetricSource', 'object': 'V2beta1ObjectMetricSource', 'pods': 'V2beta1PodsMetricSource', 'resource': 'V2beta1ResourceMetricSource', 'type': 'str' } attribute_map = { 'container_resource': 'containerResource', 'external': 'external', 'object': 'object', 'pods': 'pods', 'resource': 'resource', 'type': 'type' } def __init__(self, container_resource=None, external=None, object=None, pods=None, resource=None, type=None, local_vars_configuration=None): # noqa: E501 """V2beta1MetricSpec - a model defined in OpenAPI""" # noqa: E501 if local_vars_configuration is None: local_vars_configuration = Configuration.get_default_copy() self.local_vars_configuration = local_vars_configuration self._container_resource = None self._external = None self._object = None self._pods = None self._resource = None self._type = None self.discriminator = None if container_resource is not None: self.container_resource = container_resource if external is not None: self.external = external if object is not None: self.object = object if pods is not None: self.pods = pods if resource is not None: self.resource = resource self.type = type @property def container_resource(self): """Gets the container_resource of this V2beta1MetricSpec. # noqa: E501 :return: The container_resource of this V2beta1MetricSpec. # noqa: E501 :rtype: V2beta1ContainerResourceMetricSource """ return self._container_resource @container_resource.setter def container_resource(self, container_resource): """Sets the container_resource of this V2beta1MetricSpec. :param container_resource: The container_resource of this V2beta1MetricSpec. # noqa: E501 :type container_resource: V2beta1ContainerResourceMetricSource """ self._container_resource = container_resource @property def external(self): """Gets the external of this V2beta1MetricSpec. # noqa: E501 :return: The external of this V2beta1MetricSpec. # noqa: E501 :rtype: V2beta1ExternalMetricSource """ return self._external @external.setter def external(self, external): """Sets the external of this V2beta1MetricSpec. :param external: The external of this V2beta1MetricSpec. # noqa: E501 :type external: V2beta1ExternalMetricSource """ self._external = external @property def object(self): """Gets the object of this V2beta1MetricSpec. # noqa: E501 :return: The object of this V2beta1MetricSpec. # noqa: E501 :rtype: V2beta1ObjectMetricSource """ return self._object @object.setter def object(self, object): """Sets the object of this V2beta1MetricSpec. :param object: The object of this V2beta1MetricSpec. # noqa: E501 :type object: V2beta1ObjectMetricSource """ self._object = object @property def pods(self): """Gets the pods of this V2beta1MetricSpec. # noqa: E501 :return: The pods of this V2beta1MetricSpec. # noqa: E501 :rtype: V2beta1PodsMetricSource """ return self._pods @pods.setter def pods(self, pods): """Sets the pods of this V2beta1MetricSpec. :param pods: The pods of this V2beta1MetricSpec. # noqa: E501 :type pods: V2beta1PodsMetricSource """ self._pods = pods @property def resource(self): """Gets the resource of this V2beta1MetricSpec. # noqa: E501 :return: The resource of this V2beta1MetricSpec. # noqa: E501 :rtype: V2beta1ResourceMetricSource """ return self._resource @resource.setter def resource(self, resource): """Sets the resource of this V2beta1MetricSpec. :param resource: The resource of this V2beta1MetricSpec. # noqa: E501 :type resource: V2beta1ResourceMetricSource """ self._resource = resource @property def type(self): """Gets the type of this V2beta1MetricSpec. # noqa: E501 type is the type of metric source. It should be one of \"ContainerResource\", \"External\", \"Object\", \"Pods\" or \"Resource\", each mapping to a matching field in the object. Note: \"ContainerResource\" type is available on when the feature-gate HPAContainerMetrics is enabled # noqa: E501 :return: The type of this V2beta1MetricSpec. # noqa: E501 :rtype: str """ return self._type @type.setter def type(self, type): """Sets the type of this V2beta1MetricSpec. type is the type of metric source. It should be one of \"ContainerResource\", \"External\", \"Object\", \"Pods\" or \"Resource\", each mapping to a matching field in the object. Note: \"ContainerResource\" type is available on when the feature-gate HPAContainerMetrics is enabled # noqa: E501 :param type: The type of this V2beta1MetricSpec. # noqa: E501 :type type: str """ if self.local_vars_configuration.client_side_validation and type is None: # noqa: E501 raise ValueError("Invalid value for `type`, must not be `None`") # noqa: E501 self._type = type def to_dict(self, serialize=False): """Returns the model properties as a dict""" result = {} def convert(x): if hasattr(x, "to_dict"): args = getfullargspec(x.to_dict).args if len(args) == 1: return x.to_dict() else: return x.to_dict(serialize) else: return x for attr, _ in six.iteritems(self.openapi_types): value = getattr(self, attr) attr = self.attribute_map.get(attr, attr) if serialize else attr if isinstance(value, list): result[attr] = list(map( lambda x: convert(x), value )) elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], convert(item[1])), value.items() )) else: result[attr] = convert(value) return result def to_str(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict()) def __repr__(self): """For `print` and `pprint`""" return self.to_str() def __eq__(self, other): """Returns true if both objects are equal""" if not isinstance(other, V2beta1MetricSpec): return False return self.to_dict() == other.to_dict() def __ne__(self, other): """Returns true if both objects are not equal""" if not isinstance(other, V2beta1MetricSpec): return True return self.to_dict() != other.to_dict()
31.366038
302
0.617661
try: from inspect import getfullargspec except ImportError: from inspect import getargspec as getfullargspec import pprint import re import six from kubernetes_asyncio.client.configuration import Configuration class V2beta1MetricSpec(object): openapi_types = { 'container_resource': 'V2beta1ContainerResourceMetricSource', 'external': 'V2beta1ExternalMetricSource', 'object': 'V2beta1ObjectMetricSource', 'pods': 'V2beta1PodsMetricSource', 'resource': 'V2beta1ResourceMetricSource', 'type': 'str' } attribute_map = { 'container_resource': 'containerResource', 'external': 'external', 'object': 'object', 'pods': 'pods', 'resource': 'resource', 'type': 'type' } def __init__(self, container_resource=None, external=None, object=None, pods=None, resource=None, type=None, local_vars_configuration=None): if local_vars_configuration is None: local_vars_configuration = Configuration.get_default_copy() self.local_vars_configuration = local_vars_configuration self._container_resource = None self._external = None self._object = None self._pods = None self._resource = None self._type = None self.discriminator = None if container_resource is not None: self.container_resource = container_resource if external is not None: self.external = external if object is not None: self.object = object if pods is not None: self.pods = pods if resource is not None: self.resource = resource self.type = type @property def container_resource(self): return self._container_resource @container_resource.setter def container_resource(self, container_resource): self._container_resource = container_resource @property def external(self): return self._external @external.setter def external(self, external): self._external = external @property def object(self): return self._object @object.setter def object(self, object): self._object = object @property def pods(self): return self._pods @pods.setter def pods(self, pods): self._pods = pods @property def resource(self): return self._resource @resource.setter def resource(self, resource): self._resource = resource @property def type(self): return self._type @type.setter def type(self, type): if self.local_vars_configuration.client_side_validation and type is None: raise ValueError("Invalid value for `type`, must not be `None`") self._type = type def to_dict(self, serialize=False): result = {} def convert(x): if hasattr(x, "to_dict"): args = getfullargspec(x.to_dict).args if len(args) == 1: return x.to_dict() else: return x.to_dict(serialize) else: return x for attr, _ in six.iteritems(self.openapi_types): value = getattr(self, attr) attr = self.attribute_map.get(attr, attr) if serialize else attr if isinstance(value, list): result[attr] = list(map( lambda x: convert(x), value )) elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], convert(item[1])), value.items() )) else: result[attr] = convert(value) return result def to_str(self): return pprint.pformat(self.to_dict()) def __repr__(self): return self.to_str() def __eq__(self, other): if not isinstance(other, V2beta1MetricSpec): return False return self.to_dict() == other.to_dict() def __ne__(self, other): if not isinstance(other, V2beta1MetricSpec): return True return self.to_dict() != other.to_dict()
true
true
f70ec900d5a3779330fc5069041b7d547061f02c
765
py
Python
backend/api/fields.py
NeutralDread/overworld
0f608f83c10b666f2ed2cbe519186cf425a5e354
[ "Apache-2.0" ]
null
null
null
backend/api/fields.py
NeutralDread/overworld
0f608f83c10b666f2ed2cbe519186cf425a5e354
[ "Apache-2.0" ]
null
null
null
backend/api/fields.py
NeutralDread/overworld
0f608f83c10b666f2ed2cbe519186cf425a5e354
[ "Apache-2.0" ]
null
null
null
_game_fields = [ 'cover.image_id', 'first_release_date', 'genres.name', 'involved_companies.developer', 'involved_companies.publisher', 'involved_companies.company.country', 'involved_companies.company.name', 'name', 'platforms.name', 'screenshots.image_id', 'slug', 'summary', 'time_to_beat.normally', 'themes.name', ] _search_fields = [ 'first_release_date', 'name', 'slug', ] _popular_fields = [ 'cover.image_id', 'name', 'popularity', ] _backdrop_fields = [ 'name', 'screenshots.image_id', 'slug', ] game_fields = ','.join(_game_fields) search_fields = ','.join(_search_fields) popular_fields = ','.join(_popular_fields) backdrop_fields = ','.join(_backdrop_fields)
19.615385
44
0.648366
_game_fields = [ 'cover.image_id', 'first_release_date', 'genres.name', 'involved_companies.developer', 'involved_companies.publisher', 'involved_companies.company.country', 'involved_companies.company.name', 'name', 'platforms.name', 'screenshots.image_id', 'slug', 'summary', 'time_to_beat.normally', 'themes.name', ] _search_fields = [ 'first_release_date', 'name', 'slug', ] _popular_fields = [ 'cover.image_id', 'name', 'popularity', ] _backdrop_fields = [ 'name', 'screenshots.image_id', 'slug', ] game_fields = ','.join(_game_fields) search_fields = ','.join(_search_fields) popular_fields = ','.join(_popular_fields) backdrop_fields = ','.join(_backdrop_fields)
true
true
f70ec9f26a955802a506cc9f3c9095b2b09eb7e3
2,934
py
Python
Ryu/topology.py
JoshuaYu-crash/C4EP2-2021
62e2e5a2fb396c598e7e8a265515c342fb78f63e
[ "MIT" ]
null
null
null
Ryu/topology.py
JoshuaYu-crash/C4EP2-2021
62e2e5a2fb396c598e7e8a265515c342fb78f63e
[ "MIT" ]
null
null
null
Ryu/topology.py
JoshuaYu-crash/C4EP2-2021
62e2e5a2fb396c598e7e8a265515c342fb78f63e
[ "MIT" ]
null
null
null
from flask import Flask, render_template, request, jsonify from pyecharts import options as opts from pyecharts.charts import Graph import json import redis from flask_cors import * r = redis.Redis(host="127.0.0.1", port=6379) app = Flask(__name__) CORS(app, supports_credentials=True) @app.route("/dockermsg", methods=["POST"]) def dockerMsg(): data = request.json host = data["host"] datalist = data["data"] # print(datalist) r.set(host, json.dumps(datalist)) return "ok" @app.route("/getdockermsg", methods=["GET"]) def getDockerMsg(): host = request.args.get("host") docker = request.args.get("dockerdata") dockers = json.loads(r.get(host)) tar = None # print(dockers) for doc in dockers: print(doc["NetworkSettings"]["Networks"]["bridge"]["IPAddress"], docker) if docker == doc["NetworkSettings"]["Networks"]["bridge"]["IPAddress"]: tar = doc break print(tar) return jsonify(tar) def graph_base() -> Graph: nodes = [] links = [] categories = [ {"symbol": "circle", 'name': 'ryu'}, {"symbol": "diamond", 'name': 'host'}, {"symbol": "roundRect", 'name': 'dockerdata'}, ] ryu = opts.GraphNode(name="RYU", symbol_size=40, category=0) # symbol='roundRect' nodes.append(ryu) doc_id = 1 for key in r.keys(): host = opts.GraphNode(name=key, symbol_size=30, category=1) # symbol='diamond' nodes.append(host) ryuHostLink = opts.GraphLink(source="RYU", target=key) links.append(ryuHostLink) dockerlist = json.loads(r.get(key)) for doc in dockerlist: docName = doc["Names"][0] docInfo = str(key, encoding='utf-8') + '/' + doc["NetworkSettings"]["Networks"]["bridge"]["IPAddress"] new_node = opts.GraphNode(name=str(doc_id) + docName, symbol_size=20, category=2, value=docInfo) nodes.append(new_node) hostDocLink = opts.GraphLink(source=key, target=str(doc_id) + docName) links.append(hostDocLink) doc_id += 1 linestyle_opts = opts.LineStyleOpts(is_show=True, width=2, curve=0.1, type_="solid", color="orange", ) g = ( Graph() .add("", nodes, links, repulsion=1000, categories=categories, label_opts=opts.LabelOpts(is_show=True, position="left", color='white'), linestyle_opts=linestyle_opts) .set_global_opts(title_opts=opts.TitleOpts(title="")) ) return g @app.route("/graphchart", methods=["GET"]) def get_bar_chart(): c = graph_base() return c.dump_options_with_quotes() if __name__ == '__main__': app.run(host="127.0.0.1", port=5000, debug=True)
33.340909
114
0.576005
from flask import Flask, render_template, request, jsonify from pyecharts import options as opts from pyecharts.charts import Graph import json import redis from flask_cors import * r = redis.Redis(host="127.0.0.1", port=6379) app = Flask(__name__) CORS(app, supports_credentials=True) @app.route("/dockermsg", methods=["POST"]) def dockerMsg(): data = request.json host = data["host"] datalist = data["data"] r.set(host, json.dumps(datalist)) return "ok" @app.route("/getdockermsg", methods=["GET"]) def getDockerMsg(): host = request.args.get("host") docker = request.args.get("dockerdata") dockers = json.loads(r.get(host)) tar = None for doc in dockers: print(doc["NetworkSettings"]["Networks"]["bridge"]["IPAddress"], docker) if docker == doc["NetworkSettings"]["Networks"]["bridge"]["IPAddress"]: tar = doc break print(tar) return jsonify(tar) def graph_base() -> Graph: nodes = [] links = [] categories = [ {"symbol": "circle", 'name': 'ryu'}, {"symbol": "diamond", 'name': 'host'}, {"symbol": "roundRect", 'name': 'dockerdata'}, ] ryu = opts.GraphNode(name="RYU", symbol_size=40, category=0) nodes.append(ryu) doc_id = 1 for key in r.keys(): host = opts.GraphNode(name=key, symbol_size=30, category=1) nodes.append(host) ryuHostLink = opts.GraphLink(source="RYU", target=key) links.append(ryuHostLink) dockerlist = json.loads(r.get(key)) for doc in dockerlist: docName = doc["Names"][0] docInfo = str(key, encoding='utf-8') + '/' + doc["NetworkSettings"]["Networks"]["bridge"]["IPAddress"] new_node = opts.GraphNode(name=str(doc_id) + docName, symbol_size=20, category=2, value=docInfo) nodes.append(new_node) hostDocLink = opts.GraphLink(source=key, target=str(doc_id) + docName) links.append(hostDocLink) doc_id += 1 linestyle_opts = opts.LineStyleOpts(is_show=True, width=2, curve=0.1, type_="solid", color="orange", ) g = ( Graph() .add("", nodes, links, repulsion=1000, categories=categories, label_opts=opts.LabelOpts(is_show=True, position="left", color='white'), linestyle_opts=linestyle_opts) .set_global_opts(title_opts=opts.TitleOpts(title="")) ) return g @app.route("/graphchart", methods=["GET"]) def get_bar_chart(): c = graph_base() return c.dump_options_with_quotes() if __name__ == '__main__': app.run(host="127.0.0.1", port=5000, debug=True)
true
true
f70eca4f5372f1f42a5b19b9379ac5026d17bfca
829
py
Python
app/core/migrations/0010_requestlogs.py
shravands/django-restapi-recipe
c21d01ab3b0d92d249e638a1b503ea54dd6d69bd
[ "MIT" ]
null
null
null
app/core/migrations/0010_requestlogs.py
shravands/django-restapi-recipe
c21d01ab3b0d92d249e638a1b503ea54dd6d69bd
[ "MIT" ]
null
null
null
app/core/migrations/0010_requestlogs.py
shravands/django-restapi-recipe
c21d01ab3b0d92d249e638a1b503ea54dd6d69bd
[ "MIT" ]
null
null
null
# Generated by Django 3.0.3 on 2020-06-09 08:55 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('core', '0009_user_username'), ] operations = [ migrations.CreateModel( name='RequestLogs', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('ip_address', models.CharField(blank=True, max_length=255)), ('user_id', models.IntegerField(blank=True)), ('method_type', models.CharField(blank=True, max_length=50)), ('request_path', models.CharField(blank=True, max_length=255)), ('response_code', models.CharField(blank=True, max_length=15)), ], ), ]
33.16
114
0.589867
from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('core', '0009_user_username'), ] operations = [ migrations.CreateModel( name='RequestLogs', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('ip_address', models.CharField(blank=True, max_length=255)), ('user_id', models.IntegerField(blank=True)), ('method_type', models.CharField(blank=True, max_length=50)), ('request_path', models.CharField(blank=True, max_length=255)), ('response_code', models.CharField(blank=True, max_length=15)), ], ), ]
true
true