hexsha
stringlengths
40
40
size
int64
3
1.03M
ext
stringclasses
10 values
lang
stringclasses
1 value
max_stars_repo_path
stringlengths
3
972
max_stars_repo_name
stringlengths
6
130
max_stars_repo_head_hexsha
stringlengths
40
78
max_stars_repo_licenses
sequencelengths
1
10
max_stars_count
int64
1
191k
⌀
max_stars_repo_stars_event_min_datetime
stringlengths
24
24
⌀
max_stars_repo_stars_event_max_datetime
stringlengths
24
24
⌀
max_issues_repo_path
stringlengths
3
972
max_issues_repo_name
stringlengths
6
130
max_issues_repo_head_hexsha
stringlengths
40
78
max_issues_repo_licenses
sequencelengths
1
10
max_issues_count
int64
1
116k
⌀
max_issues_repo_issues_event_min_datetime
stringlengths
24
24
⌀
max_issues_repo_issues_event_max_datetime
stringlengths
24
24
⌀
max_forks_repo_path
stringlengths
3
972
max_forks_repo_name
stringlengths
6
130
max_forks_repo_head_hexsha
stringlengths
40
78
max_forks_repo_licenses
sequencelengths
1
10
max_forks_count
int64
1
105k
⌀
max_forks_repo_forks_event_min_datetime
stringlengths
24
24
⌀
max_forks_repo_forks_event_max_datetime
stringlengths
24
24
⌀
content
stringlengths
3
1.03M
avg_line_length
float64
1.13
941k
max_line_length
int64
2
941k
alphanum_fraction
float64
0
1
117cd8a1fc1b6cb288bd166e5b1de8219dbef713
1,968
py
Python
parser/python/validate.py
jnorwood/NNEF-Tools
5eb3755b5322040d42893e41b15093337abe04ce
[ "Apache-2.0" ]
null
null
null
parser/python/validate.py
jnorwood/NNEF-Tools
5eb3755b5322040d42893e41b15093337abe04ce
[ "Apache-2.0" ]
null
null
null
parser/python/validate.py
jnorwood/NNEF-Tools
5eb3755b5322040d42893e41b15093337abe04ce
[ "Apache-2.0" ]
null
null
null
# Copyright (c) 2017 The Khronos Group Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import nnef import argparse if __name__ == '__main__': ap = argparse.ArgumentParser() ap.add_argument('path', type=str, help='path to the model to validate') ap.add_argument('--stdlib', type=str, help='file name of alternate standard operation definitions ' '(defaults to all-primitive definitions)', default='') ap.add_argument('--lower', type=str, help='comma separated list of operations to lower (if defined as compound)', default='') ap.add_argument('--shapes', action="store_true", help='perform shape validation as well') args = ap.parse_args() stdlib = '' if args.stdlib: try: with open(args.stdlib) as file: stdlib = file.read() except FileNotFoundError as e: print('Could not open file: ' + args.stdlib) exit(-1) try: graph = nnef.load_graph(args.path, stdlib=stdlib, lowered=args.lower.split(',')) except nnef.Error as err: print('Parse error: ' + str(err)) exit(-1) if args.shapes: try: nnef.infer_shapes(graph) except nnef.Error as err: print('Shape error: ' + str(err)) exit(-1) print(nnef.format_graph(graph.name, graph.inputs, graph.outputs, graph.operations)) print('Validation succeeded')
36.444444
117
0.643293
82654da5b66d8c54bff2b307bcff5b804bd3119c
18,195
py
Python
armi/physics/neutronics/latticePhysics/latticePhysicsInterface.py
celikten/armi
4e100dd514a59caa9c502bd5a0967fd77fdaf00e
[ "Apache-2.0" ]
null
null
null
armi/physics/neutronics/latticePhysics/latticePhysicsInterface.py
celikten/armi
4e100dd514a59caa9c502bd5a0967fd77fdaf00e
[ "Apache-2.0" ]
null
null
null
armi/physics/neutronics/latticePhysics/latticePhysicsInterface.py
celikten/armi
4e100dd514a59caa9c502bd5a0967fd77fdaf00e
[ "Apache-2.0" ]
null
null
null
# Copyright 2019 TerraPower, LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """" Lattice Physics Interface Parent classes for codes responsible for generating broad-group cross sections """ import os import shutil from armi import nuclearDataIO from armi import interfaces, runLog from armi.utils import codeTiming from armi.physics import neutronics from armi.physics.neutronics.const import CONF_CROSS_SECTION from armi.utils.customExceptions import important LATTICE_PHYSICS = "latticePhysics" @important def SkippingXsGen_BuChangedLessThanTolerance(tolerance): return "Skipping XS Generation this cycle because median block burnups changes less than {}%".format( tolerance ) def setBlockNeutronVelocities(r, neutronVelocities): """ Set the ``mgNeutronVelocity`` parameter for each block using the ``neutronVelocities`` dictionary data. Parameters ---------- neutronVelocities : dict Dictionary that is keyed with the ``representativeBlock`` XS IDs with values of multigroup neutron velocity data computed by MC2. Raises ------ ValueError Multi-group neutron velocities was not computed during the cross section calculation. """ for b in r.core.getBlocks(): xsID = b.getMicroSuffix() if xsID not in neutronVelocities: raise ValueError( "Cannot assign multi-group neutron velocity to {} because it does not exist in " "the neutron velocities dictionary with keys: {}. The XS library does not contain " "data for the {} xsid.".format(b, neutronVelocities.keys(), xsID) ) b.p.mgNeutronVelocity = neutronVelocities[b.getMicroSuffix()] class LatticePhysicsInterface(interfaces.Interface): """Class for interacting with lattice physics codes.""" function = LATTICE_PHYSICS def __init__(self, r, cs): interfaces.Interface.__init__(self, r, cs) # Set to True by default, but should be disabled when perturbed cross sections are generated. self._updateBlockNeutronVelocities = True # Geometry options available through the lattice physics interfaces self._ZERO_DIMENSIONAL_GEOM = "0D" self._ONE_DIMENSIONAL_GEOM = "1D" self._TWO_DIMENSIONAL_GEOM = "2D" self._SLAB_MODEL = " slab" self._CYLINDER_MODEL = " cylinder" self._HEX_MODEL = " hex" self._burnupTolerance = self.cs["tolerateBurnupChange"] self._oldXsIdsAndBurnup = {} self.executablePath = self._getExecutablePath() self.executableRoot = os.path.dirname(self.executablePath) self.includeGammaXS = neutronics.gammaTransportIsRequested( cs ) or neutronics.gammaXsAreRequested(cs) def _getExecutablePath(self): raise NotImplementedError @codeTiming.timed def interactBOC(self, cycle=0): """ Run the lattice physics code if ``genXS`` is set and update burnup groups. Generate new cross sections based off the case settings and the current state of the reactor. Notes ----- :py:meth:`armi.physics.fuelCycle.fuelHandlers.FuelHandler.interactBOC` also calls this if the ``runLatticePhysicsBeforeShuffling``setting is True. This happens because branch searches may need XS. """ self.updateXSLibrary(cycle) def updateXSLibrary(self, cycle): """ Update the current XS library, either by creating or reloading one. Parameters ---------- cycle : int The cycle that is being processed. Used to name the library. See Also -------- computeCrossSections : run lattice physics on the current reactor state no matter weather needed or not. """ runLog.important("Preparing XS for cycle {}".format(cycle)) representativeBlocks, xsIds = self._getBlocksAndXsIds() if self._newLibraryShouldBeCreated(cycle, representativeBlocks, xsIds): if self.cs["clearXS"]: self.clearXS() self.computeCrossSections( blockList=representativeBlocks, xsLibrarySuffix=self._getSuffix(cycle) ) self._renameExistingLibrariesForCycle(cycle) else: self.readExistingXSLibraries(cycle) self._checkInputs() def _renameExistingLibrariesForCycle(self, cycle): """Copy the existing neutron and/or gamma libraries into cycle-dependent files.""" shutil.copy(neutronics.ISOTXS, nuclearDataIO.getExpectedISOTXSFileName(cycle)) if self.includeGammaXS: shutil.copy( neutronics.GAMISO, nuclearDataIO.getExpectedGAMISOFileName( cycle=cycle, suffix=self._getSuffix(cycle) ), ) shutil.copy( neutronics.PMATRX, nuclearDataIO.getExpectedPMATRXFileName( cycle=cycle, suffix=self._getSuffix(cycle) ), ) def _checkInputs(self): pass def readExistingXSLibraries(self, cycle): raise NotImplementedError def makeCycleXSFilesAsBaseFiles(self, cycle): raise NotImplementedError @staticmethod def _copyLibraryFilesForCycle(cycle, libFiles): runLog.extra("Current library files: {}".format(libFiles)) for baseName, cycleName in libFiles.items(): if not os.path.exists(cycleName): if not os.path.exists(baseName): raise ValueError( "Neither {} nor {} libraries exist. Either the " "current cycle library for cycle {} should exist " "or a base library is required to continue.".format( cycleName, baseName, cycle ) ) runLog.info( "Existing library {} for cycle {} does not exist. " "The active library is {}".format(cycleName, cycle, baseName) ) else: runLog.info("Using {} as an active library".format(baseName)) if cycleName != baseName: shutil.copy(cycleName, baseName) def _readGammaBinaries(self, lib, gamisoFileName, pmatrxFileName): raise NotImplementedError( "Gamma cross sections not implemented in {}".format(self.cs["xsKernel"]) ) def _writeGammaBinaries(self, lib, gamisoFileName, pmatrxFileName): raise NotImplementedError( "Gamma cross sections not implemented in {}".format(self.cs["xsKernel"]) ) def _getSuffix(self, cycle): # pylint: disable=unused-argument, no-self-use return "" def interactCoupled(self, iteration): """ Runs on secondary coupled iterations. After a coupled iteration, the cross sections need to be regenerated. This will bring in the spectral effects of changes in densities, as well as changes in Doppler. Parameters ---------- iteration : int This is unused since cross sections are generated on a per-cycle basis. """ self.r.core.lib = None self.updateXSLibrary(self.r.p.cycle) def clearXS(self): raise NotImplementedError def interactEOC(self, cycle=None): """ Interact at the end of a cycle. Force updating cross sections at the start of the next cycle. """ self.r.core.lib = None def computeCrossSections( self, baseList=None, forceSerial=False, xsLibrarySuffix="", blockList=None ): """ Prepare a batch of inputs, execute them, and store results on reactor library. Parameters ---------- baseList : list a user-specified set of bases that will be run instead of calculating all of them forceSerial : bool, optional Will run on 1 processor in sequence instead of on many in parallel Useful for optimization/batch runs where every processor is on a different branch xsLibrarySuffix : str, optional A book-keeping suffix used in Doppler calculations blockList : list, optional List of blocks for which to generate cross sections. If None, representative blocks will be determined """ self.r.core.lib = self._generateXsLibrary( baseList, forceSerial, xsLibrarySuffix, blockList ) def _generateXsLibrary( self, baseList, forceSerial, xsLibrarySuffix, blockList, writers=None, purgeFP=True, ): raise NotImplementedError def _executeLatticePhysicsCalculation(self, returnedFromWriters, forceSerial): raise NotImplementedError def generateLatticePhysicsInputs( self, baseList, xsLibrarySuffix, blockList, xsWriters=None ): """ Write input files for the generation of cross section libraries. Parameters ---------- baseList : list A list of cross-section id strings (e.g. AA, BC) that will be generated. Default: all in reactor xsLibrarySuffix : str A suffix added to the end of the XS file names such as 'voided' for voided XS. Default: Empty blockList : list The blocks to write inputs for. xsWriters : list, optional The specified writers to write the input files Returns ------- returnedFromWriters: list A list of what this specific writer instance returns for each representative block. It is the responsibility of the subclassed interface to implement. In many cases, it is the executing agent. See Also -------- :py:meth:`terrapower.physics.neutronics.mc2.mc2Writers.Mc2V2Writer.write` :py:meth:`armi.physics.neutronics.latticePhysics.serpentWriters.SerpentWriter.write` """ returnedFromWriters = [] baseList = set(baseList or []) representativeBlocks = blockList or self.getRepresentativeBlocks() for repBlock in representativeBlocks: xsId = repBlock.getMicroSuffix() if not baseList or xsId in baseList: # write the step number to the info log runLog.info( "Creating input writer(s) for {0} with {1:65s} BU (%FIMA): {2:10.2f}".format( xsId, repBlock, repBlock.p.percentBu ) ) writers = self.getWriters(repBlock, xsLibrarySuffix, xsWriters) for writer in writers: fromWriter = writer.write() returnedFromWriters.append(fromWriter) return returnedFromWriters def getWriters(self, representativeBlock, xsLibrarySuffix, writers=None): """ Return valid lattice physics writer subclass(es). Parameters ---------- representativeBlock : Block A representative block object that can be created from a block collection. xsLibrarySuffix : str A suffix added to the end of the XS file names such as 'voided' for voided XS. Default: Empty writers : list of lattice physics writer objects, optional If the writers are known, they can be provided and constructed. Returns ------- writers : list A list of writers for the provided representative block. """ xsID = representativeBlock.getMicroSuffix() if writers: # Construct the writers that are provided writers = [ w( representativeBlock, r=self.r, externalCodeInterface=self, xsLibrarySuffix=xsLibrarySuffix, ) for w in writers ] else: geom = self.cs[CONF_CROSS_SECTION][xsID].geometry writers = self._getGeomDependentWriters( representativeBlock, xsID, geom, xsLibrarySuffix ) return writers def _getGeomDependentWriters( self, representativeBlock, xsID, geom, xsLibrarySuffix ): raise NotImplementedError def getReader(self): raise NotImplementedError def _newLibraryShouldBeCreated(self, cycle, representativeBlockList, xsIDs): """ Determines whether the cross section generator should be executed at this cycle. Criteria include: #. genXS setting is turned on #. We are beyond any requested skipCycles (restart cycles) #. The blocks have changed burnup beyond the burnup threshold #. Lattice physics kernel (e.g. MC2) hasn't already been executed for this cycle (possible if it runs during fuel handling) """ executeXSGen = bool(self.cs["genXS"] and cycle >= self.cs["skipCycles"]) idsChangedBurnup = self._checkBurnupThresholds(representativeBlockList) if executeXSGen and not idsChangedBurnup: executeXSGen = False if self.r.core._lib is not None: # pylint: disable=protected-access # justification=r.core.lib property can raise exception or load pre-generated # ISOTXS, but the interface should have responsibilty of loading # XS's have already generated for this cycle (maybe during fuel management). Should we update due to # changes that occurred during fuel management? missing = set(xsIDs) - set(self.r.core.lib.xsIDs) if missing and not executeXSGen: runLog.warning( "Even though XS generation is not activated, new XS {0} are needed. " "Perhaps a booster came in.".format(missing) ) elif missing: runLog.important( "New XS sets {0} will be generated for this cycle".format(missing) ) else: runLog.important( "No new XS needed for this cycle. {0} exist. Skipping".format( self.r.core.lib.xsIDs ) ) executeXSGen = False # no newXs return executeXSGen def _checkBurnupThresholds(self, blockList): """ Check to see if burnup has changed meaningfully. If there are, then the xs sets should be regenerated. Otherwise then go ahead and skip xs generation. This is motivated by the idea that during very long explicit equilibrium runs, it might save time to turn off xs generation at a certain point. Parameters ---------- blockList: iterable List of all blocks to examine Returns ------- idsChangedBurnup: bool flag regarding whether or not burnup changed substantially """ idsChangedBurnup = True if self._burnupTolerance > 0: idsChangedBurnup = False for b in blockList: xsID = b.getMicroSuffix() if xsID not in self._oldXsIdsAndBurnup: # Looks like a new ID was found that was not in the old ID's # have to regenerate the cross-sections this time around self._oldXsIdsAndBurnup[xsID] = b.p.percentBu idsChangedBurnup = True else: # The id was found. Now it is time to compare the burnups to determine # if there has been enough meaningful change between the runs buOld = self._oldXsIdsAndBurnup[xsID] buNow = b.p.percentBu if abs(buOld - buNow) > self._burnupTolerance: idsChangedBurnup = True # update the oldXs burnup to be the about to be newly generated xsBurnup self._oldXsIdsAndBurnup[xsID] = buNow runLog.important( "Burnup has changed in xsID {} from {} to {}. " "Recalculating Cross-sections".format(xsID, buOld, buNow) ) if not idsChangedBurnup: SkippingXsGen_BuChangedLessThanTolerance(self._burnupTolerance) return idsChangedBurnup def _getProcessesPerNode(self): raise NotImplementedError def getRepresentativeBlocks(self): """Return a list of all blocks in the problem.""" xsGroupManager = self.getInterface("xsGroups") return xsGroupManager.representativeBlocks.values() # OrderedDict def _getBlocksAndXsIds(self): """Return blocks and their xsIds.""" blocks = self.getRepresentativeBlocks() return blocks, [b.getMicroSuffix() for b in blocks] def updatePhysicsCouplingControl(self): """ Disable XS update in equilibrium cases after a while. Notes ----- This is only relevant for equilibrium cases. We have to turn off XS updates after several cyclics or else the number densities will never converge. """ if self.r.core.p.cyclics >= self.cs["numCyclicsBeforeStoppingXS"]: self.enabled(False) runLog.important( "Disabling {} because numCyclics={}".format(self, self.r.core.p.cyclics) )
37.985386
112
0.614784
afa51d355786df98b0025450846ad5820c6447a0
761
py
Python
code/exercises/07_DataStructures/ex_07_05_sequence.py
chiachang100/LearnToCodeWithPython
fe16115cb3be612d5abd8ffdbd6a14a37d6b4d52
[ "Apache-2.0" ]
null
null
null
code/exercises/07_DataStructures/ex_07_05_sequence.py
chiachang100/LearnToCodeWithPython
fe16115cb3be612d5abd8ffdbd6a14a37d6b4d52
[ "Apache-2.0" ]
null
null
null
code/exercises/07_DataStructures/ex_07_05_sequence.py
chiachang100/LearnToCodeWithPython
fe16115cb3be612d5abd8ffdbd6a14a37d6b4d52
[ "Apache-2.0" ]
null
null
null
#!/usr/bin/python # Filename: ex_sequence.py shoplist = ['apple', 'mango', 'carrot', 'banana'] name = 'swaroop' # Indexing or 'Subscription' operation print('Item 0 is', shoplist[0]) print('Item 1 is', shoplist[1]) print('Item 2 is', shoplist[2]) print('Item 3 is', shoplist[3]) print('Item -1 is', shoplist[-1]) print('Item -2 is', shoplist[-2]) print('Character 0 is', name[0]) # Slicing on a list print('Item 1 to 3 is', shoplist[1:3]) print('Item 2 to end is', shoplist[2:]) print('Item 1 to -1 is', shoplist[1:-1]) print('Item start to end is', shoplist[:]) # Slicing on a string print('characters 1 to 3 is', name[1:3]) print('characters 2 to end is', name[2:]) print('characters 1 to -1 is', name[1:-1]) print('characters start to end is', name[:])
27.178571
49
0.658344
7727c3a4fe50654ffd803485c0e0c94d2c96db3c
121,553
py
Python
salt/utils/parsers.py
eyj/salt
5232d7f60cb78591d8bad25d796b0a945fc6350c
[ "Apache-2.0" ]
2
2015-09-21T14:13:30.000Z
2016-02-12T11:33:46.000Z
salt/utils/parsers.py
eyj/salt
5232d7f60cb78591d8bad25d796b0a945fc6350c
[ "Apache-2.0" ]
null
null
null
salt/utils/parsers.py
eyj/salt
5232d7f60cb78591d8bad25d796b0a945fc6350c
[ "Apache-2.0" ]
null
null
null
# -*- coding: utf-8 -*- ''' :codeauthor: :email:`Pedro Algarvio (pedro@algarvio.me)` salt.utils.parsers ~~~~~~~~~~~~~~~~~~ This is where all the black magic happens on all of salt's CLI tools. ''' # pylint: disable=missing-docstring,protected-access,too-many-ancestors,too-few-public-methods # pylint: disable=attribute-defined-outside-init,no-self-use # Import python libs from __future__ import absolute_import, print_function import os import sys import types import signal import getpass import logging import optparse import traceback import yaml from functools import partial # Import salt libs import salt.config as config import salt.defaults.exitcodes import salt.log.setup as log import salt.syspaths as syspaths import salt.version as version import salt.utils import salt.utils.args import salt.utils.xdg import salt.utils.jid from salt.utils import kinds from salt.defaults import DEFAULT_TARGET_DELIM from salt.utils.validate.path import is_writeable from salt.utils.verify import verify_files import salt.exceptions import salt.ext.six as six from salt.ext.six.moves import range # pylint: disable=import-error,redefined-builtin def _sorted(mixins_or_funcs): return sorted( mixins_or_funcs, key=lambda mf: getattr(mf, '_mixin_prio_', 1000) ) class MixInMeta(type): # This attribute here won't actually do anything. But, if you need to # specify an order or a dependency within the mix-ins, please define the # attribute on your own MixIn _mixin_prio_ = 0 def __new__(mcs, name, bases, attrs): instance = super(MixInMeta, mcs).__new__(mcs, name, bases, attrs) if not hasattr(instance, '_mixin_setup'): raise RuntimeError( 'Don\'t subclass {0} in {1} if you\'re not going to use it ' 'as a salt parser mix-in.'.format(mcs.__name__, name) ) return instance class OptionParserMeta(MixInMeta): def __new__(mcs, name, bases, attrs): instance = super(OptionParserMeta, mcs).__new__(mcs, name, bases, attrs) if not hasattr(instance, '_mixin_setup_funcs'): instance._mixin_setup_funcs = [] if not hasattr(instance, '_mixin_process_funcs'): instance._mixin_process_funcs = [] if not hasattr(instance, '_mixin_after_parsed_funcs'): instance._mixin_after_parsed_funcs = [] if not hasattr(instance, '_mixin_before_exit_funcs'): instance._mixin_before_exit_funcs = [] for base in _sorted(bases + (instance,)): func = getattr(base, '_mixin_setup', None) if func is not None and func not in instance._mixin_setup_funcs: instance._mixin_setup_funcs.append(func) func = getattr(base, '_mixin_after_parsed', None) if func is not None and func not in \ instance._mixin_after_parsed_funcs: instance._mixin_after_parsed_funcs.append(func) func = getattr(base, '_mixin_before_exit', None) if func is not None and func not in \ instance._mixin_before_exit_funcs: instance._mixin_before_exit_funcs.append(func) # Mark process_<opt> functions with the base priority for sorting for func in dir(base): if not func.startswith('process_'): continue func = getattr(base, func) if getattr(func, '_mixin_prio_', None) is not None: # Function already has the attribute set, don't override it continue if six.PY2: func.__func__._mixin_prio_ = getattr( base, '_mixin_prio_', 1000 ) else: func._mixin_prio_ = getattr( base, '_mixin_prio_', 1000 ) return instance class CustomOption(optparse.Option, object): def take_action(self, action, dest, *args, **kwargs): # see https://github.com/python/cpython/blob/master/Lib/optparse.py#L786 self.explicit = True return optparse.Option.take_action(self, action, dest, *args, **kwargs) class OptionParser(optparse.OptionParser, object): VERSION = version.__saltstack_version__.formatted_version usage = '%prog' epilog = ('You can find additional help about %prog issuing "man %prog" ' 'or on http://docs.saltstack.com') description = None # Private attributes _mixin_prio_ = 100 # Setup multiprocessing logging queue listener _setup_mp_logging_listener_ = False def __init__(self, *args, **kwargs): kwargs.setdefault('version', '%prog {0}'.format(self.VERSION)) kwargs.setdefault('usage', self.usage) if self.description: kwargs.setdefault('description', self.description) if self.epilog: kwargs.setdefault('epilog', self.epilog) kwargs.setdefault('option_class', CustomOption) optparse.OptionParser.__init__(self, *args, **kwargs) if self.epilog and '%prog' in self.epilog: self.epilog = self.epilog.replace('%prog', self.get_prog_name()) def add_option_group(self, *args, **kwargs): option_group = optparse.OptionParser.add_option_group(self, *args, **kwargs) option_group.option_class = CustomOption return option_group def parse_args(self, args=None, values=None): options, args = optparse.OptionParser.parse_args(self, args, values) if 'args_stdin' in options.__dict__ and options.args_stdin is True: # Read additional options and/or arguments from stdin and combine # them with the options and arguments from the command line. new_inargs = sys.stdin.readlines() new_inargs = [arg.rstrip('\r\n') for arg in new_inargs] new_options, new_args = optparse.OptionParser.parse_args( self, new_inargs) options.__dict__.update(new_options.__dict__) args.extend(new_args) if options.versions_report: self.print_versions_report() self.options, self.args = options, args # Let's get some proper sys.stderr logging as soon as possible!!! # This logging handler will be removed once the proper console or # logfile logging is setup. log.setup_temp_logger( getattr(self.options, 'log_level', 'error') ) # Gather and run the process_<option> functions in the proper order process_option_funcs = [] for option_key in options.__dict__: process_option_func = getattr( self, 'process_{0}'.format(option_key), None ) if process_option_func is not None: process_option_funcs.append(process_option_func) for process_option_func in _sorted(process_option_funcs): try: process_option_func() except Exception as err: # pylint: disable=broad-except logging.getLogger(__name__).exception(err) self.error( 'Error while processing {0}: {1}'.format( process_option_func, traceback.format_exc(err) ) ) # Run the functions on self._mixin_after_parsed_funcs for mixin_after_parsed_func in self._mixin_after_parsed_funcs: # pylint: disable=no-member try: mixin_after_parsed_func(self) except Exception as err: # pylint: disable=broad-except logging.getLogger(__name__).exception(err) self.error( 'Error while processing {0}: {1}'.format( mixin_after_parsed_func, traceback.format_exc(err) ) ) if self.config.get('conf_file', None) is not None: # pylint: disable=no-member logging.getLogger(__name__).debug( 'Configuration file path: {0}'.format( self.config['conf_file'] # pylint: disable=no-member ) ) # Retain the standard behavior of optparse to return options and args return options, args def _populate_option_list(self, option_list, add_help=True): optparse.OptionParser._populate_option_list( self, option_list, add_help=add_help ) for mixin_setup_func in self._mixin_setup_funcs: # pylint: disable=no-member mixin_setup_func(self) def _add_version_option(self): optparse.OptionParser._add_version_option(self) self.add_option( '--versions-report', '-V', action='store_true', help='Show program\'s dependencies version number and exit.' ) def print_versions_report(self, file=sys.stdout): # pylint: disable=redefined-builtin print('\n'.join(version.versions_report()), file=file) self.exit(salt.defaults.exitcodes.EX_OK) def exit(self, status=0, msg=None): # Run the functions on self._mixin_after_parsed_funcs for mixin_before_exit_func in self._mixin_before_exit_funcs: # pylint: disable=no-member try: mixin_before_exit_func(self) except Exception as err: # pylint: disable=broad-except logger = logging.getLogger(__name__) logger.exception(err) logger.error( 'Error while processing {0}: {1}'.format( mixin_before_exit_func, traceback.format_exc(err) ) ) if self._setup_mp_logging_listener_ is True: # Stop the logging queue listener process log.shutdown_multiprocessing_logging_listener(daemonizing=True) if isinstance(msg, six.string_types) and msg and msg[-1] != '\n': msg = '{0}\n'.format(msg) optparse.OptionParser.exit(self, status, msg) def error(self, msg): """error(msg : string) Print a usage message incorporating 'msg' to stderr and exit. This keeps option parsing exit status uniform for all parsing errors. """ self.print_usage(sys.stderr) self.exit(salt.defaults.exitcodes.EX_USAGE, '{0}: error: {1}\n'.format(self.get_prog_name(), msg)) class MergeConfigMixIn(six.with_metaclass(MixInMeta, object)): ''' This mix-in will simply merge the CLI-passed options, by overriding the configuration file loaded settings. This mix-in should run last. ''' _mixin_prio_ = six.MAXSIZE def _mixin_setup(self): if not hasattr(self, 'setup_config') and not hasattr(self, 'config'): # No configuration was loaded on this parser. # There's nothing to do here. return # Add an additional function that will merge the shell options with # the config options and if needed override them self._mixin_after_parsed_funcs.append(self.__merge_config_with_cli) def __merge_config_with_cli(self, *args): # pylint: disable=unused-argument # Merge parser options for option in self.option_list: if option.dest is None: # --version does not have dest attribute set for example. # All options defined by us, even if not explicitly(by kwarg), # will have the dest attribute set continue # Get the passed value from shell. If empty get the default one default = self.defaults.get(option.dest) value = getattr(self.options, option.dest, default) if option.dest not in self.config: # There's no value in the configuration file if value is not None: # There's an actual value, add it to the config self.config[option.dest] = value elif value is not None and getattr(option, "explicit", False): # Only set the value in the config file IF it was explicitly # specified by the user, this makes it possible to tweak settings # on the configuration files bypassing the shell option flags' # defaults self.config[option.dest] = value elif option.dest in self.config: # Let's update the option value with the one from the # configuration file. This allows the parsers to make use of # the updated value by using self.options.<option> setattr(self.options, option.dest, self.config[option.dest]) # Merge parser group options if any for group in self.option_groups: for option in group.option_list: if option.dest is None: continue # Get the passed value from shell. If empty get the default one default = self.defaults.get(option.dest) value = getattr(self.options, option.dest, default) if option.dest not in self.config: # There's no value in the configuration file if value is not None: # There's an actual value, add it to the config self.config[option.dest] = value elif value is not None and getattr(option, "explicit", False): # Only set the value in the config file IF it was explicitly # specified by the user, this makes it possible to tweak # settings on the configuration files bypassing the shell # option flags' defaults self.config[option.dest] = value elif option.dest in self.config: # Let's update the option value with the one from the # configuration file. This allows the parsers to make use # of the updated value by using self.options.<option> setattr(self.options, option.dest, self.config[option.dest]) class SaltfileMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = -20 def _mixin_setup(self): self.add_option( '--saltfile', default=None, help='Specify the path to a Saltfile. If not passed, one will be ' 'searched for in the current working directory.' ) def process_saltfile(self): if self.options.saltfile is None: # No one passed a Saltfile as an option, environment variable!? self.options.saltfile = os.environ.get('SALT_SALTFILE', None) if self.options.saltfile is None: # If we're here, no one passed a Saltfile either to the CLI tool or # as an environment variable. # Is there a Saltfile in the current directory? try: # cwd may not exist if it was removed but salt was run from it saltfile = os.path.join(os.getcwd(), 'Saltfile') except OSError: saltfile = '' if os.path.isfile(saltfile): self.options.saltfile = saltfile else: saltfile = self.options.saltfile if not self.options.saltfile: # There's still no valid Saltfile? No need to continue... return if not os.path.isfile(self.options.saltfile): self.error( '\'{0}\' file does not exist.\n'.format(self.options.saltfile ) ) # Make sure we have an absolute path self.options.saltfile = os.path.abspath(self.options.saltfile) # Make sure we let the user know that we will be loading a Saltfile logging.getLogger(__name__).info( 'Loading Saltfile from \'{0}\''.format(self.options.saltfile) ) try: saltfile_config = config._read_conf_file(saltfile) except salt.exceptions.SaltConfigurationError as error: self.error(error.message) self.exit(salt.defaults.exitcodes.EX_GENERIC, '{0}: error: {1}\n'.format(self.get_prog_name(), error.message)) if not saltfile_config: # No configuration was loaded from the Saltfile return if self.get_prog_name() not in saltfile_config: # There's no configuration specific to the CLI tool. Stop! return # We just want our own configuration cli_config = saltfile_config[self.get_prog_name()] # If there are any options, who's names match any key from the loaded # Saltfile, we need to update its default value for option in self.option_list: if option.dest is None: # --version does not have dest attribute set for example. continue if option.dest not in cli_config: # If we don't have anything in Saltfile for this option, let's # continue processing right now continue # Get the passed value from shell. If empty get the default one default = self.defaults.get(option.dest) value = getattr(self.options, option.dest, default) if value != default: # The user passed an argument, we won't override it with the # one from Saltfile, if any continue # We reached this far! Set the Saltfile value on the option setattr(self.options, option.dest, cli_config[option.dest]) option.explicit = True # Let's also search for options referred in any option groups for group in self.option_groups: for option in group.option_list: if option.dest is None: continue if option.dest not in cli_config: # If we don't have anything in Saltfile for this option, # let's continue processing right now continue # Get the passed value from shell. If empty get the default one default = self.defaults.get(option.dest) value = getattr(self.options, option.dest, default) if value != default: # The user passed an argument, we won't override it with # the one from Saltfile, if any continue setattr(self.options, option.dest, cli_config[option.dest]) option.explicit = True # Any left over value in the saltfile can now be safely added for key in cli_config: setattr(self.options, key, cli_config[key]) class HardCrashMixin(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 40 _config_filename_ = None def _mixin_setup(self): hard_crash = os.environ.get('SALT_HARD_CRASH', False) self.add_option( '--hard-crash', action='store_true', default=hard_crash, help=('Raise any original exception rather than exiting gracefully. ' 'Default: %default.') ) class ConfigDirMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = -10 _config_filename_ = None _default_config_dir_ = syspaths.CONFIG_DIR _default_config_dir_env_var_ = 'SALT_CONFIG_DIR' def _mixin_setup(self): config_dir = os.environ.get(self._default_config_dir_env_var_, None) if not config_dir: config_dir = self._default_config_dir_ logging.getLogger(__name__).debug('SYSPATHS setup as: {0}'.format(syspaths.CONFIG_DIR)) self.add_option( '-c', '--config-dir', default=config_dir, help=('Pass in an alternative configuration directory. Default: ' '\'%default\'.') ) def process_config_dir(self): if not os.path.isdir(self.options.config_dir): # No logging is configured yet sys.stderr.write( 'WARNING: CONFIG \'{0}\' directory does not exist.\n'.format( self.options.config_dir ) ) # Make sure we have an absolute path self.options.config_dir = os.path.abspath(self.options.config_dir) if hasattr(self, 'setup_config'): if not hasattr(self, 'config'): self.config = {} try: self.config.update(self.setup_config()) except (IOError, OSError) as exc: self.error( 'Failed to load configuration: {0}'.format(exc) ) def get_config_file_path(self, configfile=None): if configfile is None: configfile = self._config_filename_ return os.path.join(self.options.config_dir, configfile) class LogLevelMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 10 _default_logging_level_ = 'warning' _default_logging_logfile_ = None _logfile_config_setting_name_ = 'log_file' _loglevel_config_setting_name_ = 'log_level' _logfile_loglevel_config_setting_name_ = 'log_level_logfile' # pylint: disable=invalid-name _skip_console_logging_config_ = False def _mixin_setup(self): if self._default_logging_logfile_ is None: # This is an attribute available for programmers, so, raise a # RuntimeError to let them know about the proper usage. raise RuntimeError( 'Please set {0}._default_logging_logfile_'.format( self.__class__.__name__ ) ) group = self.logging_options_group = optparse.OptionGroup( self, 'Logging Options', 'Logging options which override any settings defined on the ' 'configuration files.' ) self.add_option_group(group) if not getattr(self, '_skip_console_logging_config_', False): group.add_option( '-l', '--log-level', choices=list(log.LOG_LEVELS), help='Console logging log level. One of {0}. ' 'Default: \'{1}\'.'.format( ', '.join([repr(l) for l in log.SORTED_LEVEL_NAMES]), getattr(self, '_default_logging_level_', 'warning') ) ) group.add_option( '--log-file', default=None, help='Log file path. Default: \'{0}\'.'.format( self._default_logging_logfile_ ) ) group.add_option( '--log-file-level', dest=self._logfile_loglevel_config_setting_name_, choices=list(log.LOG_LEVELS), help='Logfile logging log level. One of {0}. ' 'Default: \'{1}\'.'.format( ', '.join([repr(l) for l in log.SORTED_LEVEL_NAMES]), getattr(self, '_default_logging_level_', 'warning') ) ) def process_log_level(self): if not self.options.log_level: cli_log_level = 'cli_{0}_log_level'.format( self.get_prog_name().replace('-', '_') ) if self.config.get(cli_log_level, None) is not None: self.options.log_level = self.config.get(cli_log_level) elif self.config.get(self._loglevel_config_setting_name_, None): self.options.log_level = self.config.get( self._loglevel_config_setting_name_ ) else: self.options.log_level = self._default_logging_level_ # Setup extended logging right before the last step self._mixin_after_parsed_funcs.append(self.__setup_extended_logging) # Setup the console and log file configuration before the MP logging # listener because the MP logging listener may need that config. self._mixin_after_parsed_funcs.append(self.__setup_logfile_logger_config) self._mixin_after_parsed_funcs.append(self.__setup_console_logger_config) # Setup the multiprocessing log queue listener if enabled self._mixin_after_parsed_funcs.append(self._setup_mp_logging_listener) # Setup the console as the last _mixin_after_parsed_func to run self._mixin_after_parsed_funcs.append(self.__setup_console_logger) def process_log_file(self): if not self.options.log_file: cli_setting_name = 'cli_{0}_log_file'.format( self.get_prog_name().replace('-', '_') ) if self.config.get(cli_setting_name, None) is not None: # There's a configuration setting defining this log file path, # i.e., `key_log_file` if the cli tool is `salt-key` self.options.log_file = self.config.get(cli_setting_name) elif self.config.get(self._logfile_config_setting_name_, None): # Is the regular log file setting set? self.options.log_file = self.config.get( self._logfile_config_setting_name_ ) else: # Nothing is set on the configuration? Let's use the cli tool # defined default self.options.log_file = self._default_logging_logfile_ def process_log_file_level(self): if not self.options.log_file_level: cli_setting_name = 'cli_{0}_log_file_level'.format( self.get_prog_name().replace('-', '_') ) if self.config.get(cli_setting_name, None) is not None: # There's a configuration setting defining this log file # logging level, i.e., `key_log_file_level` if the cli tool is # `salt-key` self.options.log_file_level = self.config.get(cli_setting_name) elif self.config.get( self._logfile_loglevel_config_setting_name_, None): # Is the regular log file level setting set? self.options.log_file_level = self.config.get( self._logfile_loglevel_config_setting_name_ ) else: # Nothing is set on the configuration? Let's use the cli tool # defined default self.options.log_level = self._default_logging_level_ def __setup_logfile_logger_config(self, *args): # pylint: disable=unused-argument if self._logfile_loglevel_config_setting_name_ in self.config and not \ self.config.get(self._logfile_loglevel_config_setting_name_): # Remove it from config so it inherits from log_level self.config.pop(self._logfile_loglevel_config_setting_name_) loglevel = self.config.get( self._logfile_loglevel_config_setting_name_, self.config.get( # From the config setting self._loglevel_config_setting_name_, # From the console setting self.config['log_level'] ) ) cli_log_path = 'cli_{0}_log_file'.format( self.get_prog_name().replace('-', '_') ) if cli_log_path in self.config and not self.config.get(cli_log_path): # Remove it from config so it inherits from log_level_logfile self.config.pop(cli_log_path) if self._logfile_config_setting_name_ in self.config and not \ self.config.get(self._logfile_config_setting_name_): # Remove it from config so it inherits from log_file self.config.pop(self._logfile_config_setting_name_) logfile = self.config.get( # First from the config cli setting cli_log_path, self.config.get( # From the config setting self._logfile_config_setting_name_ ) ) if self.config['verify_env']: # Verify the logfile if it was explicitly set but do not try to # verify the default if logfile is not None and not logfile.startswith(('tcp://', 'udp://', 'file://')): # Logfile is not using Syslog, verify current_umask = os.umask(0o027) verify_files([logfile], self.config['user']) os.umask(current_umask) if logfile is None: # Use the default setting if the logfile wasn't explicity set logfile = self._default_logging_logfile_ cli_log_file_fmt = 'cli_{0}_log_file_fmt'.format( self.get_prog_name().replace('-', '_') ) if cli_log_file_fmt in self.config and not \ self.config.get(cli_log_file_fmt): # Remove it from config so it inherits from log_fmt_logfile self.config.pop(cli_log_file_fmt) if self.config.get('log_fmt_logfile', None) is None: # Remove it from config so it inherits from log_fmt_console self.config.pop('log_fmt_logfile', None) log_file_fmt = self.config.get( cli_log_file_fmt, self.config.get( 'cli_{0}_log_fmt'.format( self.get_prog_name().replace('-', '_') ), self.config.get( 'log_fmt_logfile', self.config.get( 'log_fmt_console', self.config.get( 'log_fmt', config._DFLT_LOG_FMT_CONSOLE ) ) ) ) ) cli_log_file_datefmt = 'cli_{0}_log_file_datefmt'.format( self.get_prog_name().replace('-', '_') ) if cli_log_file_datefmt in self.config and not \ self.config.get(cli_log_file_datefmt): # Remove it from config so it inherits from log_datefmt_logfile self.config.pop(cli_log_file_datefmt) if self.config.get('log_datefmt_logfile', None) is None: # Remove it from config so it inherits from log_datefmt_console self.config.pop('log_datefmt_logfile', None) if self.config.get('log_datefmt_console', None) is None: # Remove it from config so it inherits from log_datefmt self.config.pop('log_datefmt_console', None) log_file_datefmt = self.config.get( cli_log_file_datefmt, self.config.get( 'cli_{0}_log_datefmt'.format( self.get_prog_name().replace('-', '_') ), self.config.get( 'log_datefmt_logfile', self.config.get( 'log_datefmt_console', self.config.get( 'log_datefmt', '%Y-%m-%d %H:%M:%S' ) ) ) ) ) if not is_writeable(logfile, check_parent=True): # Since we're not be able to write to the log file or its parent # directory (if the log file does not exit), are we the same user # as the one defined in the configuration file? current_user = salt.utils.get_user() if self.config['user'] != current_user: # Yep, not the same user! # Is the current user in ACL? acl = self.config.get('publisher_acl') or self.config.get('client_acl', {}) if salt.utils.check_whitelist_blacklist(current_user, whitelist=six.iterkeys(acl)): # Yep, the user is in ACL! # Let's write the logfile to its home directory instead. xdg_dir = salt.utils.xdg.xdg_config_dir() user_salt_dir = (xdg_dir if os.path.isdir(xdg_dir) else os.path.expanduser('~/.salt')) if not os.path.isdir(user_salt_dir): os.makedirs(user_salt_dir, 0o750) logfile_basename = os.path.basename( self._default_logging_logfile_ ) logging.getLogger(__name__).debug( 'The user \'{0}\' is not allowed to write to \'{1}\'. ' 'The log file will be stored in ' '\'~/.salt/\'{2}\'.log\''.format( current_user, logfile, logfile_basename ) ) logfile = os.path.join( user_salt_dir, '{0}.log'.format(logfile_basename) ) # If we haven't changed the logfile path and it's not writeable, # salt will fail once we try to setup the logfile logging. # Save the settings back to the configuration self.config[self._logfile_config_setting_name_] = logfile self.config[self._logfile_loglevel_config_setting_name_] = loglevel self.config['log_fmt_logfile'] = log_file_fmt self.config['log_datefmt_logfile'] = log_file_datefmt def setup_logfile_logger(self): logfile = self.config[self._logfile_config_setting_name_] loglevel = self.config[self._logfile_loglevel_config_setting_name_] log_file_fmt = self.config['log_fmt_logfile'] log_file_datefmt = self.config['log_datefmt_logfile'] log.setup_logfile_logger( logfile, loglevel, log_format=log_file_fmt, date_format=log_file_datefmt ) for name, level in six.iteritems(self.config['log_granular_levels']): log.set_logger_level(name, level) def __setup_extended_logging(self, *args): # pylint: disable=unused-argument log.setup_extended_logging(self.config) def _get_mp_logging_listener_queue(self): return log.get_multiprocessing_logging_queue() def _setup_mp_logging_listener(self, *args): # pylint: disable=unused-argument if self._setup_mp_logging_listener_: log.setup_multiprocessing_logging_listener( self.config, self._get_mp_logging_listener_queue() ) def __setup_console_logger_config(self, *args): # pylint: disable=unused-argument # Since we're not going to be a daemon, setup the console logger cli_log_fmt = 'cli_{0}_log_fmt'.format( self.get_prog_name().replace('-', '_') ) if cli_log_fmt in self.config and not self.config.get(cli_log_fmt): # Remove it from config so it inherits from log_fmt_console self.config.pop(cli_log_fmt) logfmt = self.config.get( cli_log_fmt, self.config.get( 'log_fmt_console', self.config.get( 'log_fmt', config._DFLT_LOG_FMT_CONSOLE ) ) ) cli_log_datefmt = 'cli_{0}_log_datefmt'.format( self.get_prog_name().replace('-', '_') ) if cli_log_datefmt in self.config and not \ self.config.get(cli_log_datefmt): # Remove it from config so it inherits from log_datefmt_console self.config.pop(cli_log_datefmt) if self.config.get('log_datefmt_console', None) is None: # Remove it from config so it inherits from log_datefmt self.config.pop('log_datefmt_console', None) datefmt = self.config.get( cli_log_datefmt, self.config.get( 'log_datefmt_console', self.config.get( 'log_datefmt', '%Y-%m-%d %H:%M:%S' ) ) ) # Save the settings back to the configuration self.config['log_fmt_console'] = logfmt self.config['log_datefmt_console'] = datefmt def __setup_console_logger(self, *args): # pylint: disable=unused-argument # If daemon is set force console logger to quiet if getattr(self.options, 'daemon', False) is True: return log.setup_console_logger( self.config['log_level'], log_format=self.config['log_fmt_console'], date_format=self.config['log_datefmt_console'] ) for name, level in six.iteritems(self.config['log_granular_levels']): log.set_logger_level(name, level) class RunUserMixin(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 20 def _mixin_setup(self): self.add_option( '-u', '--user', help='Specify user to run {0}.'.format(self.get_prog_name()) ) class DaemonMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 30 def _mixin_setup(self): self.add_option( '-d', '--daemon', default=False, action='store_true', help='Run the {0} as a daemon.'.format(self.get_prog_name()) ) self.add_option( '--pid-file', dest='pidfile', default=os.path.join( syspaths.PIDFILE_DIR, '{0}.pid'.format(self.get_prog_name()) ), help=('Specify the location of the pidfile. Default: \'%default\'.') ) def _mixin_before_exit(self): if hasattr(self, 'config') and self.config.get('pidfile', ''): # We've loaded and merged options into the configuration, it's safe # to query about the pidfile if self.check_pidfile(): os.unlink(self.config['pidfile']) def set_pidfile(self): from salt.utils.process import set_pidfile set_pidfile(self.config['pidfile'], self.config['user']) def check_pidfile(self): ''' Report whether a pidfile exists ''' from salt.utils.process import check_pidfile return check_pidfile(self.config['pidfile']) def get_pidfile(self): ''' Return a pid contained in a pidfile ''' from salt.utils.process import get_pidfile return get_pidfile(self.config['pidfile']) def daemonize_if_required(self): if self.options.daemon: if self._setup_mp_logging_listener_ is True: # Stop the logging queue listener for the current process # We'll restart it once forked log.shutdown_multiprocessing_logging_listener(daemonizing=True) # Late import so logging works correctly salt.utils.daemonize() # Setup the multiprocessing log queue listener if enabled self._setup_mp_logging_listener() def check_running(self): ''' Check if a pid file exists and if it is associated with a running process. ''' # There is no os.getppid method for windows if salt.utils.is_windows(): from salt.utils.win_functions import get_parent_pid ppid = get_parent_pid() else: ppid = os.getppid() if self.check_pidfile(): pid = self.get_pidfile() if not salt.utils.is_windows(): if self.check_pidfile() and self.is_daemonized(pid) and not os.getppid() == pid: return True else: # We have no os.getppid() on Windows. Best effort. if self.check_pidfile() and self.is_daemonized(pid): return True return False def is_daemonized(self, pid): from salt.utils.process import os_is_running return os_is_running(pid) # Common methods for scripts which can daemonize def _install_signal_handlers(self): signal.signal(signal.SIGTERM, self._handle_signals) signal.signal(signal.SIGINT, self._handle_signals) def prepare(self): self.parse_args() def start(self): self.prepare() self._install_signal_handlers() def _handle_signals(self, signum, sigframe): # pylint: disable=unused-argument msg = self.__class__.__name__ if signum == signal.SIGINT: msg += ' received a SIGINT.' elif signum == signal.SIGTERM: msg += ' received a SIGTERM.' logging.getLogger(__name__).warning('{0} Exiting.'.format(msg)) self.shutdown(exitmsg='{0} Exited.'.format(msg)) def shutdown(self, exitcode=0, exitmsg=None): self.exit(exitcode, exitmsg) class PidfileMixin(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 40 def _mixin_setup(self): salt.utils.warn_until( 'Nitrogen', 'Please stop sub-classing PidfileMix and instead subclass ' 'DaemonMixIn which contains the same behavior. PidfileMixin ' 'will be supported until Salt {version}.' ) try: self.add_option( '--pid-file', dest='pidfile', default=os.path.join( syspaths.PIDFILE_DIR, '{0}.pid'.format(self.get_prog_name()) ), help=('Specify the location of the pidfile. Default: \'%default\'.') ) # Since there was no colision with DaemonMixin, let's add the # pidfile mixin methods. This is used using types.MethodType # because if we had defined these at the class level, they would # have overridden the exact same methods from the DaemonMixin. def set_pidfile(self): from salt.utils.process import set_pidfile set_pidfile(self.config['pidfile'], self.config['user']) self.set_pidfile = types.MethodType(set_pidfile, self) def check_pidfile(self): ''' Report whether a pidfile exists ''' from salt.utils.process import check_pidfile return check_pidfile(self.config['pidfile']) self.check_pidfile = types.MethodType(check_pidfile, self) def get_pidfile(self): ''' Return a pid contained in a pidfile ''' from salt.utils.process import get_pidfile return get_pidfile(self.config['pidfile']) self.get_pidfile = types.MethodType(get_pidfile, self) except optparse.OptionConflictError: # The option was already added by the DaemonMixin pass class TargetOptionsMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 20 selected_target_option = None def _mixin_setup(self): group = self.target_options_group = optparse.OptionGroup( self, 'Target Options', 'Target selection options.' ) self.add_option_group(group) group.add_option( '-E', '--pcre', default=False, action='store_true', help=('Instead of using shell globs to evaluate the target ' 'servers, use pcre regular expressions.') ) group.add_option( '-L', '--list', default=False, action='store_true', help=('Instead of using shell globs to evaluate the target ' 'servers, take a comma or space delimited list of ' 'servers.') ) group.add_option( '-G', '--grain', default=False, action='store_true', help=('Instead of using shell globs to evaluate the target ' 'use a grain value to identify targets, the syntax ' 'for the target is the grain key followed by a glob' 'expression: "os:Arch*".') ) group.add_option( '-P', '--grain-pcre', default=False, action='store_true', help=('Instead of using shell globs to evaluate the target ' 'use a grain value to identify targets, the syntax ' 'for the target is the grain key followed by a pcre ' 'regular expression: "os:Arch.*".') ) group.add_option( '-N', '--nodegroup', default=False, action='store_true', help=('Instead of using shell globs to evaluate the target ' 'use one of the predefined nodegroups to identify a ' 'list of targets.') ) group.add_option( '-R', '--range', default=False, action='store_true', help=('Instead of using shell globs to evaluate the target ' 'use a range expression to identify targets. ' 'Range expressions look like %cluster.') ) group = self.additional_target_options_group = optparse.OptionGroup( self, 'Additional Target Options', 'Additional options for minion targeting.' ) self.add_option_group(group) group.add_option( '--delimiter', default=DEFAULT_TARGET_DELIM, help=('Change the default delimiter for matching in multi-level ' 'data structures. Default: \'%default\'.') ) self._create_process_functions() def _create_process_functions(self): for option in self.target_options_group.option_list: def process(opt): if getattr(self.options, opt.dest): self.selected_target_option = opt.dest funcname = 'process_{0}'.format(option.dest) if not hasattr(self, funcname): setattr(self, funcname, partial(process, option)) def _mixin_after_parsed(self): group_options_selected = [ option for option in self.target_options_group.option_list if getattr(self.options, option.dest) is True ] if len(group_options_selected) > 1: self.error( 'The options {0} are mutually exclusive. Please only choose ' 'one of them'.format('/'.join( [option.get_opt_string() for option in group_options_selected])) ) self.config['selected_target_option'] = self.selected_target_option class ExtendedTargetOptionsMixIn(TargetOptionsMixIn): def _mixin_setup(self): TargetOptionsMixIn._mixin_setup(self) group = self.target_options_group group.add_option( '-C', '--compound', default=False, action='store_true', help=('The compound target option allows for multiple target ' 'types to be evaluated, allowing for greater granularity in ' 'target matching. The compound target is space delimited, ' 'targets other than globs are preceded with an identifier ' 'matching the specific targets argument type: salt ' '\'G@os:RedHat and webser* or E@database.*\'.') ) group.add_option( '-I', '--pillar', default=False, dest='pillar_target', action='store_true', help=('Instead of using shell globs to evaluate the target ' 'use a pillar value to identify targets, the syntax ' 'for the target is the pillar key followed by a glob ' 'expression: "role:production*".') ) group.add_option( '-J', '--pillar-pcre', default=False, action='store_true', help=('Instead of using shell globs to evaluate the target ' 'use a pillar value to identify targets, the syntax ' 'for the target is the pillar key followed by a pcre ' 'regular expression: "role:prod.*".') ) group.add_option( '-S', '--ipcidr', default=False, action='store_true', help=('Match based on Subnet (CIDR notation) or IP address.') ) self._create_process_functions() def process_pillar_target(self): if self.options.pillar_target: self.selected_target_option = 'pillar' class TimeoutMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 10 def _mixin_setup(self): if not hasattr(self, 'default_timeout'): raise RuntimeError( 'You need to define the \'default_timeout\' attribute ' 'on {0}'.format(self.__class__.__name__) ) self.add_option( '-t', '--timeout', type=int, default=self.default_timeout, help=('Change the timeout, if applicable, for the running ' 'command (in seconds). Default: %default.') ) class ArgsStdinMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 10 def _mixin_setup(self): self.add_option( '--args-stdin', default=False, dest='args_stdin', action='store_true', help=('Read additional options and/or arguments from stdin. ' 'Each entry is newline separated.') ) class ProxyIdMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio = 40 def _mixin_setup(self): self.add_option( '--proxyid', default=None, dest='proxyid', help=('Id for this proxy.') ) class OutputOptionsMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 40 _include_text_out_ = False selected_output_option = None def _mixin_setup(self): group = self.output_options_group = optparse.OptionGroup( self, 'Output Options', 'Configure your preferred output format.' ) self.add_option_group(group) group.add_option( '--out', '--output', dest='output', help=( 'Print the output from the \'{0}\' command using the ' 'specified outputter.'.format( self.get_prog_name(), ) ) ) group.add_option( '--out-indent', '--output-indent', dest='output_indent', default=None, type=int, help=('Print the output indented by the provided value in spaces. ' 'Negative values disables indentation. Only applicable in ' 'outputters that support indentation.') ) group.add_option( '--out-file', '--output-file', dest='output_file', default=None, help='Write the output to the specified file.' ) group.add_option( '--out-file-append', '--output-file-append', action='store_true', dest='output_file_append', default=False, help='Append the output to the specified file.' ) group.add_option( '--no-color', '--no-colour', default=False, action='store_true', help='Disable all colored output.' ) group.add_option( '--force-color', '--force-colour', default=False, action='store_true', help='Force colored output.' ) group.add_option( '--state-output', '--state_output', default=None, help=('Override the configured state_output value for minion ' 'output. One of \'full\', \'terse\', \'mixed\', \'changes\' or \'filter\'. ' 'Default: \'%default\'.') ) group.add_option( '--state-verbose', '--state_verbose', default=None, help=('Override the configured state_verbose value for minion ' 'output. Set to True or False. Default: %default.') ) for option in self.output_options_group.option_list: def process(opt): default = self.defaults.get(opt.dest) if getattr(self.options, opt.dest, default) is False: return self.selected_output_option = opt.dest funcname = 'process_{0}'.format(option.dest) if not hasattr(self, funcname): setattr(self, funcname, partial(process, option)) def process_output(self): self.selected_output_option = self.options.output def process_output_file(self): if self.options.output_file is not None and self.options.output_file_append is False: if os.path.isfile(self.options.output_file): try: with salt.utils.fopen(self.options.output_file, 'w') as ofh: # Make this a zero length filename instead of removing # it. This way we keep the file permissions. ofh.write('') except (IOError, OSError) as exc: self.error( '{0}: Access denied: {1}'.format( self.options.output_file, exc ) ) def process_state_verbose(self): if self.options.state_verbose == "True" or self.options.state_verbose == "true": self.options.state_verbose = True elif self.options.state_verbose == "False" or self.options.state_verbose == "false": self.options.state_verbose = False def _mixin_after_parsed(self): group_options_selected = [ option for option in self.output_options_group.option_list if ( getattr(self.options, option.dest) and (option.dest.endswith('_out') or option.dest == 'output')) ] if len(group_options_selected) > 1: self.error( 'The options {0} are mutually exclusive. Please only choose ' 'one of them'.format('/'.join([ option.get_opt_string() for option in group_options_selected ])) ) self.config['selected_output_option'] = self.selected_output_option class ExecutionOptionsMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 10 def _mixin_setup(self): group = self.execution_group = optparse.OptionGroup( self, 'Execution Options', # Include description here as a string ) group.add_option( '-L', '--location', default=None, help='Specify which region to connect to.' ) group.add_option( '-a', '--action', default=None, help='Perform an action that may be specific to this cloud ' 'provider. This argument requires one or more instance ' 'names to be specified.' ) group.add_option( '-f', '--function', nargs=2, default=None, metavar='<FUNC-NAME> <PROVIDER>', help='Perform an function that may be specific to this cloud ' 'provider, that does not apply to an instance. This ' 'argument requires a provider to be specified (i.e.: nova).' ) group.add_option( '-p', '--profile', default=None, help='Create an instance using the specified profile.' ) group.add_option( '-m', '--map', default=None, help='Specify a cloud map file to use for deployment. This option ' 'may be used alone, or in conjunction with -Q, -F, -S or -d. ' 'The map can also be filtered by a list of VM names.' ) group.add_option( '-H', '--hard', default=False, action='store_true', help='Delete all VMs that are not defined in the map file. ' 'CAUTION!!! This operation can irrevocably destroy VMs! It ' 'must be explicitly enabled in the cloud config file.' ) group.add_option( '-d', '--destroy', default=False, action='store_true', help='Destroy the specified instance(s).' ) group.add_option( '--no-deploy', default=True, dest='deploy', action='store_false', help='Don\'t run a deploy script after instance creation.' ) group.add_option( '-P', '--parallel', default=False, action='store_true', help='Build all of the specified instances in parallel.' ) group.add_option( '-u', '--update-bootstrap', default=False, action='store_true', help='Update salt-bootstrap to the latest stable bootstrap release.' ) group.add_option( '-y', '--assume-yes', default=False, action='store_true', help='Default "yes" in answer to all confirmation questions.' ) group.add_option( '-k', '--keep-tmp', default=False, action='store_true', help='Do not remove files from /tmp/ after deploy.sh finishes.' ) group.add_option( '--show-deploy-args', default=False, action='store_true', help='Include the options used to deploy the minion in the data ' 'returned.' ) group.add_option( '--script-args', default=None, help='Script arguments to be fed to the bootstrap script when ' 'deploying the VM.' ) group.add_option( '-b', '--bootstrap', nargs=1, default=False, metavar='<HOST> [MINION_ID] [OPTIONS...]', help='Bootstrap an existing machine.' ) self.add_option_group(group) def process_function(self): if self.options.function: self.function_name, self.function_provider = self.options.function if self.function_provider.startswith('-') or \ '=' in self.function_provider: self.error( '--function expects two arguments: <function-name> ' '<provider>' ) class CloudQueriesMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 20 selected_query_option = None def _mixin_setup(self): group = self.cloud_queries_group = optparse.OptionGroup( self, 'Query Options', # Include description here as a string ) group.add_option( '-Q', '--query', default=False, action='store_true', help=('Execute a query and return some information about the ' 'nodes running on configured cloud providers.') ) group.add_option( '-F', '--full-query', default=False, action='store_true', help=('Execute a query and return all information about the ' 'nodes running on configured cloud providers.') ) group.add_option( '-S', '--select-query', default=False, action='store_true', help=('Execute a query and return select information about ' 'the nodes running on configured cloud providers.') ) group.add_option( '--list-providers', default=False, action='store_true', help='Display a list of configured providers.' ) group.add_option( '--list-profiles', default=None, action='store', help='Display a list of configured profiles. Pass in a cloud ' 'provider to view the provider\'s associated profiles, ' 'such as digital_ocean, or pass in "all" to list all the ' 'configured profiles.' ) self.add_option_group(group) self._create_process_functions() def _create_process_functions(self): for option in self.cloud_queries_group.option_list: def process(opt): if getattr(self.options, opt.dest): query = 'list_nodes' if opt.dest == 'full_query': query += '_full' elif opt.dest == 'select_query': query += '_select' elif opt.dest == 'list_providers': query = 'list_providers' if self.args: self.error( '\'--list-providers\' does not accept any ' 'arguments' ) elif opt.dest == 'list_profiles': query = 'list_profiles' option_dict = vars(self.options) if option_dict.get('list_profiles') == '--list-providers': self.error( '\'--list-profiles\' does not accept ' '\'--list-providers\' as an argument' ) self.selected_query_option = query funcname = 'process_{0}'.format(option.dest) if not hasattr(self, funcname): setattr(self, funcname, partial(process, option)) def _mixin_after_parsed(self): group_options_selected = [ option for option in self.cloud_queries_group.option_list if getattr(self.options, option.dest) is not False and getattr(self.options, option.dest) is not None ] if len(group_options_selected) > 1: self.error( 'The options {0} are mutually exclusive. Please only choose ' 'one of them'.format('/'.join([ option.get_opt_string() for option in group_options_selected ])) ) self.config['selected_query_option'] = self.selected_query_option class CloudProvidersListsMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 30 def _mixin_setup(self): group = self.providers_listings_group = optparse.OptionGroup( self, 'Cloud Providers Listings', # Include description here as a string ) group.add_option( '--list-locations', default=None, help=('Display a list of locations available in configured cloud ' 'providers. Pass the cloud provider that available ' 'locations are desired on, aka "linode", or pass "all" to ' 'list locations for all configured cloud providers.') ) group.add_option( '--list-images', default=None, help=('Display a list of images available in configured cloud ' 'providers. Pass the cloud provider that available images ' 'are desired on, aka "linode", or pass "all" to list images ' 'for all configured cloud providers.') ) group.add_option( '--list-sizes', default=None, help=('Display a list of sizes available in configured cloud ' 'providers. Pass the cloud provider that available sizes ' 'are desired on, aka "AWS", or pass "all" to list sizes ' 'for all configured cloud providers.') ) self.add_option_group(group) def _mixin_after_parsed(self): list_options_selected = [ option for option in self.providers_listings_group.option_list if getattr(self.options, option.dest) is not None ] if len(list_options_selected) > 1: self.error( 'The options {0} are mutually exclusive. Please only choose ' 'one of them'.format( '/'.join([ option.get_opt_string() for option in list_options_selected ]) ) ) class ProfilingPMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 130 def _mixin_setup(self): group = self.profiling_group = optparse.OptionGroup( self, 'Profiling support', # Include description here as a string ) group.add_option( '--profiling-path', dest='profiling_path', default='/tmp/stats', help=('Folder that will hold all stats generations path. Default: \'%default\'.') ) group.add_option( '--enable-profiling', dest='profiling_enabled', default=False, action='store_true', help=('Enable generating profiling stats. See also: --profiling-path.') ) self.add_option_group(group) class CloudCredentialsMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 30 def _mixin_setup(self): group = self.cloud_credentials_group = optparse.OptionGroup( self, 'Cloud Credentials', # Include description here as a string ) group.add_option( '--set-password', default=None, nargs=2, metavar='<USERNAME> <PROVIDER>', help=('Configure password for a cloud provider and save it to the keyring. ' 'PROVIDER can be specified with or without a driver, for example: ' '"--set-password bob rackspace" or more specific ' '"--set-password bob rackspace:openstack" ' 'Deprecated.') ) self.add_option_group(group) def process_set_password(self): if self.options.set_password: raise RuntimeError( 'This functionality is not supported; ' 'please see the keyring module at http://docs.saltstack.com/en/latest/topics/sdb/' ) class EAuthMixIn(six.with_metaclass(MixInMeta, object)): _mixin_prio_ = 30 def _mixin_setup(self): group = self.eauth_group = optparse.OptionGroup( self, 'External Authentication', # Include description here as a string ) group.add_option( '-a', '--auth', '--eauth', '--external-auth', default='', dest='eauth', help=('Specify an external authentication system to use.') ) group.add_option( '-T', '--make-token', default=False, dest='mktoken', action='store_true', help=('Generate and save an authentication token for re-use. The ' 'token is generated and made available for the period ' 'defined in the Salt Master.') ) group.add_option( '--username', dest='username', nargs=1, help=('Username for external authentication.') ) group.add_option( '--password', dest='password', nargs=1, help=('Password for external authentication.') ) self.add_option_group(group) class MasterOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, MergeConfigMixIn, LogLevelMixIn, RunUserMixin, DaemonMixIn, SaltfileMixIn)): description = 'The Salt master, used to control the Salt minions.' # ConfigDirMixIn config filename attribute _config_filename_ = 'master' # LogLevelMixIn attributes _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'master') _setup_mp_logging_listener_ = True def setup_config(self): return config.master_config(self.get_config_file_path()) class MinionOptionParser(six.with_metaclass(OptionParserMeta, MasterOptionParser)): # pylint: disable=no-init description = ( 'The Salt minion, receives commands from a remote Salt master.' ) # ConfigDirMixIn config filename attribute _config_filename_ = 'minion' # LogLevelMixIn attributes _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'minion') _setup_mp_logging_listener_ = True def setup_config(self): opts = config.minion_config(self.get_config_file_path(), # pylint: disable=no-member cache_minion_id=True, ignore_config_errors=False) # Optimization: disable multiprocessing logging if running as a # daemon, without engines and without multiprocessing if not opts.get('engines') and not opts.get('multiprocessing', True) \ and self.options.daemon: # pylint: disable=no-member self._setup_mp_logging_listener_ = False return opts class ProxyMinionOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, ProxyIdMixIn, ConfigDirMixIn, MergeConfigMixIn, LogLevelMixIn, RunUserMixin, DaemonMixIn, SaltfileMixIn)): # pylint: disable=no-init description = ( 'The Salt proxy minion, connects to and controls devices not able to run a minion. ' 'Receives commands from a remote Salt master.' ) # ConfigDirMixIn config filename attribute _config_filename_ = 'proxy' # LogLevelMixIn attributes _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'proxy') def setup_config(self): try: minion_id = self.values.proxyid except AttributeError: minion_id = None return config.minion_config(self.get_config_file_path(), cache_minion_id=False, minion_id=minion_id) class SyndicOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, MergeConfigMixIn, LogLevelMixIn, RunUserMixin, DaemonMixIn, SaltfileMixIn)): description = ( 'A seamless master of masters. Scale Salt to thousands of hosts or ' 'across many different networks.' ) # ConfigDirMixIn config filename attribute _config_filename_ = 'master' # LogLevelMixIn attributes _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'master') _setup_mp_logging_listener_ = True def setup_config(self): return config.syndic_config( self.get_config_file_path(), self.get_config_file_path('minion')) class SaltCMDOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, MergeConfigMixIn, TimeoutMixIn, ExtendedTargetOptionsMixIn, OutputOptionsMixIn, LogLevelMixIn, HardCrashMixin, SaltfileMixIn, ArgsStdinMixIn, EAuthMixIn)): default_timeout = 5 usage = '%prog [options] \'<target>\' <function> [arguments]' # ConfigDirMixIn config filename attribute _config_filename_ = 'master' # LogLevelMixIn attributes _default_logging_level_ = 'warning' _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'master') _loglevel_config_setting_name_ = 'cli_salt_log_file' try: os.getcwd() except OSError: sys.exit("Cannot access current working directory. Exiting!") def _mixin_setup(self): self.add_option( '-s', '--static', default=False, action='store_true', help=('Return the data from minions as a group after they ' 'all return.') ) self.add_option( '-p', '--progress', default=False, action='store_true', help=('Display a progress graph. Requires "progressbar" python package.') ) self.add_option( '--failhard', default=False, action='store_true', help=('Stop batch execution upon first "bad" return.') ) self.add_option( '--async', default=False, dest='async', action='store_true', help=('Run the salt command but don\'t wait for a reply.') ) self.add_option( '--subset', default=0, type=int, help=('Execute the routine on a random subset of the targeted ' 'minions. The minions will be verified that they have the ' 'named function before executing.') ) self.add_option( '-v', '--verbose', default=False, action='store_true', help=('Turn on command verbosity, display jid and active job ' 'queries.') ) self.add_option( '--hide-timeout', dest='show_timeout', default=True, action='store_false', help=('Hide minions that timeout.') ) self.add_option( '--show-jid', default=False, action='store_true', help=('Display jid without the additional output of --verbose.') ) self.add_option( '-b', '--batch', '--batch-size', default='', dest='batch', help=('Execute the salt job in batch mode, pass either the number ' 'of minions to batch at a time, or the percentage of ' 'minions to have running.') ) self.add_option( '--batch-wait', default=0, dest='batch_wait', type=float, help=('Wait the specified time in seconds after each job is done ' 'before freeing the slot in the batch for the next one.') ) self.add_option( '--return', default='', metavar='RETURNER', help=('Set an alternative return method. By default salt will ' 'send the return data from the command back to the master, ' 'but the return data can be redirected into any number of ' 'systems, databases or applications.') ) self.add_option( '--return_config', default='', metavar='RETURNER_CONF', help=('Set an alternative return method. By default salt will ' 'send the return data from the command back to the master, ' 'but the return data can be redirected into any number of ' 'systems, databases or applications.') ) self.add_option( '--return_kwargs', default={}, metavar='RETURNER_KWARGS', help=('Set any returner options at the command line.') ) self.add_option( '--module-executors', dest='module_executors', default=None, metavar='EXECUTOR_LIST', help=('Set an alternative list of executors to override the one ' 'set in minion config.') ) self.add_option( '-d', '--doc', '--documentation', dest='doc', default=False, action='store_true', help=('Return the documentation for the specified module or for ' 'all modules if none are specified.') ) self.add_option( '--args-separator', dest='args_separator', default=',', help=('Set the special argument used as a delimiter between ' 'command arguments of compound commands. This is useful ' 'when one wants to pass commas as arguments to ' 'some of the commands in a compound command.') ) self.add_option( '--summary', dest='cli_summary', default=False, action='store_true', help=('Display summary information about a salt command.') ) self.add_option( '--metadata', default='', metavar='METADATA', help=('Pass metadata into Salt, used to search jobs.') ) self.add_option( '--output-diff', dest='state_output_diff', action='store_true', default=False, help=('Report only those states that have changed.') ) self.add_option( '--config-dump', dest='config_dump', action='store_true', default=False, help=('Dump the master configuration values') ) def _mixin_after_parsed(self): if len(self.args) <= 1 and not self.options.doc: try: self.print_help() except Exception: # pylint: disable=broad-except # We get an argument that Python's optparser just can't deal # with. Perhaps stdout was redirected, or a file glob was # passed in. Regardless, we're in an unknown state here. sys.stdout.write('Invalid options passed. Please try -h for ' 'help.') # Try to warn if we can. sys.exit(salt.defaults.exitcodes.EX_GENERIC) # Dump the master configuration file, exit normally at the end. if self.options.config_dump: cfg = config.master_config(self.get_config_file_path()) sys.stdout.write(yaml.dump(cfg, default_flow_style=False)) sys.exit(salt.defaults.exitcodes.EX_OK) if self.options.doc: # Include the target if not self.args: self.args.insert(0, '*') if len(self.args) < 2: # Include the function self.args.insert(1, 'sys.doc') if self.args[1] != 'sys.doc': self.args.insert(1, 'sys.doc') if len(self.args) > 3: self.error('You can only get documentation for one method at one time.') if self.options.list: try: if ',' in self.args[0]: self.config['tgt'] = self.args[0].replace(' ', '').split(',') else: self.config['tgt'] = self.args[0].split() except IndexError: self.exit(42, '\nCannot execute command without defining a target.\n\n') else: try: self.config['tgt'] = self.args[0] except IndexError: self.exit(42, '\nCannot execute command without defining a target.\n\n') # Detect compound command and set up the data for it if self.args: try: if ',' in self.args[1]: self.config['fun'] = self.args[1].split(',') self.config['arg'] = [[]] cmd_index = 0 if (self.args[2:].count(self.options.args_separator) == len(self.config['fun']) - 1): # new style parsing: standalone argument separator for arg in self.args[2:]: if arg == self.options.args_separator: cmd_index += 1 self.config['arg'].append([]) else: self.config['arg'][cmd_index].append(arg) else: # old style parsing: argument separator can be inside args for arg in self.args[2:]: if self.options.args_separator in arg: sub_args = arg.split(self.options.args_separator) for sub_arg_index, sub_arg in enumerate(sub_args): if sub_arg: self.config['arg'][cmd_index].append(sub_arg) if sub_arg_index != len(sub_args) - 1: cmd_index += 1 self.config['arg'].append([]) else: self.config['arg'][cmd_index].append(arg) if len(self.config['fun']) != len(self.config['arg']): self.exit(42, 'Cannot execute compound command without ' 'defining all arguments.\n') # parse the args and kwargs before sending to the publish # interface for i in range(len(self.config['arg'])): self.config['arg'][i] = salt.utils.args.parse_input( self.config['arg'][i]) else: self.config['fun'] = self.args[1] self.config['arg'] = self.args[2:] # parse the args and kwargs before sending to the publish # interface self.config['arg'] = \ salt.utils.args.parse_input(self.config['arg']) except IndexError: self.exit(42, '\nIncomplete options passed.\n\n') def setup_config(self): return config.client_config(self.get_config_file_path()) class SaltCPOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, OutputOptionsMixIn, ConfigDirMixIn, MergeConfigMixIn, TimeoutMixIn, TargetOptionsMixIn, LogLevelMixIn, HardCrashMixin, SaltfileMixIn)): description = ( 'salt-cp is NOT intended to broadcast large files, it is intended to ' 'handle text files.\nsalt-cp can be used to distribute configuration ' 'files.' ) default_timeout = 5 usage = '%prog [options] \'<target>\' SOURCE DEST' # ConfigDirMixIn config filename attribute _config_filename_ = 'master' # LogLevelMixIn attributes _default_logging_level_ = 'warning' _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'master') _loglevel_config_setting_name_ = 'cli_salt_cp_log_file' def _mixin_after_parsed(self): # salt-cp needs arguments if len(self.args) <= 1: self.print_help() self.error('Insufficient arguments') if self.options.list: if ',' in self.args[0]: self.config['tgt'] = self.args[0].split(',') else: self.config['tgt'] = self.args[0].split() else: self.config['tgt'] = self.args[0] self.config['src'] = self.args[1:-1] self.config['dest'] = self.args[-1] def setup_config(self): return config.master_config(self.get_config_file_path()) class SaltKeyOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, MergeConfigMixIn, LogLevelMixIn, OutputOptionsMixIn, RunUserMixin, HardCrashMixin, SaltfileMixIn, EAuthMixIn)): description = 'Salt key is used to manage Salt authentication keys' usage = '%prog [options]' # ConfigDirMixIn config filename attribute _config_filename_ = 'master' # LogLevelMixIn attributes _skip_console_logging_config_ = True _logfile_config_setting_name_ = 'key_logfile' _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'key') def _mixin_setup(self): actions_group = optparse.OptionGroup(self, 'Actions') actions_group.set_conflict_handler('resolve') actions_group.add_option( '-l', '--list', default='', metavar='ARG', help=('List the public keys. The args ' '\'pre\', \'un\', and \'unaccepted\' will list ' 'unaccepted/unsigned keys. ' '\'acc\' or \'accepted\' will list accepted/signed keys. ' '\'rej\' or \'rejected\' will list rejected keys. ' '\'den\' or \'denied\' will list denied keys. ' 'Finally, \'all\' will list all keys.') ) actions_group.add_option( '-L', '--list-all', default=False, action='store_true', help='List all public keys. Deprecated: use "--list all".' ) actions_group.add_option( '-a', '--accept', default='', help='Accept the specified public key (use --include-rejected and ' '--include-denied to match rejected and denied keys in ' 'addition to pending keys). Globs are supported.', ) actions_group.add_option( '-A', '--accept-all', default=False, action='store_true', help='Accept all pending keys.' ) actions_group.add_option( '-r', '--reject', default='', help='Reject the specified public key. Use --include-accepted and ' '--include-denied to match accepted and denied keys in ' 'addition to pending keys. Globs are supported.' ) actions_group.add_option( '-R', '--reject-all', default=False, action='store_true', help='Reject all pending keys.' ) actions_group.add_option( '--include-all', default=False, action='store_true', help='Include rejected/accepted keys when accepting/rejecting. ' 'Deprecated: use "--include-rejected" and "--include-accepted".' ) actions_group.add_option( '--include-accepted', default=False, action='store_true', help='Include accepted keys when rejecting.' ) actions_group.add_option( '--include-rejected', default=False, action='store_true', help='Include rejected keys when accepting.' ) actions_group.add_option( '--include-denied', default=False, action='store_true', help='Include denied keys when accepting/rejecting.' ) actions_group.add_option( '-p', '--print', default='', help='Print the specified public key.' ) actions_group.add_option( '-P', '--print-all', default=False, action='store_true', help='Print all public keys.' ) actions_group.add_option( '-d', '--delete', default='', help='Delete the specified key. Globs are supported.' ) actions_group.add_option( '-D', '--delete-all', default=False, action='store_true', help='Delete all keys.' ) actions_group.add_option( '-f', '--finger', default='', help='Print the specified key\'s fingerprint.' ) actions_group.add_option( '-F', '--finger-all', default=False, action='store_true', help='Print all keys\' fingerprints.' ) self.add_option_group(actions_group) self.add_option( '-q', '--quiet', default=False, action='store_true', help='Suppress output.' ) self.add_option( '-y', '--yes', default=False, action='store_true', help='Answer "Yes" to all questions presented. Default: %default.' ) self.add_option( '--rotate-aes-key', default=True, help=('Setting this to False prevents the master from refreshing ' 'the key session when keys are deleted or rejected, this ' 'lowers the security of the key deletion/rejection operation. ' 'Default: %default.') ) key_options_group = optparse.OptionGroup( self, 'Key Generation Options' ) self.add_option_group(key_options_group) key_options_group.add_option( '--gen-keys', default='', help='Set a name to generate a keypair for use with salt.' ) key_options_group.add_option( '--gen-keys-dir', default='.', help=('Set the directory to save the generated keypair, only ' 'works with "gen_keys_dir" option. Default: \'%default\'.') ) key_options_group.add_option( '--keysize', default=2048, type=int, help=('Set the keysize for the generated key, only works with ' 'the "--gen-keys" option, the key size must be 2048 or ' 'higher, otherwise it will be rounded up to 2048. ' 'Default: %default.') ) key_options_group.add_option( '--gen-signature', default=False, action='store_true', help=('Create a signature file of the masters public-key named ' 'master_pubkey_signature. The signature can be send to a ' 'minion in the masters auth-reply and enables the minion ' 'to verify the masters public-key cryptographically. ' 'This requires a new signing-key-pair which can be auto-created ' 'with the --auto-create parameter.') ) key_options_group.add_option( '--priv', default='', type=str, help=('The private-key file to create a signature with.') ) key_options_group.add_option( '--signature-path', default='', type=str, help=('The path where the signature file should be written.') ) key_options_group.add_option( '--pub', default='', type=str, help=('The public-key file to create a signature for.') ) key_options_group.add_option( '--auto-create', default=False, action='store_true', help=('Auto-create a signing key-pair if it does not yet exist.') ) def process_config_dir(self): if self.options.gen_keys: # We're generating keys, override the default behavior of this # function if we don't have any access to the configuration # directory. if not os.access(self.options.config_dir, os.R_OK): if not os.path.isdir(self.options.gen_keys_dir): # This would be done at a latter stage, but we need it now # so no errors are thrown os.makedirs(self.options.gen_keys_dir) self.options.config_dir = self.options.gen_keys_dir super(SaltKeyOptionParser, self).process_config_dir() # Don't change its mixin priority! process_config_dir._mixin_prio_ = ConfigDirMixIn._mixin_prio_ def setup_config(self): keys_config = config.master_config(self.get_config_file_path()) if self.options.gen_keys: # Since we're generating the keys, some defaults can be assumed # or tweaked keys_config['key_logfile'] = os.devnull keys_config['pki_dir'] = self.options.gen_keys_dir return keys_config def process_rotate_aes_key(self): if hasattr(self.options, 'rotate_aes_key') and isinstance(self.options.rotate_aes_key, str): if self.options.rotate_aes_key.lower() == 'true': self.options.rotate_aes_key = True elif self.options.rotate_aes_key.lower() == 'false': self.options.rotate_aes_key = False def process_list(self): # Filter accepted list arguments as soon as possible if not self.options.list: return if not self.options.list.startswith(('acc', 'pre', 'un', 'rej', 'den', 'all')): self.error( '\'{0}\' is not a valid argument to \'--list\''.format( self.options.list ) ) def process_keysize(self): if self.options.keysize < 2048: self.error('The minimum value for keysize is 2048') elif self.options.keysize > 32768: self.error('The maximum value for keysize is 32768') def process_gen_keys_dir(self): # Schedule __create_keys_dir() to run if there's a value for # --create-keys-dir self._mixin_after_parsed_funcs.append(self.__create_keys_dir) # pylint: disable=no-member def _mixin_after_parsed(self): # It was decided to always set this to info, since it really all is # info or error. self.config['loglevel'] = 'info' def __create_keys_dir(self, *args): # pylint: disable=unused-argument if not os.path.isdir(self.config['gen_keys_dir']): os.makedirs(self.config['gen_keys_dir']) class SaltCallOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, MergeConfigMixIn, LogLevelMixIn, OutputOptionsMixIn, HardCrashMixin, SaltfileMixIn, ArgsStdinMixIn, ProfilingPMixIn)): description = ('Salt call is used to execute module functions locally ' 'on a minion') usage = '%prog [options] <function> [arguments]' # ConfigDirMixIn config filename attribute _config_filename_ = 'minion' # LogLevelMixIn attributes _default_logging_level_ = 'info' _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'minion') def _mixin_setup(self): self.add_option( '-g', '--grains', dest='grains_run', default=False, action='store_true', help='Return the information generated by the salt grains.' ) self.add_option( '-m', '--module-dirs', default=[], action='append', help=('Specify an additional directory to pull modules from. ' 'Multiple directories can be provided by passing ' '`-m/--module-dirs` multiple times.') ) self.add_option( '-d', '--doc', '--documentation', dest='doc', default=False, action='store_true', help=('Return the documentation for the specified module or for ' 'all modules if none are specified.') ) self.add_option( '--master', default='', dest='master', help=('Specify the master to use. The minion must be ' 'authenticated with the master. If this option is omitted, ' 'the master options from the minion config will be used. ' 'If multi masters are set up the first listed master that ' 'responds will be used.') ) self.add_option( '--return', default='', metavar='RETURNER', help=('Set salt-call to pass the return data to one or many ' 'returner interfaces.') ) self.add_option( '--local', default=False, action='store_true', help='Run salt-call locally, as if there was no master running.' ) self.add_option( '--file-root', default=None, help='Set this directory as the base file root.' ) self.add_option( '--pillar-root', default=None, help='Set this directory as the base pillar root.' ) self.add_option( '--states-dir', default=None, help='Set this directory to search for additional states.' ) self.add_option( '--retcode-passthrough', default=False, action='store_true', help=('Exit with the salt call retcode and not the salt binary ' 'retcode.') ) self.add_option( '--metadata', default=False, dest='print_metadata', action='store_true', help=('Print out the execution metadata as well as the return. ' 'This will print out the outputter data, the return code, ' 'etc.') ) self.add_option( '--set-metadata', dest='metadata', default=None, metavar='METADATA', help=('Pass metadata into Salt, used to search jobs.') ) self.add_option( '--id', default='', dest='id', help=('Specify the minion id to use. If this option is omitted, ' 'the id option from the minion config will be used.') ) self.add_option( '--skip-grains', default=False, action='store_true', help=('Do not load grains.') ) self.add_option( '--refresh-grains-cache', default=False, action='store_true', help=('Force a refresh of the grains cache.') ) self.add_option( '-t', '--timeout', default=60, dest='auth_timeout', type=int, help=('Change the timeout, if applicable, for the running ' 'command. Default: %default.') ) self.add_option( '--output-diff', dest='state_output_diff', action='store_true', default=False, help=('Report only those states that have changed.') ) def _mixin_after_parsed(self): if not self.args and not self.options.grains_run and not self.options.doc: self.print_help() self.error('Requires function, --grains or --doc') elif len(self.args) >= 1: if self.options.grains_run: self.error('-g/--grains does not accept any arguments') if self.options.doc and len(self.args) > 1: self.error('You can only get documentation for one method at one time') self.config['fun'] = self.args[0] self.config['arg'] = self.args[1:] def setup_config(self): opts = config.minion_config(self.get_config_file_path(), cache_minion_id=True) if opts.get('transport') == 'raet': if not self._find_raet_minion(opts): # must create caller minion opts['__role'] = kinds.APPL_KIND_NAMES[kinds.applKinds.caller] return opts def _find_raet_minion(self, opts): ''' Returns true if local RAET Minion is available ''' yardname = 'manor' dirpath = opts['sock_dir'] role = opts.get('id') if not role: emsg = ("Missing role required to setup RAET SaltCaller.") logging.getLogger(__name__).error(emsg + "\n") raise ValueError(emsg) kind = opts.get('__role') # application kind 'master', 'minion', etc if kind not in kinds.APPL_KINDS: emsg = ("Invalid application kind = '{0}' for RAET SaltCaller.".format(kind)) logging.getLogger(__name__).error(emsg + "\n") raise ValueError(emsg) if kind in [kinds.APPL_KIND_NAMES[kinds.applKinds.minion], kinds.APPL_KIND_NAMES[kinds.applKinds.caller], ]: lanename = "{0}_{1}".format(role, kind) else: emsg = ("Unsupported application kind '{0}' for RAET SaltCaller.".format(kind)) logging.getLogger(__name__).error(emsg + '\n') raise ValueError(emsg) if kind == kinds.APPL_KIND_NAMES[kinds.applKinds.minion]: # minion check from raet.lane.yarding import Yard ha, dirpath = Yard.computeHa(dirpath, lanename, yardname) # pylint: disable=invalid-name if (os.path.exists(ha) and not os.path.isfile(ha) and not os.path.isdir(ha)): # minion manor yard return True return False def process_module_dirs(self): for module_dir in self.options.module_dirs: # Provide some backwards compatibility with previous comma # delimited format if ',' in module_dir: self.config.setdefault('module_dirs', []).extend( os.path.abspath(x) for x in module_dir.split(',')) continue self.config.setdefault('module_dirs', []).append(os.path.abspath(module_dir)) class SaltRunOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, MergeConfigMixIn, TimeoutMixIn, LogLevelMixIn, HardCrashMixin, SaltfileMixIn, OutputOptionsMixIn, ArgsStdinMixIn, ProfilingPMixIn, EAuthMixIn)): default_timeout = 1 usage = '%prog [options]' # ConfigDirMixIn config filename attribute _config_filename_ = 'master' # LogLevelMixIn attributes _default_logging_level_ = 'warning' _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'master') _loglevel_config_setting_name_ = 'cli_salt_run_log_file' def _mixin_setup(self): self.add_option( '-d', '--doc', '--documentation', dest='doc', default=False, action='store_true', help=('Display documentation for runners, pass a runner or ' 'runner.function to see documentation on only that runner ' 'or function.') ) self.add_option( '--async', default=False, action='store_true', help=('Start the runner operation and immediately return control.') ) group = self.output_options_group = optparse.OptionGroup( self, 'Output Options', 'Configure your preferred output format.' ) self.add_option_group(group) group.add_option( '--quiet', default=False, action='store_true', help='Do not display the results of the run.' ) def _mixin_after_parsed(self): if self.options.doc and len(self.args) > 1: self.error('You can only get documentation for one method at one time') if len(self.args) > 0: self.config['fun'] = self.args[0] else: self.config['fun'] = '' if len(self.args) > 1: self.config['arg'] = self.args[1:] else: self.config['arg'] = [] def setup_config(self): return config.client_config(self.get_config_file_path()) class SaltSSHOptionParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, MergeConfigMixIn, LogLevelMixIn, TargetOptionsMixIn, OutputOptionsMixIn, SaltfileMixIn, HardCrashMixin)): usage = '%prog [options]' # ConfigDirMixIn config filename attribute _config_filename_ = 'master' # LogLevelMixIn attributes _default_logging_level_ = 'warning' _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'ssh') _loglevel_config_setting_name_ = 'cli_salt_run_log_file' def _mixin_setup(self): self.add_option( '-r', '--raw', '--raw-shell', dest='raw_shell', default=False, action='store_true', help=('Don\'t execute a salt routine on the targets, execute a ' 'raw shell command.') ) self.add_option( '--roster', dest='roster', default='flat', help=('Define which roster system to use, this defines if a ' 'database backend, scanner, or custom roster system is ' 'used. Default: \'flat\'.') ) self.add_option( '--roster-file', dest='roster_file', default='', help=('Define an alternative location for the default roster ' 'file location. The default roster file is called roster ' 'and is found in the same directory as the master config ' 'file.') ) self.add_option( '--refresh', '--refresh-cache', dest='refresh_cache', default=False, action='store_true', help=('Force a refresh of the master side data cache of the ' 'target\'s data. This is needed if a target\'s grains have ' 'been changed and the auto refresh timeframe has not been ' 'reached.') ) self.add_option( '--max-procs', dest='ssh_max_procs', default=25, type=int, help='Set the number of concurrent minions to communicate with. ' 'This value defines how many processes are opened up at a ' 'time to manage connections, the more running processes the ' 'faster communication should be. Default: %default.' ) self.add_option( '--extra-filerefs', dest='extra_filerefs', default=None, help='Pass in extra files to include in the state tarball.' ) self.add_option('--min-extra-modules', dest='min_extra_mods', default=None, help='One or comma-separated list of extra Python modules' 'to be included into Minimal Salt.') self.add_option( '--thin-extra-modules', dest='thin_extra_mods', default=None, help='One or comma-separated list of extra Python modules' 'to be included into Thin Salt.') self.add_option( '-v', '--verbose', default=False, action='store_true', help=('Turn on command verbosity, display jid.') ) self.add_option( '-s', '--static', default=False, action='store_true', help=('Return the data from minions as a group after they ' 'all return.') ) self.add_option( '-w', '--wipe', default=False, action='store_true', dest='ssh_wipe', help='Remove the deployment of the salt files when done executing.', ) self.add_option( '-W', '--rand-thin-dir', default=False, action='store_true', help=('Select a random temp dir to deploy on the remote system. ' 'The dir will be cleaned after the execution.')) self.add_option( '-t', '--regen-thin', '--thin', dest='regen_thin', default=False, action='store_true', help=('Trigger a thin tarball regeneration. This is needed if ' 'custom grains/modules/states have been added or updated.')) self.add_option( '--python2-bin', default='python2', help='Path to a python2 binary which has salt installed.' ) self.add_option( '--python3-bin', default='python3', help='Path to a python3 binary which has salt installed.' ) self.add_option( '--jid', default=None, help='Pass a JID to be used instead of generating one.' ) ports_group = optparse.OptionGroup( self, 'Port Forwarding Options', 'Parameters for setting up SSH port forwarding.' ) ports_group.add_option( '--remote-port-forwards', dest='ssh_remote_port_forwards', help='Setup remote port forwarding using the same syntax as with ' 'the -R parameter of ssh. A comma separated list of port ' 'forwarding definitions will be translated into multiple ' '-R parameters.' ) self.add_option_group(ports_group) auth_group = optparse.OptionGroup( self, 'Authentication Options', 'Parameters affecting authentication.' ) auth_group.add_option( '--priv', dest='ssh_priv', help='Ssh private key file.' ) auth_group.add_option( '-i', '--ignore-host-keys', dest='ignore_host_keys', default=False, action='store_true', help='By default ssh host keys are honored and connections will ' 'ask for approval. Use this option to disable ' 'StrictHostKeyChecking.' ) auth_group.add_option( '--no-host-keys', dest='no_host_keys', default=False, action='store_true', help='Removes all host key checking functionality from SSH session.' ) auth_group.add_option( '--user', dest='ssh_user', default='root', help='Set the default user to attempt to use when ' 'authenticating.' ) auth_group.add_option( '--passwd', dest='ssh_passwd', default='', help='Set the default password to attempt to use when ' 'authenticating.' ) auth_group.add_option( '--askpass', dest='ssh_askpass', default=False, action='store_true', help='Interactively ask for the SSH password with no echo - avoids ' 'password in process args and stored in history.' ) auth_group.add_option( '--key-deploy', dest='ssh_key_deploy', default=False, action='store_true', help='Set this flag to attempt to deploy the authorized ssh key ' 'with all minions. This combined with --passwd can make ' 'initial deployment of keys very fast and easy.' ) auth_group.add_option( '--identities-only', dest='ssh_identities_only', default=False, action='store_true', help='Use the only authentication identity files configured in the ' 'ssh_config files. See IdentitiesOnly flag in man ssh_config.' ) auth_group.add_option( '--sudo', dest='ssh_sudo', default=False, action='store_true', help='Run command via sudo.' ) self.add_option_group(auth_group) scan_group = optparse.OptionGroup( self, 'Scan Roster Options', 'Parameters affecting scan roster.' ) scan_group.add_option( '--scan-ports', default='22', dest='ssh_scan_ports', help='Comma-separated list of ports to scan in the scan roster.', ) scan_group.add_option( '--scan-timeout', default=0.01, dest='ssh_scan_timeout', help='Scanning socket timeout for the scan roster.', ) self.add_option_group(scan_group) def _mixin_after_parsed(self): if not self.args: self.print_help() self.error('Insufficient arguments') if self.options.list: if ',' in self.args[0]: self.config['tgt'] = self.args[0].split(',') else: self.config['tgt'] = self.args[0].split() else: self.config['tgt'] = self.args[0] self.config['argv'] = self.args[1:] if not self.config['argv'] or not self.config['tgt']: self.print_help() self.error('Insufficient arguments') if self.options.ssh_askpass: self.options.ssh_passwd = getpass.getpass('Password: ') def setup_config(self): return config.master_config(self.get_config_file_path()) def process_jid(self): if self.options.jid is not None: if not salt.utils.jid.is_jid(self.options.jid): self.error('\'{0}\' is not a valid JID'.format(self.options.jid)) class SaltCloudParser(six.with_metaclass(OptionParserMeta, OptionParser, LogLevelMixIn, MergeConfigMixIn, OutputOptionsMixIn, ConfigDirMixIn, CloudQueriesMixIn, ExecutionOptionsMixIn, CloudProvidersListsMixIn, CloudCredentialsMixIn, HardCrashMixin, SaltfileMixIn)): # ConfigDirMixIn attributes _config_filename_ = 'cloud' # LogLevelMixIn attributes _default_logging_level_ = 'info' _logfile_config_setting_name_ = 'log_file' _loglevel_config_setting_name_ = 'log_level_logfile' _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'cloud') def print_versions_report(self, file=sys.stdout): # pylint: disable=redefined-builtin print('\n'.join(version.versions_report(include_salt_cloud=True)), file=file) self.exit(salt.defaults.exitcodes.EX_OK) def parse_args(self, args=None, values=None): try: # Late import in order not to break setup from salt.cloud import libcloudfuncs libcloudfuncs.check_libcloud_version() except ImportError as exc: self.error(exc) return super(SaltCloudParser, self).parse_args(args, values) def _mixin_after_parsed(self): if 'DUMP_SALT_CLOUD_CONFIG' in os.environ: import pprint print('Salt cloud configuration dump(INCLUDES SENSIBLE DATA):') pprint.pprint(self.config) self.exit(salt.defaults.exitcodes.EX_OK) if self.args: self.config['names'] = self.args def setup_config(self): try: return config.cloud_config(self.get_config_file_path()) except salt.exceptions.SaltCloudConfigError as exc: self.error(exc) class SPMParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, LogLevelMixIn, MergeConfigMixIn, SaltfileMixIn)): ''' The cli parser object used to fire up the salt spm system. ''' description = 'SPM is used to manage 3rd party formulas and other Salt components' usage = '%prog [options] <function> [arguments]' # ConfigDirMixIn config filename attribute _config_filename_ = 'spm' # LogLevelMixIn attributes _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'spm') def _mixin_setup(self): self.add_option( '-y', '--assume-yes', default=False, action='store_true', help='Default "yes" in answer to all confirmation questions.' ) self.add_option( '-f', '--force', default=False, action='store_true', help='Default "yes" in answer to all confirmation questions.' ) self.add_option( '-v', '--verbose', default=False, action='store_true', help='Display more detailed information.' ) def _mixin_after_parsed(self): # spm needs arguments if len(self.args) <= 1: if not self.args or self.args[0] not in ('update_repo',): self.print_help() self.error('Insufficient arguments') def setup_config(self): return salt.config.spm_config(self.get_config_file_path()) class SaltAPIParser(six.with_metaclass(OptionParserMeta, OptionParser, ConfigDirMixIn, LogLevelMixIn, DaemonMixIn, MergeConfigMixIn)): ''' The Salt API cli parser object used to fire up the salt api system. ''' # ConfigDirMixIn config filename attribute _config_filename_ = 'master' # LogLevelMixIn attributes _default_logging_logfile_ = os.path.join(syspaths.LOGS_DIR, 'api') def setup_config(self): return salt.config.api_config(self.get_config_file_path()) # pylint: disable=no-member
38.735819
110
0.543787
026962c395159a92f5db4dc41caa8e06f959058f
443
py
Python
ml/data_utils.py
wds-seu/Aceso
70fca2b3ff810fa32b7ebe6edf0d264da126d521
[ "MIT" ]
1
2019-10-08T10:54:52.000Z
2019-10-08T10:54:52.000Z
ml/data_utils.py
wds-seu/Aceso
70fca2b3ff810fa32b7ebe6edf0d264da126d521
[ "MIT" ]
null
null
null
ml/data_utils.py
wds-seu/Aceso
70fca2b3ff810fa32b7ebe6edf0d264da126d521
[ "MIT" ]
null
null
null
#!/usr/bin/env python # -*- coding:utf-8 _*- from sklearn.model_selection import train_test_split from utils.DataHelper import DataHelper from config import DefaultConfig opt = DefaultConfig() def build_data(): x_text, y, vocabulary, vocabulary_inv = DataHelper(opt.train_data_root, train=True).load_text_data(opt.use_umls) x_train, x_val, y_train, y_val = train_test_split(x_text, y, test_size=0.3, random_state=1, shuffle=True)
34.076923
116
0.772009
161fc216cd97b68346d08db33a110ef3a04fb9bd
5,738
py
Python
betka/utils.py
phracek/betka
12c920bd76d33f81bbf8cda6f27d673c8826cf9e
[ "MIT" ]
1
2020-11-05T21:16:28.000Z
2020-11-05T21:16:28.000Z
betka/utils.py
phracek/betka
12c920bd76d33f81bbf8cda6f27d673c8826cf9e
[ "MIT" ]
16
2020-03-20T11:23:27.000Z
2022-03-08T17:09:11.000Z
betka/utils.py
phracek/betka
12c920bd76d33f81bbf8cda6f27d673c8826cf9e
[ "MIT" ]
1
2020-03-11T09:29:48.000Z
2020-03-11T09:29:48.000Z
# MIT License # # Copyright (c) 2020 SCL team at Red Hat # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. from contextlib import contextmanager import logging import shutil import os import json import jinja2 import subprocess from pathlib import Path from betka.constants import HOME logger = logging.getLogger(__name__) def run_cmd(cmd, return_output=False, ignore_error=False, shell=False, **kwargs): """ Run provided command on host system using the same user as invoked this code. Raises subprocess.CalledProcessError if it fails. :param cmd: list or str :param return_output: bool, return output of the command :param ignore_error: bool, do not fail in case nonzero return code :param shell: bool, run command in shell :param kwargs: pass keyword arguments to subprocess.check_* functions; for more info, please check `help(subprocess.Popen)` :return: None or str """ logger.debug("command: %r", cmd) try: if return_output: return subprocess.check_output( cmd, stderr=subprocess.STDOUT, universal_newlines=True, shell=shell, **kwargs, ) else: return subprocess.check_call(cmd, shell=shell, **kwargs) except subprocess.CalledProcessError as cpe: if ignore_error: if return_output: return cpe.output else: return cpe.returncode else: logger.error(f"failed with code {cpe.returncode} and output:\n{cpe.output}") raise cpe def text_from_template(template_dir, template_filename, template_data): """ Create text based on template in path template_dir/template_filename :param template_dir: string, directory containing templates :param template_filename: template for text in jinja :param template_data: dict, data for substitution in template :return: string """ if not os.path.exists(os.path.join(template_dir, template_filename)): raise FileNotFoundError("Path to template not found.") template_loader = jinja2.FileSystemLoader(searchpath=template_dir) template_env = jinja2.Environment(loader=template_loader) template = template_env.get_template(template_filename) output_text = template.render(template_data=template_data) logger.debug("Text from template created:") logger.debug(output_text) return output_text def copy_upstream2downstream(src_parent: Path, dest_parent: Path): """Copies content from upstream repo to downstream repo Copies all files/dirs/symlinks from upstream source to dist-git one by one, while removing previous if exists. :param src_parent: path to source directory :param dest_parent: path to destination directory """ for f in src_parent.iterdir(): if f.name.startswith(".git"): continue dest = dest_parent / f.name src = src_parent / f.name logger.debug(f"Copying {str(src)} to {str(dest)}.") # First remove the dest only if it is not symlink. if dest.is_dir() and not dest.is_symlink(): logger.debug("rmtree %s", dest) shutil.rmtree(dest) else: if dest.exists(): dest.unlink() # Now copy the src to dest if src.is_symlink() or not src.is_dir(): logger.debug("cp %s %s", src, dest) shutil.copy2(src, dest, follow_symlinks=False) else: logger.debug("cp -r %s %s", src, dest) shutil.copytree(src, dest, symlinks=True) def clean_directory(path: Path): """ Function cleans directory except itself :param path: directory path which is cleaned """ for d in path.iterdir(): src = path / d if src.is_dir(): logger.debug("rmtree %s", str(src)) shutil.rmtree(src) else: src.unlink() def list_dir_content(dir_name: Path): """ Lists all content of dir_name :param dir_name: Directory for showing files """ logger.info("Look for a content in '%s' directory", str(dir_name)) for f in dir_name.rglob("*"): if str(f).startswith(".git"): continue logger.debug(f"{f.parent / f.name}") def load_config_json(): with open(f"{HOME}/config.json") as config_file: data = json.load(config_file) return data @contextmanager def cwd(path): """ Switch to Path directory and once action is done returns back :param path: :return: """ prev_cwd = Path.cwd() os.chdir(path) try: yield finally: os.chdir(prev_cwd)
32.788571
89
0.66626
5d0138a06f2443ef3fc96673c74b3944b37a3723
1,782
py
Python
torch_glow/tests/nodes/avgpool1d_test.py
YaronBenAtar/glow
a13706a4239fa7eaf059c670dc573e3eb0768f86
[ "Apache-2.0" ]
2,838
2018-05-02T16:57:22.000Z
2022-03-31T14:35:26.000Z
torch_glow/tests/nodes/avgpool1d_test.py
YaronBenAtar/glow
a13706a4239fa7eaf059c670dc573e3eb0768f86
[ "Apache-2.0" ]
4,149
2018-05-02T17:50:14.000Z
2022-03-31T23:56:43.000Z
torch_glow/tests/nodes/avgpool1d_test.py
LaudateCorpus1/glow-1
cda5383b1609ebad1a3631ca77b41b8a863443d4
[ "Apache-2.0" ]
685
2018-05-02T16:54:09.000Z
2022-03-24T01:12:24.000Z
# Copyright (c) Glow Contributors. See CONTRIBUTORS file. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from __future__ import absolute_import, division, print_function, unicode_literals import torch import torch.nn.functional as F from tests import utils class SimpleAvgPool1dModule(torch.nn.Module): def __init__(self, kernel_size, stride=None, padding=0): super(SimpleAvgPool1dModule, self).__init__() self.kernel_size = kernel_size self.padding = padding self.stride = stride def forward(self, inputs): return F.avg_pool1d( inputs, self.kernel_size, padding=self.padding, stride=self.stride ) class TestAvgPool1d(utils.TorchGlowTestCase): def test_avg_pool1d_basic(self): """Basic test of the PyTorch avg_pool1d Node on Glow.""" inputs = torch.randn(1, 4, 9) utils.compare_tracing_methods( SimpleAvgPool1dModule(3), inputs, fusible_ops={"aten::avg_pool1d"} ) def test_avg_pool1d_with_args(self): """Test of the PyTorch avg_pool1d Node with arguments on Glow.""" inputs = torch.randn(1, 4, 10) utils.compare_tracing_methods( SimpleAvgPool1dModule(3, stride=7), inputs, fusible_ops={"aten::avg_pool1d"} )
34.269231
88
0.708193
b1c5f4cfef4234cf455db73d71bf20f2ddeb87c5
15,306
py
Python
python/ccxt/async_support/btcbox.py
atoll6/ccxt
3c5fd65a32073bfbe1a7e0f5a02d7d3f93850780
[ "MIT" ]
1
2018-08-20T09:38:13.000Z
2018-08-20T09:38:13.000Z
python/ccxt/async_support/btcbox.py
atoll6/ccxt
3c5fd65a32073bfbe1a7e0f5a02d7d3f93850780
[ "MIT" ]
null
null
null
python/ccxt/async_support/btcbox.py
atoll6/ccxt
3c5fd65a32073bfbe1a7e0f5a02d7d3f93850780
[ "MIT" ]
1
2021-09-16T08:33:31.000Z
2021-09-16T08:33:31.000Z
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.async_support.base.exchange import Exchange # ----------------------------------------------------------------------------- try: basestring # Python 3 except NameError: basestring = str # Python 2 import json from ccxt.base.errors import ExchangeError from ccxt.base.errors import AuthenticationError from ccxt.base.errors import PermissionDenied from ccxt.base.errors import InsufficientFunds from ccxt.base.errors import InvalidOrder from ccxt.base.errors import OrderNotFound from ccxt.base.errors import DDoSProtection from ccxt.base.errors import InvalidNonce from ccxt.base.precise import Precise class btcbox(Exchange): def describe(self): return self.deep_extend(super(btcbox, self).describe(), { 'id': 'btcbox', 'name': 'BtcBox', 'countries': ['JP'], 'rateLimit': 1000, 'version': 'v1', 'has': { 'cancelOrder': True, 'CORS': None, 'createOrder': True, 'fetchBalance': True, 'fetchOpenOrders': True, 'fetchOrder': True, 'fetchOrderBook': True, 'fetchOrders': True, 'fetchTicker': True, 'fetchTickers': None, 'fetchTrades': True, }, 'urls': { 'logo': 'https://user-images.githubusercontent.com/51840849/87327317-98c55400-c53c-11ea-9a11-81f7d951cc74.jpg', 'api': 'https://www.btcbox.co.jp/api', 'www': 'https://www.btcbox.co.jp/', 'doc': 'https://blog.btcbox.jp/en/archives/8762', 'fees': 'https://support.btcbox.co.jp/hc/en-us/articles/360001235694-Fees-introduction', }, 'api': { 'public': { 'get': [ 'depth', 'orders', 'ticker', ], }, 'private': { 'post': [ 'balance', 'trade_add', 'trade_cancel', 'trade_list', 'trade_view', 'wallet', ], }, }, 'markets': { 'BTC/JPY': {'id': 'btc', 'symbol': 'BTC/JPY', 'base': 'BTC', 'quote': 'JPY', 'baseId': 'btc', 'quoteId': 'jpy', 'taker': 0.05 / 100, 'maker': 0.05 / 100}, 'ETH/JPY': {'id': 'eth', 'symbol': 'ETH/JPY', 'base': 'ETH', 'quote': 'JPY', 'baseId': 'eth', 'quoteId': 'jpy', 'taker': 0.10 / 100, 'maker': 0.10 / 100}, 'LTC/JPY': {'id': 'ltc', 'symbol': 'LTC/JPY', 'base': 'LTC', 'quote': 'JPY', 'baseId': 'ltc', 'quoteId': 'jpy', 'taker': 0.10 / 100, 'maker': 0.10 / 100}, 'BCH/JPY': {'id': 'bch', 'symbol': 'BCH/JPY', 'base': 'BCH', 'quote': 'JPY', 'baseId': 'bch', 'quoteId': 'jpy', 'taker': 0.10 / 100, 'maker': 0.10 / 100}, }, 'exceptions': { '104': AuthenticationError, '105': PermissionDenied, '106': InvalidNonce, '107': InvalidOrder, # price should be an integer '200': InsufficientFunds, '201': InvalidOrder, # amount too small '202': InvalidOrder, # price should be [0 : 1000000] '203': OrderNotFound, '401': OrderNotFound, # cancel canceled, closed or non-existent order '402': DDoSProtection, }, }) async def fetch_balance(self, params={}): await self.load_markets() response = await self.privatePostBalance(params) result = {'info': response} codes = list(self.currencies.keys()) for i in range(0, len(codes)): code = codes[i] currency = self.currency(code) currencyId = currency['id'] free = currencyId + '_balance' if free in response: account = self.account() used = currencyId + '_lock' account['free'] = self.safe_string(response, free) account['used'] = self.safe_string(response, used) result[code] = account return self.parse_balance(result) async def fetch_order_book(self, symbol, limit=None, params={}): await self.load_markets() market = self.market(symbol) request = {} numSymbols = len(self.symbols) if numSymbols > 1: request['coin'] = market['baseId'] response = await self.publicGetDepth(self.extend(request, params)) return self.parse_order_book(response, symbol) def parse_ticker(self, ticker, market=None): timestamp = self.milliseconds() symbol = None if market is not None: symbol = market['symbol'] last = self.safe_number(ticker, 'last') return { 'symbol': symbol, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'high': self.safe_number(ticker, 'high'), 'low': self.safe_number(ticker, 'low'), 'bid': self.safe_number(ticker, 'buy'), 'bidVolume': None, 'ask': self.safe_number(ticker, 'sell'), 'askVolume': None, 'vwap': None, 'open': None, 'close': last, 'last': last, 'previousClose': None, 'change': None, 'percentage': None, 'average': None, 'baseVolume': self.safe_number(ticker, 'vol'), 'quoteVolume': self.safe_number(ticker, 'volume'), 'info': ticker, } async def fetch_ticker(self, symbol, params={}): await self.load_markets() market = self.market(symbol) request = {} numSymbols = len(self.symbols) if numSymbols > 1: request['coin'] = market['baseId'] response = await self.publicGetTicker(self.extend(request, params)) return self.parse_ticker(response, market) def parse_trade(self, trade, market=None): timestamp = self.safe_timestamp(trade, 'date') symbol = None if market is not None: symbol = market['symbol'] id = self.safe_string(trade, 'tid') priceString = self.safe_string(trade, 'price') amountString = self.safe_string(trade, 'amount') price = self.parse_number(priceString) amount = self.parse_number(amountString) cost = self.parse_number(Precise.string_mul(priceString, amountString)) type = None side = self.safe_string(trade, 'type') return { 'info': trade, 'id': id, 'order': None, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'symbol': symbol, 'type': type, 'side': side, 'takerOrMaker': None, 'price': price, 'amount': amount, 'cost': cost, 'fee': None, } async def fetch_trades(self, symbol, since=None, limit=None, params={}): await self.load_markets() market = self.market(symbol) request = {} numSymbols = len(self.symbols) if numSymbols > 1: request['coin'] = market['baseId'] response = await self.publicGetOrders(self.extend(request, params)) return self.parse_trades(response, market, since, limit) async def create_order(self, symbol, type, side, amount, price=None, params={}): await self.load_markets() market = self.market(symbol) request = { 'amount': amount, 'price': price, 'type': side, 'coin': market['baseId'], } response = await self.privatePostTradeAdd(self.extend(request, params)) # # { # "result":true, # "id":"11" # } # return self.parse_order(response, market) async def cancel_order(self, id, symbol=None, params={}): await self.load_markets() # a special case for btcbox – default symbol is BTC/JPY if symbol is None: symbol = 'BTC/JPY' market = self.market(symbol) request = { 'id': id, 'coin': market['baseId'], } response = await self.privatePostTradeCancel(self.extend(request, params)) # # {"result":true, "id":"11"} # return self.parse_order(response, market) def parse_order_status(self, status): statuses = { # TODO: complete list 'part': 'open', # partially or not at all executed 'all': 'closed', # fully executed 'cancelled': 'canceled', 'closed': 'closed', # never encountered, seems to be bug in the doc 'no': 'closed', # not clarified in the docs... } return self.safe_string(statuses, status, status) def parse_order(self, order, market=None): # # { # "id":11, # "datetime":"2014-10-21 10:47:20", # "type":"sell", # "price":42000, # "amount_original":1.2, # "amount_outstanding":1.2, # "status":"closed", # "trades":[] # } # id = self.safe_string(order, 'id') datetimeString = self.safe_string(order, 'datetime') timestamp = None if datetimeString is not None: timestamp = self.parse8601(order['datetime'] + '+09:00') # Tokyo time amount = self.safe_number(order, 'amount_original') remaining = self.safe_number(order, 'amount_outstanding') price = self.safe_number(order, 'price') # status is set by fetchOrder method only status = self.parse_order_status(self.safe_string(order, 'status')) # fetchOrders do not return status, use heuristic if status is None: if remaining is not None and remaining == 0: status = 'closed' trades = None # todo: self.parse_trades(order['trades']) symbol = None if market is not None: symbol = market['symbol'] side = self.safe_string(order, 'type') return self.safe_order({ 'id': id, 'clientOrderId': None, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'lastTradeTimestamp': None, 'amount': amount, 'remaining': remaining, 'filled': None, 'side': side, 'type': None, 'timeInForce': None, 'postOnly': None, 'status': status, 'symbol': symbol, 'price': price, 'stopPrice': None, 'cost': None, 'trades': trades, 'fee': None, 'info': order, 'average': None, }) async def fetch_order(self, id, symbol=None, params={}): await self.load_markets() # a special case for btcbox – default symbol is BTC/JPY if symbol is None: symbol = 'BTC/JPY' market = self.market(symbol) request = self.extend({ 'id': id, 'coin': market['baseId'], }, params) response = await self.privatePostTradeView(self.extend(request, params)) return self.parse_order(response, market) async def fetch_orders_by_type(self, type, symbol=None, since=None, limit=None, params={}): await self.load_markets() # a special case for btcbox – default symbol is BTC/JPY if symbol is None: symbol = 'BTC/JPY' market = self.market(symbol) request = { 'type': type, # 'open' or 'all' 'coin': market['baseId'], } response = await self.privatePostTradeList(self.extend(request, params)) orders = self.parse_orders(response, market, since, limit) # status(open/closed/canceled) is None # btcbox does not return status, but we know it's 'open' as we queried for open orders if type == 'open': for i in range(0, len(orders)): orders[i]['status'] = 'open' return orders async def fetch_orders(self, symbol=None, since=None, limit=None, params={}): return await self.fetch_orders_by_type('all', symbol, since, limit, params) async def fetch_open_orders(self, symbol=None, since=None, limit=None, params={}): return await self.fetch_orders_by_type('open', symbol, since, limit, params) def nonce(self): return self.milliseconds() def sign(self, path, api='public', method='GET', params={}, headers=None, body=None): url = self.urls['api'] + '/' + self.version + '/' + path if api == 'public': if params: url += '?' + self.urlencode(params) else: self.check_required_credentials() nonce = str(self.nonce()) query = self.extend({ 'key': self.apiKey, 'nonce': nonce, }, params) request = self.urlencode(query) secret = self.hash(self.encode(self.secret)) query['signature'] = self.hmac(self.encode(request), self.encode(secret)) body = self.urlencode(query) headers = { 'Content-Type': 'application/x-www-form-urlencoded', } return {'url': url, 'method': method, 'body': body, 'headers': headers} def handle_errors(self, httpCode, reason, url, method, headers, body, response, requestHeaders, requestBody): if response is None: return # resort to defaultErrorHandler # typical error response: {"result":false,"code":"401"} if httpCode >= 400: return # resort to defaultErrorHandler result = self.safe_value(response, 'result') if result is None or result is True: return # either public API(no error codes expected) or success code = self.safe_value(response, 'code') feedback = self.id + ' ' + body self.throw_exactly_matched_exception(self.exceptions, code, feedback) raise ExchangeError(feedback) # unknown message async def request(self, path, api='public', method='GET', params={}, headers=None, body=None, config={}, context={}): response = await self.fetch2(path, api, method, params, headers, body, config, context) if isinstance(response, basestring): # sometimes the exchange returns whitespace prepended to json response = self.strip(response) if not self.is_json_encoded_object(response): raise ExchangeError(self.id + ' ' + response) response = json.loads(response) return response
39.65285
170
0.531948
dbc4bc7d40e97ad6c5ca88fbb4c028df75f0f9d1
6,860
py
Python
raasbot/api/views.py
kneeraazon01/raasbot
932648a6017cc5f888038ad3c54bb59c35deadf8
[ "Apache-2.0" ]
null
null
null
raasbot/api/views.py
kneeraazon01/raasbot
932648a6017cc5f888038ad3c54bb59c35deadf8
[ "Apache-2.0" ]
null
null
null
raasbot/api/views.py
kneeraazon01/raasbot
932648a6017cc5f888038ad3c54bb59c35deadf8
[ "Apache-2.0" ]
null
null
null
import json from django.http import JsonResponse from raasbot.settings import BASE_DIR from random import choice from os import path from core.models import ( User, Bot, ScrapedUser, UserMessageData, Message, Text, LogData ) def get_essentials(request): data = json.loads(request.body) token = data.get('token') user = User.objects.get(token=token) return (data, token, user) # Create your views here. def add_scraped_users(request): try: data, token, user = get_essentials(request) users = data.get('users') #for i in users: # a, b = ScrapedUser.objects.get_or_create(scrap_id = i['id'], name=i['username'], user=user, person_id = f"{i['id']}_{user.token}") # print(a, b) ScrapedUser.objects.bulk_create( [ScrapedUser(scrap_id = i['id'], name=i['username'], user=user, person_id = f"{i['id']}_{user.token}") for i in users], ignore_conflicts=True ) return JsonResponse({ "success": True, "message": "Users scraped and saved in database" }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def get_next_user(request): try: data, token, user = get_essentials(request) target = user.scraped_users.filter(texted=False).first() if not target: return JsonResponse({ 'success': True, 'continue': False }) target.texted=True target.save() return JsonResponse({ "success": True, "continue": True, "id": target.scrap_id, "username": target.name, "person_id": target.person_id }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def get_init_data(request): try: data, token, user = get_essentials(request) text_data = user.text_data.first() restricted_roles = text_data.roles.all() return JsonResponse({ "success": True, "guild_id": text_data.guild, "channel_id": text_data.channel, "invite_link": text_data.guild_invite, "restricted_roles": [role.name for role in restricted_roles] }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def get_text(request): try: data, token, user = get_essentials(request) message = user.dms.first() with open(path.join(BASE_DIR, 'texts.json'), 'r', encoding='utf-8') as texts: text = [] var = json.loads(texts.read()) text.append(choice(var.get('greetings'))) text.append(choice(var.get('relevance'))) text.append(choice(var.get("approach")).replace('<topic>', message.topic).replace('<invite>', message.guild_invite).replace('<info>', message.invite_info)) text.append(choice(var.get("link")).replace('<invite>', message.guild_invite).replace('<topic>', message.topic).replace('<info>', message.invite_info)) text.append(choice(var.get("info")).replace('<info>', message.invite_info).replace('<topic>', message.topic).replace('<invite>', message.guild_invite)) text.append(choice(var.get("ending"))) text.append(choice(var.get("conclusion"))) return JsonResponse({ "success": True, "text": text }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def get_tokens(request): try: data, token, user = get_essentials(request) bots = user.bots.filter(is_alive=True) return JsonResponse({ "success": True, "tokens": [bot.token for bot in bots] }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def should_scrape(request): try: data, token, user = get_essentials(request) count = user.scraped_users.filter(texted=False).count() return JsonResponse({ 'success': True, 'scrape': not (count > 1000) }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def log_result(request): try: data, token, user = get_essentials(request) target_id = data.get('target') target = ScrapedUser.objects.get(user=user, scrap_id=target_id) LogData.objects.create( user = user, target = target, success = data.get('success') ) return JsonResponse({ 'success': True }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def log_dup_result(request): try: data, token, user = get_essentials(request) target_id = data.get('target') target = ScrapedUser.objects.get(person_id=target_id) LogData.objects.create( user = user, target = target, success = data.get('success') ) return JsonResponse({ 'success': True }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def mark_dead(request): try: data, token, user = get_essentials(request) bot_token = data.get('bot_token') bot = Bot.objects.get(token=bot_token) bot.is_alive = False bot.save() return JsonResponse({ 'success': True }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" }) def add_bulk_token(request): try: data, token, user = get_essentials(request) tokens = data.get('tokens') Bot.objects.bulk_create( [Bot(token = bot, is_alive=True, user=user) for bot in tokens], ignore_conflicts=True ) return JsonResponse({ 'success': True }) except Exception as e: print(e) return JsonResponse({ "success": False, "message": f"{e}" })
31.46789
168
0.528134
947964009551475601d73b7ef519a2e87594ff86
39
py
Python
packages/jsii-rosetta/test/translations/calls/method_call.py
NGL321/jsii
a31ebf5ef676391d97f2286edc21e5859c38c96c
[ "Apache-2.0" ]
1,639
2019-07-05T07:21:00.000Z
2022-03-31T09:55:01.000Z
packages/jsii-rosetta/test/translations/calls/method_call.py
NGL321/jsii
a31ebf5ef676391d97f2286edc21e5859c38c96c
[ "Apache-2.0" ]
2,704
2019-07-01T23:10:28.000Z
2022-03-31T23:40:12.000Z
packages/jsii-rosetta/test/translations/calls/method_call.py
NGL321/jsii
a31ebf5ef676391d97f2286edc21e5859c38c96c
[ "Apache-2.0" ]
146
2019-07-02T14:36:25.000Z
2022-03-26T00:21:27.000Z
some_object.call_some_function(1, 2, 3)
39
39
0.820513
README.md exists but content is empty. Use the Edit dataset card button to edit it.
Downloads last month
70
Edit dataset card