idx
int64
0
63k
question
stringlengths
61
4.03k
target
stringlengths
6
1.23k
54,900
def draw ( self ) : str_wires = [ [ "-" ] * 7 * self . depth ] str_wires [ 0 ] [ 0 ] = "0" str_wires [ 0 ] [ 1 ] = " o" str_spaces = [ ] for i in range ( 1 , self . dimension ) : str_wires . append ( [ "-" ] * 7 * self . depth ) str_spaces . append ( [ " " ] * 7 * self . depth ) str_wires [ i ] [ 0 ] = str ( i ) str_wi...
Return an ASCII representation of the network .
54,901
def getWorkersName ( data ) : names = [ fichier for fichier in data . keys ( ) ] names . sort ( ) try : names . remove ( "broker" ) except ValueError : pass return names
Returns the list of the names of the workers sorted alphabetically
54,902
def importData ( directory ) : dataTask = OrderedDict ( ) dataQueue = OrderedDict ( ) for fichier in sorted ( os . listdir ( directory ) ) : try : with open ( "{directory}/{fichier}" . format ( ** locals ( ) ) , 'rb' ) as f : fileName , fileType = fichier . rsplit ( '-' , 1 ) if fileType == "QUEUE" : dataQueue [ fileNa...
Parse the input files and return two dictionnaries
54,903
def getTimes ( dataTasks ) : global begin_time start_time , end_time = float ( 'inf' ) , 0 for fichier , vals in dataTask . items ( ) : try : if hasattr ( vals , 'values' ) : tmp_start_time = min ( [ a [ 'start_time' ] for a in vals . values ( ) ] ) [ 0 ] if tmp_start_time < start_time : start_time = tmp_start_time tmp...
Get the start time and the end time of data in milliseconds
54,904
def WorkersDensity ( dataTasks ) : start_time , end_time = getTimes ( dataTasks ) graphdata = [ ] for name in getWorkersName ( dataTasks ) : vals = dataTasks [ name ] if hasattr ( vals , 'values' ) : workerdata = [ ] print ( "Plotting density map for {}" . format ( name ) ) try : for graphtime in timeRange ( start_time...
Return the worker density data for the graph .
54,905
def plotDensity ( dataTask , filename ) : def format_time ( x , pos = None ) : start_time , end_time = [ ( a - begin_time ) / 1000 for a in getTimes ( dataTask ) ] return int ( end_time * x / DENSITY_MAP_TIME_AXIS_LENGTH ) graphdata = WorkersDensity ( dataTask ) if len ( graphdata ) : fig = plt . figure ( ) ax = fig . ...
Plot the worker density graph
54,906
def plotBrokerQueue ( dataTask , filename ) : print ( "Plotting broker queue length for {0}." . format ( filename ) ) plt . figure ( ) plt . subplot ( 211 ) for fichier , vals in dataTask . items ( ) : if type ( vals ) == list : timestamps = list ( map ( datetime . fromtimestamp , map ( int , list ( zip ( * vals ) ) [ ...
Generates the broker queue length graphic .
54,907
def getWorkerInfo ( dataTask ) : workertime = [ ] workertasks = [ ] for fichier , vals in dataTask . items ( ) : if hasattr ( vals , 'values' ) : totaltime = sum ( [ a [ 'executionTime' ] for a in vals . values ( ) ] ) totaltasks = sum ( [ 1 for a in vals . values ( ) ] ) workertime . append ( totaltime ) workertasks ....
Returns the total execution time and task quantity by worker
54,908
def timelines ( fig , y , xstart , xstop , color = 'b' ) : fig . hlines ( y , xstart , xstop , color , lw = 4 ) fig . vlines ( xstart , y + 0.03 , y - 0.03 , color , lw = 2 ) fig . vlines ( xstop , y + 0.03 , y - 0.03 , color , lw = 2 )
Plot timelines at y from xstart to xstop with given color .
54,909
def plotTimeline ( dataTask , filename ) : fig = plt . figure ( ) ax = fig . gca ( ) worker_names = [ x for x in dataTask . keys ( ) if "broker" not in x ] min_time = getMinimumTime ( dataTask ) ystep = 1. / ( len ( worker_names ) + 1 ) y = 0 for worker , vals in dataTask . items ( ) : if "broker" in worker : continue ...
Build a timeline
54,910
def setWorker ( self , * args , ** kwargs ) : try : la = self . LAUNCHING_ARGUMENTS ( * args , ** kwargs ) except TypeError as e : scoop . logger . error ( ( "addWorker failed to convert args %s and kwargs %s " "to namedtuple (requires %s arguments (names %s)" ) % ( args , kwargs , len ( self . LAUNCHING_ARGUMENTS . _f...
Add a worker assignation Arguments and order to pass are defined in LAUNCHING_ARGUMENTS Using named args is advised .
54,911
def _WorkerCommand_environment ( self ) : worker = self . workersArguments c = [ ] if worker . prolog : c . extend ( [ "source" , worker . prolog , "&&" , ] ) if worker . pythonPath and not self . isLocal ( ) : c . extend ( [ "env" , "PYTHONPATH={0}:$PYTHONPATH" . format ( worker . pythonPath ) , ] ) elif worker . pyth...
Return list of shell commands to prepare the environment for bootstrap .
54,912
def _WorkerCommand_launcher ( self ) : return [ self . workersArguments . pythonExecutable , '-m' , 'scoop.launch.__main__' , str ( self . workerAmount ) , str ( self . workersArguments . verbose ) , ]
Return list commands to start the bootstrap process
54,913
def _WorkerCommand_options ( self ) : worker = self . workersArguments c = [ ] if self . hostname == worker . brokerHostname : broker = "127.0.0.1" else : broker = worker . brokerHostname if worker . nice is not None : c . extend ( [ '--nice' , str ( worker . nice ) ] ) c . extend ( [ '--size' , str ( worker . size ) ]...
Return list of options for bootstrap
54,914
def _WorkerCommand_executable ( self ) : worker = self . workersArguments c = [ ] if worker . executable : c . append ( worker . executable ) if worker . args : if self . isLocal ( ) : c . extend ( [ '{0}' . format ( a ) for a in worker . args ] ) else : c . extend ( [ '"{0}"' . format ( a . replace ( '"' , '\\\"' ) ) ...
Return executable and any options to be executed by bootstrap
54,915
def _getWorkerCommandList ( self ) : c = [ ] c . extend ( self . _WorkerCommand_environment ( ) ) c . extend ( self . _WorkerCommand_launcher ( ) ) c . extend ( self . _WorkerCommand_options ( ) ) c . extend ( self . _WorkerCommand_executable ( ) ) return c
Generate the workerCommand as list
54,916
def launch ( self , tunnelPorts = None ) : if self . isLocal ( ) : c = self . _getWorkerCommandList ( ) self . subprocesses . append ( subprocess . Popen ( c ) ) else : BASE_SSH [ 0 ] = self . ssh_executable sshCmd = BASE_SSH if not self . rsh else BASE_RSH if tunnelPorts is not None : sshCmd += [ '-R {0}:127.0.0.1:{0}...
Launch every worker assigned on this host .
54,917
def _switch ( self , future ) : scoop . _control . current = self assert self . greenlet is not None , ( "No greenlet to switch to:" "\n{0}" . format ( self . __dict__ ) ) return self . greenlet . switch ( future )
Switch greenlet .
54,918
def cancel ( self ) : if self in scoop . _control . execQueue . movable : self . exceptionValue = CancelledError ( ) scoop . _control . futureDict [ self . id ] . _delete ( ) scoop . _control . execQueue . remove ( self ) return True return False
If the call is currently being executed or sent for remote execution then it cannot be cancelled and the method will return False otherwise the call will be cancelled and the method will return True .
54,919
def done ( self ) : try : scoop . _control . execQueue . remove ( self ) scoop . _control . execQueue . socket . sendFuture ( self ) except ValueError as e : pass scoop . _control . execQueue . updateQueue ( ) return self . _ended ( )
Returns True if the call was successfully cancelled or finished running False otherwise . This function updates the executionQueue so it receives all the awaiting message .
54,920
def add_done_callback ( self , callable_ , inCallbackType = CallbackType . standard , inCallbackGroup = None ) : self . callback . append ( callbackEntry ( callable_ , inCallbackType , inCallbackGroup ) ) if self . _ended ( ) : self . callback [ - 1 ] . func ( self )
Attach a callable to the future that will be called when the future is cancelled or finishes running . Callable will be called with the future as its only argument .
54,921
def append ( self , future ) : if future . _ended ( ) and future . index is None : self . inprogress . add ( future ) elif future . _ended ( ) and future . index is not None : self . ready . append ( future ) elif future . greenlet is not None : self . inprogress . add ( future ) else : self . movable . append ( future...
Append a future to the queue .
54,922
def askForPreviousFutures ( self ) : if time . time ( ) < self . lastStatus + POLLING_TIME / 1000 : return self . lastStatus = time . time ( ) for future in scoop . _control . futureDict . values ( ) : if scoop . IS_ORIGIN and future . id == ( scoop . worker , 0 ) : continue if future not in self . inprogress : self . ...
Request a status for every future to the broker .
54,923
def pop ( self ) : self . updateQueue ( ) if self . timelen ( self ) < self . lowwatermark : self . requestFuture ( ) if len ( self . ready ) != 0 : return self . ready . popleft ( ) elif len ( self . movable ) != 0 : return self . movable . popleft ( ) else : self . lastStatus = time . time ( ) while len ( self ) == 0...
Pop the next future from the queue ; in progress futures have priority over those that have not yet started ; higher level futures have priority over lower level ones ;
54,924
def flush ( self ) : for elem in self : if elem . id [ 0 ] != scoop . worker : elem . _delete ( ) self . socket . sendFuture ( elem ) self . ready . clear ( ) self . movable . clear ( )
Empty the local queue and send its elements to be executed remotely .
54,925
def updateQueue ( self ) : for future in self . socket . recvFuture ( ) : if future . _ended ( ) : try : thisFuture = scoop . _control . futureDict [ future . id ] except KeyError : scoop . logger . warn ( '{0}: Received an unexpected future: ' '{1}' . format ( scoop . worker , future . id ) ) continue thisFuture . res...
Process inbound communication buffer . Updates the local queue with elements from the broker .
54,926
def sendResult ( self , future ) : future . greenlet = None assert future . _ended ( ) , "The results are not valid" self . socket . sendResult ( future )
Send back results to broker for distribution to parent task .
54,927
def shutdown ( self ) : self . socket . shutdown ( ) if scoop : if scoop . DEBUG : from scoop import _debug _debug . writeWorkerDebug ( scoop . _control . debug_stats , scoop . _control . QueueLength , )
Shutdown the ressources used by the queue
54,928
def redirectSTDOUTtoDebugFile ( ) : import sys kwargs = { } if sys . version_info >= ( 3 , ) : kwargs [ "encoding" ] = "utf8" sys . stdout = open ( os . path . join ( getDebugDirectory ( ) , "{0}.stdout" . format ( getDebugIdentifier ( ) ) , ) , "w" , 1 , ** kwargs ) sys . stderr = open ( os . path . join ( getDebugDir...
Redirects the stdout and stderr of the current process to a file .
54,929
def writeWorkerDebug ( debugStats , queueLength , path_suffix = "" ) : createDirectory ( path_suffix ) origin_prefix = "origin-" if scoop . IS_ORIGIN else "" statsFilename = os . path . join ( getDebugDirectory ( ) , path_suffix , "{1}worker-{0}-STATS" . format ( getDebugIdentifier ( ) , origin_prefix ) ) lengthFilenam...
Serialize the execution data using pickle and writes it into the debug directory .
54,930
def main ( ) : parser = makeParser ( ) args = parser . parse_args ( ) hosts = utils . getHosts ( args . hostfile , args . hosts ) if args . n : n = args . n else : n = utils . getWorkerQte ( hosts ) assert n >= 0 , ( "Scoop couldn't determine the number of worker to start.\n" "Use the '-n' flag to set it manually." ) i...
Execution of the SCOOP module . Parses its command - line arguments and launch needed resources .
54,931
def initLogging ( self ) : verbose_levels = { 0 : logging . WARNING , 1 : logging . INFO , 2 : logging . DEBUG , } logging . basicConfig ( level = verbose_levels [ self . verbose ] , format = "[%(asctime)-15s] %(module)-9s %(levelname)-7s %(message)s" ) return logging . getLogger ( self . __class__ . __name__ )
Configures the logger .
54,932
def divideHosts ( self , hosts , qty ) : maximumWorkers = sum ( host [ 1 ] for host in hosts ) if qty > maximumWorkers : index = 0 while qty > maximumWorkers : hosts [ index ] = ( hosts [ index ] [ 0 ] , hosts [ index ] [ 1 ] + 1 ) index = ( index + 1 ) % len ( hosts ) maximumWorkers += 1 elif qty < maximumWorkers : wh...
Divide processes among hosts .
54,933
def showHostDivision ( self , headless ) : scoop . logger . info ( 'Worker d--istribution: ' ) for worker , number in self . worker_hosts : first_worker = ( worker == self . worker_hosts [ 0 ] [ 0 ] ) scoop . logger . info ( ' {0}:\t{1} {2}' . format ( worker , number - 1 if first_worker or headless else str ( number...
Show the worker distribution over the hosts .
54,934
def setWorkerInfo ( self , hostname , workerAmount , origin ) : scoop . logger . debug ( 'Initialising {0}{1} worker {2} [{3}].' . format ( "local" if hostname in utils . localHostnames else "remote" , " origin" if origin else "" , self . workersLeft , hostname , ) ) add_args , add_kwargs = self . _setWorker_args ( ori...
Sets the worker information for the current host .
54,935
def close ( self ) : if self . debug : time . sleep ( 10 ) for host in self . workers : host . close ( ) for broker in self . brokers : try : broker . close ( ) except AttributeError : pass scoop . logger . info ( 'Finished cleaning spawned subprocesses.' )
Subprocess cleanup .
54,936
def processConfig ( self , worker_config ) : self . config [ 'headless' ] |= worker_config . get ( "headless" , False ) if self . config [ 'headless' ] : if not self . discovery_thread : self . discovery_thread = discovery . Advertise ( port = "," . join ( str ( a ) for a in self . getPorts ( ) ) , )
Update the pool configuration with a worker configuration .
54,937
def main ( self ) : if self . args is None : self . parse ( ) self . log = utils . initLogging ( self . verbose ) if self . args . workingDirectory : os . chdir ( self . args . workingDirectory ) if not self . args . brokerHostname : self . log . info ( "Discovering SCOOP Brokers on network..." ) pools = discovery . Se...
Bootstrap an arbitrary script . If no agruments were passed use discovery module to search and connect to a broker .
54,938
def makeParser ( self ) : self . parser = argparse . ArgumentParser ( description = 'Starts the executable.' , prog = ( "{0} -m scoop.bootstrap" ) . format ( sys . executable ) ) self . parser . add_argument ( '--origin' , help = "To specify that the worker is the origin" , action = 'store_true' ) self . parser . add_a...
Generate the argparse parser object containing the bootloader accepted parameters
54,939
def parse ( self ) : if self . parser is None : self . makeParser ( ) self . args = self . parser . parse_args ( ) self . verbose = self . args . verbose
Generate a argparse parser and parse the command - line arguments
54,940
def setScoop ( self ) : scoop . IS_RUNNING = True scoop . IS_ORIGIN = self . args . origin scoop . BROKER = BrokerInfo ( self . args . brokerHostname , self . args . taskPort , self . args . metaPort , self . args . externalBrokerHostname if self . args . externalBrokerHostname else self . args . brokerHostname , ) sco...
Setup the SCOOP constants .
54,941
def myFunc ( parameter ) : print ( 'Hello World from {0}!' . format ( scoop . worker ) ) print ( shared . getConst ( 'myVar' ) [ 2 ] ) return parameter + 1
This function will be executed on the remote host even if it was not available at launch .
54,942
def sendResult ( self , future ) : future = copy . copy ( future ) future . callable = future . args = future . kargs = future . greenlet = None if not future . sendResultBack : future . resultValue = None self . _sendReply ( future . id . worker , pickle . dumps ( future , pickle . HIGHEST_PROTOCOL , ) , )
Send a terminated future back to its parent .
54,943
def getSize ( string ) : try : with urllib . request . urlopen ( string , None , 1 ) as f : return sum ( len ( line ) for line in f ) except ( urllib . error . URLError , socket . timeout ) as e : return 0
This functions opens a web sites and then calculate the total size of the page in bytes . This is for the sake of the example . Do not use this technique in real code as it is not a very bright way to do this .
54,944
def getValue ( words ) : value = 0 for word in words : for letter in word : value += shared . getConst ( 'lettersValue' ) [ letter ] return value
Computes the sum of the values of the words .
54,945
def _run_module_code ( code , init_globals = None , mod_name = None , mod_fname = None , mod_loader = None , pkg_name = None ) : with _ModifiedArgv0 ( mod_fname ) : with _TempModule ( mod_name ) as temp_module : mod_globals = temp_module . module . __dict__ _run_code ( code , mod_globals , init_globals , mod_name , mod...
Helper to run code in new namespace with sys modified
54,946
def run_module ( mod_name , init_globals = None , run_name = None , alter_sys = False ) : mod_name , loader , code , fname = _get_module_details ( mod_name ) if run_name is None : run_name = mod_name pkg_name = mod_name . rpartition ( '.' ) [ 0 ] if alter_sys : return _run_module_code ( code , init_globals , run_name ,...
Execute a module s code without importing it
54,947
def _get_importer ( path_name ) : cache = sys . path_importer_cache try : importer = cache [ path_name ] except KeyError : cache [ path_name ] = None for hook in sys . path_hooks : try : importer = hook ( path_name ) break except ImportError : pass else : try : importer = imp . NullImporter ( path_name ) except ImportE...
Python version of PyImport_GetImporter C API function
54,948
def run_path ( path_name , init_globals = None , run_name = None ) : if run_name is None : run_name = "<run_path>" importer = _get_importer ( path_name ) if isinstance ( importer , imp . NullImporter ) : code = _get_code_from_file ( path_name ) return _run_module_code ( code , init_globals , run_name , path_name ) else...
Execute code located at the specified filesystem location
54,949
def maxTreeDepthDivide ( rootValue , currentDepth = 0 , parallelLevel = 2 ) : thisRoot = shared . getConst ( 'myTree' ) . search ( rootValue ) if currentDepth >= parallelLevel : return thisRoot . maxDepth ( currentDepth ) else : if not any ( [ thisRoot . left , thisRoot . right ] ) : return currentDepth if not all ( [ ...
Finds a tree node that represents rootValue and computes the max depth of this tree branch . This function will emit new futures until currentDepth = parallelLevel
54,950
def insert ( self , value ) : if not self . payload or value == self . payload : self . payload = value else : if value <= self . payload : if self . left : self . left . insert ( value ) else : self . left = BinaryTreeNode ( value ) else : if self . right : self . right . insert ( value ) else : self . right = BinaryT...
Insert a value in the tree
54,951
def maxDepth ( self , currentDepth = 0 ) : if not any ( ( self . left , self . right ) ) : return currentDepth result = 0 for child in ( self . left , self . right ) : if child : result = max ( result , child . maxDepth ( currentDepth + 1 ) ) return result
Compute the depth of the longest branch of the tree
54,952
def search ( self , value ) : if self . payload == value : return self else : if value <= self . payload : if self . left : return self . left . search ( value ) else : if self . right : return self . right . search ( value ) return None
Find an element in the tree
54,953
def createZMQSocket ( self , sock_type ) : sock = self . ZMQcontext . socket ( sock_type ) sock . setsockopt ( zmq . LINGER , LINGER_TIME ) sock . setsockopt ( zmq . IPV4ONLY , 0 ) sock . setsockopt ( zmq . SNDHWM , 0 ) sock . setsockopt ( zmq . RCVHWM , 0 ) try : sock . setsockopt ( zmq . IMMEDIATE , 1 ) except : pass...
Create a socket of the given sock_type and deactivate message dropping
54,954
def _reportFutures ( self ) : try : while True : time . sleep ( scoop . TIME_BETWEEN_STATUS_REPORTS ) fids = set ( x . id for x in scoop . _control . execQueue . movable ) fids . update ( set ( x . id for x in scoop . _control . execQueue . ready ) ) fids . update ( set ( x . id for x in scoop . _control . execQueue . ...
Sends futures status updates to broker at intervals of scoop . TIME_BETWEEN_STATUS_REPORTS seconds . Is intended to be run by a separate thread .
54,955
def _sendReply ( self , destination , fid , * args ) : self . addPeer ( destination ) try : self . direct_socket . send_multipart ( [ destination , REPLY , ] + list ( args ) , flags = zmq . NOBLOCK ) except zmq . error . ZMQError as e : scoop . logger . debug ( "{0}: Could not send result directly to peer {1}, routing ...
Send a REPLY directly to its destination . If it doesn t work launch it back to the broker .
54,956
def _startup ( rootFuture , * args , ** kargs ) : import greenlet global _controller _controller = greenlet . greenlet ( control . runController ) try : result = _controller . switch ( rootFuture , * args , ** kargs ) except scoop . _comm . Shutdown : result = None control . execQueue . shutdown ( ) return result
Initializes the SCOOP environment .
54,957
def _recursiveReduce ( mapFunc , reductionFunc , scan , * iterables ) : if iterables : half = min ( len ( x ) // 2 for x in iterables ) data_left = [ list ( x ) [ : half ] for x in iterables ] data_right = [ list ( x ) [ half : ] for x in iterables ] else : data_left = data_right = [ [ ] ] out_futures = [ None , None ]...
Generates the recursive reduction tree . Used by mapReduce .
54,958
def _createFuture ( func , * args , ** kwargs ) : assert callable ( func ) , ( "The provided func parameter is not a callable." ) if scoop . IS_ORIGIN and "SCOOP_WORKER" not in sys . modules : sys . modules [ "SCOOP_WORKER" ] = sys . modules [ "__main__" ] lambdaType = type ( lambda : None ) funcIsLambda = isinstance (...
Helper function to create a future .
54,959
def _waitAny ( * children ) : n = len ( children ) for index , future in enumerate ( children ) : if future . exceptionValue : raise future . exceptionValue if future . _ended ( ) : future . _delete ( ) yield future n -= 1 else : future . index = index future = control . current while n > 0 : future . stopWatch . halt ...
Waits on any child Future created by the calling Future .
54,960
def wait ( fs , timeout = - 1 , return_when = ALL_COMPLETED ) : DoneAndNotDoneFutures = namedtuple ( 'DoneAndNotDoneFutures' , 'done not_done' ) if timeout < 0 : if return_when == FIRST_COMPLETED : next ( _waitAny ( * fs ) ) elif return_when in [ ALL_COMPLETED , FIRST_EXCEPTION ] : for _ in _waitAll ( * fs ) : pass don...
Wait for the futures in the given sequence to complete . Using this function may prevent a worker from executing .
54,961
def advertiseBrokerWorkerDown ( exctype , value , traceback ) : if not scoop . SHUTDOWN_REQUESTED : execQueue . shutdown ( ) sys . __excepthook__ ( exctype , value , traceback )
Hook advertizing the broker if an impromptu shutdown is occuring .
54,962
def delFutureById ( futureId , parentId ) : try : del futureDict [ futureId ] except KeyError : pass try : toDel = [ a for a in futureDict [ parentId ] . children if a . id == futureId ] for f in toDel : del futureDict [ parentId ] . children [ f ] except KeyError : pass
Delete future on id basis
54,963
def delFuture ( afuture ) : try : del futureDict [ afuture . id ] except KeyError : pass try : del futureDict [ afuture . parentId ] . children [ afuture ] except KeyError : pass
Delete future afuture
54,964
def runFuture ( future ) : global debug_stats global QueueLength if scoop . DEBUG : init_debug ( ) debug_stats [ future . id ] [ 'start_time' ] . append ( time . time ( ) ) future . waitTime = future . stopWatch . get ( ) future . stopWatch . reset ( ) try : uniqueReference = [ cb . groupID for cb in future . callback ...
Callable greenlet in charge of running tasks .
54,965
def runController ( callable_ , * args , ** kargs ) : global execQueue rootId = ( - 1 , 0 ) if execQueue is None : execQueue = FutureQueue ( ) sys . excepthook = advertiseBrokerWorkerDown if scoop . DEBUG : from scoop import _debug _debug . redirectSTDOUTtoDebugFile ( ) headless = scoop . CONFIGURATION . get ( "headles...
Callable greenlet implementing controller logic .
54,966
def mode ( self ) : mu = self . mean ( ) sigma = self . std ( ) ret_val = math . exp ( mu - sigma ** 2 ) if math . isnan ( ret_val ) : ret_val = float ( "inf" ) return ret_val
Computes the mode of a log - normal distribution built with the stats data .
54,967
def median ( self ) : mu = self . mean ( ) ret_val = math . exp ( mu ) if math . isnan ( ret_val ) : ret_val = float ( "inf" ) return ret_val
Computes the median of a log - normal distribution built with the stats data .
54,968
def _decode_string ( buf , pos ) : for i in range ( pos , len ( buf ) ) : if buf [ i : i + 1 ] == _compat_bytes ( '\x00' ) : try : return ( buf [ pos : i ] . decode ( _CHARSET ) , i + 1 ) except UnicodeDecodeError : raise MinusconfError ( 'Not a valid ' + _CHARSET + ' string: ' + repr ( buf [ pos : i ] ) ) raise Minusc...
Decodes a string in the buffer buf starting at position pos . Returns a tupel of the read string and the next byte to read .
54,969
def _find_sock ( ) : if socket . has_ipv6 : try : return socket . socket ( socket . AF_INET6 , socket . SOCK_DGRAM ) except socket . gaierror : pass return socket . socket ( socket . AF_INET , socket . SOCK_DGRAM )
Create a UDP socket
54,970
def _compat_inet_pton ( family , addr ) : if family == socket . AF_INET : res = _compat_bytes ( '' ) parts = addr . split ( '.' ) if len ( parts ) != 4 : raise ValueError ( 'Expected 4 dot-separated numbers' ) for part in parts : intval = int ( part , 10 ) if intval < 0 or intval > 0xff : raise ValueError ( "Invalid in...
socket . inet_pton for platforms that don t have it
54,971
def start_blocking ( self ) : self . _cav_started . clear ( ) self . start ( ) self . _cav_started . wait ( )
Start the advertiser in the background but wait until it is ready
54,972
def _send_queries ( self ) : res = 0 addrs = _resolve_addrs ( self . addresses , self . port , self . ignore_senderrors , [ self . _sock . family ] ) for addr in addrs : try : self . _send_query ( addr [ 1 ] ) res += 1 except : if not self . ignore_senderrors : raise return res
Sends queries to multiple addresses . Returns the number of successful queries .
54,973
def clean ( self ) : raise forms . ValidationError ( self . error_messages [ 'invalid_login' ] , code = 'invalid_login' , params = { 'username' : self . username_field . verbose_name } )
Always raise the default error message because we don t care what they entered here .
54,974
def types ( ** requirements ) : def predicate ( args ) : for name , kind in sorted ( requirements . items ( ) ) : assert hasattr ( args , name ) , "missing required argument `%s`" % name if not isinstance ( kind , tuple ) : kind = ( kind , ) if not any ( isinstance ( getattr ( args , name ) , k ) for k in kind ) : retu...
Specify a precondition based on the types of the function s arguments .
54,975
def ensure ( arg1 , arg2 = None ) : assert ( isinstance ( arg1 , str ) and isfunction ( arg2 ) ) or ( isfunction ( arg1 ) and arg2 is None ) description = "" predicate = lambda x : x if isinstance ( arg1 , str ) : description = arg1 predicate = arg2 else : description = get_function_source ( arg1 ) predicate = arg1 ret...
Specify a precondition described by description and tested by predicate .
54,976
def invariant ( arg1 , arg2 = None ) : desc = "" predicate = lambda x : x if isinstance ( arg1 , str ) : desc = arg1 predicate = arg2 else : desc = get_function_source ( arg1 ) predicate = arg1 def invariant ( c ) : def check ( name , func ) : exceptions = ( "__getitem__" , "__setitem__" , "__lt__" , "__le__" , "__eq__...
Specify a class invariant described by description and tested by predicate .
54,977
def mkpassword ( length = 16 , chars = None , punctuation = None ) : if chars is None : chars = string . ascii_letters + string . digits data = [ random . choice ( chars ) for _ in range ( length ) ] if punctuation : data = data [ : - punctuation ] for _ in range ( punctuation ) : data . append ( random . choice ( PUNC...
Generates a random ascii string - useful to generate authinfos
54,978
def disk_check_size ( ctx , param , value ) : if value : if isinstance ( value , tuple ) : val = value [ 1 ] else : val = value if val % 1024 : raise click . ClickException ( 'Size must be a multiple of 1024.' ) return value
Validation callback for disk size parameter .
54,979
def create ( cls , fqdn , flags , algorithm , public_key ) : fqdn = fqdn . lower ( ) params = { 'flags' : flags , 'algorithm' : algorithm , 'public_key' : public_key , } result = cls . call ( 'domain.dnssec.create' , fqdn , params ) return result
Create a dnssec key .
54,980
def from_name ( cls , name ) : snps = cls . list ( { 'name' : name } ) if len ( snps ) == 1 : return snps [ 0 ] [ 'id' ] elif not snps : return raise DuplicateResults ( 'snapshot profile name %s is ambiguous.' % name )
Retrieve a snapshot profile accsociated to a name .
54,981
def list ( cls , options = None , target = None ) : options = options or { } result = [ ] if not target or target == 'paas' : for profile in cls . safe_call ( 'paas.snapshotprofile.list' , options ) : profile [ 'target' ] = 'paas' result . append ( ( profile [ 'id' ] , profile ) ) if not target or target == 'vm' : for ...
List all snapshot profiles .
54,982
def records ( cls , fqdn , sort_by = None , text = False ) : meta = cls . get_fqdn_info ( fqdn ) url = meta [ 'domain_records_href' ] kwargs = { } if text : kwargs = { 'headers' : { 'Accept' : 'text/plain' } } return cls . json_get ( cls . get_sort_url ( url , sort_by ) , ** kwargs )
Display records information about a domain .
54,983
def add_record ( cls , fqdn , name , type , value , ttl ) : data = { "rrset_name" : name , "rrset_type" : type , "rrset_values" : value , } if ttl : data [ 'rrset_ttl' ] = int ( ttl ) meta = cls . get_fqdn_info ( fqdn ) url = meta [ 'domain_records_href' ] return cls . json_post ( url , data = json . dumps ( data ) )
Create record for a domain .
54,984
def update_record ( cls , fqdn , name , type , value , ttl , content ) : data = { "rrset_name" : name , "rrset_type" : type , "rrset_values" : value , } if ttl : data [ 'rrset_ttl' ] = int ( ttl ) meta = cls . get_fqdn_info ( fqdn ) if content : url = meta [ 'domain_records_href' ] kwargs = { 'headers' : { 'Content-Typ...
Update all records for a domain .
54,985
def del_record ( cls , fqdn , name , type ) : meta = cls . get_fqdn_info ( fqdn ) url = meta [ 'domain_records_href' ] delete_url = url if name : delete_url = '%s/%s' % ( delete_url , name ) if type : delete_url = '%s/%s' % ( delete_url , type ) return cls . json_delete ( delete_url )
Delete record for a domain .
54,986
def keys ( cls , fqdn , sort_by = None ) : meta = cls . get_fqdn_info ( fqdn ) url = meta [ 'domain_keys_href' ] return cls . json_get ( cls . get_sort_url ( url , sort_by ) )
Display keys information about a domain .
54,987
def keys_info ( cls , fqdn , key ) : return cls . json_get ( '%s/domains/%s/keys/%s' % ( cls . api_url , fqdn , key ) )
Retrieve key information .
54,988
def keys_create ( cls , fqdn , flag ) : data = { "flags" : flag , } meta = cls . get_fqdn_info ( fqdn ) url = meta [ 'domain_keys_href' ] ret , headers = cls . json_post ( url , data = json . dumps ( data ) , return_header = True ) return cls . json_get ( headers [ 'location' ] )
Create new key entry for a domain .
54,989
def list ( gandi , datacenter , id , subnet , gateway ) : output_keys = [ 'name' , 'state' , 'dc' ] if id : output_keys . append ( 'id' ) if subnet : output_keys . append ( 'subnet' ) if gateway : output_keys . append ( 'gateway' ) datacenters = gandi . datacenter . list ( ) vlans = gandi . vlan . list ( datacenter ) f...
List vlans .
54,990
def info ( gandi , resource , ip ) : output_keys = [ 'name' , 'state' , 'dc' , 'subnet' , 'gateway' ] datacenters = gandi . datacenter . list ( ) vlan = gandi . vlan . info ( resource ) gateway = vlan [ 'gateway' ] if not ip : output_vlan ( gandi , vlan , datacenters , output_keys , justify = 11 ) return vlan gateway_e...
Display information about a vlan .
54,991
def create ( gandi , name , datacenter , subnet , gateway , background ) : try : gandi . datacenter . is_opened ( datacenter , 'iaas' ) except DatacenterLimited as exc : gandi . echo ( '/!\ Datacenter %s will be closed on %s, ' 'please consider using another datacenter.' % ( datacenter , exc . date ) ) result = gandi ....
Create a new vlan
54,992
def update ( gandi , resource , name , gateway , create , bandwidth ) : params = { } if name : params [ 'name' ] = name vlan_id = gandi . vlan . usable_id ( resource ) try : if gateway : IP ( gateway ) params [ 'gateway' ] = gateway except ValueError : vm = gandi . iaas . info ( gateway ) ips = [ ip for sublist in [ [ ...
Update a vlan
54,993
def list_migration_choice ( cls , datacenter ) : datacenter_id = cls . usable_id ( datacenter ) dc_list = cls . list ( ) available_dcs = [ dc for dc in dc_list if dc [ 'id' ] == datacenter_id ] [ 0 ] [ 'can_migrate_to' ] choices = [ dc for dc in dc_list if dc [ 'id' ] in available_dcs ] return choices
List available datacenters for migration from given datacenter .
54,994
def is_opened ( cls , dc_code , type_ ) : options = { 'dc_code' : dc_code , '%s_opened' % type_ : True } datacenters = cls . safe_call ( 'hosting.datacenter.list' , options ) if not datacenters : options = { 'iso' : dc_code , '%s_opened' % type_ : True } datacenters = cls . safe_call ( 'hosting.datacenter.list' , optio...
List opened datacenters for given type .
54,995
def filtered_list ( cls , name = None , obj = None ) : options = { } if name : options [ 'id' ] = cls . usable_id ( name ) def obj_ok ( dc , obj ) : if not obj or obj [ 'datacenter_id' ] == dc [ 'id' ] : return True return False return [ x for x in cls . list ( options ) if obj_ok ( x , obj ) ]
List datacenters matching name and compatible with obj
54,996
def from_iso ( cls , iso ) : result = cls . list ( { 'sort_by' : 'id ASC' } ) dc_isos = { } for dc in result : if dc [ 'iso' ] not in dc_isos : dc_isos [ dc [ 'iso' ] ] = dc [ 'id' ] return dc_isos . get ( iso )
Retrieve the first datacenter id associated to an ISO .
54,997
def from_name ( cls , name ) : result = cls . list ( ) dc_names = { } for dc in result : dc_names [ dc [ 'name' ] ] = dc [ 'id' ] return dc_names . get ( name )
Retrieve datacenter id associated to a name .
54,998
def from_country ( cls , country ) : result = cls . list ( { 'sort_by' : 'id ASC' } ) dc_countries = { } for dc in result : if dc [ 'country' ] not in dc_countries : dc_countries [ dc [ 'country' ] ] = dc [ 'id' ] return dc_countries . get ( country )
Retrieve the first datacenter id associated to a country .
54,999
def from_dc_code ( cls , dc_code ) : result = cls . list ( ) dc_codes = { } for dc in result : if dc . get ( 'dc_code' ) : dc_codes [ dc [ 'dc_code' ] ] = dc [ 'id' ] return dc_codes . get ( dc_code )
Retrieve the datacenter id associated to a dc_code