idx
int64
0
251k
question
stringlengths
53
3.53k
target
stringlengths
5
1.23k
len_question
int64
20
893
len_target
int64
3
238
241,300
def get_poll_option_formset ( self , formset_class ) : if self . request . forum_permission_handler . can_create_polls ( self . get_forum ( ) , self . request . user , ) : return formset_class ( * * self . get_poll_option_formset_kwargs ( ) )
Returns an instance of the poll option formset to be used in the view .
77
16
241,301
def get_poll_option_formset_kwargs ( self ) : kwargs = { 'prefix' : 'poll' , } if self . request . method in ( 'POST' , 'PUT' ) : kwargs . update ( { 'data' : self . request . POST , 'files' : self . request . FILES , } ) else : topic = self . get_topic ( ) poll_option_queryset = TopicPollOption . objects . filter ( po...
Returns the keyword arguments for instantiating the poll option formset .
137
13
241,302
def _remove_exts ( self , string ) : # If the user hasn't disabled it, we will detect image extensions, such as .png, .jpg, etc. # We'll remove them from the string before hashing. # This ensures that /Bear.png and /Bear.bmp will send back the same image, in different formats. if string . lower ( ) . endswith ( ( '.png...
Sets the string to create the Robohash
178
10
241,303
def _get_list_of_files ( self , path ) : chosen_files = [ ] # Get a list of all subdirectories directories = [ ] for root , dirs , files in natsort . natsorted ( os . walk ( path , topdown = False ) ) : for name in dirs : if name [ : 1 ] is not '.' : directories . append ( os . path . join ( root , name ) ) directories...
Go through each subdirectory of path and choose one file from each to use in our hash . Continue to increase self . iter so we use a different slot of randomness each time .
273
37
241,304
def collect_members ( module_to_name ) : members = { } for module , module_name in module_to_name . items ( ) : all_names = getattr ( module , "__all__" , None ) for name , member in inspect . getmembers ( module ) : if ( ( inspect . isfunction ( member ) or inspect . isclass ( member ) ) and not _always_drop_symbol_re...
Collect all symbols from a list of modules .
281
9
241,305
def _get_anchor ( module_to_name , fullname ) : if not _anchor_re . match ( fullname ) : raise ValueError ( "'%s' is not a valid anchor" % fullname ) anchor = fullname for module_name in module_to_name . values ( ) : if fullname . startswith ( module_name + "." ) : rest = fullname [ len ( module_name ) + 1 : ] # Use th...
Turn a full member name into an anchor .
128
9
241,306
def write_libraries ( dir , libraries ) : files = [ open ( os . path . join ( dir , k ) , "w" ) for k , _ in libraries ] # Document mentioned symbols for all libraries for f , ( _ , v ) in zip ( files , libraries ) : v . write_markdown_to_file ( f ) # Document symbols that no library mentioned. We do this after writing...
Write a list of libraries to disk .
133
8
241,307
def write_markdown_to_file ( self , f ) : print ( "---" , file = f ) print ( "---" , file = f ) print ( "<!-- This file is machine generated: DO NOT EDIT! -->" , file = f ) print ( "" , file = f ) print ( "# TensorFlow Python reference documentation" , file = f ) print ( "" , file = f ) fullname_f = lambda name : self ...
Writes this index to file f .
337
8
241,308
def _should_include_member ( self , name , member ) : # Always exclude symbols matching _always_drop_symbol_re. if _always_drop_symbol_re . match ( name ) : return False # Finally, exclude any specifically-excluded symbols. if name in self . _exclude_symbols : return False return True
Returns True if this member should be included in the document .
76
12
241,309
def get_imported_modules ( self , module ) : for name , member in inspect . getmembers ( module ) : if inspect . ismodule ( member ) : yield name , member
Returns the list of modules imported from module .
39
9
241,310
def get_class_members ( self , cls_name , cls ) : for name , member in inspect . getmembers ( cls ) : # Only show methods and properties presently. In Python 3, # methods register as isfunction. is_method = inspect . ismethod ( member ) or inspect . isfunction ( member ) if not ( is_method or isinstance ( member , prop...
Returns the list of class members to document in cls .
144
12
241,311
def _generate_signature_for_function ( self , func ) : args_list = [ ] argspec = inspect . getargspec ( func ) first_arg_with_default = ( len ( argspec . args or [ ] ) - len ( argspec . defaults or [ ] ) ) for arg in argspec . args [ : first_arg_with_default ] : if arg == "self" : # Python documentation typically skips...
Given a function returns a string representing its args .
351
10
241,312
def _remove_docstring_indent ( self , docstring ) : docstring = docstring or "" lines = docstring . strip ( ) . split ( "\n" ) min_indent = len ( docstring ) for l in lines [ 1 : ] : l = l . rstrip ( ) if l : i = 0 while i < len ( l ) and l [ i ] == " " : i += 1 if i < min_indent : min_indent = i for i in range ( 1 , l...
Remove indenting .
157
4
241,313
def _print_formatted_docstring ( self , docstring , f ) : lines = self . _remove_docstring_indent ( docstring ) # Output the lines, identifying "Args" and other section blocks. i = 0 def _at_start_of_section ( ) : """Returns the header if lines[i] is at start of a docstring section.""" l = lines [ i ] match = _section_...
Formats the given docstring as Markdown and prints it to f .
449
15
241,314
def _print_function ( self , f , prefix , fullname , func ) : heading = prefix + " `" + fullname if not isinstance ( func , property ) : heading += self . _generate_signature_for_function ( func ) heading += "` {#%s}" % _get_anchor ( self . _module_to_name , fullname ) print ( heading , file = f ) print ( "" , file = f...
Prints the given function to f .
128
8
241,315
def _write_member_markdown_to_file ( self , f , prefix , name , member ) : if ( inspect . isfunction ( member ) or inspect . ismethod ( member ) or isinstance ( member , property ) ) : print ( "- - -" , file = f ) print ( "" , file = f ) self . _print_function ( f , prefix , name , member ) print ( "" , file = f ) elif...
Print member to f .
227
5
241,316
def _write_class_markdown_to_file ( self , f , name , cls ) : # Build the list of class methods to document. methods = dict ( self . get_class_members ( name , cls ) ) # Used later to check if any methods were called out in the class # docstring. num_methods = len ( methods ) try : self . _write_docstring_markdown_to_f...
Write the class doc to f .
331
7
241,317
def write_markdown_to_file ( self , f ) : print ( "---" , file = f ) print ( "---" , file = f ) print ( "<!-- This file is machine generated: DO NOT EDIT! -->" , file = f ) print ( "" , file = f ) # TODO(touts): Do not insert these. Let the doc writer put them in # the module docstring explicitly. print ( "#" , self . ...
Prints this library to file f .
170
8
241,318
def write_other_members ( self , f , catch_all = False ) : if catch_all : names = self . _members . items ( ) else : names = inspect . getmembers ( self . _module ) leftovers = [ ] for name , _ in names : if name in self . _members and name not in self . _documented : leftovers . append ( name ) if leftovers : print ( ...
Writes the leftover members to f .
191
8
241,319
def assert_no_leftovers ( self ) : leftovers = [ ] for name in self . _members . keys ( ) : if name in self . _members and name not in self . _documented : leftovers . append ( name ) if leftovers : raise RuntimeError ( "%s: undocumented members: %s" % ( self . _title , ", " . join ( leftovers ) ) )
Generate an error if there are leftover members .
85
10
241,320
def start_http_server ( self , port , host = '0.0.0.0' , endpoint = None ) : if self . should_start_http_server ( ) : pc_start_http_server ( port , host , registry = self . registry )
Start an HTTP server for exposing the metrics if the should_start_http_server function says we should otherwise just return . Uses the implementation from prometheus_client rather than a Flask app .
59
39
241,321
def init_app ( self , app ) : if self . path : self . register_endpoint ( self . path , app ) if self . _export_defaults : self . export_defaults ( self . buckets , self . group_by , self . _defaults_prefix , app )
This callback can be used to initialize an application for the use with this prometheus reporter setup .
64
19
241,322
def register_endpoint ( self , path , app = None ) : if is_running_from_reloader ( ) and not os . environ . get ( 'DEBUG_METRICS' ) : return if app is None : app = self . app or current_app @ app . route ( path ) @ self . do_not_track ( ) def prometheus_metrics ( ) : # import these here so they don't clash with our own...
Register the metrics endpoint on the Flask application .
234
9
241,323
def start_http_server ( self , port , host = '0.0.0.0' , endpoint = '/metrics' ) : if is_running_from_reloader ( ) : return app = Flask ( 'prometheus-flask-exporter-%d' % port ) self . register_endpoint ( endpoint , app ) def run_app ( ) : app . run ( host = host , port = port ) thread = threading . Thread ( target = r...
Start an HTTP server for exposing the metrics . This will be an individual Flask application not the one registered with this class .
120
24
241,324
def histogram ( self , name , description , labels = None , * * kwargs ) : return self . _track ( Histogram , lambda metric , time : metric . observe ( time ) , kwargs , name , description , labels , registry = self . registry )
Use a Histogram to track the execution time and invocation count of the method .
58
16
241,325
def summary ( self , name , description , labels = None , * * kwargs ) : return self . _track ( Summary , lambda metric , time : metric . observe ( time ) , kwargs , name , description , labels , registry = self . registry )
Use a Summary to track the execution time and invocation count of the method .
56
15
241,326
def gauge ( self , name , description , labels = None , * * kwargs ) : return self . _track ( Gauge , lambda metric , time : metric . dec ( ) , kwargs , name , description , labels , registry = self . registry , before = lambda metric : metric . inc ( ) )
Use a Gauge to track the number of invocations in progress for the method .
67
17
241,327
def counter ( self , name , description , labels = None , * * kwargs ) : return self . _track ( Counter , lambda metric , time : metric . inc ( ) , kwargs , name , description , labels , registry = self . registry )
Use a Counter to track the total number of invocations of the method .
55
15
241,328
def _track ( metric_type , metric_call , metric_kwargs , name , description , labels , registry , before = None ) : if labels is not None and not isinstance ( labels , dict ) : raise TypeError ( 'labels needs to be a dictionary of {labelname: callable}' ) label_names = labels . keys ( ) if labels else tuple ( ) parent_...
Internal method decorator logic .
495
6
241,329
def do_not_track ( ) : def decorator ( f ) : @ functools . wraps ( f ) def func ( * args , * * kwargs ) : request . prom_do_not_track = True return f ( * args , * * kwargs ) return func return decorator
Decorator to skip the default metrics collection for the method .
66
13
241,330
def info ( self , name , description , labelnames = None , labelvalues = None , * * labels ) : if labels and labelnames : raise ValueError ( 'Cannot have labels defined as `dict` ' 'and collections of names and values' ) if labelnames is None and labels : labelnames = labels . keys ( ) elif labelnames and labelvalues :...
Report any information as a Prometheus metric . This will create a Gauge with the initial value of 1 .
149
21
241,331
def inve ( env , command , * args , * * kwargs ) : # we don't strictly need to restore the environment, since pew runs in # its own process, but it feels like the right thing to do with temp_environ ( ) : os . environ [ 'VIRTUAL_ENV' ] = str ( workon_home / env ) os . environ [ 'PATH' ] = compute_path ( env ) unsetenv ...
Run a command in the given virtual environment .
206
9
241,332
def ls_cmd ( argv ) : parser = argparse . ArgumentParser ( ) p_group = parser . add_mutually_exclusive_group ( ) p_group . add_argument ( '-b' , '--brief' , action = 'store_false' ) p_group . add_argument ( '-l' , '--long' , action = 'store_true' ) args = parser . parse_args ( argv ) lsvirtualenv ( args . long )
List available environments .
107
4
241,333
def workon_cmd ( argv ) : parser = argparse . ArgumentParser ( prog = 'pew workon' ) parser . add_argument ( 'envname' , nargs = '?' ) parser . add_argument ( '-n' , '--no-cd' , action = 'store_true' , help = ( 'Do not change working directory to project directory after ' 'activating virtualenv.' ) ) args = parser . pa...
List or change working virtual environments .
218
7
241,334
def add_cmd ( argv ) : parser = argparse . ArgumentParser ( ) parser . add_argument ( '-d' , dest = 'remove' , action = 'store_true' ) parser . add_argument ( 'dirs' , nargs = '+' ) args = parser . parse_args ( argv ) extra_paths = sitepackages_dir ( ) / '_virtualenv_path_extensions.pth' new_paths = [ os . path . abspa...
Add the specified directories to the Python path for the currently active virtualenv .
346
15
241,335
def lssitepackages_cmd ( argv ) : site = sitepackages_dir ( ) print ( * sorted ( site . iterdir ( ) ) , sep = os . linesep ) extra_paths = site / '_virtualenv_path_extensions.pth' if extra_paths . exists ( ) : print ( 'from _virtualenv_path_extensions.pth:' ) with extra_paths . open ( ) as extra : print ( '' . join ( e...
Show the content of the site - packages directory of the current virtualenv .
117
15
241,336
def toggleglobalsitepackages_cmd ( argv ) : quiet = argv == [ '-q' ] site = sitepackages_dir ( ) ngsp_file = site . parent / 'no-global-site-packages.txt' if ngsp_file . exists ( ) : ngsp_file . unlink ( ) if not quiet : print ( 'Enabled global site-packages' ) else : with ngsp_file . open ( 'w' ) : if not quiet : prin...
Toggle the current virtualenv between having and not having access to the global site - packages .
122
19
241,337
def cp_cmd ( argv ) : parser = argparse . ArgumentParser ( ) parser . add_argument ( 'source' ) parser . add_argument ( 'target' , nargs = '?' ) parser . add_argument ( '-d' , '--dont-activate' , action = 'store_false' , default = True , dest = 'activate' , help = "After \ creation, continue with the existing shell (do...
Duplicate the named virtualenv to make a new one .
144
13
241,338
def rename_cmd ( argv ) : parser = argparse . ArgumentParser ( ) parser . add_argument ( 'source' ) parser . add_argument ( 'target' ) pargs = parser . parse_args ( argv ) copy_virtualenv_project ( pargs . source , pargs . target ) return rmvirtualenvs ( [ pargs . source ] )
Rename a virtualenv
80
5
241,339
def setproject_cmd ( argv ) : args = dict ( enumerate ( argv ) ) project = os . path . abspath ( args . get ( 1 , '.' ) ) env = args . get ( 0 , os . environ . get ( 'VIRTUAL_ENV' ) ) if not env : sys . exit ( 'pew setproject [virtualenv] [project_path]' ) if not ( workon_home / env ) . exists ( ) : sys . exit ( "Envir...
Given a virtualenv directory and a project directory set the \ virtualenv up to be associated with the project .
160
22
241,340
def getproject_cmd ( argv ) : # Parse command line arguments parser = argparse . ArgumentParser ( description = "Print an environment's project directory." , ) parser . add_argument ( 'envname' , nargs = '?' , default = os . environ . get ( 'VIRTUAL_ENV' ) , help = ( 'The name of the environment to return the project d...
Print a virtualenv s project directory if set .
240
10
241,341
def mkproject_cmd ( argv ) : if '-l' in argv or '--list' in argv : templates = [ t . name [ 9 : ] for t in workon_home . glob ( "template_*" ) ] print ( "Available project templates:" , * templates , sep = '\n' ) return parser = mkvirtualenv_argparser ( ) parser . add_argument ( 'envname' ) parser . add_argument ( '-t'...
Create a new project directory and its associated virtualenv .
406
11
241,342
def mktmpenv_cmd ( argv ) : parser = mkvirtualenv_argparser ( ) env = '.' while ( workon_home / env ) . exists ( ) : env = hex ( random . getrandbits ( 64 ) ) [ 2 : - 1 ] args , rest = parser . parse_known_args ( argv ) mkvirtualenv ( env , args . python , args . packages , requirements = args . requirements , rest = r...
Create a temporary virtualenv .
146
6
241,343
def inall_cmd ( argv ) : envs = lsenvs ( ) errors = False for env in envs : print ( "\n%s:" % env ) try : inve ( env , * argv ) except CalledProcessError as e : errors = True err ( e ) sys . exit ( errors )
Run a command in each virtualenv .
68
8
241,344
def in_cmd ( argv ) : if len ( argv ) == 1 : return workon_cmd ( argv ) parse_envname ( argv , lambda : sys . exit ( 'You must provide a valid virtualenv to target' ) ) return inve ( * argv )
Run a command in the given virtualenv .
62
9
241,345
def restore_cmd ( argv ) : if len ( argv ) < 1 : sys . exit ( 'You must provide a valid virtualenv to target' ) env = argv [ 0 ] path = workon_home / env py = path / env_bin_dir / ( 'python.exe' if windows else 'python' ) exact_py = py . resolve ( ) . name return check_call ( [ sys . executable , "-m" , "virtualenv" , ...
Try to restore a broken virtualenv by reinstalling the same python version on top of it
124
19
241,346
def dir_cmd ( argv ) : env = parse_envname ( argv , lambda : sys . exit ( 'You must provide a valid virtualenv to target' ) ) print ( workon_home / env )
Print the path for the virtualenv directory
47
8
241,347
def install_cmd ( argv ) : installer = InstallCommand ( ) options , versions = installer . parser . parse_args ( argv ) if len ( versions ) != 1 : installer . parser . print_help ( ) sys . exit ( 1 ) else : try : actual_installer = PythonInstaller . get_installer ( versions [ 0 ] , options ) return actual_installer . i...
Use Pythonz to download and build the specified Python version
99
11
241,348
def version_cmd ( argv ) : import pkg_resources try : __version__ = pkg_resources . get_distribution ( 'pew' ) . version except pkg_resources . DistributionNotFound : __version__ = 'unknown' print ( 'Setuptools has some issues here, failed to get our own package.' , file = sys . stderr ) print ( __version__ )
Prints current pew version
87
6
241,349
def csstext_to_pairs ( csstext ) : # The lock is required to avoid ``cssutils`` concurrency # issues documented in issue #65 with csstext_to_pairs . _lock : return sorted ( [ ( prop . name . strip ( ) , format_value ( prop ) ) for prop in cssutils . parseStyle ( csstext ) ] , key = itemgetter ( 0 ) , )
csstext_to_pairs takes css text and make it to list of tuple of key value .
94
23
241,350
def merge_styles ( inline_style , new_styles , classes , remove_unset_properties = False ) : # building classes styles = OrderedDict ( [ ( "" , OrderedDict ( ) ) ] ) for pc in set ( classes ) : styles [ pc ] = OrderedDict ( ) for i , style in enumerate ( new_styles ) : for k , v in style : styles [ classes [ i ] ] [ k ...
This will merge all new styles where the order is important The last one will override the first When that is done it will apply old inline style again The old inline style is always important and override all new ones . The inline style must be valid .
448
49
241,351
def make_important ( bulk ) : return ";" . join ( "%s !important" % p if not p . endswith ( "!important" ) else p for p in bulk . split ( ";" ) )
makes every property in a string !important .
47
9
241,352
def capitalize_float_margin ( css_body ) : def _capitalize_property ( match ) : return "{0}:{1}{2}" . format ( match . group ( "property" ) . capitalize ( ) , match . group ( "value" ) , match . group ( "terminator" ) , ) return _lowercase_margin_float_rule . sub ( _capitalize_property , css_body )
Capitalize float and margin CSS property names
92
8
241,353
def _load_external ( self , url ) : if url . startswith ( "//" ) : # then we have to rely on the base_url if self . base_url and "https://" in self . base_url : url = "https:" + url else : url = "http:" + url if url . startswith ( "http://" ) or url . startswith ( "https://" ) : css_body = self . _load_external_url ( u...
loads an external stylesheet from a remote url or local path
244
12
241,354
def _css_rules_to_string ( self , rules ) : lines = [ ] for item in rules : if isinstance ( item , tuple ) : k , v = item lines . append ( "%s {%s}" % ( k , make_important ( v ) ) ) # media rule else : for rule in item . cssRules : if isinstance ( rule , ( cssutils . css . csscomment . CSSComment , cssutils . css . css...
given a list of css rules returns a css string
173
12
241,355
def check_workers ( self ) : if time . time ( ) - self . _worker_alive_time > 5 : self . _worker_alive_time = time . time ( ) # join processes if they are now gone, it should not do anything bad # if the process is still running [ worker . join ( ) for worker in self . _workers if not worker . is_alive ( ) ] self . _wo...
Kill workers that have been pending for a while and check if all workers are alive .
143
17
241,356
def kill_all ( self ) : while self . _num_workers > 0 and self . _worker_backend_socket . poll ( 1000 ) : msg = self . _worker_backend_socket . recv_pyobj ( ) self . _worker_backend_socket . send_pyobj ( None ) self . _num_workers -= 1 self . report ( f'Kill {msg[1:]}' ) # join all processes [ worker . join ( ) for wor...
Kill all workers
110
3
241,357
def _install ( self , name , autoinstall ) : import importlib import pkg_resources spam_spec = importlib . util . find_spec ( name ) reinstall = False if spam_spec is not None : if self . _version : mod = importlib . __import__ ( name ) if hasattr ( mod , '__version__' ) : ver = mod . __version__ else : try : ver = pkg...
Check existence of Python module and install it using command pip install if necessary .
730
15
241,358
def execute_task ( task_id , verbosity = None , runmode = 'run' , sigmode = None , monitor_interval = 5 , resource_monitor_interval = 60 ) : tf = TaskFile ( task_id ) # this will automatically create a pulse file tf . status = 'running' # write result file try : signal . signal ( signal . SIGTERM , signal_handler ) res...
Execute single or master task return a dictionary
442
9
241,359
def textMD5 ( text ) : m = hash_md5 ( ) if isinstance ( text , str ) : m . update ( text . encode ( ) ) else : m . update ( text ) return m . hexdigest ( )
Get md5 of a piece of text
51
8
241,360
def objectMD5 ( obj ) : if hasattr ( obj , 'target_name' ) : return obj . target_name ( ) try : return textMD5 ( pickle . dumps ( obj ) ) except : return ''
Get md5 of an object
48
6
241,361
def fileMD5 ( filename , partial = True ) : filesize = os . path . getsize ( filename ) # calculate md5 for specified file md5 = hash_md5 ( ) block_size = 2 ** 20 # buffer of 1M try : # 2**24 = 16M if ( not partial ) or filesize < 2 ** 24 : with open ( filename , 'rb' ) as f : while True : data = f . read ( block_size ...
Calculate partial MD5 basically the first and last 8M of the file for large files . This should signicicantly reduce the time spent on the creation and comparison of file signature when dealing with large bioinformat ics datasets .
226
49
241,362
def target_signature ( self ) : if self . exists ( ) : if not self . _md5 : self . _md5 = fileMD5 ( self ) return ( os . path . getmtime ( self ) , os . path . getsize ( self ) , self . _md5 ) elif ( self + '.zapped' ) . is_file ( ) : with open ( self + '.zapped' ) as sig : line = sig . readline ( ) _ , mtime , size , ...
Return file signature
169
3
241,363
def validate ( self , sig = None ) : if sig is not None : sig_mtime , sig_size , sig_md5 = sig else : try : with open ( self . sig_file ( ) ) as sig : sig_mtime , sig_size , sig_md5 = sig . read ( ) . strip ( ) . split ( ) except : return False if not self . exists ( ) : if ( self + '.zapped' ) . is_file ( ) : with ope...
Check if file matches its signature
200
6
241,364
def write_sig ( self ) : if not self . _md5 : self . _md5 = fileMD5 ( self ) with open ( self . sig_file ( ) , 'w' ) as sig : sig . write ( f'{os.path.getmtime(self)}\t{os.path.getsize(self)}\t{self._md5}' )
Write signature to sig store
86
5
241,365
def remove_targets ( self , type , kept = None ) : if kept is None : kept = [ i for i , x in enumerate ( self . _targets ) if not isinstance ( x , type ) ] if len ( kept ) == len ( self . _targets ) : return self self . _targets = [ self . _targets [ x ] for x in kept ] self . _labels = [ self . _labels [ x ] for x in ...
Remove targets of certain type
263
5
241,366
def resolve_remote ( self ) : for idx , target in enumerate ( self . _targets ) : if isinstance ( target , remote ) : resolved = target . resolve ( ) if isinstance ( resolved , str ) : resolved = interpolate ( resolved , env . sos_dict . dict ( ) ) self . _targets [ idx ] = file_target ( resolved ) . set ( * * target ....
If target is of remote type resolve it
97
8
241,367
def _handle_paired_with ( self , paired_with ) : if paired_with is None or not paired_with : var_name = [ ] var_value = [ ] elif isinstance ( paired_with , str ) : var_name = [ '_' + paired_with ] if paired_with not in env . sos_dict : raise ValueError ( f'Variable {paired_with} does not exist.' ) var_value = [ env . s...
Handle input option paired_with
359
6
241,368
def _handle_group_with ( self , group_with ) : if group_with is None or not group_with : var_name = [ ] var_value = [ ] elif isinstance ( group_with , str ) : var_name = [ '_' + group_with ] if group_with not in env . sos_dict : raise ValueError ( f'Variable {group_with} does not exist.' ) var_value = [ env . sos_dict ...
Handle input option group_with
346
6
241,369
def _handle_extract_pattern ( self , pattern ) : if pattern is None or not pattern : patterns = [ ] elif isinstance ( pattern , str ) : patterns = [ pattern ] elif isinstance ( pattern , Iterable ) : patterns = pattern else : raise ValueError ( f'Unacceptable value for parameter pattern: {pattern}' ) # for pattern in p...
Handle input option pattern
144
4
241,370
def write ( self ) : if not self . output_files . valid ( ) : raise ValueError ( f'Cannot write signature with undetermined output {self.output_files}' ) else : if 'TARGET' in env . config [ 'SOS_DEBUG' ] or 'ALL' in env . config [ 'SOS_DEBUG' ] : env . log_to_file ( 'TARGET' , f'write signature {self.sig_id} with outp...
Write signature file with signature of script input output and dependent files . Because local input and output files can only be determined after the execution of workflow . They are not part of the construction .
324
37
241,371
def clear_output ( output = None ) : for target in env . sos_dict [ '_output' ] if output is None else output : if isinstance ( target , file_target ) and target . exists ( ) : try : target . unlink ( ) except Exception as e : env . logger . warning ( f'Failed to remove {target}: {e}' )
Remove file targets in _output when a step fails to complete
82
12
241,372
def add_forward_workflow ( self , dag , sections , satisfies = None ) : dag . new_forward_workflow ( ) if 'DAG' in env . config [ 'SOS_DEBUG' ] or 'ALL' in env . config [ 'SOS_DEBUG' ] : env . log_to_file ( 'DAG' , f'Adding mini-workflow with {len(sections)} sections' ) default_input : sos_targets = sos_targets ( [ ] )...
Add a forward - workflow return number of nodes added
583
10
241,373
def initialize_dag ( self , targets : Optional [ List [ str ] ] = [ ] , nested : bool = False ) -> SoS_DAG : self . reset_dict ( ) dag = SoS_DAG ( name = self . md5 ) targets = sos_targets ( targets ) self . add_forward_workflow ( dag , self . workflow . sections ) # if self . resolve_dangling_targets ( dag , targets )...
Create a DAG by analyzing sections statically .
236
9
241,374
def short_repr ( obj , noneAsNA = False ) : if obj is None : return 'unspecified' if noneAsNA else 'None' elif isinstance ( obj , str ) and len ( obj ) > 80 : return '{}...{}' . format ( obj [ : 60 ] . replace ( '\n' , '\\n' ) , obj [ - 20 : ] . replace ( '\n' , '\\n' ) ) elif isinstance ( obj , ( str , int , float , b...
Return a short representation of obj for clarity .
518
9
241,375
def tail_of_file ( filename , n , ansi2html = False ) : avg_line_length = 74 to_read = n with open ( filename ) as f : while 1 : try : f . seek ( - ( avg_line_length * to_read ) , 2 ) except IOError : # woops. apparently file is smaller than what we want # to step back, go to the beginning instead f . seek ( 0 ) pos = ...
Reads a n lines from f with an offset of offset lines .
190
14
241,376
def sample_lines ( lines , n ) : if len ( lines ) <= n : return '' . join ( lines ) else : m = len ( lines ) return '' . join ( [ lines [ x * m // n + m // ( 2 * n ) ] for x in range ( n ) ] )
Draw a sample of n lines from filename largely evenly .
64
11
241,377
def set ( self , key , value ) : if hasattr ( value , 'labels' ) : if 'VARIABLE' in env . config [ 'SOS_DEBUG' ] or 'ALL' in env . config [ 'SOS_DEBUG' ] : env . log_to_file ( 'VARIABLE' , f"Set {key} to {short_repr(value)} with labels {short_repr(value.labels)}" ) else : if 'VARIABLE' in env . config [ 'SOS_DEBUG' ] o...
A short cut to set value to key without triggering any logging or warning message .
194
16
241,378
def update ( self , obj ) : self . _dict . update ( obj ) for k , v in obj . items ( ) : # if k.isupper(): # self._check_readonly(k, v) if env . verbosity > 2 : self . _log ( k , v )
Redefine update to trigger logging message
64
8
241,379
def execute_substep ( stmt , global_def , global_vars , task = '' , task_params = '' , proc_vars = { } , shared_vars = [ ] , config = { } ) : assert not env . zmq_context . closed assert 'workflow_id' in proc_vars assert 'step_id' in proc_vars assert '_input' in proc_vars assert '_output' in proc_vars assert '_depends'...
Execute a substep with specific input etc
383
9
241,380
def files ( self ) : try : cur = self . conn . cursor ( ) cur . execute ( 'SELECT id, item FROM workflows WHERE entry_type = "tracked_files"' ) return [ ( x [ 0 ] , eval ( x [ 1 ] ) ) for x in cur . fetchall ( ) ] except sqlite3 . DatabaseError as e : env . logger . warning ( f'Failed to get files from signature databa...
Listing files related to workflows related to current directory
102
11
241,381
def find_executable ( self ) : if 'DAG' in env . config [ 'SOS_DEBUG' ] or 'ALL' in env . config [ 'SOS_DEBUG' ] : env . log_to_file ( 'DAG' , 'find_executable' ) for node in self . nodes ( ) : # if it has not been executed if node . _status is None : with_dependency = False for edge in self . in_edges ( node ) : if ed...
Find an executable node which means nodes that has not been completed and has no input dependency .
384
18
241,382
def dangling ( self , targets : sos_targets ) : existing = [ ] missing = [ ] if env . config [ 'trace_existing' ] : for x in self . _all_depends_files . keys ( ) : if x not in self . _all_output_files : if x . target_exists ( ) : existing . append ( x ) else : missing . append ( x ) else : missing = [ x for x in self ....
returns 1 . missing targets which are missing from the DAG or from the provided targets 2 . existing targets of provided target list not in DAG
183
30
241,383
def subgraph_from ( self , targets : sos_targets ) : if 'DAG' in env . config [ 'SOS_DEBUG' ] or 'ALL' in env . config [ 'SOS_DEBUG' ] : env . log_to_file ( 'DAG' , 'create subgraph' ) # first, find all nodes with targets subnodes = [ ] for node in self . nodes ( ) : if node . _output_targets . valid ( ) and any ( x in...
Trim DAG to keep only nodes that produce targets
184
11
241,384
def build ( self ) : # right now we do not worry about status of nodes # connecting the output to the input of other nodes # # NOTE: This is implemented in the least efficient way just for # testing. It has to be re-implemented. # # refer to http://stackoverflow.com/questions/33494376/networkx-add-edges-to-graph-from-n...
Connect nodes according to status of targets
681
7
241,385
def monitor_tasks ( self , tasks = None , status = None , age = None ) : self . engine_ready . wait ( ) if not tasks : tasks = self . task_status . keys ( ) else : tasks = [ x for x in tasks if x in self . task_status ] # we only monitor running tasks with threading . Lock ( ) : for task in tasks : if self . task_statu...
Start monitoring specified or all tasks
335
6
241,386
def _submit_task_with_template ( self , task_ids ) : runtime = self . config runtime . update ( { 'workdir' : os . getcwd ( ) , 'cur_dir' : os . getcwd ( ) , # for backward compatibility 'verbosity' : env . verbosity , 'sig_mode' : env . config . get ( 'sig_mode' , 'default' ) , 'run_mode' : env . config . get ( 'run_m...
Submit tasks by interpolating a shell script defined in job_template
562
13
241,387
def is_type_hint ( stmt : str ) -> bool : if stmt . count ( '=' ) > 1 : return False if ':' not in stmt : return False # # action: if not stmt . split ( ':' ) [ 1 ] . strip ( ) : return False # # action: int # # or # # input: variable # if '=' not in stmt : action , par = [ x . strip ( ) for x in stmt . split ( ':' , 1...
Try to differentiate
394
3
241,388
def indented_script ( self ) -> bool : # get all leading space, tab and newline leading = INDENTED . match ( self . _script ) return 0 if leading is None else len ( leading . group ( 2 ) )
check self . _script and see if it is indented
50
12
241,389
def category ( self ) -> Optional [ str ] : if self . statements : if self . statements [ - 1 ] [ 0 ] == ':' : # a hack. ... to avoid calling isValid recursively def validDirective ( ) : if not self . values : return True if self . values [ - 1 ] . strip ( ) . endswith ( ',' ) : return False try : compile ( 'func(' + '...
Determine the category of existing statement
154
8
241,390
def isValid ( self ) -> bool : if not self . values : return True try : if self . category ( ) == 'directive' : # we add func() because the expression can be multi-line and # can have keyword-argument like options # # However, python considers # # func('value', ) # # a valid syntax but we do want , to continue to the n...
Determine if the statement expression or directive is valid . Otherwise the parser will continue until a valid multi - line expression or statement can be found .
393
30
241,391
def extend ( self , line : str ) -> None : if self . category ( ) == 'directive' : self . add_directive ( None , line ) elif self . category ( ) == 'script' : self . _script += line else : self . add_statement ( line )
Extend the current directive expression or script
63
8
241,392
def add_statement ( self , line : str , lineno : Optional [ int ] = None ) -> None : # there can be only one statement block if self . category ( ) != 'statements' : self . values = [ line ] else : self . values . append ( line ) if self . statements and self . statements [ - 1 ] [ 0 ] == '!' : self . statements [ - 1 ...
statements are regular python statements
119
6
241,393
def get_tokens ( self ) -> str : def _get_tokens ( statement ) : return [ x [ 1 ] for x in generate_tokens ( StringIO ( statement ) . readline ) if x [ 1 ] not in ( '' , '\n' ) ] tokens : List = [ ] for statement in self . statements : tokens . extend ( _get_tokens ( statement [ 2 ] if statement [ 0 ] == ':' else state...
Get tokens after input statement
135
5
241,394
def show ( self ) : textWidth = max ( 60 , shutil . get_terminal_size ( ( 80 , 20 ) ) . columns ) text = f' {self.step_name() + ":":<21} ' + self . comment print ( '\n' . join ( textwrap . wrap ( text , width = textWidth , initial_indent = '' , subsequent_indent = ' ' * 24 ) ) ) local_parameters = { x : y for x , y in ...
Output for command sos show
227
6
241,395
def extend ( self , workflow : 'SoS_Workflow' ) -> None : # all sections are simply appended ... # but we will need to make sure that the new workflow is # executed after the previous one. if not workflow . sections : return if not self . sections : self . sections = workflow . sections return section = workflow . sect...
Append another workflow to existing one to created a combined workflow
265
12
241,396
def add_comment ( self , line : str ) -> None : # the rule is like # # # comment line --> add to last comment # blank line --> clears last comment # [ ] --> use last comment # parameter: --> use last comment # All others: clear last comment self . _last_comment += ( ' ' if self . _last_comment else '' ) + line . lstrip...
Keeping track of last comment for section and parameter
91
9
241,397
def workflow ( self , workflow_name : Optional [ str ] = None , use_default : bool = True ) -> SoS_Workflow : if workflow_name is None and not use_default : return SoS_Workflow ( self . content , '' , '' , self . sections , self . global_stmts ) allowed_steps = None if not workflow_name : wf_name = '' else : # if consi...
Return a workflow with name_step + name_step specified in wf_name This function might be called recursively because of nested workflow .
526
30
241,398
def print_help ( self , script_name : str ) : textWidth = max ( 60 , shutil . get_terminal_size ( ( 80 , 20 ) ) . columns ) if len ( script_name ) > 20 : print ( f'usage: sos run {script_name}' ) print ( ' [workflow_name | -t targets] [options] [workflow_options]' ) else : print ( f'usage: sos run {script_name} [workfl...
print a help message from the script
438
7
241,399
def glob_wildcards ( pattern : str , files : Optional [ List [ str ] ] = None ) -> Dict [ str , Union [ List [ Any ] , List [ str ] ] ] : pattern = os . path . normpath ( pattern ) if sys . platform == 'win32' : # we perform path matching with / slash only pattern = pattern . replace ( '\\' , '/' ) first_wildcard = re ...
Glob the values of the wildcards by matching the given pattern to the filesystem . Returns a named tuple with a list of values for each wildcard .
326
31