idx int64 0 251k | question stringlengths 53 3.53k | target stringlengths 5 1.23k | len_question int64 20 893 | len_target int64 3 238 |
|---|---|---|---|---|
8,500 | def export_to_directory_crtomo ( self , directory , norrec = 'norrec' ) : exporter_crtomo . write_files_to_directory ( self . data , directory , norrec = norrec ) | Export the sEIT data into data files that can be read by CRTomo . | 51 | 18 |
8,501 | def export_to_crtomo_seit_manager ( self , grid ) : import crtomo g = self . data . groupby ( 'frequency' ) seit_data = { } for name , item in g : print ( name , item . shape , item . size ) if item . shape [ 0 ] > 0 : seit_data [ name ] = item [ [ 'a' , 'b' , 'm' , 'n' , 'r' , 'rpha' ] ] . values seit = crtomo . eitMa... | Return a ready - initialized seit - manager object from the CRTomo tools . This function only works if the crtomo_tools are installed . | 139 | 31 |
8,502 | def get_tape ( self , start = 0 , end = 10 ) : self . tape_start = start self . tape_end = end self . tape_length = end - start tmp = '\n' + "|" + str ( start ) + "| " for i in xrange ( len ( self . tape [ start : end ] ) ) : if i == self . cur_cell : tmp += "[" + str ( self . tape [ i ] ) + "] " else : tmp += ":" + st... | Pretty prints the tape values | 141 | 5 |
8,503 | def import_sip04 ( self , filename , timestep = None ) : df = reda_sip04 . import_sip04_data ( filename ) if timestep is not None : print ( 'adding timestep' ) df [ 'timestep' ] = timestep self . _add_to_container ( df ) print ( 'Summary:' ) self . _describe_data ( df ) | SIP04 data import | 93 | 5 |
8,504 | def check_dataframe ( self , dataframe ) : if dataframe is None : return None # is this a DataFrame if not isinstance ( dataframe , pd . DataFrame ) : raise Exception ( 'The provided dataframe object is not a pandas.DataFrame' ) for column in self . required_columns : if column not in dataframe : raise Exception ( 'Req... | Check the given dataframe for the required type and columns | 101 | 11 |
8,505 | def reduce_duplicate_frequencies ( self ) : group_keys = [ 'frequency' , ] if 'timestep' in self . data . columns : group_keys = group_keys + [ 'timestep' , ] g = self . data . groupby ( group_keys ) def group_apply ( item ) : y = item [ [ 'zt_1' , 'zt_2' , 'zt_3' ] ] . values . flatten ( ) zt_imag_std = np . std ( y .... | In case multiple frequencies were measured average them and compute std min max values for zt . | 455 | 18 |
8,506 | def _load_class ( class_path ) : parts = class_path . rsplit ( '.' , 1 ) module = __import__ ( parts [ 0 ] , fromlist = parts [ 1 ] ) return getattr ( module , parts [ 1 ] ) | Load the module and return the required class . | 55 | 9 |
8,507 | def rev_comp ( seq , molecule = 'dna' ) : if molecule == 'dna' : nuc_dict = { "A" : "T" , "B" : "V" , "C" : "G" , "D" : "H" , "G" : "C" , "H" : "D" , "K" : "M" , "M" : "K" , "N" : "N" , "R" : "Y" , "S" : "S" , "T" : "A" , "V" : "B" , "W" : "W" , "Y" : "R" } elif molecule == 'rna' : nuc_dict = { "A" : "U" , "B" : "V" , ... | DNA|RNA seq - > reverse complement | 360 | 8 |
8,508 | def from_json ( cls , key , scopes , subject = None ) : credentials_type = key [ 'type' ] if credentials_type != 'service_account' : raise ValueError ( 'key: expected type service_account ' '(got %s)' % credentials_type ) email = key [ 'client_email' ] key = OpenSSL . crypto . load_privatekey ( OpenSSL . crypto . FILET... | Alternate constructor intended for using JSON format of private key . | 126 | 12 |
8,509 | def from_pkcs12 ( cls , key , email , scopes , subject = None , passphrase = PKCS12_PASSPHRASE ) : key = OpenSSL . crypto . load_pkcs12 ( key , passphrase ) . get_privatekey ( ) return cls ( key = key , email = email , scopes = scopes , subject = subject ) | Alternate constructor intended for using . p12 files . | 84 | 11 |
8,510 | def issued_at ( self ) : issued_at = self . _issued_at if issued_at is None : self . _issued_at = int ( time . time ( ) ) return self . _issued_at | Time when access token was requested as seconds since epoch . | 47 | 11 |
8,511 | def access_token ( self ) : if ( self . _access_token is None or self . expiration_time <= int ( time . time ( ) ) ) : resp = self . make_access_request ( ) self . _access_token = resp . json ( ) [ 'access_token' ] return self . _access_token | Stores always valid OAuth2 access token . | 72 | 10 |
8,512 | def make_access_request ( self ) : del self . issued_at assertion = b'.' . join ( ( self . header ( ) , self . claims ( ) , self . signature ( ) ) ) post_data = { 'grant_type' : GRANT_TYPE , 'assertion' : assertion , } resp = requests . post ( AUDIENCE , post_data ) if resp . status_code != 200 : raise AuthenticationEr... | Makes an OAuth2 access token request with crafted JWT and signature . | 99 | 16 |
8,513 | def authorized_request ( self , method , url , * * kwargs ) : headers = kwargs . pop ( 'headers' , { } ) if headers . get ( 'Authorization' ) or kwargs . get ( 'auth' ) : raise ValueError ( "Found custom Authorization header, " "method call would override it." ) headers [ 'Authorization' ] = 'Bearer ' + self . access_t... | Shortcut for requests . request with proper Authorization header . | 111 | 11 |
8,514 | def import_txt ( filename , * * kwargs ) : # read in text file into a buffer with open ( filename , 'r' ) as fid : text = fid . read ( ) strings_to_replace = { 'Mixed / non conventional' : 'Mixed/non-conventional' , 'Date' : 'Date Time AM-PM' , } for key in strings_to_replace . keys ( ) : text = text . replace ( key , ... | Import Syscal measurements from a text file exported as Spreadsheet . | 403 | 14 |
8,515 | def import_bin ( filename , * * kwargs ) : metadata , data_raw = _import_bin ( filename ) skip_rows = kwargs . get ( 'skip_rows' , 0 ) if skip_rows > 0 : data_raw . drop ( data_raw . index [ range ( 0 , skip_rows ) ] , inplace = True ) data_raw = data_raw . reset_index ( ) if kwargs . get ( 'check_meas_nums' , True ) :... | Read a . bin file generated by the IRIS Instruments Syscal Pro System and return a curated dataframe for further processing . This dataframe contains only information currently deemed important . Use the function reda . importers . iris_syscal_pro_binary . _import_bin to extract ALL information from a given . bin file ... | 755 | 70 |
8,516 | def call_and_notificate ( args , opts ) : # store starttime stctime = time . clock ( ) stttime = time . time ( ) stdtime = datetime . datetime . now ( ) # call subprocess exit_code , output = call ( args ) # calculate delta cdelta = time . clock ( ) - stctime tdelta = time . time ( ) - stttime endtime = datetime . date... | Execute specified arguments and send notification email | 335 | 8 |
8,517 | def get_thumbnail_format ( self ) : if self . field . thumbnail_format : # Over-ride was given, use that instead. return self . field . thumbnail_format . lower ( ) else : # Use the existing extension from the file. filename_split = self . name . rsplit ( '.' , 1 ) return filename_split [ - 1 ] | Determines the target thumbnail type either by looking for a format override specified at the model level or by using the format the user uploaded . | 78 | 28 |
8,518 | def save ( self , name , content , save = True ) : super ( ImageWithThumbsFieldFile , self ) . save ( name , content , save ) try : self . generate_thumbs ( name , content ) except IOError , exc : if 'cannot identify' in exc . message or 'bad EPS header' in exc . message : raise UploadedImageIsUnreadableError ( "We wer... | Handles some extra logic to generate the thumbnails when the original file is uploaded . | 110 | 17 |
8,519 | def delete ( self , save = True ) : for thumb in self . field . thumbs : thumb_name , thumb_options = thumb thumb_filename = self . _calc_thumb_filename ( thumb_name ) self . storage . delete ( thumb_filename ) super ( ImageWithThumbsFieldFile , self ) . delete ( save ) | Deletes the original plus any thumbnails . Fails silently if there are errors deleting the thumbnails . | 73 | 21 |
8,520 | def dump_edn_val ( v ) : if isinstance ( v , ( str , unicode ) ) : return json . dumps ( v ) elif isinstance ( v , E ) : return unicode ( v ) else : return dumps ( v ) | edn simple value dump | 55 | 5 |
8,521 | def tx_schema ( self , * * kwargs ) : for s in self . schema . schema : tx = self . tx ( s , * * kwargs ) | Builds the data structure edn and puts it in the db | 38 | 13 |
8,522 | def tx ( self , * args , * * kwargs ) : if 0 == len ( args ) : return TX ( self ) ops = [ ] for op in args : if isinstance ( op , list ) : ops += op elif isinstance ( op , ( str , unicode ) ) : ops . append ( op ) if 'debug' in kwargs : pp ( ops ) tx_proc = "[ %s ]" % "" . join ( ops ) x = self . rest ( 'POST' , self .... | Executes a raw tx string or get a new TX object to work with . | 132 | 16 |
8,523 | def e ( self , eid ) : ta = datetime . datetime . now ( ) rs = self . rest ( 'GET' , self . uri_db + '-/entity' , data = { 'e' : int ( eid ) } , parse = True ) tb = datetime . datetime . now ( ) - ta print cl ( '<<< fetched entity %s in %sms' % ( eid , tb . microseconds / 1000.0 ) , 'cyan' ) return rs | Get an Entity | 114 | 3 |
8,524 | def retract ( self , e , a , v ) : ta = datetime . datetime . now ( ) ret = u"[:db/retract %i :%s %s]" % ( e , a , dump_edn_val ( v ) ) rs = self . tx ( ret ) tb = datetime . datetime . now ( ) - ta print cl ( '<<< retracted %s,%s,%s in %sms' % ( e , a , v , tb . microseconds / 1000.0 ) , 'cyan' ) return rs | redact the value of an attribute | 124 | 7 |
8,525 | def datoms ( self , index = 'aevt' , e = '' , a = '' , v = '' , limit = 0 , offset = 0 , chunk = 100 , start = '' , end = '' , since = '' , as_of = '' , history = '' , * * kwargs ) : assert index in [ 'aevt' , 'eavt' , 'avet' , 'vaet' ] , "non-existant index" data = { 'index' : index , 'a' : ':{0}' . format ( a ) if a ... | Returns a lazy generator that will only fetch groups of datoms at the chunk size specified . | 396 | 18 |
8,526 | def debug ( self , defn , args , kwargs , fmt = None , color = 'green' ) : ta = datetime . datetime . now ( ) rs = defn ( * args , * * kwargs ) tb = datetime . datetime . now ( ) - ta fmt = fmt or "processed {defn} in {ms}ms" logmsg = fmt . format ( ms = tb . microseconds / 1000.0 , defn = defn ) "terminal output" prin... | debug timing colored terminal output | 133 | 5 |
8,527 | def find ( self , * args , * * kwargs ) : return Query ( * args , db = self , schema = self . schema ) | new query builder on current db | 31 | 6 |
8,528 | def hashone ( self ) : rs = self . one ( ) if not rs : return { } else : finds = " " . join ( self . _find ) . split ( ' ' ) return dict ( zip ( ( x . replace ( '?' , '' ) for x in finds ) , rs ) ) | execute query get back | 65 | 4 |
8,529 | def all ( self ) : query , inputs = self . _toedn ( ) return self . db . q ( query , inputs = inputs , limit = self . _limit , offset = self . _offset , history = self . _history ) | execute query get all list of lists | 52 | 7 |
8,530 | def _toedn ( self ) : finds = u"" inputs = u"" wheres = u"" args = [ ] ": in and args" for a , b in self . _input : inputs += " {0}" . format ( a ) args . append ( dump_edn_val ( b ) ) if inputs : inputs = u":in ${0}" . format ( inputs ) " :where " for where in self . _where : if isinstance ( where , ( str , unicode ) ... | prepare the query for the rest api | 291 | 8 |
8,531 | def add ( self , * args , * * kwargs ) : assert self . resp is None , "Transaction already committed" entity , av_pairs , args = None , [ ] , list ( args ) if len ( args ) : if isinstance ( args [ 0 ] , ( int , long ) ) : " first arg is an entity or tempid" entity = E ( args [ 0 ] , tx = self ) elif isinstance ( args [... | Accumulate datums for the transaction | 565 | 8 |
8,532 | def resolve ( self ) : assert isinstance ( self . resp , dict ) , "Transaction in uncommitted or failed state" rids = [ ( v ) for k , v in self . resp [ 'tempids' ] . items ( ) ] self . txid = self . resp [ 'tx-data' ] [ 0 ] [ 'tx' ] rids . reverse ( ) for t in self . tmpents : pos = self . tmpents . index ( t ) t . _e... | Resolve one or more tempids . Automatically takes place after transaction is executed . | 140 | 17 |
8,533 | def get_usage ( self ) : resp = requests . get ( FITNESS_URL , timeout = 30 ) resp . raise_for_status ( ) soup = BeautifulSoup ( resp . text , "html5lib" ) eastern = pytz . timezone ( 'US/Eastern' ) output = [ ] for item in soup . findAll ( "div" , { "class" : "barChart" } ) : data = [ x . strip ( ) for x in item . get... | Get fitness locations and their current usage . | 307 | 8 |
8,534 | def search ( self , keyword ) : params = { "source" : "map" , "description" : keyword } data = self . _request ( ENDPOINTS [ 'SEARCH' ] , params ) data [ 'result_data' ] = [ res for res in data [ 'result_data' ] if isinstance ( res , dict ) ] return data | Return all buildings related to the provided query . | 79 | 9 |
8,535 | def compute_K_numerical ( dataframe , settings = None , keep_dir = None ) : inversion_code = reda . rcParams . get ( 'geom_factor.inversion_code' , 'crtomo' ) if inversion_code == 'crtomo' : import reda . utils . geom_fac_crtomo as geom_fac_crtomo if keep_dir is not None : keep_dir = os . path . abspath ( keep_dir ) K ... | Use a finite - element modeling code to infer geometric factors for meshes with topography or irregular electrode spacings . | 167 | 22 |
8,536 | def _get_object_key ( self , p_object ) : matched_key = None matched_index = None if hasattr ( p_object , self . _searchNames [ 0 ] ) : return getattr ( p_object , self . _searchNames [ 0 ] ) for x in xrange ( len ( self . _searchNames ) ) : key = self . _searchNames [ x ] if hasattr ( p_object , key ) : matched_key = ... | Get key from object | 168 | 4 |
8,537 | def correct ( self , temp , we_t ) : if not PIDTempComp . in_range ( temp ) : return None n_t = self . cf_t ( temp ) if n_t is None : return None we_c = we_t * n_t return we_c | Compute weC from weT | 63 | 7 |
8,538 | def compute_norrec_differences ( df , keys_diff ) : raise Exception ( 'This function is depreciated!' ) print ( 'computing normal-reciprocal differences' ) # df.sort_index(level='norrec') def norrec_diff ( x ) : """compute norrec_diff""" if x . shape [ 0 ] != 2 : return np . nan else : return np . abs ( x . iloc [ 1 ] ... | DO NOT USE ANY MORE - DEPRECIATED! | 282 | 11 |
8,539 | def _normalize_abmn ( abmn ) : abmn_2d = np . atleast_2d ( abmn ) abmn_normalized = np . hstack ( ( np . sort ( abmn_2d [ : , 0 : 2 ] , axis = 1 ) , np . sort ( abmn_2d [ : , 2 : 4 ] , axis = 1 ) , ) ) return abmn_normalized | return a normalized version of abmn | 94 | 7 |
8,540 | def assign_norrec_diffs ( df , diff_list ) : extra_dims = [ x for x in ( 'timestep' , 'frequency' , 'id' ) if x in df . columns ] g = df . groupby ( extra_dims ) def subrow ( row ) : if row . size == 2 : return row . iloc [ 1 ] - row . iloc [ 0 ] else : return np . nan for diffcol in diff_list : diff = g [ diffcol ] . ... | Compute and write the difference between normal and reciprocal values for all columns specified in the diff_list parameter . | 211 | 22 |
8,541 | def handle_authenticated_user ( self , response ) : current_user = get_user ( self . request ) ulogin , registered = ULoginUser . objects . get_or_create ( uid = response [ 'uid' ] , network = response [ 'network' ] , defaults = { 'identity' : response [ 'identity' ] , 'user' : current_user } ) if not registered : ulog... | Handles the ULogin response if user is already authenticated | 182 | 11 |
8,542 | def form_valid ( self , form ) : response = self . ulogin_response ( form . cleaned_data [ 'token' ] , self . request . get_host ( ) ) if 'error' in response : return render ( self . request , self . error_template_name , { 'json' : response } ) if user_is_authenticated ( get_user ( self . request ) ) : user , identity... | The request from ulogin service is correct | 192 | 8 |
8,543 | def ulogin_response ( self , token , host ) : response = requests . get ( settings . TOKEN_URL , params = { 'token' : token , 'host' : host } ) content = response . content if sys . version_info >= ( 3 , 0 ) : content = content . decode ( 'utf8' ) return json . loads ( content ) | Makes a request to ULOGIN | 78 | 8 |
8,544 | def initialise_parsimonious_states ( tree , feature , states ) : ps_feature_down = get_personalized_feature_name ( feature , BU_PARS_STATES ) ps_feature = get_personalized_feature_name ( feature , PARS_STATES ) all_states = set ( states ) for node in tree . traverse ( ) : state = getattr ( node , feature , set ( ) ) if... | Initializes the bottom - up state arrays for tips based on their states given by the feature . | 152 | 19 |
8,545 | def uppass ( tree , feature ) : ps_feature = get_personalized_feature_name ( feature , BU_PARS_STATES ) for node in tree . traverse ( 'postorder' ) : if not node . is_leaf ( ) : children_states = get_most_common_states ( getattr ( child , ps_feature ) for child in node . children ) node_states = getattr ( node , ps_fea... | UPPASS traverses the tree starting from the tips and going up till the root and assigns to each parent node a state based on the states of its child nodes . | 133 | 33 |
8,546 | def parsimonious_acr ( tree , character , prediction_method , states , num_nodes , num_tips ) : initialise_parsimonious_states ( tree , character , states ) uppass ( tree , character ) results = [ ] result = { STATES : states , NUM_NODES : num_nodes , NUM_TIPS : num_tips } logger = logging . getLogger ( 'pastml' ) def ... | Calculates parsimonious states on the tree and stores them in the corresponding feature . | 756 | 18 |
8,547 | def balance_to_ringchart_items ( balance , account = '' , show = SHOW_CREDIT ) : show = show if show else SHOW_CREDIT # cannot show all in ring chart rcis = [ ] for item in balance : subaccount = item [ 'account_fragment' ] if not account else ':' . join ( ( account , item [ 'account_fragment' ] ) ) ch = balance_to_rin... | Convert a balance data structure into RingChartItem objects . | 216 | 12 |
8,548 | def log_to_file ( log_path , log_urllib = False , limit = None ) : log_path = log_path file_handler = logging . FileHandler ( log_path ) if limit : file_handler = RotatingFileHandler ( log_path , mode = 'a' , maxBytes = limit * 1024 * 1024 , backupCount = 2 , encoding = None , delay = 0 ) fmt = '[%(asctime)s %(filename... | Add file_handler to logger | 217 | 6 |
8,549 | def session_context ( fn ) : @ functools . wraps ( fn ) def wrap ( * args , * * kwargs ) : session = args [ 0 ] . Session ( ) # obtain from self result = fn ( * args , session = session , * * kwargs ) session . close ( ) return result return wrap | Handles session setup and teardown | 70 | 8 |
8,550 | def _syscal_write_electrode_coords ( fid , spacing , N ) : fid . write ( '# X Y Z\n' ) for i in range ( 0 , N ) : fid . write ( '{0} {1} {2} {3}\n' . format ( i + 1 , i * spacing , 0 , 0 ) ) | helper function that writes out electrode positions to a file descriptor | 79 | 12 |
8,551 | def _syscal_write_quadpoles ( fid , quadpoles ) : fid . write ( '# A B M N\n' ) for nr , quadpole in enumerate ( quadpoles ) : fid . write ( '{0} {1} {2} {3} {4}\n' . format ( nr , quadpole [ 0 ] , quadpole [ 1 ] , quadpole [ 2 ] , quadpole [ 3 ] ) ) | helper function that writes the actual measurement configurations to a file descriptor . | 100 | 14 |
8,552 | def syscal_save_to_config_txt ( filename , configs , spacing = 1 ) : print ( 'Number of measurements: ' , configs . shape [ 0 ] ) number_of_electrodes = configs . max ( ) . astype ( int ) with open ( filename , 'w' ) as fid : _syscal_write_electrode_coords ( fid , spacing , number_of_electrodes ) _syscal_write_quadpole... | Write configurations to a Syscal ascii file that can be read by the Electre Pro program . | 118 | 22 |
8,553 | def setup ( use_latex = False , overwrite = False ) : # just make sure we can access matplotlib as mpl import matplotlib as mpl # general settings if overwrite : mpl . rcParams [ "lines.linewidth" ] = 2.0 mpl . rcParams [ "lines.markeredgewidth" ] = 3.0 mpl . rcParams [ "lines.markersize" ] = 3.0 mpl . rcParams [ "font... | Set up matplotlib imports and settings . | 292 | 9 |
8,554 | def load_seit_data ( directory , frequency_file = 'frequencies.dat' , data_prefix = 'volt_' , * * kwargs ) : frequencies = np . loadtxt ( directory + os . sep + frequency_file ) data_files = sorted ( glob ( directory + os . sep + data_prefix + '*' ) ) # check that the number of frequencies matches the number of data fi... | Load sEIT data from data directory . This function loads data previously exported from reda using reda . exporters . crtomo . write_files_to_directory | 189 | 36 |
8,555 | def get_diagonalisation ( frequencies , rate_matrix = None ) : Q = get_normalised_generator ( frequencies , rate_matrix ) d , A = np . linalg . eig ( Q ) return d , A , np . linalg . inv ( A ) | Normalises and diagonalises the rate matrix . | 64 | 9 |
8,556 | def get_normalised_generator ( frequencies , rate_matrix = None ) : if rate_matrix is None : n = len ( frequencies ) rate_matrix = np . ones ( shape = ( n , n ) , dtype = np . float64 ) - np . eye ( n ) generator = rate_matrix * frequencies generator -= np . diag ( generator . sum ( axis = 1 ) ) mu = - generator . diag... | Calculates the normalised generator from the rate matrix and character state frequencies . | 108 | 16 |
8,557 | def get_pij_matrix ( t , diag , A , A_inv ) : return A . dot ( np . diag ( np . exp ( diag * t ) ) ) . dot ( A_inv ) | Calculates the probability matrix of substitutions i - > j over time t given the normalised generator diagonalisation . | 49 | 24 |
8,558 | def split_arguments ( args ) : prev = False for i , value in enumerate ( args [ 1 : ] ) : if value . startswith ( '-' ) : prev = True elif prev : prev = False else : return args [ : i + 1 ] , args [ i + 1 : ] return args , [ ] | Split specified arguments to two list . | 71 | 7 |
8,559 | def parse_arguments ( args , config ) : import notify from conf import config_to_options opts = config_to_options ( config ) usage = ( "%(prog)s " "[-h] [-t TO_ADDR] [-f FROM_ADDR] [-e ENCODING] [-s SUBJECT]\n" " " "[-o HOST] [-p PORT] [--username USERNAME] [--password PASSWORD]\n" " " "[--setup] [--check] COMMAND ARGU... | Parse specified arguments via config | 663 | 6 |
8,560 | def should_require_authentication ( self , url ) : return ( not self . routes # require auth for all URLs or any ( route . match ( url ) for route in self . routes ) ) | Returns True if we should require authentication for the URL given | 42 | 11 |
8,561 | def authenticate ( self , environ ) : try : hd = parse_dict_header ( environ [ 'HTTP_AUTHORIZATION' ] ) except ( KeyError , ValueError ) : return False return self . credentials_valid ( hd [ 'response' ] , environ [ 'REQUEST_METHOD' ] , environ [ 'httpauth.uri' ] , hd [ 'nonce' ] , hd [ 'Digest username' ] , ) | Returns True if the credentials passed in the Authorization header are valid False otherwise . | 103 | 15 |
8,562 | def next ( self ) : try : return self . dict_to_xn ( self . csvreader . next ( ) ) except MetadataException : # row was metadata; proceed to next row return next ( self ) | Return the next transaction object . | 47 | 6 |
8,563 | def parse_date ( self , date ) : if self . date_format is not None : return datetime . datetime . strptime ( date , self . date_format ) . date ( ) if re . match ( '\d{8}$' , date ) : # assume YYYYMMDD return datetime . date ( * map ( int , ( date [ : 4 ] , date [ 4 : 6 ] , date [ 6 : ] ) ) ) try : # split by '-' or '/... | Parse the date and return a datetime object | 235 | 10 |
8,564 | def create ( self , uri , buffer = "queue" , interval = 10 ) : return self . _http_client . put_json ( "subscriptions/{}" . format ( self . short_name ) , { "subscription" : { "uri" : uri , "buffer" : buffer , "interval" : interval , } } ) | Create a subscription with this short name and the provided parameters | 78 | 11 |
8,565 | def read_pal_version ( ) : verfile = os . path . join ( "cextern" , "pal" , "configure.ac" ) verstring = "-1.-1.-1" for line in open ( verfile ) : if line . startswith ( "AC_INIT" ) : # Version will be in string [nn.mm.pp] match = re . search ( r"\[(\d+\.\d+\.\d+)\]" , line ) if match : verstring = match . group ( 1 ) ... | Scans the PAL configure . ac looking for the version number . | 148 | 13 |
8,566 | def _reset_model ( self , response ) : # pylint: disable=no-member # Reset the model to the initial state self . _provision_done = False # Set back the provision flag self . _changes . clear ( ) # Clear the changes # Process the raw data from the update response fields = self . process_raw_data ( response ) # Update th... | Update the fields value with the received information . | 106 | 9 |
8,567 | def is_ready ( self ) : if not self . provisioning_state : raise exception . ServiceException ( "The object doesn't contain " "`provisioningState`." ) elif self . provisioning_state == constant . FAILED : raise exception . ServiceException ( "Failed to complete the required operation." ) elif self . provisioning_state ... | Check if the current model is ready to be used . | 127 | 11 |
8,568 | def _get_all ( cls , parent_id = None , grandparent_id = None ) : client = cls . _get_client ( ) endpoint = cls . _endpoint . format ( resource_id = "" , parent_id = parent_id or "" , grandparent_id = grandparent_id or "" ) resources = [ ] while True : response = client . get_resource ( endpoint ) for raw_data in respo... | Retrives all the required resources . | 173 | 8 |
8,569 | def get ( cls , resource_id = None , parent_id = None , grandparent_id = None ) : if not resource_id : return cls . _get_all ( parent_id , grandparent_id ) else : return cls . _get ( resource_id , parent_id , grandparent_id ) | Retrieves the required resources . | 72 | 7 |
8,570 | def refresh ( self ) : client = self . _get_client ( ) endpoint = self . _endpoint . format ( resource_id = self . resource_id or "" , parent_id = self . parent_id or "" , grandparent_id = self . grandparent_id or "" ) response = client . get_resource ( endpoint ) self . _reset_model ( response ) | Get the latest representation of the current model . | 83 | 9 |
8,571 | def commit ( self , if_match = None , wait = True , timeout = None ) : if not self . _changes : LOG . debug ( "No changes available for %s: %s" , self . __class__ . __name__ , self . resource_id ) return LOG . debug ( "Apply all the changes on the current %s: %s" , self . __class__ . __name__ , self . resource_id ) cli... | Apply all the changes on the current model . | 337 | 9 |
8,572 | def _set_fields ( self , fields ) : super ( _BaseHNVModel , self ) . _set_fields ( fields ) if not self . resource_ref : endpoint = self . _endpoint . format ( resource_id = self . resource_id , parent_id = self . parent_id , grandparent_id = self . grandparent_id ) self . resource_ref = re . sub ( "(/networking/v[0-9]... | Set or update the fields value . | 106 | 7 |
8,573 | def get_resource ( self ) : references = { "resource_id" : None , "parent_id" : None , "grandparent_id" : None } for model_cls , regexp in self . _regexp . iteritems ( ) : match = regexp . search ( self . resource_ref ) if match is not None : references . update ( match . groupdict ( ) ) return model_cls . get ( * * re... | Return the associated resource . | 129 | 5 |
8,574 | def _get_nr_bins ( count ) : if count <= 30 : # use the square-root choice, used by Excel and Co k = np . ceil ( np . sqrt ( count ) ) else : # use Sturges' formula k = np . ceil ( np . log2 ( count ) ) + 1 return int ( k ) | depending on the number of data points compute a best guess for an optimal number of bins | 75 | 17 |
8,575 | def plot_histograms ( ertobj , keys , * * kwargs ) : # you can either provide a DataFrame or an ERT object if isinstance ( ertobj , pd . DataFrame ) : df = ertobj else : df = ertobj . data if df . shape [ 0 ] == 0 : raise Exception ( 'No data present, cannot plot' ) if isinstance ( keys , str ) : keys = [ keys , ] figu... | Generate histograms for one or more keys in the given container . | 717 | 14 |
8,576 | def plot_histograms_extra_dims ( dataobj , keys , * * kwargs ) : if isinstance ( dataobj , pd . DataFrame ) : df_raw = dataobj else : df_raw = dataobj . data if kwargs . get ( 'subquery' , False ) : df = df_raw . query ( kwargs . get ( 'subquery' ) ) else : df = df_raw split_timestamps = True if split_timestamps : grou... | Produce histograms grouped by the extra dimensions . | 784 | 10 |
8,577 | def parse_substring ( allele , pred , max_len = None ) : result = "" pos = 0 if max_len is None : max_len = len ( allele ) else : max_len = min ( max_len , len ( allele ) ) while pos < max_len and pred ( allele [ pos ] ) : result += allele [ pos ] pos += 1 return result , allele [ pos : ] | Extract substring of letters for which predicate is True | 87 | 11 |
8,578 | def fetch ( self ) : if not self . local_path : self . make_local_path ( ) fetcher = BookFetcher ( self ) fetcher . fetch ( ) | just pull files from PG | 39 | 5 |
8,579 | def make ( self ) : logger . debug ( "preparing to add all git files" ) num_added = self . local_repo . add_all_files ( ) if num_added : self . local_repo . commit ( "Initial import from Project Gutenberg" ) file_handler = NewFilesHandler ( self ) file_handler . add_new_files ( ) num_added = self . local_repo . add_all... | turn fetched files into a local repo make auxiliary files | 129 | 11 |
8,580 | def push ( self ) : self . github_repo . create_and_push ( ) self . _repo = self . github_repo . repo return self . _repo | create a github repo and push the local repo into it | 40 | 11 |
8,581 | def tag ( self , version = 'bump' , message = '' ) : self . clone_from_github ( ) self . github_repo . tag ( version , message = message ) | tag and commit | 41 | 3 |
8,582 | def format_title ( self ) : def asciify ( _title ) : _title = unicodedata . normalize ( 'NFD' , unicode ( _title ) ) ascii = True out = [ ] ok = u"1234567890qwertyuiopasdfghjklzxcvbnmQWERTYUIOPASDFGHJKLZXCVBNM- '," for ch in _title : if ch in ok : out . append ( ch ) elif unicodedata . category ( ch ) [ 0 ] == ( "L" ) ... | Takes a string and sanitizes it for Github s url name format | 407 | 15 |
8,583 | def _request ( self , path , method , body = None ) : url = '/' . join ( [ _SERVER , path ] ) ( resp , content ) = _HTTP . request ( url , method , headers = self . _headers , body = body ) content_type = resp . get ( 'content-type' ) if content_type and content_type . startswith ( 'application/json' ) : content = json... | Make a request from the API . | 113 | 7 |
8,584 | def put ( self , path , payload ) : body = json . dumps ( payload ) return self . _request ( path , 'PUT' , body ) | Make a PUT request from the API . | 32 | 9 |
8,585 | def post ( self , path , payload ) : body = json . dumps ( payload ) return self . _request ( path , 'POST' , body ) | Make a POST request from the API . | 32 | 8 |
8,586 | def create_child ( self , modules ) : binder = self . _binder . create_child ( ) return Injector ( modules , binder = binder , stage = self . _stage ) | Create a new injector that inherits the state from this injector . | 44 | 15 |
8,587 | def validate ( self , message , schema_name ) : err = None try : jsonschema . validate ( message , self . schemas [ schema_name ] ) except KeyError : msg = ( f'Schema "{schema_name}" was not found (available: ' f'{", ".join(self.schemas.keys())})' ) err = { 'msg' : msg } except jsonschema . ValidationError as e : msg =... | Validate a message given a schema . | 163 | 8 |
8,588 | def compose ( * functions ) : def inner ( func1 , func2 ) : return lambda * x , * * y : func1 ( func2 ( * x , * * y ) ) return functools . reduce ( inner , functions ) | evaluates functions from right to left . | 51 | 8 |
8,589 | def validate_instance ( instance ) : excludes = settings . AUTOMATED_LOGGING [ 'exclude' ] [ 'model' ] for excluded in excludes : if ( excluded in [ instance . _meta . app_label . lower ( ) , instance . __class__ . __name__ . lower ( ) ] or instance . __module__ . lower ( ) . startswith ( excluded ) ) : return False re... | Validating if the instance should be logged or is excluded | 90 | 11 |
8,590 | def get_current_user ( ) : thread_local = AutomatedLoggingMiddleware . thread_local if hasattr ( thread_local , 'current_user' ) : user = thread_local . current_user if isinstance ( user , AnonymousUser ) : user = None else : user = None return user | Get current user object from middleware | 67 | 7 |
8,591 | def get_current_environ ( ) : thread_local = AutomatedLoggingMiddleware . thread_local if hasattr ( thread_local , 'request_uri' ) : request_uri = thread_local . request_uri else : request_uri = None if hasattr ( thread_local , 'application' ) : application = thread_local . application application = Application . objec... | Get current application and path object from middleware | 161 | 9 |
8,592 | def processor ( status , sender , instance , updated = None , addition = '' ) : logger = logging . getLogger ( __name__ ) if validate_instance ( instance ) : user = get_current_user ( ) application = instance . _meta . app_label model_name = instance . __class__ . __name__ level = settings . AUTOMATED_LOGGING [ 'loglev... | This is the standard logging processor . | 234 | 7 |
8,593 | def parents ( self ) : parents = [ ] if self . parent is None : return [ ] category = self while category . parent is not None : parents . append ( category . parent ) category = category . parent return parents [ : : - 1 ] | Returns a list of all the current category s parents . | 52 | 11 |
8,594 | def root_parent ( self , category = None ) : return next ( filter ( lambda c : c . is_root , self . hierarchy ( ) ) ) | Returns the topmost parent of the current category . | 33 | 10 |
8,595 | def active ( self ) -> bool : states = self . _client . get_state ( self . _state_url ) [ 'states' ] for state in states : state = state [ 'State' ] if int ( state [ 'Id' ] ) == self . _state_id : # yes, the ZM API uses the *string* "1" for this... return state [ 'IsActive' ] == "1" return False | Indicate if this RunState is currently active . | 94 | 10 |
8,596 | def to_reasonable_unit ( value , units , round_digits = 2 ) : def to_unit ( unit ) : return float ( value ) / unit [ 1 ] exponents = [ abs ( Decimal ( to_unit ( u ) ) . adjusted ( ) - 1 ) for u in units ] best = min ( enumerate ( exponents ) , key = itemgetter ( 1 ) ) [ 0 ] return dict ( val = round ( to_unit ( units [... | Convert a value to the most reasonable unit . | 131 | 10 |
8,597 | def get_text ( self ) : done_units = to_reasonable_unit ( self . done , self . units ) current = round ( self . current / done_units [ 'multiplier' ] , 2 ) percent = int ( self . current * 100 / self . done ) return '{0:.2f} of {1:.2f} {2} ({3}%)' . format ( current , done_units [ 'val' ] , done_units [ 'label' ] , per... | Return extended progress bar text | 110 | 5 |
8,598 | def add_progress ( self , delta , done = None ) : if done is not None : self . done = done self . bar . current = max ( min ( self . done , self . current + delta ) , 0 ) self . rate_display . set_text ( self . rate_text ) self . remaining_time_display . set_text ( self . remaining_time_text ) return self . current == ... | Add to the current progress amount | 92 | 6 |
8,599 | async def valid_token_set ( self ) : is_valid = False if self . _auth_client . token : # Account for a token near expiration now = datetime . datetime . utcnow ( ) skew = datetime . timedelta ( seconds = 60 ) if self . _auth_client . expiry > ( now + skew ) : is_valid = True return is_valid | Check for validity of token and refresh if none or expired . | 85 | 12 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.