idx
int64
0
251k
question
stringlengths
53
3.53k
target
stringlengths
5
1.23k
len_question
int64
20
893
len_target
int64
3
238
243,800
def tar_open ( f ) : if isinstance ( f , six . string_types ) : return tarfile . open ( name = f ) else : return tarfile . open ( fileobj = f )
Open either a filename or a file - like object as a TarFile .
44
15
243,801
def copy_from_server_to_local ( dataset_remote_dir , dataset_local_dir , remote_fname , local_fname ) : log . debug ( "Copying file `{}` to a local directory `{}`." . format ( remote_fname , dataset_local_dir ) ) head , tail = os . path . split ( local_fname ) head += os . path . sep if not os . path . exists ( head ) ...
Copies a remote file locally .
514
7
243,802
def convert_to_one_hot ( y ) : max_value = max ( y ) min_value = min ( y ) length = len ( y ) one_hot = numpy . zeros ( ( length , ( max_value - min_value + 1 ) ) ) one_hot [ numpy . arange ( length ) , y ] = 1 return one_hot
converts y into one hot reprsentation .
81
10
243,803
def convert_binarized_mnist ( directory , output_directory , output_filename = 'binarized_mnist.hdf5' ) : output_path = os . path . join ( output_directory , output_filename ) h5file = h5py . File ( output_path , mode = 'w' ) train_set = numpy . loadtxt ( os . path . join ( directory , TRAIN_FILE ) ) . reshape ( ( - 1 ...
Converts the binarized MNIST dataset to HDF5 .
340
14
243,804
def fill_subparser ( subparser ) : url = 'http://www.cs.toronto.edu/~kriz/cifar-10-python.tar.gz' filename = 'cifar-10-python.tar.gz' subparser . set_defaults ( urls = [ url ] , filenames = [ filename ] ) return default_downloader
Sets up a subparser to download the CIFAR - 10 dataset file .
83
17
243,805
def convert_celeba_aligned_cropped ( directory , output_directory , output_filename = OUTPUT_FILENAME ) : output_path = os . path . join ( output_directory , output_filename ) h5file = _initialize_conversion ( directory , output_path , ( 218 , 178 ) ) features_dataset = h5file [ 'features' ] image_file_path = os . path...
Converts the aligned and cropped CelebA dataset to HDF5 .
257
14
243,806
def convert_celeba ( which_format , directory , output_directory , output_filename = None ) : if which_format not in ( 'aligned_cropped' , '64' ) : raise ValueError ( "CelebA format needs to be either " "'aligned_cropped' or '64'." ) if not output_filename : output_filename = 'celeba_{}.hdf5' . format ( which_format ) ...
Converts the CelebA dataset to HDF5 .
146
11
243,807
def disk_usage ( path ) : st = os . statvfs ( path ) total = st . f_blocks * st . f_frsize used = ( st . f_blocks - st . f_bfree ) * st . f_frsize return total , used
Return free usage about the given path in bytes .
61
10
243,808
def safe_mkdir ( folder_name , force_perm = None ) : if os . path . exists ( folder_name ) : return intermediary_folders = folder_name . split ( os . path . sep ) # Remove invalid elements from intermediary_folders if intermediary_folders [ - 1 ] == "" : intermediary_folders = intermediary_folders [ : - 1 ] if force_pe...
Create the specified folder .
221
5
243,809
def check_enough_space ( dataset_local_dir , remote_fname , local_fname , max_disk_usage = 0.9 ) : storage_need = os . path . getsize ( remote_fname ) storage_total , storage_used = disk_usage ( dataset_local_dir ) # Instead of only looking if there's enough space, we ensure we do not # go over max disk usage level to ...
Check if the given local folder has enough space .
123
10
243,810
def convert_cifar100 ( directory , output_directory , output_filename = 'cifar100.hdf5' ) : output_path = os . path . join ( output_directory , output_filename ) h5file = h5py . File ( output_path , mode = "w" ) input_file = os . path . join ( directory , 'cifar-100-python.tar.gz' ) tar_file = tarfile . open ( input_fi...
Converts the CIFAR - 100 dataset to HDF5 .
807
14
243,811
def verify_axis_labels ( self , expected , actual , source_name ) : if not getattr ( self , '_checked_axis_labels' , False ) : self . _checked_axis_labels = defaultdict ( bool ) if not self . _checked_axis_labels [ source_name ] : if actual is None : log . warning ( "%s instance could not verify (missing) axis " "expec...
Verify that axis labels for a given source are as expected .
174
13
243,812
def get_data ( self , request = None ) : if request is None : raise ValueError data = [ [ ] for _ in self . sources ] for i in range ( request ) : try : for source_data , example in zip ( data , next ( self . child_epoch_iterator ) ) : source_data . append ( example ) except StopIteration : # If some data has been extr...
Get data from the dataset .
157
6
243,813
def _producer_wrapper ( f , port , addr = 'tcp://127.0.0.1' ) : try : context = zmq . Context ( ) socket = context . socket ( zmq . PUSH ) socket . connect ( ':' . join ( [ addr , str ( port ) ] ) ) f ( socket ) finally : # Works around a Python 3.x bug. context . destroy ( )
A shim that sets up a socket and starts the producer callable .
91
15
243,814
def _spawn_producer ( f , port , addr = 'tcp://127.0.0.1' ) : process = Process ( target = _producer_wrapper , args = ( f , port , addr ) ) process . start ( ) return process
Start a process that sends results on a PUSH socket .
56
12
243,815
def producer_consumer ( producer , consumer , addr = 'tcp://127.0.0.1' , port = None , context = None ) : context_created = False if context is None : context_created = True context = zmq . Context ( ) try : consumer_socket = context . socket ( zmq . PULL ) if port is None : port = consumer_socket . bind_to_random_port...
A producer - consumer pattern .
148
6
243,816
def main ( args = None ) : built_in_datasets = dict ( downloaders . all_downloaders ) if fuel . config . extra_downloaders : for name in fuel . config . extra_downloaders : extra_datasets = dict ( importlib . import_module ( name ) . all_downloaders ) if any ( key in built_in_datasets for key in extra_datasets . keys (...
Entry point for fuel - download script .
427
8
243,817
def fill_subparser ( subparser ) : filenames = [ 'train-images-idx3-ubyte.gz' , 'train-labels-idx1-ubyte.gz' , 't10k-images-idx3-ubyte.gz' , 't10k-labels-idx1-ubyte.gz' ] urls = [ 'http://yann.lecun.com/exdb/mnist/' + f for f in filenames ] subparser . set_defaults ( urls = urls , filenames = filenames ) return default...
Sets up a subparser to download the MNIST dataset files .
144
14
243,818
def main ( args = None ) : parser = argparse . ArgumentParser ( description = 'Extracts metadata from a Fuel-converted HDF5 file.' ) parser . add_argument ( "filename" , help = "HDF5 file to analyze" ) args = parser . parse_args ( ) with h5py . File ( args . filename , 'r' ) as h5file : interface_version = h5file . att...
Entry point for fuel - info script .
270
8
243,819
def convert_silhouettes ( size , directory , output_directory , output_filename = None ) : if size not in ( 16 , 28 ) : raise ValueError ( 'size must be 16 or 28' ) if output_filename is None : output_filename = 'caltech101_silhouettes{}.hdf5' . format ( size ) output_file = os . path . join ( output_directory , output...
Convert the CalTech 101 Silhouettes Datasets .
527
13
243,820
def cross_validation ( scheme_class , num_examples , num_folds , strict = True , * * kwargs ) : if strict and num_examples % num_folds != 0 : raise ValueError ( ( "{} examples are not divisible in {} evenly-sized " + "folds. To allow this, have a look at the " + "`strict` argument." ) . format ( num_examples , num_fold...
Return pairs of schemes to be used for cross - validation .
222
12
243,821
def main ( args = None ) : built_in_datasets = dict ( converters . all_converters ) if fuel . config . extra_converters : for name in fuel . config . extra_converters : extra_datasets = dict ( importlib . import_module ( name ) . all_converters ) if any ( key in built_in_datasets for key in extra_datasets . keys ( ) ) ...
Entry point for fuel - convert script .
865
8
243,822
def refresh_lock ( lock_file ) : unique_id = '%s_%s_%s' % ( os . getpid ( ) , '' . join ( [ str ( random . randint ( 0 , 9 ) ) for i in range ( 10 ) ] ) , hostname ) try : lock_write = open ( lock_file , 'w' ) lock_write . write ( unique_id + '\n' ) lock_write . close ( ) except Exception : # In some strange case, this...
Refresh an existing lock .
172
6
243,823
def get_lock ( lock_dir , * * kw ) : if not hasattr ( get_lock , 'n_lock' ) : # Initialization. get_lock . n_lock = 0 if not hasattr ( get_lock , 'lock_is_enabled' ) : # Enable lock by default. get_lock . lock_is_enabled = True get_lock . lock_dir = lock_dir get_lock . unlocker = Unlocker ( get_lock . lock_dir ) else :...
Obtain lock on compilation directory .
494
7
243,824
def release_lock ( ) : get_lock . n_lock -= 1 assert get_lock . n_lock >= 0 # Only really release lock once all lock requests have ended. if get_lock . lock_is_enabled and get_lock . n_lock == 0 : get_lock . start_time = None get_lock . unlocker . unlock ( )
Release lock on compilation directory .
78
6
243,825
def release_readlock ( lockdir_name ) : # Make sure the lock still exists before deleting it if os . path . exists ( lockdir_name ) and os . path . isdir ( lockdir_name ) : os . rmdir ( lockdir_name )
Release a previously obtained readlock .
59
7
243,826
def get_readlock ( pid , path ) : timestamp = int ( time . time ( ) * 1e6 ) lockdir_name = "%s.readlock.%i.%i" % ( path , pid , timestamp ) os . mkdir ( lockdir_name ) # Register function to release the readlock at the end of the script atexit . register ( release_readlock , lockdir_name = lockdir_name )
Obtain a readlock on a file .
94
9
243,827
def unlock ( self ) : # If any error occurs, we assume this is because someone else tried to # unlock this directory at the same time. # Note that it is important not to have both remove statements within # the same try/except block. The reason is that while the attempt to # remove the file may fail (e.g. because for s...
Remove current lock .
142
4
243,828
def filename_from_url ( url , path = None ) : r = requests . get ( url , stream = True ) if 'Content-Disposition' in r . headers : filename = re . findall ( r'filename=([^;]+)' , r . headers [ 'Content-Disposition' ] ) [ 0 ] . strip ( '"\"' ) else : filename = os . path . basename ( urllib . parse . urlparse ( url ) . ...
Parses a URL to determine a file name .
106
11
243,829
def download ( url , file_handle , chunk_size = 1024 ) : r = requests . get ( url , stream = True ) total_length = r . headers . get ( 'content-length' ) if total_length is None : maxval = UnknownLength else : maxval = int ( total_length ) name = file_handle . name with progress_bar ( name = name , maxval = maxval ) as...
Downloads a given URL to a specific file .
135
10
243,830
def default_downloader ( directory , urls , filenames , url_prefix = None , clear = False ) : # Parse file names from URL if not provided for i , url in enumerate ( urls ) : filename = filenames [ i ] if not filename : filename = filename_from_url ( url ) if not filename : raise ValueError ( "no filename available for ...
Downloads or clears files from URLs and filenames .
245
12
243,831
def find_in_data_path ( filename ) : for path in config . data_path : path = os . path . expanduser ( os . path . expandvars ( path ) ) file_path = os . path . join ( path , filename ) if os . path . isfile ( file_path ) : return file_path raise IOError ( "{} not found in Fuel's data path" . format ( filename ) )
Searches for a file within Fuel s data path .
92
12
243,832
def lazy_property_factory ( lazy_property ) : def lazy_property_getter ( self ) : if not hasattr ( self , '_' + lazy_property ) : self . load ( ) if not hasattr ( self , '_' + lazy_property ) : raise ValueError ( "{} wasn't loaded" . format ( lazy_property ) ) return getattr ( self , '_' + lazy_property ) def lazy_prop...
Create properties that perform lazy loading of attributes .
135
9
243,833
def do_not_pickle_attributes ( * lazy_properties ) : def wrap_class ( cls ) : if not hasattr ( cls , 'load' ) : raise ValueError ( "no load method implemented" ) # Attach the lazy loading properties to the class for lazy_property in lazy_properties : setattr ( cls , lazy_property , property ( * lazy_property_factory ( ...
r Decorator to assign non - pickable properties .
252
12
243,834
def sorted_fancy_indexing ( indexable , request ) : if len ( request ) > 1 : indices = numpy . argsort ( request ) data = numpy . empty ( shape = ( len ( request ) , ) + indexable . shape [ 1 : ] , dtype = indexable . dtype ) data [ indices ] = indexable [ numpy . array ( request ) [ indices ] , ... ] else : data = ind...
Safe fancy indexing .
100
5
243,835
def slice_to_numerical_args ( slice_ , num_examples ) : start = slice_ . start if slice_ . start is not None else 0 stop = slice_ . stop if slice_ . stop is not None else num_examples step = slice_ . step if slice_ . step is not None else 1 return start , stop , step
Translate a slice s attributes into numerical attributes .
77
10
243,836
def get_list_representation ( self ) : if self . is_list : return self . list_or_slice else : return self [ list ( range ( self . num_examples ) ) ]
Returns this subset s representation as a list of indices .
44
11
243,837
def index_within_subset ( self , indexable , subset_request , sort_indices = False ) : # Translate the request within the context of this subset to a # request to the indexable object if isinstance ( subset_request , numbers . Integral ) : request , = self [ [ subset_request ] ] else : request = self [ subset_request ]...
Index an indexable object within the context of this subset .
264
12
243,838
def num_examples ( self ) : if self . is_list : return len ( self . list_or_slice ) else : start , stop , step = self . slice_to_numerical_args ( self . list_or_slice , self . original_num_examples ) return stop - start
The number of examples this subset spans .
68
8
243,839
def get_epoch_iterator ( self , * * kwargs ) : if not self . _fresh_state : self . next_epoch ( ) else : self . _fresh_state = False return super ( DataStream , self ) . get_epoch_iterator ( * * kwargs )
Get an epoch iterator for the data stream .
66
9
243,840
def fill_subparser ( subparser ) : sets = [ 'train' , 'valid' , 'test' ] urls = [ 'http://www.cs.toronto.edu/~larocheh/public/datasets/' + 'binarized_mnist/binarized_mnist_{}.amat' . format ( s ) for s in sets ] filenames = [ 'binarized_mnist_{}.amat' . format ( s ) for s in sets ] subparser . set_defaults ( urls = url...
Sets up a subparser to download the binarized MNIST dataset files .
137
17
243,841
def download ( directory , youtube_id , clear = False ) : filepath = os . path . join ( directory , '{}.m4a' . format ( youtube_id ) ) if clear : os . remove ( filepath ) return if not PAFY_AVAILABLE : raise ImportError ( "pafy is required to download YouTube videos" ) url = 'https://www.youtube.com/watch?v={}' . forma...
Download the audio of a YouTube video .
134
8
243,842
def fill_subparser ( subparser ) : subparser . add_argument ( '--youtube-id' , type = str , required = True , help = ( "The YouTube ID of the video from which to extract audio, " "usually an 11-character string." ) ) return download
Sets up a subparser to download audio of YouTube videos .
61
13
243,843
def convert_youtube_audio ( directory , output_directory , youtube_id , channels , sample , output_filename = None ) : input_file = os . path . join ( directory , '{}.m4a' . format ( youtube_id ) ) wav_filename = '{}.wav' . format ( youtube_id ) wav_file = os . path . join ( directory , wav_filename ) ffmpeg_not_availa...
Converts downloaded YouTube audio to HDF5 format .
409
11
243,844
def fill_subparser ( subparser ) : subparser . add_argument ( '--youtube-id' , type = str , required = True , help = ( "The YouTube ID of the video from which to extract audio, " "usually an 11-character string." ) ) subparser . add_argument ( '--channels' , type = int , default = 1 , help = ( "The number of audio chan...
Sets up a subparser to convert YouTube audio files .
179
12
243,845
def convert_ilsvrc2012 ( directory , output_directory , output_filename = 'ilsvrc2012.hdf5' , shuffle_seed = config . default_seed ) : devkit_path = os . path . join ( directory , DEVKIT_ARCHIVE ) train , valid , test = [ os . path . join ( directory , fn ) for fn in IMAGE_TARS ] n_train , valid_groundtruth , n_test , ...
Converter for data from the ILSVRC 2012 competition .
341
14
243,846
def fill_subparser ( subparser ) : subparser . add_argument ( "--shuffle-seed" , help = "Seed to use for randomizing order of the " "training set on disk." , default = config . default_seed , type = int , required = False ) return convert_ilsvrc2012
Sets up a subparser to convert the ILSVRC2012 dataset files .
69
17
243,847
def read_metadata_mat_file ( meta_mat ) : mat = loadmat ( meta_mat , squeeze_me = True ) synsets = mat [ 'synsets' ] new_dtype = numpy . dtype ( [ ( 'ILSVRC2012_ID' , numpy . int16 ) , ( 'WNID' , ( 'S' , max ( map ( len , synsets [ 'WNID' ] ) ) ) ) , ( 'wordnet_height' , numpy . int8 ) , ( 'gloss' , ( 'S' , max ( map (...
Read ILSVRC2012 metadata from the distributed MAT file .
431
13
243,848
def multiple_paths_parser ( value ) : if isinstance ( value , six . string_types ) : value = value . split ( os . path . pathsep ) return value
Parses data_path argument .
40
8
243,849
def add_config ( self , key , type_ , default = NOT_SET , env_var = None ) : self . config [ key ] = { 'type' : type_ } if env_var is not None : self . config [ key ] [ 'env_var' ] = env_var if default is not NOT_SET : self . config [ key ] [ 'default' ] = default
Add a configuration setting .
86
5
243,850
def send_arrays ( socket , arrays , stop = False ) : if arrays : # The buffer protocol only works on contiguous arrays arrays = [ numpy . ascontiguousarray ( array ) for array in arrays ] if stop : headers = { 'stop' : True } socket . send_json ( headers ) else : headers = [ header_data_from_array_1_0 ( array ) for arr...
Send NumPy arrays using the buffer interface and some metadata .
139
12
243,851
def recv_arrays ( socket ) : headers = socket . recv_json ( ) if 'stop' in headers : raise StopIteration arrays = [ ] for header in headers : data = socket . recv ( copy = False ) buf = buffer_ ( data ) array = numpy . frombuffer ( buf , dtype = numpy . dtype ( header [ 'descr' ] ) ) array . shape = header [ 'shape' ] ...
Receive a list of NumPy arrays .
139
9
243,852
def start_server ( data_stream , port = 5557 , hwm = 10 ) : logging . basicConfig ( level = 'INFO' ) context = zmq . Context ( ) socket = context . socket ( zmq . PUSH ) socket . set_hwm ( hwm ) socket . bind ( 'tcp://*:{}' . format ( port ) ) it = data_stream . get_epoch_iterator ( ) logger . info ( 'server started' )...
Start a data processing server .
187
6
243,853
def create_images ( raw_data_directory : str , destination_directory : str , stroke_thicknesses : List [ int ] , canvas_width : int = None , canvas_height : int = None , staff_line_spacing : int = 14 , staff_line_vertical_offsets : List [ int ] = None , random_position_on_canvas : bool = False ) -> dict : all_symbol_fi...
Creates a visual representation of the Homus Dataset by parsing all text - files and the symbols as specified by the parameters by drawing lines that connect the points from each stroke of each symbol .
800
40
243,854
def extract_and_render_all_symbol_masks ( self , raw_data_directory : str , destination_directory : str ) : print ( "Extracting Symbols from Muscima++ Dataset..." ) xml_files = self . get_all_xml_file_paths ( raw_data_directory ) crop_objects = self . load_crop_objects_from_xml_files ( xml_files ) self . render_masks_o...
Extracts all symbols from the raw XML documents and generates individual symbols from the masks
120
17
243,855
def invert_images ( self , image_directory : str , image_file_ending : str = "*.bmp" ) : image_paths = [ y for x in os . walk ( image_directory ) for y in glob ( os . path . join ( x [ 0 ] , image_file_ending ) ) ] for image_path in tqdm ( image_paths , desc = "Inverting all images in directory {0}" . format ( image_di...
In - situ converts the white on black images of a directory to black on white images
181
17
243,856
def create_capitan_images ( self , raw_data_directory : str , destination_directory : str , stroke_thicknesses : List [ int ] ) -> None : symbols = self . load_capitan_symbols ( raw_data_directory ) self . draw_capitan_stroke_images ( symbols , destination_directory , stroke_thicknesses ) self . draw_capitan_score_imag...
Creates a visual representation of the Capitan strokes by parsing all text - files and the symbols as specified by the parameters by drawing lines that connect the points from each stroke of each symbol .
97
38
243,857
def draw_capitan_stroke_images ( self , symbols : List [ CapitanSymbol ] , destination_directory : str , stroke_thicknesses : List [ int ] ) -> None : total_number_of_symbols = len ( symbols ) * len ( stroke_thicknesses ) output = "Generating {0} images with {1} symbols in {2} different stroke thicknesses ({3})" . form...
Creates a visual representation of the Capitan strokes by drawing lines that connect the points from each stroke of each symbol .
389
24
243,858
def overlap ( r1 : 'Rectangle' , r2 : 'Rectangle' ) : h_overlaps = ( r1 . left <= r2 . right ) and ( r1 . right >= r2 . left ) v_overlaps = ( r1 . bottom >= r2 . top ) and ( r1 . top <= r2 . bottom ) return h_overlaps and v_overlaps
Overlapping rectangles overlap both horizontally & vertically
90
10
243,859
def extract_symbols ( self , raw_data_directory : str , destination_directory : str ) : print ( "Extracting Symbols from Audiveris OMR Dataset..." ) all_xml_files = [ y for x in os . walk ( raw_data_directory ) for y in glob ( os . path . join ( x [ 0 ] , '*.xml' ) ) ] all_image_files = [ y for x in os . walk ( raw_dat...
Extracts the symbols from the raw XML documents and matching images of the Audiveris OMR dataset into individual symbols
218
24
243,860
def initialize_from_string ( content : str ) -> 'HomusSymbol' : if content is None or content is "" : return None lines = content . splitlines ( ) min_x = sys . maxsize max_x = 0 min_y = sys . maxsize max_y = 0 symbol_name = lines [ 0 ] strokes = [ ] for stroke_string in lines [ 1 : ] : stroke = [ ] for point_string in...
Create and initializes a new symbol from a string
284
10
243,861
def draw_into_bitmap ( self , export_path : ExportPath , stroke_thickness : int , margin : int = 0 ) -> None : self . draw_onto_canvas ( export_path , stroke_thickness , margin , self . dimensions . width + 2 * margin , self . dimensions . height + 2 * margin )
Draws the symbol in the original size that it has plus an optional margin
75
15
243,862
def update_locals ( locals_instance , instance_iterator , * args , * * kwargs ) : # http://stackoverflow.com/a/4526709/564709 # http://stackoverflow.com/a/511059/564709 for instance in instance_iterator ( ) : locals_instance . update ( { type ( instance ) . __name__ : instance . __class__ } )
import all of the detector classes into the local namespace to make it easy to do things like import scrubadub . detectors . NameDetector without having to add each new Detector or Filth
92
39
243,863
def iter_filth_clss ( ) : return iter_subclasses ( os . path . dirname ( os . path . abspath ( __file__ ) ) , Filth , _is_abstract_filth , )
Iterate over all of the filths that are included in this sub - package . This is a convenience method for capturing all new Filth that are added over time .
50
34
243,864
def iter_filths ( ) : for filth_cls in iter_filth_clss ( ) : if issubclass ( filth_cls , RegexFilth ) : m = next ( re . finditer ( r"\s+" , "fake pattern string" ) ) yield filth_cls ( m ) else : yield filth_cls ( )
Iterate over all instances of filth
84
8
243,865
def _update_content ( self , other_filth ) : if self . end < other_filth . beg or other_filth . end < self . beg : raise exceptions . FilthMergeError ( "a_filth goes from [%s, %s) and b_filth goes from [%s, %s)" % ( self . beg , self . end , other_filth . beg , other_filth . end ) ) # get the text over lap correct if s...
this updates the bounds text and placeholder for the merged filth
291
12
243,866
def add_detector ( self , detector_cls ) : if not issubclass ( detector_cls , detectors . base . Detector ) : raise TypeError ( ( '"%(detector_cls)s" is not a subclass of Detector' ) % locals ( ) ) # TODO: should add tests to make sure filth_cls is actually a proper # filth_cls name = detector_cls . filth_cls . type if...
Add a Detector to scrubadub
164
8
243,867
def clean ( self , text , * * kwargs ) : if sys . version_info < ( 3 , 0 ) : # Only in Python 2. In 3 every string is a Python 2 unicode if not isinstance ( text , unicode ) : raise exceptions . UnicodeRequired clean_chunks = [ ] filth = Filth ( ) for next_filth in self . iter_filth ( text ) : clean_chunks . append ( t...
This is the master method that cleans all of the filth out of the dirty dirty text . All keyword arguments to this function are passed through to the Filth . replace_with method to fine - tune how the Filth is cleaned .
168
48
243,868
def iter_filth ( self , text ) : # currently doing this by aggregating all_filths and then sorting # inline instead of with a Filth.__cmp__ method, which is apparently # much slower http://stackoverflow.com/a/988728/564709 # # NOTE: we could probably do this in a more efficient way by iterating # over all detectors sim...
Iterate over the different types of filth that can exist .
317
13
243,869
async def download_file ( self , Bucket , Key , Filename , ExtraArgs = None , Callback = None , Config = None ) : with open ( Filename , 'wb' ) as open_file : await download_fileobj ( self , Bucket , Key , open_file , ExtraArgs = ExtraArgs , Callback = Callback , Config = Config )
Download an S3 object to a file .
78
9
243,870
async def download_fileobj ( self , Bucket , Key , Fileobj , ExtraArgs = None , Callback = None , Config = None ) : try : resp = await self . get_object ( Bucket = Bucket , Key = Key ) except ClientError as err : if err . response [ 'Error' ] [ 'Code' ] == 'NoSuchKey' : # Convert to 404 so it looks the same when boto3....
Download an object from S3 to a file - like object .
199
13
243,871
async def upload_fileobj ( self , Fileobj : BinaryIO , Bucket : str , Key : str , ExtraArgs : Optional [ Dict [ str , Any ] ] = None , Callback : Optional [ Callable [ [ int ] , None ] ] = None , Config : Optional [ S3TransferConfig ] = None ) : if not ExtraArgs : ExtraArgs = { } # I was debating setting up a queue etc...
Upload a file - like object to S3 .
555
10
243,872
async def upload_file ( self , Filename , Bucket , Key , ExtraArgs = None , Callback = None , Config = None ) : with open ( Filename , 'rb' ) as open_file : await upload_fileobj ( self , open_file , Bucket , Key , ExtraArgs = ExtraArgs , Callback = Callback , Config = Config )
Upload a file to an S3 object .
78
9
243,873
def _create_action ( factory_self , action_model , resource_name , service_context , is_load = False ) : # Create the action in in this closure but before the ``do_action`` # method below is invoked, which allows instances of the resource # to share the ServiceAction instance. action = AIOServiceAction ( action_model ,...
Creates a new method which makes a request to the underlying AWS service .
490
15
243,874
def from_der_private_key ( data : bytes , password : Optional [ str ] = None ) -> _RSAPrivateKey : return serialization . load_der_private_key ( data , password , default_backend ( ) )
Convert private key in DER encoding to a Private key object
53
13
243,875
async def get_object ( self , Bucket : str , Key : str , * * kwargs ) -> dict : if self . _s3_client is None : await self . setup ( ) # Ok so if we are doing a range get. We need to align the range start/end with AES block boundaries # 9223372036854775806 is 8EiB so I have no issue with hardcoding it. # We pass the act...
S3 GetObject . Takes same args as Boto3 documentation
619
13
243,876
async def put_object ( self , Body : Union [ bytes , IO ] , Bucket : str , Key : str , Metadata : Dict = None , * * kwargs ) : if self . _s3_client is None : await self . setup ( ) if hasattr ( Body , 'read' ) : if inspect . iscoroutinefunction ( Body . read ) : Body = await Body . read ( ) else : Body = Body . read ( ...
PutObject . Takes same args as Boto3 documentation
709
11
243,877
def histogram1d ( x , bins , range , weights = None ) : nx = bins if not np . isscalar ( bins ) : raise TypeError ( 'bins should be an integer' ) xmin , xmax = range if not np . isfinite ( xmin ) : raise ValueError ( "xmin should be finite" ) if not np . isfinite ( xmax ) : raise ValueError ( "xmax should be finite" ) ...
Compute a 1D histogram assuming equally spaced bins .
189
12
243,878
def histogram2d ( x , y , bins , range , weights = None ) : if isinstance ( bins , numbers . Integral ) : nx = ny = bins else : nx , ny = bins if not np . isscalar ( nx ) or not np . isscalar ( ny ) : raise TypeError ( 'bins should be an iterable of two integers' ) ( xmin , xmax ) , ( ymin , ymax ) = range if not np . ...
Compute a 2D histogram assuming equally spaced bins .
351
12
243,879
def to_networkx ( self ) : return nx_util . to_networkx ( self . session . get ( self . __url ) . json ( ) )
Return this network in NetworkX graph object .
36
9
243,880
def to_dataframe ( self , extra_edges_columns = [ ] ) : return df_util . to_dataframe ( self . session . get ( self . __url ) . json ( ) , edges_attr_cols = extra_edges_columns )
Return this network in pandas DataFrame .
61
9
243,881
def add_node ( self , node_name , dataframe = False ) : if node_name is None : return None return self . add_nodes ( [ node_name ] , dataframe = dataframe )
Add a single node to the network .
46
8
243,882
def add_nodes ( self , node_name_list , dataframe = False ) : res = self . session . post ( self . __url + 'nodes' , data = json . dumps ( node_name_list ) , headers = HEADERS ) check_response ( res ) nodes = res . json ( ) if dataframe : return pd . DataFrame ( nodes ) . set_index ( [ 'SUID' ] ) else : return { node [...
Add new nodes to the network
117
6
243,883
def add_edge ( self , source , target , interaction = '-' , directed = True , dataframe = True ) : new_edge = { 'source' : source , 'target' : target , 'interaction' : interaction , 'directed' : directed } return self . add_edges ( [ new_edge ] , dataframe = dataframe )
Add a single edge from source to target .
76
9
243,884
def get_views ( self ) : url = self . __url + 'views' return self . session . get ( url ) . json ( )
Get views as a list of SUIDs
31
8
243,885
def diffuse_advanced ( self , heatColumnName = None , time = None , verbose = False ) : PARAMS = set_param ( [ "heatColumnName" , "time" ] , [ heatColumnName , time ] ) response = api ( url = self . __url + "/diffuse_advanced" , PARAMS = PARAMS , method = "POST" , verbose = verbose ) return response
Diffusion will send the selected network view and its selected nodes to a web - based REST service to calculate network propagation . Results are returned and represented by columns in the node table . Columns are created for each execution of Diffusion and their names are returned in the response .
90
55
243,886
def to_networkx ( cyjs , directed = True ) : if directed : g = nx . MultiDiGraph ( ) else : g = nx . MultiGraph ( ) network_data = cyjs [ DATA ] if network_data is not None : for key in network_data . keys ( ) : g . graph [ key ] = network_data [ key ] nodes = cyjs [ ELEMENTS ] [ NODES ] edges = cyjs [ ELEMENTS ] [ EDG...
Convert Cytoscape . js - style JSON object into NetworkX object .
177
17
243,887
def dialog ( self = None , wid = None , text = None , title = None , url = None , debug = False , verbose = False ) : PARAMS = set_param ( [ "id" , "text" , "title" , "url" , "debug" ] , [ wid , text , title , url , debug ] ) response = api ( url = self . __url + "/dialog?" , PARAMS = PARAMS , method = "GET" , verbose ...
Launch and HTML browser in a separate window .
110
9
243,888
def hide ( self , wid , verbose = False ) : PARAMS = { "id" : wid } response = api ( url = self . __url + "/hide?" , PARAMS = PARAMS , method = "GET" , verbose = verbose ) return response
Hide and HTML browser in the Results Panel .
58
9
243,889
def show ( self , wid = None , text = None , title = None , url = None , verbose = False ) : PARAMS = { } for p , v in zip ( [ "id" , "text" , "title" , "url" ] , [ wid , text , title , url ] ) : if v : PARAMS [ p ] = v response = api ( url = self . __url + "/show?" , PARAMS = PARAMS , method = "GET" , verbose = verbos...
Launch an HTML browser in the Results Panel .
113
9
243,890
def check_response ( res ) : try : res . raise_for_status ( ) # Alternative is res.ok except Exception as exc : # Bad response code, e.g. if adding an edge with nodes that doesn't exist try : err_info = res . json ( ) err_msg = err_info [ 'message' ] # or 'localizeMessage' except ValueError : err_msg = res . text [ : 4...
Check HTTP response and raise exception if response is not OK .
158
12
243,891
def from_dataframe ( df , source_col = 'source' , target_col = 'target' , interaction_col = 'interaction' , name = 'From DataFrame' , edge_attr_cols = [ ] ) : network = cyjs . get_empty_network ( name = name ) nodes = set ( ) if edge_attr_cols is None : edge_attr_cols = [ ] for index , row in df . iterrows ( ) : s = ro...
Utility to convert Pandas DataFrame object into Cytoscape . js JSON
264
17
243,892
def to_dataframe ( network , interaction = 'interaction' , default_interaction = '-' , edges_attr_cols = [ ] ) : edges = network [ 'elements' ] [ 'edges' ] if edges_attr_cols is None : edges_attr_cols = [ ] edges_attr_cols = sorted ( edges_attr_cols ) network_array = [ ] # the set avoids duplicates valid_extra_cols = s...
Utility to convert a Cytoscape dictionary into a Pandas Dataframe .
284
17
243,893
def render ( network , style = DEF_STYLE , layout_algorithm = DEF_LAYOUT , background = DEF_BACKGROUND_COLOR , height = DEF_HEIGHT , width = DEF_WIDTH , style_file = STYLE_FILE , def_nodes = DEF_NODES , def_edges = DEF_EDGES ) : from jinja2 import Template from IPython . core . display import display , HTML STYLES = se...
Render network data with embedded Cytoscape . js widget .
330
13
243,894
def create_attribute ( self , column = None , listType = None , namespace = None , network = None , atype = None , verbose = False ) : network = check_network ( self , network , verbose = verbose ) PARAMS = set_param ( [ "column" , "listType" , "namespace" , "network" , "type" ] , [ column , listType , namespace , netw...
Creates a new edge column .
132
7
243,895
def get ( self , edge = None , network = None , sourceNode = None , targetNode = None , atype = None , verbose = False ) : network = check_network ( self , network , verbose = verbose ) PARAMS = set_param ( [ "edge" , "network" , "sourceNode" , "targetNode" , "type" ] , [ edge , network , sourceNode , targetNode , atyp...
Returns the SUID of an edge that matches the passed parameters . If multiple edges are found only one will be returned and a warning will be reported in the Cytoscape Task History dialog .
131
39
243,896
def add_edge ( self , isDirected = None , name = None , network = None , sourceName = None , targetName = None , verbose = False ) : network = check_network ( self , network , verbose = verbose ) PARAMS = set_param ( [ "isDirected" , "name" , "network" , "sourceName" , "targetName" ] , [ isDirected , name , network , s...
Add a new edge between two existing nodes in a network . The names of the nodes must be specified and much match the value in the name column for each node .
138
33
243,897
def create ( self , edgeList = None , excludeEdges = None , networkName = None , nodeList = None , source = None , verbose = False ) : network = check_network ( self , source , verbose = verbose ) PARAMS = set_param ( [ "edgeList" , "excludeEdges" , "networkName" , "nodeList" , "source" ] , [ edgeList , excludeEdges , ...
Create a new network from a list of nodes and edges in an existing source network . The SUID of the network and view are returned .
139
28
243,898
def create_empty ( self , name = None , renderers = None , RootNetworkList = None , verbose = False ) : PARAMS = set_param ( [ "name" , "renderers" , "RootNetworkList" ] , [ name , renderers , RootNetworkList ] ) response = api ( url = self . __url + "/create empty" , PARAMS = PARAMS , method = "POST" , verbose = verbo...
Create a new empty network . The new network may be created as part of an existing network collection or a new network collection .
99
25
243,899
def list ( self , verbose = False ) : response = api ( url = self . __url + "/list" , method = "POST" , verbose = verbose ) return response
List all of the networks in the current session .
40
10