idx int64 0 63k | question stringlengths 61 4.03k | target stringlengths 6 1.23k |
|---|---|---|
55,900 | def build_routename ( cls , name , routename_prefix = None ) : if routename_prefix is None : routename_prefix = 'api_{}' . format ( cls . __name__ . replace ( 'Resource' , '' ) . lower ( ) ) routename_prefix = routename_prefix . rstrip ( '_' ) return '_' . join ( [ routename_prefix , name ] ) | Given a name & an optional routename_prefix this generates a name for a URL . |
55,901 | def add_views ( cls , config , rule_prefix , routename_prefix = None ) : methods = ( 'GET' , 'POST' , 'PUT' , 'DELETE' ) config . add_route ( cls . build_routename ( 'list' , routename_prefix ) , rule_prefix ) config . add_view ( cls . as_list ( ) , route_name = cls . build_routename ( 'list' , routename_prefix ) , req... | A convenience method for registering the routes and views in pyramid . |
55,902 | def deserialize ( self , body ) : try : if isinstance ( body , bytes ) : return json . loads ( body . decode ( 'utf-8' ) ) return json . loads ( body ) except ValueError : raise BadRequest ( 'Request body is not valid JSON' ) | The low - level deserialization . |
55,903 | def convert_mnist ( directory , output_directory , output_filename = None , dtype = None ) : if not output_filename : if dtype : output_filename = 'mnist_{}.hdf5' . format ( dtype ) else : output_filename = 'mnist.hdf5' output_path = os . path . join ( output_directory , output_filename ) h5file = h5py . File ( output_... | Converts the MNIST dataset to HDF5 . |
55,904 | def fill_subparser ( subparser ) : subparser . add_argument ( "--dtype" , help = "dtype to save to; by default, images will be " + "returned in their original unsigned byte format" , choices = ( 'float32' , 'float64' , 'bool' ) , type = str , default = None ) return convert_mnist | Sets up a subparser to convert the MNIST dataset files . |
55,905 | def read_mnist_images ( filename , dtype = None ) : with gzip . open ( filename , 'rb' ) as f : magic , number , rows , cols = struct . unpack ( '>iiii' , f . read ( 16 ) ) if magic != MNIST_IMAGE_MAGIC : raise ValueError ( "Wrong magic number reading MNIST image file" ) array = numpy . frombuffer ( f . read ( ) , dtyp... | Read MNIST images from the original ubyte file format . |
55,906 | def read_mnist_labels ( filename ) : with gzip . open ( filename , 'rb' ) as f : magic , _ = struct . unpack ( '>ii' , f . read ( 8 ) ) if magic != MNIST_LABEL_MAGIC : raise ValueError ( "Wrong magic number reading MNIST label file" ) array = numpy . frombuffer ( f . read ( ) , dtype = 'uint8' ) array = array . reshape... | Read MNIST labels from the original ubyte file format . |
55,907 | def prepare_hdf5_file ( hdf5_file , n_train , n_valid , n_test ) : n_total = n_train + n_valid + n_test splits = create_splits ( n_train , n_valid , n_test ) hdf5_file . attrs [ 'split' ] = H5PYDataset . create_split_array ( splits ) vlen_dtype = h5py . special_dtype ( vlen = numpy . dtype ( 'uint8' ) ) hdf5_file . cre... | Create datasets within a given HDF5 file . |
55,908 | def process_train_set ( hdf5_file , train_archive , patch_archive , n_train , wnid_map , shuffle_seed = None ) : producer = partial ( train_set_producer , train_archive = train_archive , patch_archive = patch_archive , wnid_map = wnid_map ) consumer = partial ( image_consumer , hdf5_file = hdf5_file , num_expected = n_... | Process the ILSVRC2010 training set . |
55,909 | def image_consumer ( socket , hdf5_file , num_expected , shuffle_seed = None , offset = 0 ) : with progress_bar ( 'images' , maxval = num_expected ) as pb : if shuffle_seed is None : index_gen = iter ( xrange ( num_expected ) ) else : rng = numpy . random . RandomState ( shuffle_seed ) index_gen = iter ( rng . permutat... | Fill an HDF5 file with incoming images from a socket . |
55,910 | def process_other_set ( hdf5_file , which_set , image_archive , patch_archive , groundtruth , offset ) : producer = partial ( other_set_producer , image_archive = image_archive , patch_archive = patch_archive , groundtruth = groundtruth , which_set = which_set ) consumer = partial ( image_consumer , hdf5_file = hdf5_fi... | Process the validation or test set . |
55,911 | def load_from_tar_or_patch ( tar , image_filename , patch_images ) : patched = True image_bytes = patch_images . get ( os . path . basename ( image_filename ) , None ) if image_bytes is None : patched = False try : image_bytes = tar . extractfile ( image_filename ) . read ( ) numpy . array ( Image . open ( io . BytesIO... | Do everything necessary to process an image inside a TAR . |
55,912 | def read_devkit ( f ) : with tar_open ( f ) as tar : meta_mat = tar . extractfile ( DEVKIT_META_PATH ) synsets , cost_matrix = read_metadata_mat_file ( meta_mat ) raw_valid_groundtruth = numpy . loadtxt ( tar . extractfile ( DEVKIT_VALID_GROUNDTRUTH_PATH ) , dtype = numpy . int16 ) return synsets , cost_matrix , raw_va... | Read relevant information from the development kit archive . |
55,913 | def extract_patch_images ( f , which_set ) : if which_set not in ( 'train' , 'valid' , 'test' ) : raise ValueError ( 'which_set must be one of train, valid, or test' ) which_set = 'val' if which_set == 'valid' else which_set patch_images = { } with tar_open ( f ) as tar : for info_obj in tar : if not info_obj . name . ... | Extracts a dict of the patch images for ILSVRC2010 . |
55,914 | def convert_cifar10 ( directory , output_directory , output_filename = 'cifar10.hdf5' ) : output_path = os . path . join ( output_directory , output_filename ) h5file = h5py . File ( output_path , mode = 'w' ) input_file = os . path . join ( directory , DISTRIBUTION_FILE ) tar_file = tarfile . open ( input_file , 'r:gz... | Converts the CIFAR - 10 dataset to HDF5 . |
55,915 | def check_exists ( required_files ) : def function_wrapper ( f ) : @ wraps ( f ) def wrapped ( directory , * args , ** kwargs ) : missing = [ ] for filename in required_files : if not os . path . isfile ( os . path . join ( directory , filename ) ) : missing . append ( filename ) if len ( missing ) > 0 : raise MissingI... | Decorator that checks if required files exist before running . |
55,916 | def fill_hdf5_file ( h5file , data ) : split_names = set ( split_tuple [ 0 ] for split_tuple in data ) for name in split_names : lengths = [ len ( split_tuple [ 2 ] ) for split_tuple in data if split_tuple [ 0 ] == name ] if not all ( le == lengths [ 0 ] for le in lengths ) : raise ValueError ( "split '{}' has sources ... | Fills an HDF5 file in a H5PYDataset - compatible manner . |
55,917 | def progress_bar ( name , maxval , prefix = 'Converting' ) : widgets = [ '{} {}: ' . format ( prefix , name ) , Percentage ( ) , ' ' , Bar ( marker = '=' , left = '[' , right = ']' ) , ' ' , ETA ( ) ] bar = ProgressBar ( widgets = widgets , max_value = maxval , fd = sys . stdout ) . start ( ) try : yield bar finally : ... | Manages a progress bar for a conversion . |
55,918 | def convert_iris ( directory , output_directory , output_filename = 'iris.hdf5' ) : classes = { b'Iris-setosa' : 0 , b'Iris-versicolor' : 1 , b'Iris-virginica' : 2 } data = numpy . loadtxt ( os . path . join ( directory , 'iris.data' ) , converters = { 4 : lambda x : classes [ x ] } , delimiter = ',' ) features = data ... | Convert the Iris dataset to HDF5 . |
55,919 | def fill_subparser ( subparser ) : urls = ( [ None ] * len ( ALL_FILES ) ) filenames = list ( ALL_FILES ) subparser . set_defaults ( urls = urls , filenames = filenames ) subparser . add_argument ( '-P' , '--url-prefix' , type = str , default = None , help = "URL prefix to prepend to the filenames of " "non-public file... | Sets up a subparser to download the ILSVRC2012 dataset files . |
55,920 | def _get_target_index ( self ) : return ( self . index + self . source_window * ( not self . overlapping ) + self . offset ) | Return the index where the target window starts . |
55,921 | def _get_end_index ( self ) : return max ( self . index + self . source_window , self . _get_target_index ( ) + self . target_window ) | Return the end of both windows . |
55,922 | def convert_svhn ( which_format , directory , output_directory , output_filename = None ) : if which_format not in ( 1 , 2 ) : raise ValueError ( "SVHN format needs to be either 1 or 2." ) if not output_filename : output_filename = 'svhn_format_{}.hdf5' . format ( which_format ) if which_format == 1 : return convert_sv... | Converts the SVHN dataset to HDF5 . |
55,923 | def open_ ( filename , mode = 'r' , encoding = None ) : if filename . endswith ( '.gz' ) : if six . PY2 : zf = io . BufferedReader ( gzip . open ( filename , mode ) ) if encoding : return codecs . getreader ( encoding ) ( zf ) else : return zf else : return io . BufferedReader ( gzip . open ( filename , mode , encoding... | Open a text file with encoding and optional gzip compression . |
55,924 | def tar_open ( f ) : if isinstance ( f , six . string_types ) : return tarfile . open ( name = f ) else : return tarfile . open ( fileobj = f ) | Open either a filename or a file - like object as a TarFile . |
55,925 | def copy_from_server_to_local ( dataset_remote_dir , dataset_local_dir , remote_fname , local_fname ) : log . debug ( "Copying file `{}` to a local directory `{}`." . format ( remote_fname , dataset_local_dir ) ) head , tail = os . path . split ( local_fname ) head += os . path . sep if not os . path . exists ( head ) ... | Copies a remote file locally . |
55,926 | def convert_to_one_hot ( y ) : max_value = max ( y ) min_value = min ( y ) length = len ( y ) one_hot = numpy . zeros ( ( length , ( max_value - min_value + 1 ) ) ) one_hot [ numpy . arange ( length ) , y ] = 1 return one_hot | converts y into one hot reprsentation . |
55,927 | def convert_binarized_mnist ( directory , output_directory , output_filename = 'binarized_mnist.hdf5' ) : output_path = os . path . join ( output_directory , output_filename ) h5file = h5py . File ( output_path , mode = 'w' ) train_set = numpy . loadtxt ( os . path . join ( directory , TRAIN_FILE ) ) . reshape ( ( - 1 ... | Converts the binarized MNIST dataset to HDF5 . |
55,928 | def fill_subparser ( subparser ) : url = 'http://www.cs.toronto.edu/~kriz/cifar-10-python.tar.gz' filename = 'cifar-10-python.tar.gz' subparser . set_defaults ( urls = [ url ] , filenames = [ filename ] ) return default_downloader | Sets up a subparser to download the CIFAR - 10 dataset file . |
55,929 | def convert_celeba_aligned_cropped ( directory , output_directory , output_filename = OUTPUT_FILENAME ) : output_path = os . path . join ( output_directory , output_filename ) h5file = _initialize_conversion ( directory , output_path , ( 218 , 178 ) ) features_dataset = h5file [ 'features' ] image_file_path = os . path... | Converts the aligned and cropped CelebA dataset to HDF5 . |
55,930 | def convert_celeba ( which_format , directory , output_directory , output_filename = None ) : if which_format not in ( 'aligned_cropped' , '64' ) : raise ValueError ( "CelebA format needs to be either " "'aligned_cropped' or '64'." ) if not output_filename : output_filename = 'celeba_{}.hdf5' . format ( which_format ) ... | Converts the CelebA dataset to HDF5 . |
55,931 | def disk_usage ( path ) : st = os . statvfs ( path ) total = st . f_blocks * st . f_frsize used = ( st . f_blocks - st . f_bfree ) * st . f_frsize return total , used | Return free usage about the given path in bytes . |
55,932 | def safe_mkdir ( folder_name , force_perm = None ) : if os . path . exists ( folder_name ) : return intermediary_folders = folder_name . split ( os . path . sep ) if intermediary_folders [ - 1 ] == "" : intermediary_folders = intermediary_folders [ : - 1 ] if force_perm : force_perm_path = folder_name . split ( os . pa... | Create the specified folder . |
55,933 | def check_enough_space ( dataset_local_dir , remote_fname , local_fname , max_disk_usage = 0.9 ) : storage_need = os . path . getsize ( remote_fname ) storage_total , storage_used = disk_usage ( dataset_local_dir ) return ( ( storage_used + storage_need ) < ( storage_total * max_disk_usage ) ) | Check if the given local folder has enough space . |
55,934 | def convert_cifar100 ( directory , output_directory , output_filename = 'cifar100.hdf5' ) : output_path = os . path . join ( output_directory , output_filename ) h5file = h5py . File ( output_path , mode = "w" ) input_file = os . path . join ( directory , 'cifar-100-python.tar.gz' ) tar_file = tarfile . open ( input_fi... | Converts the CIFAR - 100 dataset to HDF5 . |
55,935 | def verify_axis_labels ( self , expected , actual , source_name ) : if not getattr ( self , '_checked_axis_labels' , False ) : self . _checked_axis_labels = defaultdict ( bool ) if not self . _checked_axis_labels [ source_name ] : if actual is None : log . warning ( "%s instance could not verify (missing) axis " "expec... | Verify that axis labels for a given source are as expected . |
55,936 | def get_data ( self , request = None ) : if request is None : raise ValueError data = [ [ ] for _ in self . sources ] for i in range ( request ) : try : for source_data , example in zip ( data , next ( self . child_epoch_iterator ) ) : source_data . append ( example ) except StopIteration : if not self . strictness and... | Get data from the dataset . |
55,937 | def _producer_wrapper ( f , port , addr = 'tcp://127.0.0.1' ) : try : context = zmq . Context ( ) socket = context . socket ( zmq . PUSH ) socket . connect ( ':' . join ( [ addr , str ( port ) ] ) ) f ( socket ) finally : context . destroy ( ) | A shim that sets up a socket and starts the producer callable . |
55,938 | def _spawn_producer ( f , port , addr = 'tcp://127.0.0.1' ) : process = Process ( target = _producer_wrapper , args = ( f , port , addr ) ) process . start ( ) return process | Start a process that sends results on a PUSH socket . |
55,939 | def producer_consumer ( producer , consumer , addr = 'tcp://127.0.0.1' , port = None , context = None ) : context_created = False if context is None : context_created = True context = zmq . Context ( ) try : consumer_socket = context . socket ( zmq . PULL ) if port is None : port = consumer_socket . bind_to_random_port... | A producer - consumer pattern . |
55,940 | def convert_dogs_vs_cats ( directory , output_directory , output_filename = 'dogs_vs_cats.hdf5' ) : output_path = os . path . join ( output_directory , output_filename ) h5file = h5py . File ( output_path , mode = 'w' ) dtype = h5py . special_dtype ( vlen = numpy . dtype ( 'uint8' ) ) hdf_features = h5file . create_dat... | Converts the Dogs vs . Cats dataset to HDF5 . |
55,941 | def main ( args = None ) : built_in_datasets = dict ( downloaders . all_downloaders ) if fuel . config . extra_downloaders : for name in fuel . config . extra_downloaders : extra_datasets = dict ( importlib . import_module ( name ) . all_downloaders ) if any ( key in built_in_datasets for key in extra_datasets . keys (... | Entry point for fuel - download script . |
55,942 | def fill_subparser ( subparser ) : filenames = [ 'train-images-idx3-ubyte.gz' , 'train-labels-idx1-ubyte.gz' , 't10k-images-idx3-ubyte.gz' , 't10k-labels-idx1-ubyte.gz' ] urls = [ 'http://yann.lecun.com/exdb/mnist/' + f for f in filenames ] subparser . set_defaults ( urls = urls , filenames = filenames ) return default... | Sets up a subparser to download the MNIST dataset files . |
55,943 | def main ( args = None ) : parser = argparse . ArgumentParser ( description = 'Extracts metadata from a Fuel-converted HDF5 file.' ) parser . add_argument ( "filename" , help = "HDF5 file to analyze" ) args = parser . parse_args ( ) with h5py . File ( args . filename , 'r' ) as h5file : interface_version = h5file . att... | Entry point for fuel - info script . |
55,944 | def convert_silhouettes ( size , directory , output_directory , output_filename = None ) : if size not in ( 16 , 28 ) : raise ValueError ( 'size must be 16 or 28' ) if output_filename is None : output_filename = 'caltech101_silhouettes{}.hdf5' . format ( size ) output_file = os . path . join ( output_directory , output... | Convert the CalTech 101 Silhouettes Datasets . |
55,945 | def cross_validation ( scheme_class , num_examples , num_folds , strict = True , ** kwargs ) : if strict and num_examples % num_folds != 0 : raise ValueError ( ( "{} examples are not divisible in {} evenly-sized " + "folds. To allow this, have a look at the " + "`strict` argument." ) . format ( num_examples , num_folds... | Return pairs of schemes to be used for cross - validation . |
55,946 | def main ( args = None ) : built_in_datasets = dict ( converters . all_converters ) if fuel . config . extra_converters : for name in fuel . config . extra_converters : extra_datasets = dict ( importlib . import_module ( name ) . all_converters ) if any ( key in built_in_datasets for key in extra_datasets . keys ( ) ) ... | Entry point for fuel - convert script . |
55,947 | def refresh_lock ( lock_file ) : unique_id = '%s_%s_%s' % ( os . getpid ( ) , '' . join ( [ str ( random . randint ( 0 , 9 ) ) for i in range ( 10 ) ] ) , hostname ) try : lock_write = open ( lock_file , 'w' ) lock_write . write ( unique_id + '\n' ) lock_write . close ( ) except Exception : while get_lock . n_lock > 0 ... | Refresh an existing lock . |
55,948 | def get_lock ( lock_dir , ** kw ) : if not hasattr ( get_lock , 'n_lock' ) : get_lock . n_lock = 0 if not hasattr ( get_lock , 'lock_is_enabled' ) : get_lock . lock_is_enabled = True get_lock . lock_dir = lock_dir get_lock . unlocker = Unlocker ( get_lock . lock_dir ) else : if lock_dir != get_lock . lock_dir : assert ... | Obtain lock on compilation directory . |
55,949 | def release_lock ( ) : get_lock . n_lock -= 1 assert get_lock . n_lock >= 0 if get_lock . lock_is_enabled and get_lock . n_lock == 0 : get_lock . start_time = None get_lock . unlocker . unlock ( ) | Release lock on compilation directory . |
55,950 | def release_readlock ( lockdir_name ) : if os . path . exists ( lockdir_name ) and os . path . isdir ( lockdir_name ) : os . rmdir ( lockdir_name ) | Release a previously obtained readlock . |
55,951 | def get_readlock ( pid , path ) : timestamp = int ( time . time ( ) * 1e6 ) lockdir_name = "%s.readlock.%i.%i" % ( path , pid , timestamp ) os . mkdir ( lockdir_name ) atexit . register ( release_readlock , lockdir_name = lockdir_name ) | Obtain a readlock on a file . |
55,952 | def unlock ( self ) : try : self . os . remove ( self . os . path . join ( self . tmp_dir , 'lock' ) ) except Exception : pass try : self . os . rmdir ( self . tmp_dir ) except Exception : pass | Remove current lock . |
55,953 | def filename_from_url ( url , path = None ) : r = requests . get ( url , stream = True ) if 'Content-Disposition' in r . headers : filename = re . findall ( r'filename=([^;]+)' , r . headers [ 'Content-Disposition' ] ) [ 0 ] . strip ( '"\"' ) else : filename = os . path . basename ( urllib . parse . urlparse ( url ) . ... | Parses a URL to determine a file name . |
55,954 | def download ( url , file_handle , chunk_size = 1024 ) : r = requests . get ( url , stream = True ) total_length = r . headers . get ( 'content-length' ) if total_length is None : maxval = UnknownLength else : maxval = int ( total_length ) name = file_handle . name with progress_bar ( name = name , maxval = maxval ) as... | Downloads a given URL to a specific file . |
55,955 | def default_downloader ( directory , urls , filenames , url_prefix = None , clear = False ) : for i , url in enumerate ( urls ) : filename = filenames [ i ] if not filename : filename = filename_from_url ( url ) if not filename : raise ValueError ( "no filename available for URL '{}'" . format ( url ) ) filenames [ i ]... | Downloads or clears files from URLs and filenames . |
55,956 | def find_in_data_path ( filename ) : for path in config . data_path : path = os . path . expanduser ( os . path . expandvars ( path ) ) file_path = os . path . join ( path , filename ) if os . path . isfile ( file_path ) : return file_path raise IOError ( "{} not found in Fuel's data path" . format ( filename ) ) | Searches for a file within Fuel s data path . |
55,957 | def lazy_property_factory ( lazy_property ) : def lazy_property_getter ( self ) : if not hasattr ( self , '_' + lazy_property ) : self . load ( ) if not hasattr ( self , '_' + lazy_property ) : raise ValueError ( "{} wasn't loaded" . format ( lazy_property ) ) return getattr ( self , '_' + lazy_property ) def lazy_prop... | Create properties that perform lazy loading of attributes . |
55,958 | def do_not_pickle_attributes ( * lazy_properties ) : r def wrap_class ( cls ) : if not hasattr ( cls , 'load' ) : raise ValueError ( "no load method implemented" ) for lazy_property in lazy_properties : setattr ( cls , lazy_property , property ( * lazy_property_factory ( lazy_property ) ) ) if not hasattr ( cls , '__ge... | r Decorator to assign non - pickable properties . |
55,959 | def sorted_fancy_indexing ( indexable , request ) : if len ( request ) > 1 : indices = numpy . argsort ( request ) data = numpy . empty ( shape = ( len ( request ) , ) + indexable . shape [ 1 : ] , dtype = indexable . dtype ) data [ indices ] = indexable [ numpy . array ( request ) [ indices ] , ... ] else : data = ind... | Safe fancy indexing . |
55,960 | def slice_to_numerical_args ( slice_ , num_examples ) : start = slice_ . start if slice_ . start is not None else 0 stop = slice_ . stop if slice_ . stop is not None else num_examples step = slice_ . step if slice_ . step is not None else 1 return start , stop , step | Translate a slice s attributes into numerical attributes . |
55,961 | def get_list_representation ( self ) : if self . is_list : return self . list_or_slice else : return self [ list ( range ( self . num_examples ) ) ] | Returns this subset s representation as a list of indices . |
55,962 | def index_within_subset ( self , indexable , subset_request , sort_indices = False ) : if isinstance ( subset_request , numbers . Integral ) : request , = self [ [ subset_request ] ] else : request = self [ subset_request ] if isinstance ( request , numbers . Integral ) or hasattr ( request , 'step' ) : return indexabl... | Index an indexable object within the context of this subset . |
55,963 | def num_examples ( self ) : if self . is_list : return len ( self . list_or_slice ) else : start , stop , step = self . slice_to_numerical_args ( self . list_or_slice , self . original_num_examples ) return stop - start | The number of examples this subset spans . |
55,964 | def get_epoch_iterator ( self , ** kwargs ) : if not self . _fresh_state : self . next_epoch ( ) else : self . _fresh_state = False return super ( DataStream , self ) . get_epoch_iterator ( ** kwargs ) | Get an epoch iterator for the data stream . |
55,965 | def fill_subparser ( subparser ) : sets = [ 'train' , 'valid' , 'test' ] urls = [ 'http://www.cs.toronto.edu/~larocheh/public/datasets/' + 'binarized_mnist/binarized_mnist_{}.amat' . format ( s ) for s in sets ] filenames = [ 'binarized_mnist_{}.amat' . format ( s ) for s in sets ] subparser . set_defaults ( urls = url... | Sets up a subparser to download the binarized MNIST dataset files . |
55,966 | def download ( directory , youtube_id , clear = False ) : filepath = os . path . join ( directory , '{}.m4a' . format ( youtube_id ) ) if clear : os . remove ( filepath ) return if not PAFY_AVAILABLE : raise ImportError ( "pafy is required to download YouTube videos" ) url = 'https://www.youtube.com/watch?v={}' . forma... | Download the audio of a YouTube video . |
55,967 | def fill_subparser ( subparser ) : subparser . add_argument ( '--youtube-id' , type = str , required = True , help = ( "The YouTube ID of the video from which to extract audio, " "usually an 11-character string." ) ) return download | Sets up a subparser to download audio of YouTube videos . |
55,968 | def convert_youtube_audio ( directory , output_directory , youtube_id , channels , sample , output_filename = None ) : input_file = os . path . join ( directory , '{}.m4a' . format ( youtube_id ) ) wav_filename = '{}.wav' . format ( youtube_id ) wav_file = os . path . join ( directory , wav_filename ) ffmpeg_not_availa... | Converts downloaded YouTube audio to HDF5 format . |
55,969 | def fill_subparser ( subparser ) : subparser . add_argument ( '--youtube-id' , type = str , required = True , help = ( "The YouTube ID of the video from which to extract audio, " "usually an 11-character string." ) ) subparser . add_argument ( '--channels' , type = int , default = 1 , help = ( "The number of audio chan... | Sets up a subparser to convert YouTube audio files . |
55,970 | def convert_ilsvrc2012 ( directory , output_directory , output_filename = 'ilsvrc2012.hdf5' , shuffle_seed = config . default_seed ) : devkit_path = os . path . join ( directory , DEVKIT_ARCHIVE ) train , valid , test = [ os . path . join ( directory , fn ) for fn in IMAGE_TARS ] n_train , valid_groundtruth , n_test , ... | Converter for data from the ILSVRC 2012 competition . |
55,971 | def fill_subparser ( subparser ) : subparser . add_argument ( "--shuffle-seed" , help = "Seed to use for randomizing order of the " "training set on disk." , default = config . default_seed , type = int , required = False ) return convert_ilsvrc2012 | Sets up a subparser to convert the ILSVRC2012 dataset files . |
55,972 | def read_metadata_mat_file ( meta_mat ) : mat = loadmat ( meta_mat , squeeze_me = True ) synsets = mat [ 'synsets' ] new_dtype = numpy . dtype ( [ ( 'ILSVRC2012_ID' , numpy . int16 ) , ( 'WNID' , ( 'S' , max ( map ( len , synsets [ 'WNID' ] ) ) ) ) , ( 'wordnet_height' , numpy . int8 ) , ( 'gloss' , ( 'S' , max ( map (... | Read ILSVRC2012 metadata from the distributed MAT file . |
55,973 | def multiple_paths_parser ( value ) : if isinstance ( value , six . string_types ) : value = value . split ( os . path . pathsep ) return value | Parses data_path argument . |
55,974 | def add_config ( self , key , type_ , default = NOT_SET , env_var = None ) : self . config [ key ] = { 'type' : type_ } if env_var is not None : self . config [ key ] [ 'env_var' ] = env_var if default is not NOT_SET : self . config [ key ] [ 'default' ] = default | Add a configuration setting . |
55,975 | def send_arrays ( socket , arrays , stop = False ) : if arrays : arrays = [ numpy . ascontiguousarray ( array ) for array in arrays ] if stop : headers = { 'stop' : True } socket . send_json ( headers ) else : headers = [ header_data_from_array_1_0 ( array ) for array in arrays ] socket . send_json ( headers , zmq . SN... | Send NumPy arrays using the buffer interface and some metadata . |
55,976 | def recv_arrays ( socket ) : headers = socket . recv_json ( ) if 'stop' in headers : raise StopIteration arrays = [ ] for header in headers : data = socket . recv ( copy = False ) buf = buffer_ ( data ) array = numpy . frombuffer ( buf , dtype = numpy . dtype ( header [ 'descr' ] ) ) array . shape = header [ 'shape' ] ... | Receive a list of NumPy arrays . |
55,977 | def start_server ( data_stream , port = 5557 , hwm = 10 ) : logging . basicConfig ( level = 'INFO' ) context = zmq . Context ( ) socket = context . socket ( zmq . PUSH ) socket . set_hwm ( hwm ) socket . bind ( 'tcp://*:{}' . format ( port ) ) it = data_stream . get_epoch_iterator ( ) logger . info ( 'server started' )... | Start a data processing server . |
55,978 | def create_images ( raw_data_directory : str , destination_directory : str , stroke_thicknesses : List [ int ] , canvas_width : int = None , canvas_height : int = None , staff_line_spacing : int = 14 , staff_line_vertical_offsets : List [ int ] = None , random_position_on_canvas : bool = False ) -> dict : all_symbol_fi... | Creates a visual representation of the Homus Dataset by parsing all text - files and the symbols as specified by the parameters by drawing lines that connect the points from each stroke of each symbol . |
55,979 | def extract_and_render_all_symbol_masks ( self , raw_data_directory : str , destination_directory : str ) : print ( "Extracting Symbols from Muscima++ Dataset..." ) xml_files = self . get_all_xml_file_paths ( raw_data_directory ) crop_objects = self . load_crop_objects_from_xml_files ( xml_files ) self . render_masks_o... | Extracts all symbols from the raw XML documents and generates individual symbols from the masks |
55,980 | def invert_images ( self , image_directory : str , image_file_ending : str = "*.bmp" ) : image_paths = [ y for x in os . walk ( image_directory ) for y in glob ( os . path . join ( x [ 0 ] , image_file_ending ) ) ] for image_path in tqdm ( image_paths , desc = "Inverting all images in directory {0}" . format ( image_di... | In - situ converts the white on black images of a directory to black on white images |
55,981 | def create_capitan_images ( self , raw_data_directory : str , destination_directory : str , stroke_thicknesses : List [ int ] ) -> None : symbols = self . load_capitan_symbols ( raw_data_directory ) self . draw_capitan_stroke_images ( symbols , destination_directory , stroke_thicknesses ) self . draw_capitan_score_imag... | Creates a visual representation of the Capitan strokes by parsing all text - files and the symbols as specified by the parameters by drawing lines that connect the points from each stroke of each symbol . |
55,982 | def draw_capitan_stroke_images ( self , symbols : List [ CapitanSymbol ] , destination_directory : str , stroke_thicknesses : List [ int ] ) -> None : total_number_of_symbols = len ( symbols ) * len ( stroke_thicknesses ) output = "Generating {0} images with {1} symbols in {2} different stroke thicknesses ({3})" . form... | Creates a visual representation of the Capitan strokes by drawing lines that connect the points from each stroke of each symbol . |
55,983 | def overlap ( r1 : 'Rectangle' , r2 : 'Rectangle' ) : h_overlaps = ( r1 . left <= r2 . right ) and ( r1 . right >= r2 . left ) v_overlaps = ( r1 . bottom >= r2 . top ) and ( r1 . top <= r2 . bottom ) return h_overlaps and v_overlaps | Overlapping rectangles overlap both horizontally & vertically |
55,984 | def extract_symbols ( self , raw_data_directory : str , destination_directory : str ) : print ( "Extracting Symbols from Audiveris OMR Dataset..." ) all_xml_files = [ y for x in os . walk ( raw_data_directory ) for y in glob ( os . path . join ( x [ 0 ] , '*.xml' ) ) ] all_image_files = [ y for x in os . walk ( raw_dat... | Extracts the symbols from the raw XML documents and matching images of the Audiveris OMR dataset into individual symbols |
55,985 | def initialize_from_string ( content : str ) -> 'HomusSymbol' : if content is None or content is "" : return None lines = content . splitlines ( ) min_x = sys . maxsize max_x = 0 min_y = sys . maxsize max_y = 0 symbol_name = lines [ 0 ] strokes = [ ] for stroke_string in lines [ 1 : ] : stroke = [ ] for point_string in... | Create and initializes a new symbol from a string |
55,986 | def draw_into_bitmap ( self , export_path : ExportPath , stroke_thickness : int , margin : int = 0 ) -> None : self . draw_onto_canvas ( export_path , stroke_thickness , margin , self . dimensions . width + 2 * margin , self . dimensions . height + 2 * margin ) | Draws the symbol in the original size that it has plus an optional margin |
55,987 | def draw_onto_canvas ( self , export_path : ExportPath , stroke_thickness : int , margin : int , destination_width : int , destination_height : int , staff_line_spacing : int = 14 , staff_line_vertical_offsets : List [ int ] = None , bounding_boxes : dict = None , random_position_on_canvas : bool = False ) -> None : wi... | Draws the symbol onto a canvas with a fixed size |
55,988 | def update_locals ( locals_instance , instance_iterator , * args , ** kwargs ) : for instance in instance_iterator ( ) : locals_instance . update ( { type ( instance ) . __name__ : instance . __class__ } ) | import all of the detector classes into the local namespace to make it easy to do things like import scrubadub . detectors . NameDetector without having to add each new Detector or Filth |
55,989 | def iter_filth_clss ( ) : return iter_subclasses ( os . path . dirname ( os . path . abspath ( __file__ ) ) , Filth , _is_abstract_filth , ) | Iterate over all of the filths that are included in this sub - package . This is a convenience method for capturing all new Filth that are added over time . |
55,990 | def iter_filths ( ) : for filth_cls in iter_filth_clss ( ) : if issubclass ( filth_cls , RegexFilth ) : m = next ( re . finditer ( r"\s+" , "fake pattern string" ) ) yield filth_cls ( m ) else : yield filth_cls ( ) | Iterate over all instances of filth |
55,991 | def _update_content ( self , other_filth ) : if self . end < other_filth . beg or other_filth . end < self . beg : raise exceptions . FilthMergeError ( "a_filth goes from [%s, %s) and b_filth goes from [%s, %s)" % ( self . beg , self . end , other_filth . beg , other_filth . end ) ) if self . beg < other_filth . beg : ... | this updates the bounds text and placeholder for the merged filth |
55,992 | def add_detector ( self , detector_cls ) : if not issubclass ( detector_cls , detectors . base . Detector ) : raise TypeError ( ( '"%(detector_cls)s" is not a subclass of Detector' ) % locals ( ) ) name = detector_cls . filth_cls . type if name in self . _detectors : raise KeyError ( ( 'can not add Detector "%(name)s"-... | Add a Detector to scrubadub |
55,993 | def clean ( self , text , ** kwargs ) : if sys . version_info < ( 3 , 0 ) : if not isinstance ( text , unicode ) : raise exceptions . UnicodeRequired clean_chunks = [ ] filth = Filth ( ) for next_filth in self . iter_filth ( text ) : clean_chunks . append ( text [ filth . end : next_filth . beg ] ) clean_chunks . appen... | This is the master method that cleans all of the filth out of the dirty dirty text . All keyword arguments to this function are passed through to the Filth . replace_with method to fine - tune how the Filth is cleaned . |
55,994 | def iter_filth ( self , text ) : all_filths = [ ] for detector in self . _detectors . values ( ) : for filth in detector . iter_filth ( text ) : if not isinstance ( filth , Filth ) : raise TypeError ( 'iter_filth must always yield Filth' ) all_filths . append ( filth ) all_filths . sort ( key = lambda f : ( f . beg , -... | Iterate over the different types of filth that can exist . |
55,995 | async def download_file ( self , Bucket , Key , Filename , ExtraArgs = None , Callback = None , Config = None ) : with open ( Filename , 'wb' ) as open_file : await download_fileobj ( self , Bucket , Key , open_file , ExtraArgs = ExtraArgs , Callback = Callback , Config = Config ) | Download an S3 object to a file . |
55,996 | async def download_fileobj ( self , Bucket , Key , Fileobj , ExtraArgs = None , Callback = None , Config = None ) : try : resp = await self . get_object ( Bucket = Bucket , Key = Key ) except ClientError as err : if err . response [ 'Error' ] [ 'Code' ] == 'NoSuchKey' : raise ClientError ( { 'Error' : { 'Code' : '404' ... | Download an object from S3 to a file - like object . |
55,997 | async def upload_fileobj ( self , Fileobj : BinaryIO , Bucket : str , Key : str , ExtraArgs : Optional [ Dict [ str , Any ] ] = None , Callback : Optional [ Callable [ [ int ] , None ] ] = None , Config : Optional [ S3TransferConfig ] = None ) : if not ExtraArgs : ExtraArgs = { } multipart_chunksize = 8388608 if Config... | Upload a file - like object to S3 . |
55,998 | async def upload_file ( self , Filename , Bucket , Key , ExtraArgs = None , Callback = None , Config = None ) : with open ( Filename , 'rb' ) as open_file : await upload_fileobj ( self , open_file , Bucket , Key , ExtraArgs = ExtraArgs , Callback = Callback , Config = Config ) | Upload a file to an S3 object . |
55,999 | def _create_action ( factory_self , action_model , resource_name , service_context , is_load = False ) : action = AIOServiceAction ( action_model , factory = factory_self , service_context = service_context ) if is_load : async def do_action ( self , * args , ** kwargs ) : response = await action . async_call ( self , ... | Creates a new method which makes a request to the underlying AWS service . |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.