idx int64 0 63k | question stringlengths 61 4.03k | target stringlengths 6 1.23k |
|---|---|---|
1,600 | def i2c_bus_timeout ( self ) : ret = api . py_aa_i2c_bus_timeout ( self . handle , 0 ) _raise_error_if_negative ( ret ) return ret | I2C bus lock timeout in ms . |
1,601 | def i2c_master_write ( self , i2c_address , data , flags = I2C_NO_FLAGS ) : data = array . array ( 'B' , data ) status , _ = api . py_aa_i2c_write_ext ( self . handle , i2c_address , flags , len ( data ) , data ) _raise_i2c_status_code_error_if_failure ( status ) | Make an I2C write access . |
1,602 | def i2c_master_read ( self , addr , length , flags = I2C_NO_FLAGS ) : data = array . array ( 'B' , ( 0 , ) * length ) status , rx_len = api . py_aa_i2c_read_ext ( self . handle , addr , flags , length , data ) _raise_i2c_status_code_error_if_failure ( status ) del data [ rx_len : ] return bytes ( data ) | Make an I2C read access . |
1,603 | def poll ( self , timeout = None ) : if timeout is None : timeout = - 1 ret = api . py_aa_async_poll ( self . handle , timeout ) _raise_error_if_negative ( ret ) events = list ( ) for event in ( POLL_I2C_READ , POLL_I2C_WRITE , POLL_SPI , POLL_I2C_MONITOR ) : if ret & event : events . append ( event ) return events | Wait for an event to occur . |
1,604 | def enable_i2c_slave ( self , slave_address ) : ret = api . py_aa_i2c_slave_enable ( self . handle , slave_address , self . BUFFER_SIZE , self . BUFFER_SIZE ) _raise_error_if_negative ( ret ) | Enable I2C slave mode . |
1,605 | def i2c_slave_read ( self ) : data = array . array ( 'B' , ( 0 , ) * self . BUFFER_SIZE ) status , addr , rx_len = api . py_aa_i2c_slave_read_ext ( self . handle , self . BUFFER_SIZE , data ) _raise_i2c_status_code_error_if_failure ( status ) if addr == 0x80 : addr = 0x00 del data [ rx_len : ] return ( addr , bytes ( d... | Read the bytes from an I2C slave reception . |
1,606 | def i2c_slave_last_transmit_size ( self ) : ret = api . py_aa_i2c_slave_write_stats ( self . handle ) _raise_error_if_negative ( ret ) return ret | Returns the number of bytes transmitted by the slave . |
1,607 | def i2c_monitor_read ( self ) : data = array . array ( 'H' , ( 0 , ) * self . BUFFER_SIZE ) ret = api . py_aa_i2c_monitor_read ( self . handle , self . BUFFER_SIZE , data ) _raise_error_if_negative ( ret ) del data [ ret : ] return data . tolist ( ) | Retrieved any data fetched by the monitor . |
1,608 | def spi_bitrate ( self ) : ret = api . py_aa_spi_bitrate ( self . handle , 0 ) _raise_error_if_negative ( ret ) return ret | SPI bitrate in kHz . Not every bitrate is supported by the host adapter . Therefore the actual bitrate may be less than the value which is set . The slowest bitrate supported is 125kHz . Any smaller value will be rounded up to 125kHz . |
1,609 | def spi_configure ( self , polarity , phase , bitorder ) : ret = api . py_aa_spi_configure ( self . handle , polarity , phase , bitorder ) _raise_error_if_negative ( ret ) | Configure the SPI interface . |
1,610 | def spi_configure_mode ( self , spi_mode ) : if spi_mode == SPI_MODE_0 : self . spi_configure ( SPI_POL_RISING_FALLING , SPI_PHASE_SAMPLE_SETUP , SPI_BITORDER_MSB ) elif spi_mode == SPI_MODE_3 : self . spi_configure ( SPI_POL_FALLING_RISING , SPI_PHASE_SETUP_SAMPLE , SPI_BITORDER_MSB ) else : raise RuntimeError ( 'SPI ... | Configure the SPI interface by the well known SPI modes . |
1,611 | def spi_write ( self , data ) : data_out = array . array ( 'B' , data ) data_in = array . array ( 'B' , ( 0 , ) * len ( data_out ) ) ret = api . py_aa_spi_write ( self . handle , len ( data_out ) , data_out , len ( data_in ) , data_in ) _raise_error_if_negative ( ret ) return bytes ( data_in ) | Write a stream of bytes to a SPI device . |
1,612 | def spi_ss_polarity ( self , polarity ) : ret = api . py_aa_spi_master_ss_polarity ( self . handle , polarity ) _raise_error_if_negative ( ret ) | Change the ouput polarity on the SS line . |
1,613 | def edit_form ( self , obj ) : form = super ( OAISetModelView , self ) . edit_form ( obj ) del form . spec return form | Customize edit form . |
1,614 | def _schema_from_verb ( verb , partial = False ) : from . verbs import Verbs return getattr ( Verbs , verb ) ( partial = partial ) | Return an instance of schema for given verb . |
1,615 | def serialize ( pagination , ** kwargs ) : if not pagination . has_next : return token_builder = URLSafeTimedSerializer ( current_app . config [ 'SECRET_KEY' ] , salt = kwargs [ 'verb' ] , ) schema = _schema_from_verb ( kwargs [ 'verb' ] , partial = False ) data = dict ( seed = random . random ( ) , page = pagination .... | Return resumption token serializer . |
1,616 | def _deserialize ( self , value , attr , data ) : token_builder = URLSafeTimedSerializer ( current_app . config [ 'SECRET_KEY' ] , salt = data [ 'verb' ] , ) result = token_builder . loads ( value , max_age = current_app . config [ 'OAISERVER_RESUMPTION_TOKEN_EXPIRE_TIME' ] ) result [ 'token' ] = value result [ 'kwargs... | Serialize resumption token . |
1,617 | def load ( self , data , many = None , partial = None ) : result = super ( ResumptionTokenSchema , self ) . load ( data , many = many , partial = partial ) result . data . update ( result . data . get ( 'resumptionToken' , { } ) . get ( 'kwargs' , { } ) ) return result | Deserialize a data structure to an object . |
1,618 | def make_request_validator ( request ) : verb = request . values . get ( 'verb' , '' , type = str ) resumption_token = request . values . get ( 'resumptionToken' , None ) schema = Verbs if resumption_token is None else ResumptionVerbs return getattr ( schema , verb , OAISchema ) ( partial = False ) | Validate arguments in incomming request . |
1,619 | def from_iso_permissive ( datestring , use_dateutil = True ) : dateutil_available = False try : from dateutil import parser dateutil_available = True except ImportError : dateutil_available = False import datetime if dateutil_available and use_dateutil : return parser . parse ( datestring ) else : return datetime . dat... | Parse an ISO8601 - formatted datetime and return a datetime object . |
1,620 | def validate ( self , data ) : if 'verb' in data and data [ 'verb' ] != self . __class__ . __name__ : raise ValidationError ( 'This is not a valid OAI-PMH verb:{0}' . format ( data [ 'verb' ] ) , field_names = [ 'verb' ] , ) if 'from_' in data and 'until' in data and data [ 'from_' ] > data [ 'until' ] : raise Validati... | Check range between dates under keys from_ and until . |
1,621 | def sets ( self ) : if self . cache : return self . cache . get ( self . app . config [ 'OAISERVER_CACHE_KEY' ] ) | Get list of sets . |
1,622 | def sets ( self , values ) : if self . cache : self . cache . set ( self . app . config [ 'OAISERVER_CACHE_KEY' ] , values ) | Set list of sets . |
1,623 | def register_signals_oaiset ( self ) : from . models import OAISet from . receivers import after_insert_oai_set , after_update_oai_set , after_delete_oai_set listen ( OAISet , 'after_insert' , after_insert_oai_set ) listen ( OAISet , 'after_update' , after_update_oai_set ) listen ( OAISet , 'after_delete' , after_delet... | Register OAISet signals to update records . |
1,624 | def unregister_signals_oaiset ( self ) : from . models import OAISet from . receivers import after_insert_oai_set , after_update_oai_set , after_delete_oai_set if contains ( OAISet , 'after_insert' , after_insert_oai_set ) : remove ( OAISet , 'after_insert' , after_insert_oai_set ) remove ( OAISet , 'after_update' , af... | Unregister signals oaiset . |
1,625 | def extract_params ( params ) : values = [ ] if isinstance ( params , dict ) : for key , value in params . items ( ) : values . extend ( extract_params ( value ) ) elif isinstance ( params , list ) : for value in params : values . extend ( extract_params ( value ) ) else : values . append ( params ) return values | Extracts the values of a set of parameters recursing into nested dictionaries . |
1,626 | def get_list ( self , list_name , options = None ) : options = options or { } data = { 'list' : list_name } data . update ( options ) return self . api_get ( 'list' , data ) | Get detailed metadata information about a list . |
1,627 | def import_contacts ( self , email , password , include_name = False ) : data = { 'email' : email , 'password' : password } if include_name : data [ 'names' ] = 1 return self . api_post ( 'contacts' , data ) | Fetch email contacts from a user s address book on one of the major email websites . Currently supports AOL Gmail Hotmail and Yahoo! Mail . |
1,628 | def push_content ( self , title , url , images = None , date = None , expire_date = None , description = None , location = None , price = None , tags = None , author = None , site_name = None , spider = None , vars = None ) : vars = vars or { } data = { 'title' : title , 'url' : url } if images is not None : data [ 'im... | Push a new piece of content to Sailthru . |
1,629 | def delete_alert ( self , email , alert_id ) : data = { 'email' : email , 'alert_id' : alert_id } return self . api_delete ( 'alert' , data ) | delete user alert |
1,630 | def get_purchase ( self , purchase_id , purchase_key = 'sid' ) : data = { 'purchase_id' : purchase_id , 'purchase_key' : purchase_key } return self . api_get ( 'purchase' , data ) | Retrieve information about a purchase using the system s unique ID or a client s ID |
1,631 | def receive_verify_post ( self , post_params ) : if isinstance ( post_params , dict ) : required_params = [ 'action' , 'email' , 'send_id' , 'sig' ] if not self . check_for_valid_postback_actions ( required_params , post_params ) : return False else : return False if post_params [ 'action' ] != 'verify' : return False ... | Returns true if the incoming request is an authenticated verify post . |
1,632 | def receive_hardbounce_post ( self , post_params ) : if isinstance ( post_params , dict ) : required_params = [ 'action' , 'email' , 'sig' ] if not self . check_for_valid_postback_actions ( required_params , post_params ) : return False else : return False if post_params [ 'action' ] != 'hardbounce' : return False sign... | Hard bounce postbacks |
1,633 | def check_for_valid_postback_actions ( self , required_keys , post_params ) : for key in required_keys : if key not in post_params : return False return True | checks if post_params contain required keys |
1,634 | def api_get ( self , action , data , headers = None ) : return self . _api_request ( action , data , 'GET' , headers ) | Perform an HTTP GET request using the shared - secret auth hash . |
1,635 | def api_post ( self , action , data , binary_data_param = None ) : binary_data_param = binary_data_param or [ ] if binary_data_param : return self . api_post_multipart ( action , data , binary_data_param ) else : return self . _api_request ( action , data , 'POST' ) | Perform an HTTP POST request using the shared - secret auth hash . |
1,636 | def api_post_multipart ( self , action , data , binary_data_param ) : binary_data = { } data = data . copy ( ) try : file_handles = [ ] for param in binary_data_param : if param in data : binary_data [ param ] = file_handle = open ( data [ param ] , 'r' ) file_handles . append ( file_handle ) del data [ param ] json_pa... | Perform an HTTP Multipart POST request using the shared - secret auth hash . |
1,637 | def _api_request ( self , action , data , request_type , headers = None ) : if 'file' in data : file_data = { 'file' : open ( data [ 'file' ] , 'rb' ) } else : file_data = None return self . _http_request ( action , self . _prepare_json_payload ( data ) , request_type , file_data , headers ) | Make Request to Sailthru API with given data and api key format and signature hash |
1,638 | def validation_error ( exception ) : messages = getattr ( exception , 'messages' , None ) if messages is None : messages = getattr ( exception , 'data' , { 'messages' : None } ) [ 'messages' ] def extract_errors ( ) : if isinstance ( messages , dict ) : for field , message in messages . items ( ) : if field == 'verb' :... | Return formatter validation error . |
1,639 | def response ( args ) : e_tree = getattr ( xml , args [ 'verb' ] . lower ( ) ) ( ** args ) response = make_response ( etree . tostring ( e_tree , pretty_print = True , xml_declaration = True , encoding = 'UTF-8' , ) ) response . headers [ 'Content-Type' ] = 'text/xml' return response | Response endpoint . |
1,640 | def _create_percolator_mapping ( index , doc_type ) : if ES_VERSION [ 0 ] >= 5 : current_search_client . indices . put_mapping ( index = index , doc_type = doc_type , body = PERCOLATOR_MAPPING , ignore = [ 400 , 404 ] ) | Update mappings with the percolator field . |
1,641 | def _percolate_query ( index , doc_type , percolator_doc_type , document ) : if ES_VERSION [ 0 ] in ( 2 , 5 ) : results = current_search_client . percolate ( index = index , doc_type = doc_type , allow_no_indices = True , ignore_unavailable = True , body = { 'doc' : document } ) return results [ 'matches' ] elif ES_VER... | Get results for a percolate query . |
1,642 | def _new_percolator ( spec , search_pattern ) : if spec and search_pattern : query = query_string_parser ( search_pattern = search_pattern ) . to_dict ( ) for index in current_search . mappings . keys ( ) : percolator_doc_type = _get_percolator_doc_type ( index ) _create_percolator_mapping ( index , percolator_doc_type... | Create new percolator associated with the new set . |
1,643 | def _delete_percolator ( spec , search_pattern ) : if spec : for index in current_search . mappings . keys ( ) : percolator_doc_type = _get_percolator_doc_type ( index ) _create_percolator_mapping ( index , percolator_doc_type ) current_search_client . delete ( index = index , doc_type = percolator_doc_type , id = 'oai... | Delete percolator associated with the new oaiset . |
1,644 | def _build_cache ( ) : sets = current_oaiserver . sets if sets is None : sets = current_oaiserver . sets = [ oaiset . spec for oaiset in OAISet . query . filter ( OAISet . search_pattern . is_ ( None ) ) . all ( ) ] return sets | Build sets cache . |
1,645 | def get_record_sets ( record ) : record_sets = set ( record . get ( '_oai' , { } ) . get ( 'sets' , [ ] ) ) for spec in _build_cache ( ) : if spec in record_sets : yield spec index , doc_type = RecordIndexer ( ) . record_to_index ( record ) document = record . dumps ( ) percolator_doc_type = _get_percolator_doc_type ( ... | Find matching sets . |
1,646 | def _records_commit ( record_ids ) : for record_id in record_ids : record = Record . get_record ( record_id ) record . commit ( ) | Commit all records . |
1,647 | def update_affected_records ( spec = None , search_pattern = None ) : chunk_size = current_app . config [ 'OAISERVER_CELERY_TASK_CHUNK_SIZE' ] record_ids = get_affected_records ( spec = spec , search_pattern = search_pattern ) group ( update_records_sets . s ( list ( filter ( None , chunk ) ) ) for chunk in zip_longest... | Update all affected records by OAISet change . |
1,648 | def envelope ( ** kwargs ) : e_oaipmh = Element ( etree . QName ( NS_OAIPMH , 'OAI-PMH' ) , nsmap = NSMAP ) e_oaipmh . set ( etree . QName ( NS_XSI , 'schemaLocation' ) , '{0} {1}' . format ( NS_OAIPMH , NS_OAIPMH_XSD ) ) e_tree = ElementTree ( element = e_oaipmh ) if current_app . config [ 'OAISERVER_XSL_URL' ] : e_oa... | Create OAI - PMH envelope for response . |
1,649 | def error ( errors ) : e_tree , e_oaipmh = envelope ( ) for code , message in errors : e_error = SubElement ( e_oaipmh , etree . QName ( NS_OAIPMH , 'error' ) ) e_error . set ( 'code' , code ) e_error . text = message return e_tree | Create error element . |
1,650 | def verb ( ** kwargs ) : e_tree , e_oaipmh = envelope ( ** kwargs ) e_element = SubElement ( e_oaipmh , etree . QName ( NS_OAIPMH , kwargs [ 'verb' ] ) ) return e_tree , e_element | Create OAI - PMH envelope for response with verb . |
1,651 | def resumption_token ( parent , pagination , ** kwargs ) : if pagination . page == 1 and not pagination . has_next : return token = serialize ( pagination , ** kwargs ) e_resumptionToken = SubElement ( parent , etree . QName ( NS_OAIPMH , 'resumptionToken' ) ) if pagination . total : expiration_date = datetime . utcnow... | Attach resumption token element to a parent . |
1,652 | def listsets ( ** kwargs ) : e_tree , e_listsets = verb ( ** kwargs ) page = kwargs . get ( 'resumptionToken' , { } ) . get ( 'page' , 1 ) size = current_app . config [ 'OAISERVER_PAGE_SIZE' ] oai_sets = OAISet . query . paginate ( page = page , per_page = size , error_out = False ) for oai_set in oai_sets . items : e_... | Create OAI - PMH response for ListSets verb . |
1,653 | def listmetadataformats ( ** kwargs ) : cfg = current_app . config e_tree , e_listmetadataformats = verb ( ** kwargs ) if 'identifier' in kwargs : OAIIDProvider . get ( pid_value = kwargs [ 'identifier' ] ) for prefix , metadata in cfg . get ( 'OAISERVER_METADATA_FORMATS' , { } ) . items ( ) : e_metadataformat = SubEle... | Create OAI - PMH response for ListMetadataFormats verb . |
1,654 | def listidentifiers ( ** kwargs ) : e_tree , e_listidentifiers = verb ( ** kwargs ) result = get_records ( ** kwargs ) for record in result . items : pid = oaiid_fetcher ( record [ 'id' ] , record [ 'json' ] [ '_source' ] ) header ( e_listidentifiers , identifier = pid . pid_value , datestamp = record [ 'updated' ] , s... | Create OAI - PMH response for verb ListIdentifiers . |
1,655 | def listrecords ( ** kwargs ) : record_dumper = serializer ( kwargs [ 'metadataPrefix' ] ) e_tree , e_listrecords = verb ( ** kwargs ) result = get_records ( ** kwargs ) for record in result . items : pid = oaiid_fetcher ( record [ 'id' ] , record [ 'json' ] [ '_source' ] ) e_record = SubElement ( e_listrecords , etree... | Create OAI - PMH response for verb ListRecords . |
1,656 | def oaiid_fetcher ( record_uuid , data ) : pid_value = data . get ( '_oai' , { } ) . get ( 'id' ) if pid_value is None : raise PersistentIdentifierError ( ) return FetchedPID ( provider = OAIIDProvider , pid_type = OAIIDProvider . pid_type , pid_value = str ( pid_value ) , ) | Fetch a record s identifier . |
1,657 | def validate_spec ( self , key , value ) : if self . spec and self . spec != value : raise OAISetSpecUpdateError ( "Updating spec is not allowed." ) return value | Forbit updates of set identifier . |
1,658 | def add_record ( self , record ) : record . setdefault ( '_oai' , { } ) . setdefault ( 'sets' , [ ] ) assert not self . has_record ( record ) record [ '_oai' ] [ 'sets' ] . append ( self . spec ) | Add a record to the OAISet . |
1,659 | def remove_record ( self , record ) : assert self . has_record ( record ) record [ '_oai' ] [ 'sets' ] = [ s for s in record [ '_oai' ] [ 'sets' ] if s != self . spec ] | Remove a record from the OAISet . |
1,660 | def oaiserver ( sets , records ) : from invenio_db import db from invenio_oaiserver . models import OAISet from invenio_records . api import Record with db . session . begin_nested ( ) : for i in range ( sets ) : db . session . add ( OAISet ( spec = 'test{0}' . format ( i ) , name = 'Test{0}' . format ( i ) , descripti... | Initialize OAI - PMH server . |
1,661 | def serializer ( metadata_prefix ) : metadataFormats = current_app . config [ 'OAISERVER_METADATA_FORMATS' ] serializer_ = metadataFormats [ metadata_prefix ] [ 'serializer' ] if isinstance ( serializer_ , tuple ) : return partial ( import_string ( serializer_ [ 0 ] ) , ** serializer_ [ 1 ] ) return import_string ( ser... | Return etree_dumper instances . |
1,662 | def dumps_etree ( pid , record , ** kwargs ) : from dojson . contrib . to_marc21 import to_marc21 from dojson . contrib . to_marc21 . utils import dumps_etree return dumps_etree ( to_marc21 . do ( record [ '_source' ] ) , ** kwargs ) | Dump MARC21 compatible record . |
1,663 | def eprints_description ( metadataPolicy , dataPolicy , submissionPolicy = None , content = None ) : eprints = Element ( etree . QName ( NS_EPRINTS [ None ] , 'eprints' ) , nsmap = NS_EPRINTS ) eprints . set ( etree . QName ( ns [ 'xsi' ] , 'schemaLocation' ) , '{0} {1}' . format ( EPRINTS_SCHEMA_LOCATION , EPRINTS_SCH... | Generate the eprints element for the identify response . |
1,664 | def oai_identifier_description ( scheme , repositoryIdentifier , delimiter , sampleIdentifier ) : oai_identifier = Element ( etree . QName ( NS_OAI_IDENTIFIER [ None ] , 'oai_identifier' ) , nsmap = NS_OAI_IDENTIFIER ) oai_identifier . set ( etree . QName ( ns [ 'xsi' ] , 'schemaLocation' ) , '{0} {1}' . format ( OAI_I... | Generate the oai - identifier element for the identify response . |
1,665 | def friends_description ( baseURLs ) : friends = Element ( etree . QName ( NS_FRIENDS [ None ] , 'friends' ) , nsmap = NS_FRIENDS ) friends . set ( etree . QName ( ns [ 'xsi' ] , 'schemaLocation' ) , '{0} {1}' . format ( FRIENDS_SCHEMA_LOCATION , FRIENDS_SCHEMA_LOCATION_XSD ) ) for baseURL in baseURLs : friends . appen... | Generate the friends element for the identify response . |
1,666 | def after_insert_oai_set ( mapper , connection , target ) : _new_percolator ( spec = target . spec , search_pattern = target . search_pattern ) sleep ( 2 ) update_affected_records . delay ( search_pattern = target . search_pattern ) | Update records on OAISet insertion . |
1,667 | def after_update_oai_set ( mapper , connection , target ) : _delete_percolator ( spec = target . spec , search_pattern = target . search_pattern ) _new_percolator ( spec = target . spec , search_pattern = target . search_pattern ) sleep ( 2 ) update_affected_records . delay ( spec = target . spec , search_pattern = tar... | Update records on OAISet update . |
1,668 | def after_delete_oai_set ( mapper , connection , target ) : _delete_percolator ( spec = target . spec , search_pattern = target . search_pattern ) sleep ( 2 ) update_affected_records . delay ( spec = target . spec ) | Update records on OAISet deletion . |
1,669 | def query_string_parser ( search_pattern ) : if not hasattr ( current_oaiserver , 'query_parser' ) : query_parser = current_app . config [ 'OAISERVER_QUERY_PARSER' ] if isinstance ( query_parser , six . string_types ) : query_parser = import_string ( query_parser ) current_oaiserver . query_parser = query_parser return... | Elasticsearch query string parser . |
1,670 | def get_affected_records ( spec = None , search_pattern = None ) : if spec is None and search_pattern is None : raise StopIteration queries = [ ] if spec is not None : queries . append ( Q ( 'match' , ** { '_oai.sets' : spec } ) ) if search_pattern : queries . append ( query_string_parser ( search_pattern = search_patt... | Get list of affected records . |
1,671 | def get_records ( ** kwargs ) : page_ = kwargs . get ( 'resumptionToken' , { } ) . get ( 'page' , 1 ) size_ = current_app . config [ 'OAISERVER_PAGE_SIZE' ] scroll = current_app . config [ 'OAISERVER_RESUMPTION_TOKEN_EXPIRE_TIME' ] scroll_id = kwargs . get ( 'resumptionToken' , { } ) . get ( 'scroll_id' ) if scroll_id ... | Get records paginated . |
1,672 | def get_file_path ( filename , local = True , relative_to_module = None , my_dir = my_dir ) : if relative_to_module is not None : my_dir = os . path . dirname ( relative_to_module . __file__ ) user_path = result = filename if local : user_path = os . path . expanduser ( filename ) result = os . path . abspath ( user_pa... | Look for an existing path matching filename . Try to resolve relative to the module location if the path cannot by found using normal resolution . |
1,673 | def load_if_not_loaded ( widget , filenames , verbose = False , delay = 0.1 , force = False , local = True , evaluator = None ) : if evaluator is None : evaluator = EVALUATOR for filename in filenames : loaded = False if force or not filename in LOADED_JAVASCRIPT : js_text = get_text_from_file_name ( filename , local )... | Load a javascript file to the Jupyter notebook context unless it was already loaded . |
1,674 | def _set ( self , name , value ) : "Proxy to set a property of the widget element." return self . widget ( self . widget_element . _set ( name , value ) ) | Proxy to set a property of the widget element . |
1,675 | def base_url ( root ) : for attr , value in root . attrib . iteritems ( ) : if attr . endswith ( 'base' ) and 'http' in value : return value return None | Determine the base url for a root element . |
1,676 | def clean_ns ( tag ) : if '}' in tag : split = tag . split ( '}' ) return split [ 0 ] . strip ( '{' ) , split [ - 1 ] return '' , tag | Return a tag and its namespace separately . |
1,677 | def xpath ( node , query , namespaces = { } ) : if namespaces and 'None' not in namespaces : return node . xpath ( query , namespaces = namespaces ) return node . xpath ( query ) | A safe xpath that only uses namespaces if available . |
1,678 | def innertext ( node ) : if not len ( node ) : return node . text return ( node . text or '' ) + '' . join ( [ etree . tostring ( c ) for c in node ] ) + ( node . tail or '' ) | Return the inner text of a node . If a node has no sub elements this is just node . text . Otherwise it s node . text + sub - element - text + node . tail . |
1,679 | def parse ( document , clean_html = True , unix_timestamp = False , encoding = None ) : if isinstance ( clean_html , bool ) : cleaner = default_cleaner if clean_html else fake_cleaner else : cleaner = clean_html result = feedparser . FeedParserDict ( ) result [ 'feed' ] = feedparser . FeedParserDict ( ) result [ 'entri... | Parse a document and return a feedparser dictionary with attr key access . If clean_html is False the html in the feed will not be cleaned . If clean_html is True a sane version of lxml . html . clean . Cleaner will be used . If it is a Cleaner object that cleaner will be used . If unix_timestamp is True the date infor... |
1,680 | def changed_path ( self ) : "Find any changed path and update all changed modification times." result = None for path in self . paths_to_modification_times : lastmod = self . paths_to_modification_times [ path ] mod = os . path . getmtime ( path ) if mod > lastmod : result = "Watch file has been modified: " + repr ( pa... | Find any changed path and update all changed modification times . |
1,681 | def _parse_date_iso8601 ( dateString ) : m = None for _iso8601_match in _iso8601_matches : m = _iso8601_match ( dateString ) if m : break if not m : return if m . span ( ) == ( 0 , 0 ) : return params = m . groupdict ( ) ordinal = params . get ( 'ordinal' , 0 ) if ordinal : ordinal = int ( ordinal ) else : ordinal = 0 ... | Parse a variety of ISO - 8601 - compatible formats like 20040105 |
1,682 | def _parse_date_onblog ( dateString ) : m = _korean_onblog_date_re . match ( dateString ) if not m : return w3dtfdate = '%(year)s-%(month)s-%(day)sT%(hour)s:%(minute)s:%(second)s%(zonediff)s' % { 'year' : m . group ( 1 ) , 'month' : m . group ( 2 ) , 'day' : m . group ( 3 ) , 'hour' : m . group ( 4 ) , 'minute' : m . g... | Parse a string according to the OnBlog 8 - bit date format |
1,683 | def _parse_date_nate ( dateString ) : m = _korean_nate_date_re . match ( dateString ) if not m : return hour = int ( m . group ( 5 ) ) ampm = m . group ( 4 ) if ( ampm == _korean_pm ) : hour += 12 hour = str ( hour ) if len ( hour ) == 1 : hour = '0' + hour w3dtfdate = '%(year)s-%(month)s-%(day)sT%(hour)s:%(minute)s:%(... | Parse a string according to the Nate 8 - bit date format |
1,684 | def _parse_date_greek ( dateString ) : m = _greek_date_format_re . match ( dateString ) if not m : return wday = _greek_wdays [ m . group ( 1 ) ] month = _greek_months [ m . group ( 3 ) ] rfc822date = '%(wday)s, %(day)s %(month)s %(year)s %(hour)s:%(minute)s:%(second)s %(zonediff)s' % { 'wday' : wday , 'day' : m . grou... | Parse a string according to a Greek 8 - bit date format . |
1,685 | def _parse_date_hungarian ( dateString ) : m = _hungarian_date_format_re . match ( dateString ) if not m or m . group ( 2 ) not in _hungarian_months : return None month = _hungarian_months [ m . group ( 2 ) ] day = m . group ( 3 ) if len ( day ) == 1 : day = '0' + day hour = m . group ( 4 ) if len ( hour ) == 1 : hour ... | Parse a string according to a Hungarian 8 - bit date format . |
1,686 | def parse_date ( dateString ) : if not dateString : return None for handler in _date_handlers : try : date9tuple = handler ( dateString ) except ( KeyError , OverflowError , ValueError ) : continue if not date9tuple : continue if len ( date9tuple ) != 9 : continue return date9tuple return None | Parses a variety of date formats into a 9 - tuple in GMT |
1,687 | def handle_chunk_wrapper ( self , status , name , content , file_info ) : out = self . output if out is not None : with out : print ( "handling chunk " + repr ( type ( content ) ) ) self . handle_chunk ( status , name , content , file_info ) else : self . handle_chunk ( status , name , content , file_info ) | wrapper to allow output redirects for handle_chunk . |
1,688 | def handle_chunk ( self , status , name , content , file_info ) : "Handle one chunk of the file. Override this method for peicewise delivery or error handling." if status == "error" : msg = repr ( file_info . get ( "message" ) ) exc = JavaScriptError ( msg ) exc . file_info = file_info self . status = "Javascript sent... | Handle one chunk of the file . Override this method for peicewise delivery or error handling . |
1,689 | def search ( self , query , method = "lucene" , start = None , rows = None , access_token = None ) : if access_token is None : access_token = self . get_search_token_from_orcid ( ) headers = { 'Accept' : 'application/orcid+json' , 'Authorization' : 'Bearer %s' % access_token } return self . _search ( query , method , s... | Search the ORCID database . |
1,690 | def search_generator ( self , query , method = "lucene" , pagination = 10 , access_token = None ) : if access_token is None : access_token = self . get_search_token_from_orcid ( ) headers = { 'Accept' : 'application/orcid+json' , 'Authorization' : 'Bearer %s' % access_token } index = 0 while True : paginated_result = s... | Search the ORCID database with a generator . |
1,691 | def get_search_token_from_orcid ( self , scope = '/read-public' ) : payload = { 'client_id' : self . _key , 'client_secret' : self . _secret , 'scope' : scope , 'grant_type' : 'client_credentials' } url = "%s/oauth/token" % self . _endpoint headers = { 'Accept' : 'application/json' } response = requests . post ( url , ... | Get a token for searching ORCID records . |
1,692 | def get_token_from_authorization_code ( self , authorization_code , redirect_uri ) : token_dict = { "client_id" : self . _key , "client_secret" : self . _secret , "grant_type" : "authorization_code" , "code" : authorization_code , "redirect_uri" : redirect_uri , } response = requests . post ( self . _token_url , data =... | Like get_token but using an OAuth 2 authorization code . |
1,693 | def read_record_public ( self , orcid_id , request_type , token , put_code = None , accept_type = 'application/orcid+json' ) : return self . _get_info ( orcid_id , self . _get_public_info , request_type , token , put_code , accept_type ) | Get the public info about the researcher . |
1,694 | def get_user_orcid ( self , user_id , password , redirect_uri ) : response = self . _authenticate ( user_id , password , redirect_uri , '/authenticate' ) return response [ 'orcid' ] | Get the user orcid from authentication process . |
1,695 | def read_record_member ( self , orcid_id , request_type , token , put_code = None , accept_type = 'application/orcid+json' ) : return self . _get_info ( orcid_id , self . _get_member_info , request_type , token , put_code , accept_type ) | Get the member info about the researcher . |
1,696 | def get_access_tokens ( self , authorization_code ) : response = self . box_request . get_access_token ( authorization_code ) try : att = response . json ( ) except Exception , ex : raise BoxHttpResponseError ( ex ) if response . status_code >= 400 : raise BoxError ( response . status_code , att ) return att [ 'access_... | From the authorization code get the access token and the refresh token from Box . |
1,697 | def unpack_frame ( message ) : body = [ ] returned = dict ( cmd = '' , headers = { } , body = '' ) breakdown = message . split ( '\n' ) returned [ 'cmd' ] = breakdown [ 0 ] breakdown = breakdown [ 1 : ] def headD ( field ) : index = field . find ( ':' ) if index : header = field [ : index ] . strip ( ) data = field [ i... | Called to unpack a STOMP message into a dictionary . |
1,698 | def ack ( messageid , transactionid = None ) : header = 'message-id: %s' % messageid if transactionid : header = 'message-id: %s\ntransaction: %s' % ( messageid , transactionid ) return "ACK\n%s\n\n\x00\n" % header | STOMP acknowledge command . |
1,699 | def send ( dest , msg , transactionid = None ) : transheader = '' if transactionid : transheader = 'transaction: %s\n' % transactionid return "SEND\ndestination: %s\n%s\n%s\x00\n" % ( dest , transheader , msg ) | STOMP send command . |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.