idx int64 0 252k | question stringlengths 48 5.28k | target stringlengths 5 1.23k |
|---|---|---|
248,100 | def save ( self ) : new_user = super ( SignupFormExtra , self ) . save ( ) new_user . first_name = self . cleaned_data [ 'first_name' ] new_user . last_name = self . cleaned_data [ 'last_name' ] new_user . save ( ) return new_user | Override the save method to save the first and last name to the user field . |
248,101 | def save ( self , sender ) : um_to_user_list = self . cleaned_data [ 'to' ] body = self . cleaned_data [ 'body' ] msg = Message . objects . send_message ( sender , um_to_user_list , body ) return msg | Save the message and send it out into the wide world . |
248,102 | def clean_username ( self ) : try : user = get_user_model ( ) . objects . get ( username__iexact = self . cleaned_data [ 'username' ] ) except get_user_model ( ) . DoesNotExist : pass else : if userena_settings . USERENA_ACTIVATION_REQUIRED and UserenaSignup . objects . filter ( user__username__iexact = self . cleaned_... | Validate that the username is alphanumeric and is not already in use . Also validates that the username is not listed in USERENA_FORBIDDEN_USERNAMES list . |
248,103 | def save ( self ) : while True : username = sha1 ( str ( random . random ( ) ) . encode ( 'utf-8' ) ) . hexdigest ( ) [ : 5 ] try : get_user_model ( ) . objects . get ( username__iexact = username ) except get_user_model ( ) . DoesNotExist : break self . cleaned_data [ 'username' ] = username return super ( SignupFormO... | Generate a random username before falling back to parent signup form |
248,104 | def parse_file ( self , sourcepath ) : with open ( sourcepath , 'r' ) as logfile : jsonlist = logfile . readlines ( ) data = { } data [ 'entries' ] = [ ] for line in jsonlist : entry = self . parse_line ( line ) data [ 'entries' ] . append ( entry ) if self . tzone : for e in data [ 'entries' ] : e [ 'tzone' ] = self .... | Parse an object - per - line JSON file into a log data dict |
248,105 | def parse_file ( self , sourcepath ) : with open ( sourcepath , 'r' ) as logfile : jsonstr = logfile . read ( ) data = { } data [ 'entries' ] = json . loads ( jsonstr ) if self . tzone : for e in data [ 'entries' ] : e [ 'tzone' ] = self . tzone return data | Parse single JSON object into a LogData object |
248,106 | def run_job ( self ) : try : self . load_parsers ( ) self . load_filters ( ) self . load_outputs ( ) self . config_args ( ) if self . args . list_parsers : self . list_parsers ( ) if self . args . verbosemode : print ( 'Loading input files' ) self . load_inputs ( ) if self . args . verbosemode : print ( 'Running parser... | Execute a logdissect job |
248,107 | def run_parse ( self ) : parsedset = { } parsedset [ 'data_set' ] = [ ] for log in self . input_files : parsemodule = self . parse_modules [ self . args . parser ] try : if self . args . tzone : parsemodule . tzone = self . args . tzone except NameError : pass parsedset [ 'data_set' ] . append ( parsemodule . parse_fil... | Parse one or more log files |
248,108 | def run_output ( self ) : for f in logdissect . output . __formats__ : ouroutput = self . output_modules [ f ] ouroutput . write_output ( self . data_set [ 'finalized_data' ] , args = self . args ) del ( ouroutput ) if not self . args . silentmode : if self . args . verbosemode : print ( '\n==== ++++ ==== Output: ==== ... | Output finalized data |
248,109 | def config_args ( self ) : self . arg_parser . add_argument ( '--version' , action = 'version' , version = '%(prog)s ' + str ( __version__ ) ) self . arg_parser . add_argument ( '--verbose' , action = 'store_true' , dest = 'verbosemode' , help = _ ( 'set verbose terminal output' ) ) self . arg_parser . add_argument ( '... | Set config options |
248,110 | def load_inputs ( self ) : for f in self . args . files : if os . path . isfile ( f ) : fparts = str ( f ) . split ( '.' ) if fparts [ - 1 ] == 'gz' : if self . args . unzip : fullpath = os . path . abspath ( str ( f ) ) self . input_files . append ( fullpath ) else : return 0 elif fparts [ - 1 ] == 'bz2' or fparts [ -... | Load the specified inputs |
248,111 | def list_parsers ( self , * args ) : print ( '==== Available parsing modules: ====\n' ) for parser in sorted ( self . parse_modules ) : print ( self . parse_modules [ parser ] . name . ljust ( 16 ) + ': ' + self . parse_modules [ parser ] . desc ) sys . exit ( 0 ) | Return a list of available parsing modules |
248,112 | def get_utc_date ( entry ) : if entry [ 'numeric_date_stamp' ] == '0' : entry [ 'numeric_date_stamp_utc' ] = '0' return entry else : if '.' in entry [ 'numeric_date_stamp' ] : t = datetime . strptime ( entry [ 'numeric_date_stamp' ] , '%Y%m%d%H%M%S.%f' ) else : t = datetime . strptime ( entry [ 'numeric_date_stamp' ] ,... | Return datestamp converted to UTC |
248,113 | def get_local_tzone ( ) : if localtime ( ) . tm_isdst : if altzone < 0 : tzone = '+' + str ( int ( float ( altzone ) / 60 // 60 ) ) . rjust ( 2 , '0' ) + str ( int ( float ( altzone ) / 60 % 60 ) ) . ljust ( 2 , '0' ) else : tzone = '-' + str ( int ( float ( altzone ) / 60 // 60 ) ) . rjust ( 2 , '0' ) + str ( int ( fl... | Get the current time zone on the local host |
248,114 | def merge_logs ( dataset , sort = True ) : ourlog = { } ourlog [ 'entries' ] = [ ] for d in dataset : ourlog [ 'entries' ] = ourlog [ 'entries' ] + d [ 'entries' ] if sort : ourlog [ 'entries' ] . sort ( key = lambda x : x [ 'numeric_date_stamp_utc' ] ) return ourlog | Merge log dictionaries together into one log dictionary |
248,115 | def write_output ( self , data , args = None , filename = None , label = None ) : if args : if not args . outlog : return 0 if not filename : filename = args . outlog lastpath = '' with open ( str ( filename ) , 'w' ) as output_file : for entry in data [ 'entries' ] : if args . label : if entry [ 'source_path' ] == las... | Write log data to a log file |
248,116 | def write_output ( self , data , args = None , filename = None , pretty = False ) : if args : if not args . sojson : return 0 pretty = args . pretty if not filename : filename = args . sojson if pretty : logstring = json . dumps ( data [ 'entries' ] , indent = 2 , sort_keys = True , separators = ( ',' , ': ' ) ) else :... | Write log data to a single JSON object |
248,117 | def write_output ( self , data , filename = None , args = None ) : if args : if not args . linejson : return 0 if not filename : filename = args . linejson entrylist = [ ] for entry in data [ 'entries' ] : entrystring = json . dumps ( entry , sort_keys = True ) entrylist . append ( entrystring ) with open ( str ( filen... | Write log data to a file with one JSON object per line |
248,118 | def parse_file ( self , sourcepath ) : self . date_regex = re . compile ( r'{}' . format ( self . format_regex ) ) if self . backup_format_regex : self . backup_date_regex = re . compile ( r'{}' . format ( self . backup_format_regex ) ) data = { } data [ 'entries' ] = [ ] data [ 'parser' ] = self . name data [ 'source_... | Parse a file into a LogData object |
248,119 | def parse_line ( self , line ) : match = re . findall ( self . date_regex , line ) if match : fields = self . fields elif self . backup_format_regex and not match : match = re . findall ( self . backup_date_regex , line ) fields = self . backup_fields if match : entry = { } entry [ 'raw_text' ] = line entry [ 'parser' ... | Parse a line into a dictionary |
248,120 | def post_parse_action ( self , entry ) : if 'source_host' in entry . keys ( ) : host = self . ip_port_regex . findall ( entry [ 'source_host' ] ) if host : hlist = host [ 0 ] . split ( '.' ) entry [ 'source_host' ] = '.' . join ( hlist [ : 4 ] ) entry [ 'source_port' ] = hlist [ - 1 ] if 'dest_host' in entry . keys ( )... | separate hosts and ports after entry is parsed |
248,121 | def find_partition_multiplex ( graphs , partition_type , ** kwargs ) : n_layers = len ( graphs ) partitions = [ ] layer_weights = [ 1 ] * n_layers for graph in graphs : partitions . append ( partition_type ( graph , ** kwargs ) ) optimiser = Optimiser ( ) improvement = optimiser . optimise_partition_multiplex ( partiti... | Detect communities for multiplex graphs . |
248,122 | def find_partition_temporal ( graphs , partition_type , interslice_weight = 1 , slice_attr = 'slice' , vertex_id_attr = 'id' , edge_type_attr = 'type' , weight_attr = 'weight' , ** kwargs ) : G_layers , G_interslice , G = time_slices_to_layers ( graphs , interslice_weight , slice_attr = slice_attr , vertex_id_attr = ve... | Detect communities for temporal graphs . |
248,123 | def build_ext ( self ) : try : from setuptools . command . build_ext import build_ext except ImportError : from distutils . command . build_ext import build_ext buildcfg = self class custom_build_ext ( build_ext ) : def run ( self ) : if buildcfg . use_pkgconfig : detected = buildcfg . detect_from_pkgconfig ( ) else : ... | Returns a class that can be used as a replacement for the build_ext command in distutils and that will download and compile the C core of igraph if needed . |
248,124 | def Bipartite ( graph , resolution_parameter_01 , resolution_parameter_0 = 0 , resolution_parameter_1 = 0 , degree_as_node_size = False , types = 'type' , ** kwargs ) : if types is not None : if isinstance ( types , str ) : types = graph . vs [ types ] else : types = list ( types ) if set ( types ) != set ( [ 0 , 1 ] )... | Create three layers for bipartite partitions . |
248,125 | def spacing_file ( path ) : with open ( os . path . abspath ( path ) ) as f : return spacing_text ( f . read ( ) ) | Perform paranoid text spacing from file . |
248,126 | def compute ( self , text , lang = "eng" ) : params = { "lang" : lang , "text" : text , "topClustersCount" : self . _nrOfEventsToReturn } res = self . _er . jsonRequest ( "/json/getEventForText/enqueueRequest" , params ) requestId = res [ "requestId" ] for i in range ( 10 ) : time . sleep ( 1 ) res = self . _er . jsonR... | compute the list of most similar events for the given text |
248,127 | def annotate ( self , text , lang = None , customParams = None ) : params = { "lang" : lang , "text" : text } if customParams : params . update ( customParams ) return self . _er . jsonRequestAnalytics ( "/api/v1/annotate" , params ) | identify the list of entities and nonentities mentioned in the text |
248,128 | def sentiment ( self , text , method = "vocabulary" ) : assert method == "vocabulary" or method == "rnn" endpoint = method == "vocabulary" and "sentiment" or "sentimentRNN" return self . _er . jsonRequestAnalytics ( "/api/v1/" + endpoint , { "text" : text } ) | determine the sentiment of the provided text in English language |
248,129 | def semanticSimilarity ( self , text1 , text2 , distanceMeasure = "cosine" ) : return self . _er . jsonRequestAnalytics ( "/api/v1/semanticSimilarity" , { "text1" : text1 , "text2" : text2 , "distanceMeasure" : distanceMeasure } ) | determine the semantic similarity of the two provided documents |
248,130 | def extractArticleInfo ( self , url , proxyUrl = None , headers = None , cookies = None ) : params = { "url" : url } if proxyUrl : params [ "proxyUrl" ] = proxyUrl if headers : if isinstance ( headers , dict ) : headers = json . dumps ( headers ) params [ "headers" ] = headers if cookies : if isinstance ( cookies , dic... | extract all available information about an article available at url url . Returned information will include article title body authors links in the articles ... |
248,131 | def trainTopicOnTweets ( self , twitterQuery , useTweetText = True , useIdfNormalization = True , normalization = "linear" , maxTweets = 2000 , maxUsedLinks = 500 , ignoreConceptTypes = [ ] , maxConcepts = 20 , maxCategories = 10 , notifyEmailAddress = None ) : assert maxTweets < 5000 , "we can analyze at most 5000 twe... | create a new topic and train it using the tweets that match the twitterQuery |
248,132 | def trainTopicGetTrainedTopic ( self , uri , maxConcepts = 20 , maxCategories = 10 , ignoreConceptTypes = [ ] , idfNormalization = True ) : return self . _er . jsonRequestAnalytics ( "/api/v1/trainTopic" , { "action" : "getTrainedTopic" , "uri" : uri , "maxConcepts" : maxConcepts , "maxCategories" : maxCategories , "id... | retrieve topic for the topic for which you have already finished training |
248,133 | def createTopicPage1 ( ) : topic = TopicPage ( er ) topic . addKeyword ( "renewable energy" , 30 ) topic . addConcept ( er . getConceptUri ( "biofuel" ) , 50 ) topic . addConcept ( er . getConceptUri ( "solar energy" ) , 50 ) topic . addCategory ( er . getCategoryUri ( "renewable" ) , 50 ) topic . articleHasDuplicateFi... | create a topic page directly |
248,134 | def createTopicPage2 ( ) : topic = TopicPage ( er ) topic . addCategory ( er . getCategoryUri ( "renewable" ) , 50 ) topic . addKeyword ( "renewable energy" , 30 ) topic . addConcept ( er . getConceptUri ( "biofuel" ) , 50 ) topic . addConcept ( er . getConceptUri ( "solar energy" ) , 50 ) topic . restrictToSetConcepts... | create a topic page directly set the article threshold restrict results to set concepts and keywords |
248,135 | def count ( self , eventRegistry ) : self . setRequestedResult ( RequestEventArticles ( ** self . queryParams ) ) res = eventRegistry . execQuery ( self ) if "error" in res : print ( res [ "error" ] ) count = res . get ( self . queryParams [ "eventUri" ] , { } ) . get ( "articles" , { } ) . get ( "totalResults" , 0 ) r... | return the number of articles that match the criteria |
248,136 | def initWithComplexQuery ( query ) : q = QueryArticles ( ) if isinstance ( query , ComplexArticleQuery ) : q . _setVal ( "query" , json . dumps ( query . getQuery ( ) ) ) elif isinstance ( query , six . string_types ) : foo = json . loads ( query ) q . _setVal ( "query" , query ) elif isinstance ( query , dict ) : q . ... | create a query using a complex article query |
248,137 | def _getNextArticleBatch ( self ) : self . _articlePage += 1 if self . _totalPages != None and self . _articlePage > self . _totalPages : return self . setRequestedResult ( RequestArticlesInfo ( page = self . _articlePage , sortBy = self . _sortBy , sortByAsc = self . _sortByAsc , returnInfo = self . _returnInfo ) ) if... | download next batch of articles based on the article uris in the uri list |
248,138 | def initWithComplexQuery ( query ) : q = QueryEvents ( ) if isinstance ( query , ComplexEventQuery ) : q . _setVal ( "query" , json . dumps ( query . getQuery ( ) ) ) elif isinstance ( query , six . string_types ) : foo = json . loads ( query ) q . _setVal ( "query" , query ) elif isinstance ( query , dict ) : q . _set... | create a query using a complex event query |
248,139 | def count ( self , eventRegistry ) : self . setRequestedResult ( RequestEventsInfo ( ) ) res = eventRegistry . execQuery ( self ) if "error" in res : print ( res [ "error" ] ) count = res . get ( "events" , { } ) . get ( "totalResults" , 0 ) return count | return the number of events that match the criteria |
248,140 | def _setFlag ( self , name , val , defVal ) : if not hasattr ( self , "flags" ) : self . flags = { } if val != defVal : self . flags [ name ] = val | set the objects property propName if the dictKey key exists in dict and it is not the same as default value defVal |
248,141 | def _setVal ( self , name , val , defVal = None ) : if val == defVal : return if not hasattr ( self , "vals" ) : self . vals = { } self . vals [ name ] = val | set value of name to val in case the val ! = defVal |
248,142 | def _getVals ( self , prefix = "" ) : if not hasattr ( self , "vals" ) : self . vals = { } dict = { } for key in list ( self . vals . keys ( ) ) : if prefix == "" : newkey = key [ : 1 ] . lower ( ) + key [ 1 : ] if key else "" dict [ newkey ] = self . vals [ key ] else : newkey = key [ : 1 ] . upper ( ) + key [ 1 : ] i... | return the values in the vals dict in case prefix is change the first letter of the name to lowercase otherwise use prefix + name as the new name |
248,143 | def loadFromFile ( fileName ) : assert os . path . exists ( fileName ) , "File " + fileName + " does not exist" conf = json . load ( open ( fileName ) ) return ReturnInfo ( articleInfo = ArticleInfoFlags ( ** conf . get ( "articleInfo" , { } ) ) , eventInfo = EventInfoFlags ( ** conf . get ( "eventInfo" , { } ) ) , sou... | load the configuration for the ReturnInfo from a fileName |
248,144 | def loadTopicPageFromER ( self , uri ) : params = { "action" : "getTopicPageJson" , "includeConceptDescription" : True , "includeTopicPageDefinition" : True , "includeTopicPageOwner" : True , "uri" : uri } self . topicPage = self . _createEmptyTopicPage ( ) self . concept = self . eventRegistry . jsonRequest ( "/json/t... | load an existing topic page from Event Registry based on the topic page URI |
248,145 | def loadTopicPageFromFile ( self , fname ) : assert os . path . exists ( fname ) f = open ( fname , "r" , encoding = "utf-8" ) self . topicPage = json . load ( f ) | load topic page from an existing file |
248,146 | def saveTopicPageDefinitionToFile ( self , fname ) : open ( fname , "w" , encoding = "utf-8" ) . write ( json . dumps ( self . topicPage , indent = 4 , sort_keys = True ) ) | save the topic page definition to a file |
248,147 | def setArticleThreshold ( self , value ) : assert isinstance ( value , int ) assert value >= 0 self . topicPage [ "articleTreshWgt" ] = value | what is the minimum total weight that an article has to have in order to get it among the results? |
248,148 | def setEventThreshold ( self , value ) : assert isinstance ( value , int ) assert value >= 0 self . topicPage [ "eventTreshWgt" ] = value | what is the minimum total weight that an event has to have in order to get it among the results? |
248,149 | def setMaxDaysBack ( self , maxDaysBack ) : assert isinstance ( maxDaysBack , int ) , "maxDaysBack value has to be a positive integer" assert maxDaysBack >= 1 self . topicPage [ "maxDaysBack" ] = maxDaysBack | what is the maximum allowed age of the results? |
248,150 | def addConcept ( self , conceptUri , weight , label = None , conceptType = None ) : assert isinstance ( weight , ( float , int ) ) , "weight value has to be a positive or negative integer" concept = { "uri" : conceptUri , "wgt" : weight } if label != None : concept [ "label" ] = label if conceptType != None : concept [... | add a relevant concept to the topic page |
248,151 | def addKeyword ( self , keyword , weight ) : assert isinstance ( weight , ( float , int ) ) , "weight value has to be a positive or negative integer" self . topicPage [ "keywords" ] . append ( { "keyword" : keyword , "wgt" : weight } ) | add a relevant keyword to the topic page |
248,152 | def addCategory ( self , categoryUri , weight ) : assert isinstance ( weight , ( float , int ) ) , "weight value has to be a positive or negative integer" self . topicPage [ "categories" ] . append ( { "uri" : categoryUri , "wgt" : weight } ) | add a relevant category to the topic page |
248,153 | def addSource ( self , sourceUri , weight ) : assert isinstance ( weight , ( float , int ) ) , "weight value has to be a positive or negative integer" self . topicPage [ "sources" ] . append ( { "uri" : sourceUri , "wgt" : weight } ) | add a news source to the topic page |
248,154 | def addSourceLocation ( self , sourceLocationUri , weight ) : assert isinstance ( weight , ( float , int ) ) , "weight value has to be a positive or negative integer" self . topicPage [ "sourceLocations" ] . append ( { "uri" : sourceLocationUri , "wgt" : weight } ) | add a list of relevant sources by identifying them by their geographic location |
248,155 | def addSourceGroup ( self , sourceGroupUri , weight ) : assert isinstance ( weight , ( float , int ) ) , "weight value has to be a positive or negative integer" self . topicPage [ "sourceGroups" ] . append ( { "uri" : sourceGroupUri , "wgt" : weight } ) | add a list of relevant sources by specifying a whole source group to the topic page |
248,156 | def addLocation ( self , locationUri , weight ) : assert isinstance ( weight , ( float , int ) ) , "weight value has to be a positive or negative integer" self . topicPage [ "locations" ] . append ( { "uri" : locationUri , "wgt" : weight } ) | add relevant location to the topic page |
248,157 | def setLanguages ( self , languages ) : if isinstance ( languages , six . string_types ) : languages = [ languages ] for lang in languages : assert len ( lang ) == 3 , "Expected to get language in ISO3 code" self . topicPage [ "langs" ] = languages | restrict the results to the list of specified languages |
248,158 | def getArticles ( self , page = 1 , count = 100 , sortBy = "rel" , sortByAsc = False , returnInfo = ReturnInfo ( ) ) : assert page >= 1 assert count <= 100 params = { "action" : "getArticlesForTopicPage" , "resultType" : "articles" , "dataType" : self . topicPage [ "dataType" ] , "articlesCount" : count , "articlesSort... | return a list of articles that match the topic page |
248,159 | def AND ( queryArr , exclude = None ) : assert isinstance ( queryArr , list ) , "provided argument as not a list" assert len ( queryArr ) > 0 , "queryArr had an empty list" q = CombinedQuery ( ) q . setQueryParam ( "$and" , [ ] ) for item in queryArr : assert isinstance ( item , ( CombinedQuery , BaseQuery ) ) , "item ... | create a combined query with multiple items on which to perform an AND operation |
248,160 | async def start_pairing ( self ) : self . srp . initialize ( ) msg = messages . crypto_pairing ( { tlv8 . TLV_METHOD : b'\x00' , tlv8 . TLV_SEQ_NO : b'\x01' } ) resp = await self . protocol . send_and_receive ( msg , generate_identifier = False ) pairing_data = _get_pairing_data ( resp ) if tlv8 . TLV_BACK_OFF in pairi... | Start pairing procedure . |
248,161 | async def finish_pairing ( self , pin ) : self . srp . step1 ( pin ) pub_key , proof = self . srp . step2 ( self . _atv_pub_key , self . _atv_salt ) msg = messages . crypto_pairing ( { tlv8 . TLV_SEQ_NO : b'\x03' , tlv8 . TLV_PUBLIC_KEY : pub_key , tlv8 . TLV_PROOF : proof } ) resp = await self . protocol . send_and_re... | Finish pairing process . |
248,162 | async def verify_credentials ( self ) : _ , public_key = self . srp . initialize ( ) msg = messages . crypto_pairing ( { tlv8 . TLV_SEQ_NO : b'\x01' , tlv8 . TLV_PUBLIC_KEY : public_key } ) resp = await self . protocol . send_and_receive ( msg , generate_identifier = False ) resp = _get_pairing_data ( resp ) session_pu... | Verify credentials with device . |
248,163 | def lookup_tag ( name ) : return next ( ( _TAGS [ t ] for t in _TAGS if t == name ) , DmapTag ( _read_unknown , 'unknown tag' ) ) | Look up a tag based on its key . Returns a DmapTag . |
248,164 | def connect_to_apple_tv ( details , loop , protocol = None , session = None ) : service = _get_service_used_to_connect ( details , protocol ) if session is None : session = ClientSession ( loop = loop ) airplay = _setup_airplay ( loop , session , details ) if service . protocol == PROTOCOL_DMAP : return DmapAppleTV ( l... | Connect and logins to an Apple TV . |
248,165 | def add_service ( self , zeroconf , service_type , name ) : self . lock . acquire ( ) try : self . _internal_add ( zeroconf , service_type , name ) finally : self . lock . release ( ) | Handle callback from zeroconf when a service has been discovered . |
248,166 | def add_hs_service ( self , info , address ) : if self . protocol and self . protocol != PROTOCOL_DMAP : return name = info . properties [ b'Name' ] . decode ( 'utf-8' ) hsgid = info . properties [ b'hG' ] . decode ( 'utf-8' ) self . _handle_service ( address , name , conf . DmapService ( hsgid , port = info . port ) ) | Add a new device to discovered list . |
248,167 | def add_non_hs_service ( self , info , address ) : if self . protocol and self . protocol != PROTOCOL_DMAP : return name = info . properties [ b'CtlN' ] . decode ( 'utf-8' ) self . _handle_service ( address , name , conf . DmapService ( None , port = info . port ) ) | Add a new device without Home Sharing to discovered list . |
248,168 | def add_mrp_service ( self , info , address ) : if self . protocol and self . protocol != PROTOCOL_MRP : return name = info . properties [ b'Name' ] . decode ( 'utf-8' ) self . _handle_service ( address , name , conf . MrpService ( info . port ) ) | Add a new MediaRemoteProtocol device to discovered list . |
248,169 | def add_airplay_service ( self , info , address ) : name = info . name . replace ( '._airplay._tcp.local.' , '' ) self . _handle_service ( address , name , conf . AirPlayService ( info . port ) ) | Add a new AirPlay device to discovered list . |
248,170 | def usable_service ( self ) : services = self . _services for protocol in self . _supported_protocols : if protocol in services and services [ protocol ] . is_usable ( ) : return services [ protocol ] return None | Return a usable service or None if there is none . |
248,171 | def superseeded_by ( self , other_service ) : if not other_service or other_service . __class__ != self . __class__ or other_service . protocol != self . protocol or other_service . port != self . port : return False return not self . device_credentials and other_service . device_credentials | Return True if input service has login id and this has not . |
248,172 | async def print_what_is_playing ( loop ) : print ( 'Discovering devices on network...' ) atvs = await pyatv . scan_for_apple_tvs ( loop , timeout = 5 ) if not atvs : print ( 'no device found' , file = sys . stderr ) return print ( 'Connecting to {0}' . format ( atvs [ 0 ] . address ) ) atv = pyatv . connect_to_apple_tv... | Find a device and print what is playing . |
248,173 | async def start ( self , ** kwargs ) : zeroconf = kwargs [ 'zeroconf' ] self . _name = kwargs [ 'name' ] self . _pairing_guid = kwargs . get ( 'pairing_guid' , None ) or self . _generate_random_guid ( ) self . _web_server = web . Server ( self . handle_request , loop = self . _loop ) self . _server = await self . _loop... | Start the pairing server and publish service . |
248,174 | async def stop ( self , ** kwargs ) : _LOGGER . debug ( 'Shutting down pairing server' ) if self . _web_server is not None : await self . _web_server . shutdown ( ) self . _server . close ( ) if self . _server is not None : await self . _server . wait_closed ( ) | Stop pairing server and unpublish service . |
248,175 | async def handle_request ( self , request ) : service_name = request . rel_url . query [ 'servicename' ] received_code = request . rel_url . query [ 'pairingcode' ] . lower ( ) _LOGGER . info ( 'Got pairing request from %s with code %s' , service_name , received_code ) if self . _verify_pin ( received_code ) : cmpg = t... | Respond to request if PIN is correct . |
248,176 | def log_binary ( logger , message , ** kwargs ) : if logger . isEnabledFor ( logging . DEBUG ) : output = ( '{0}={1}' . format ( k , binascii . hexlify ( bytearray ( v ) ) . decode ( ) ) for k , v in sorted ( kwargs . items ( ) ) ) logger . debug ( '%s (%s)' , message , ', ' . join ( output ) ) | Log binary data if debug is enabled . |
248,177 | def _extract_command_with_args ( cmd ) : def _isint ( value ) : try : int ( value ) return True except ValueError : return False equal_sign = cmd . find ( '=' ) if equal_sign == - 1 : return cmd , [ ] command = cmd [ 0 : equal_sign ] args = cmd [ equal_sign + 1 : ] . split ( ',' ) converted = [ x if not _isint ( x ) el... | Parse input command with arguments . |
248,178 | def main ( ) : async def _run_application ( loop ) : try : return await cli_handler ( loop ) except KeyboardInterrupt : pass except SystemExit : pass except : import traceback traceback . print_exc ( file = sys . stderr ) sys . stderr . writelines ( '\n>>> An error occurred, full stack trace above\n' ) return 1 try : l... | Start the asyncio event loop and runs the application . |
248,179 | async def commands ( self ) : _print_commands ( 'Remote control' , interface . RemoteControl ) _print_commands ( 'Metadata' , interface . Metadata ) _print_commands ( 'Playing' , interface . Playing ) _print_commands ( 'AirPlay' , interface . AirPlay ) _print_commands ( 'Device' , DeviceCommands ) _print_commands ( 'Gl... | Print a list with available commands . |
248,180 | async def help ( self ) : if len ( self . args . command ) != 2 : print ( 'Which command do you want help with?' , file = sys . stderr ) return 1 iface = [ interface . RemoteControl , interface . Metadata , interface . Playing , interface . AirPlay , self . __class__ , DeviceCommands ] for cmd in iface : for key , valu... | Print help text for a command . |
248,181 | async def scan ( self ) : atvs = await pyatv . scan_for_apple_tvs ( self . loop , timeout = self . args . scan_timeout , only_usable = False ) _print_found_apple_tvs ( atvs ) return 0 | Scan for Apple TVs on the network . |
248,182 | async def cli ( self ) : print ( 'Enter commands and press enter' ) print ( 'Type help for help and exit to quit' ) while True : command = await _read_input ( self . loop , 'pyatv> ' ) if command . lower ( ) == 'exit' : break elif command == 'cli' : print ( 'Command not availble here' ) continue await _handle_device_co... | Enter commands in a simple CLI . |
248,183 | async def artwork_save ( self ) : artwork = await self . atv . metadata . artwork ( ) if artwork is not None : with open ( 'artwork.png' , 'wb' ) as file : file . write ( artwork ) else : print ( 'No artwork is currently available.' ) return 1 return 0 | Download artwork and save it to artwork . png . |
248,184 | async def push_updates ( self ) : print ( 'Press ENTER to stop' ) self . atv . push_updater . start ( ) await self . atv . login ( ) await self . loop . run_in_executor ( None , sys . stdin . readline ) self . atv . push_updater . stop ( ) return 0 | Listen for push updates . |
248,185 | async def auth ( self ) : credentials = await self . atv . airplay . generate_credentials ( ) await self . atv . airplay . load_credentials ( credentials ) try : await self . atv . airplay . start_authentication ( ) pin = await _read_input ( self . loop , 'Enter PIN on screen: ' ) await self . atv . airplay . finish_au... | Perform AirPlay device authentication . |
248,186 | async def pair ( self ) : protocol = self . atv . service . protocol if protocol == const . PROTOCOL_DMAP : await self . atv . pairing . start ( zeroconf = Zeroconf ( ) , name = self . args . remote_name , pairing_guid = self . args . pairing_guid ) elif protocol == const . PROTOCOL_MRP : await self . atv . pairing . s... | Pair pyatv as a remote control with an Apple TV . |
248,187 | def media_kind ( kind ) : if kind in [ 1 ] : return const . MEDIA_TYPE_UNKNOWN if kind in [ 3 , 7 , 11 , 12 , 13 , 18 , 32 ] : return const . MEDIA_TYPE_VIDEO if kind in [ 2 , 4 , 10 , 14 , 17 , 21 , 36 ] : return const . MEDIA_TYPE_MUSIC if kind in [ 8 , 64 ] : return const . MEDIA_TYPE_TV raise exceptions . UnknownMe... | Convert iTunes media kind to API representation . |
248,188 | def media_type_str ( mediatype ) : if mediatype == const . MEDIA_TYPE_UNKNOWN : return 'Unknown' if mediatype == const . MEDIA_TYPE_VIDEO : return 'Video' if mediatype == const . MEDIA_TYPE_MUSIC : return 'Music' if mediatype == const . MEDIA_TYPE_TV : return 'TV' return 'Unsupported' | Convert internal API media type to string . |
248,189 | def playstate ( state ) : if state is None : return const . PLAY_STATE_NO_MEDIA if state == 0 : return const . PLAY_STATE_IDLE if state == 1 : return const . PLAY_STATE_LOADING if state == 3 : return const . PLAY_STATE_PAUSED if state == 4 : return const . PLAY_STATE_PLAYING if state == 5 : return const . PLAY_STATE_FA... | Convert iTunes playstate to API representation . |
248,190 | def playstate_str ( state ) : if state == const . PLAY_STATE_NO_MEDIA : return 'No media' if state == const . PLAY_STATE_IDLE : return 'Idle' if state == const . PLAY_STATE_LOADING : return 'Loading' if state == const . PLAY_STATE_PAUSED : return 'Paused' if state == const . PLAY_STATE_PLAYING : return 'Playing' if sta... | Convert internal API playstate to string . |
248,191 | def repeat_str ( state ) : if state == const . REPEAT_STATE_OFF : return 'Off' if state == const . REPEAT_STATE_TRACK : return 'Track' if state == const . REPEAT_STATE_ALL : return 'All' return 'Unsupported' | Convert internal API repeat state to string . |
248,192 | def protocol_str ( protocol ) : if protocol == const . PROTOCOL_MRP : return 'MRP' if protocol == const . PROTOCOL_DMAP : return 'DMAP' if protocol == const . PROTOCOL_AIRPLAY : return 'AirPlay' return 'Unknown' | Convert internal API protocol to string . |
248,193 | def first ( dmap_data , * path ) : if not ( path and isinstance ( dmap_data , list ) ) : return dmap_data for key in dmap_data : if path [ 0 ] in key : return first ( key [ path [ 0 ] ] , * path [ 1 : ] ) return None | Look up a value given a path in some parsed DMAP data . |
248,194 | def pprint ( data , tag_lookup , indent = 0 ) : output = '' if isinstance ( data , dict ) : for key , value in data . items ( ) : tag = tag_lookup ( key ) if isinstance ( value , ( dict , list ) ) and tag . type is not read_bplist : output += '{0}{1}: {2}\n' . format ( indent * ' ' , key , tag ) output += pprint ( valu... | Return a pretty formatted string of parsed DMAP data . |
248,195 | def retrieve_commands ( obj ) : commands = { } for func in obj . __dict__ : if not inspect . isfunction ( obj . __dict__ [ func ] ) and not isinstance ( obj . __dict__ [ func ] , property ) : continue if func . startswith ( '_' ) : continue commands [ func ] = _get_first_sentence_in_pydoc ( obj . __dict__ [ func ] ) re... | Retrieve all commands and help texts from an API object . |
248,196 | def hash ( self ) : base = '{0}{1}{2}{3}' . format ( self . title , self . artist , self . album , self . total_time ) return hashlib . sha256 ( base . encode ( 'utf-8' ) ) . hexdigest ( ) | Create a unique hash for what is currently playing . |
248,197 | def extract_message_info ( ) : base_path = BASE_PACKAGE . replace ( '.' , '/' ) filename = os . path . join ( base_path , 'ProtocolMessage.proto' ) with open ( filename , 'r' ) as file : types_found = False for line in file : stripped = line . lstrip ( ) . rstrip ( ) if stripped == 'enum Type {' : types_found = True co... | Get information about all messages of interest . |
248,198 | def main ( ) : message_names = set ( ) packages = [ ] messages = [ ] extensions = [ ] constants = [ ] for info in extract_message_info ( ) : message_names . add ( info . title ) packages . append ( 'from {0} import {1}' . format ( BASE_PACKAGE , info . module ) ) messages . append ( 'from {0}.{1} import {2}' . format (... | Script starts somewhere around here . |
248,199 | def hkdf_expand ( salt , info , shared_secret ) : from cryptography . hazmat . primitives import hashes from cryptography . hazmat . primitives . kdf . hkdf import HKDF from cryptography . hazmat . backends import default_backend hkdf = HKDF ( algorithm = hashes . SHA512 ( ) , length = 32 , salt = salt . encode ( ) , i... | Derive encryption keys from shared secret . |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.