nwo
stringlengths
5
86
sha
stringlengths
40
40
path
stringlengths
4
189
language
stringclasses
1 value
identifier
stringlengths
1
94
parameters
stringlengths
2
4.03k
argument_list
stringclasses
1 value
return_statement
stringlengths
0
11.5k
docstring
stringlengths
1
33.2k
docstring_summary
stringlengths
0
5.15k
docstring_tokens
sequence
function
stringlengths
34
151k
function_tokens
sequence
url
stringlengths
90
278
aws/lumberyard
f85344403c1c2e77ec8c75deb2c116e97b713217
dev/Tools/Python/3.7.10/linux_x64/lib/python3.7/_pyio.py
python
IOBase._checkReadable
(self, msg=None)
Internal: raise UnsupportedOperation if file is not readable
Internal: raise UnsupportedOperation if file is not readable
[ "Internal", ":", "raise", "UnsupportedOperation", "if", "file", "is", "not", "readable" ]
def _checkReadable(self, msg=None): """Internal: raise UnsupportedOperation if file is not readable """ if not self.readable(): raise UnsupportedOperation("File or stream is not readable." if msg is None else msg)
[ "def", "_checkReadable", "(", "self", ",", "msg", "=", "None", ")", ":", "if", "not", "self", ".", "readable", "(", ")", ":", "raise", "UnsupportedOperation", "(", "\"File or stream is not readable.\"", "if", "msg", "is", "None", "else", "msg", ")" ]
https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/linux_x64/lib/python3.7/_pyio.py#L410-L415
catboost/catboost
167f64f237114a4d10b2b4ee42adb4569137debe
contrib/python/prompt-toolkit/py3/prompt_toolkit/history.py
python
ThreadedHistory.load
(self)
Like `History.load(), but call `self.load_history_strings()` in a background thread.
Like `History.load(), but call `self.load_history_strings()` in a background thread.
[ "Like", "History", ".", "load", "()", "but", "call", "self", ".", "load_history_strings", "()", "in", "a", "background", "thread", "." ]
async def load(self) -> AsyncGenerator[str, None]: """ Like `History.load(), but call `self.load_history_strings()` in a background thread. """ # Start the load thread, if this is called for the first time. if not self._load_thread: self._load_thread = threading.Thread( target=self._in_load_thread, daemon=True, ) self._load_thread.start() # Consume the `_loaded_strings` list, using asyncio. loop = get_event_loop() # Create threading Event so that we can wait for new items. event = threading.Event() event.set() self._string_load_events.append(event) items_yielded = 0 try: while True: # Wait for new items to be available. # (Use a timeout, because the executor thread is not a daemon # thread. The "slow-history.py" example would otherwise hang if # Control-C is pressed before the history is fully loaded, # because there's still this non-daemon executor thread waiting # for this event.) got_timeout = await loop.run_in_executor( None, lambda: event.wait(timeout=0.5) ) if not got_timeout: continue # Read new items (in lock). def in_executor() -> Tuple[List[str], bool]: with self._lock: new_items = self._loaded_strings[items_yielded:] done = self._loaded event.clear() return new_items, done new_items, done = await loop.run_in_executor(None, in_executor) items_yielded += len(new_items) for item in new_items: yield item if done: break finally: self._string_load_events.remove(event)
[ "async", "def", "load", "(", "self", ")", "->", "AsyncGenerator", "[", "str", ",", "None", "]", ":", "# Start the load thread, if this is called for the first time.", "if", "not", "self", ".", "_load_thread", ":", "self", ".", "_load_thread", "=", "threading", ".", "Thread", "(", "target", "=", "self", ".", "_in_load_thread", ",", "daemon", "=", "True", ",", ")", "self", ".", "_load_thread", ".", "start", "(", ")", "# Consume the `_loaded_strings` list, using asyncio.", "loop", "=", "get_event_loop", "(", ")", "# Create threading Event so that we can wait for new items.", "event", "=", "threading", ".", "Event", "(", ")", "event", ".", "set", "(", ")", "self", ".", "_string_load_events", ".", "append", "(", "event", ")", "items_yielded", "=", "0", "try", ":", "while", "True", ":", "# Wait for new items to be available.", "# (Use a timeout, because the executor thread is not a daemon", "# thread. The \"slow-history.py\" example would otherwise hang if", "# Control-C is pressed before the history is fully loaded,", "# because there's still this non-daemon executor thread waiting", "# for this event.)", "got_timeout", "=", "await", "loop", ".", "run_in_executor", "(", "None", ",", "lambda", ":", "event", ".", "wait", "(", "timeout", "=", "0.5", ")", ")", "if", "not", "got_timeout", ":", "continue", "# Read new items (in lock).", "def", "in_executor", "(", ")", "->", "Tuple", "[", "List", "[", "str", "]", ",", "bool", "]", ":", "with", "self", ".", "_lock", ":", "new_items", "=", "self", ".", "_loaded_strings", "[", "items_yielded", ":", "]", "done", "=", "self", ".", "_loaded", "event", ".", "clear", "(", ")", "return", "new_items", ",", "done", "new_items", ",", "done", "=", "await", "loop", ".", "run_in_executor", "(", "None", ",", "in_executor", ")", "items_yielded", "+=", "len", "(", "new_items", ")", "for", "item", "in", "new_items", ":", "yield", "item", "if", "done", ":", "break", "finally", ":", "self", ".", "_string_load_events", ".", "remove", "(", "event", ")" ]
https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/prompt-toolkit/py3/prompt_toolkit/history.py#L124-L179
wxWidgets/wxPython-Classic
19571e1ae65f1ac445f5491474121998c97a1bf0
wx/lib/agw/aui/framemanager.py
python
AuiPaneInfo.Dockable
(self, b=True)
return self.TopDockable(b).BottomDockable(b).LeftDockable(b).RightDockable(b)
Specifies whether a frame can be docked or not. It is the same as specifying :meth:`TopDockable` . :meth:`BottomDockable` . :meth:`LeftDockable` . :meth:`RightDockable` . :param bool `b`: whether the frame can be docked or not.
Specifies whether a frame can be docked or not. It is the same as specifying :meth:`TopDockable` . :meth:`BottomDockable` . :meth:`LeftDockable` . :meth:`RightDockable` .
[ "Specifies", "whether", "a", "frame", "can", "be", "docked", "or", "not", ".", "It", "is", "the", "same", "as", "specifying", ":", "meth", ":", "TopDockable", ".", ":", "meth", ":", "BottomDockable", ".", ":", "meth", ":", "LeftDockable", ".", ":", "meth", ":", "RightDockable", "." ]
def Dockable(self, b=True): """ Specifies whether a frame can be docked or not. It is the same as specifying :meth:`TopDockable` . :meth:`BottomDockable` . :meth:`LeftDockable` . :meth:`RightDockable` . :param bool `b`: whether the frame can be docked or not. """ return self.TopDockable(b).BottomDockable(b).LeftDockable(b).RightDockable(b)
[ "def", "Dockable", "(", "self", ",", "b", "=", "True", ")", ":", "return", "self", ".", "TopDockable", "(", "b", ")", ".", "BottomDockable", "(", "b", ")", ".", "LeftDockable", "(", "b", ")", ".", "RightDockable", "(", "b", ")" ]
https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/wx/lib/agw/aui/framemanager.py#L1609-L1617
eclipse/sumo
7132a9b8b6eea734bdec38479026b4d8c4336d03
tools/contributed/sumopy/coremodules/misc/shapeformat.py
python
facilities_to_shapefile
(facilities, filepath, dataname='facilitydata', is_access=True, parent=None, log=None)
return True
Export network edges to shapefile.
Export network edges to shapefile.
[ "Export", "network", "edges", "to", "shapefile", "." ]
def facilities_to_shapefile(facilities, filepath, dataname='facilitydata', is_access=True, parent=None, log=None): """ Export network edges to shapefile. """ net = facilities.get_net() shapedata = Shapedata(parent, dataname, name='Facilities shape data', filepath=filepath, shapetype=SHAPETYPES['PolyLine'], projparams_shape=net.get_projparams(), offset=net.get_offset(), log=log) #attrname, ftype, flen, fdigit = field attrlist = [ ('id', 'id', 'ID_FACIL', 'N', 32, 0), ('ids_roadedge_closest', 'id', 'ID_ARC', 'N', 32, 0), ('positions_roadedge_closest', 'id', 'POS_ARC', 'N', 12, 5), ('osmkeys', 'val', 'OSMKEY', 'C', 32, 0), ] print 'facilities_to_shapefile', filepath for attr in attrlist: shapedata.add_field(attr[2:]) ids_fac = facilities.get_ids() ids_shape = shapedata.add_rows(len(ids_fac)) # print ' shapedata.ID_ARC',shapedata.ID_ARC,'dir',dir(shapedata.ID_ARC) shapedata.ID_FACIL[ids_shape] = ids_fac shapedata.shapes[ids_shape] = facilities.shapes[ids_fac] for netattrname, gettype, shapeattrname, x1, x2, x3 in attrlist: if netattrname not in ('id',): getattr(shapedata, shapeattrname)[ids_shape] = getattr(facilities, netattrname)[ids_fac] shapedata.adjust_fieldlength() shapedata.export_shapefile() return True
[ "def", "facilities_to_shapefile", "(", "facilities", ",", "filepath", ",", "dataname", "=", "'facilitydata'", ",", "is_access", "=", "True", ",", "parent", "=", "None", ",", "log", "=", "None", ")", ":", "net", "=", "facilities", ".", "get_net", "(", ")", "shapedata", "=", "Shapedata", "(", "parent", ",", "dataname", ",", "name", "=", "'Facilities shape data'", ",", "filepath", "=", "filepath", ",", "shapetype", "=", "SHAPETYPES", "[", "'PolyLine'", "]", ",", "projparams_shape", "=", "net", ".", "get_projparams", "(", ")", ",", "offset", "=", "net", ".", "get_offset", "(", ")", ",", "log", "=", "log", ")", "#attrname, ftype, flen, fdigit = field", "attrlist", "=", "[", "(", "'id'", ",", "'id'", ",", "'ID_FACIL'", ",", "'N'", ",", "32", ",", "0", ")", ",", "(", "'ids_roadedge_closest'", ",", "'id'", ",", "'ID_ARC'", ",", "'N'", ",", "32", ",", "0", ")", ",", "(", "'positions_roadedge_closest'", ",", "'id'", ",", "'POS_ARC'", ",", "'N'", ",", "12", ",", "5", ")", ",", "(", "'osmkeys'", ",", "'val'", ",", "'OSMKEY'", ",", "'C'", ",", "32", ",", "0", ")", ",", "]", "print", "'facilities_to_shapefile'", ",", "filepath", "for", "attr", "in", "attrlist", ":", "shapedata", ".", "add_field", "(", "attr", "[", "2", ":", "]", ")", "ids_fac", "=", "facilities", ".", "get_ids", "(", ")", "ids_shape", "=", "shapedata", ".", "add_rows", "(", "len", "(", "ids_fac", ")", ")", "# print ' shapedata.ID_ARC',shapedata.ID_ARC,'dir',dir(shapedata.ID_ARC)", "shapedata", ".", "ID_FACIL", "[", "ids_shape", "]", "=", "ids_fac", "shapedata", ".", "shapes", "[", "ids_shape", "]", "=", "facilities", ".", "shapes", "[", "ids_fac", "]", "for", "netattrname", ",", "gettype", ",", "shapeattrname", ",", "x1", ",", "x2", ",", "x3", "in", "attrlist", ":", "if", "netattrname", "not", "in", "(", "'id'", ",", ")", ":", "getattr", "(", "shapedata", ",", "shapeattrname", ")", "[", "ids_shape", "]", "=", "getattr", "(", "facilities", ",", "netattrname", ")", "[", "ids_fac", "]", "shapedata", ".", "adjust_fieldlength", "(", ")", "shapedata", ".", "export_shapefile", "(", ")", "return", "True" ]
https://github.com/eclipse/sumo/blob/7132a9b8b6eea734bdec38479026b4d8c4336d03/tools/contributed/sumopy/coremodules/misc/shapeformat.py#L1179-L1217
adobe/chromium
cfe5bf0b51b1f6b9fe239c2a3c2f2364da9967d7
tools/code_coverage/coverage_posix.py
python
RunProgramThread._run_posix
(self)
No deadlock problem so use the simple answer. The windows solution appears to add extra buffering which we don't want on other platforms.
No deadlock problem so use the simple answer. The windows solution appears to add extra buffering which we don't want on other platforms.
[ "No", "deadlock", "problem", "so", "use", "the", "simple", "answer", ".", "The", "windows", "solution", "appears", "to", "add", "extra", "buffering", "which", "we", "don", "t", "want", "on", "other", "platforms", "." ]
def _run_posix(self): """No deadlock problem so use the simple answer. The windows solution appears to add extra buffering which we don't want on other platforms.""" self._process = subprocess.Popen(self._cmd, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) gChildPIDs.append(self._process.pid) try: while True: line = self._process.stdout.readline() if not line: # EOF break print line, self._queue.put(RunProgramThread.PROGRESS, True) except IOError: pass # If we get here the process is done. gChildPIDs.remove(self._process.pid) self._queue.put(RunProgramThread.DONE)
[ "def", "_run_posix", "(", "self", ")", ":", "self", ".", "_process", "=", "subprocess", ".", "Popen", "(", "self", ".", "_cmd", ",", "stdout", "=", "subprocess", ".", "PIPE", ",", "stderr", "=", "subprocess", ".", "STDOUT", ")", "gChildPIDs", ".", "append", "(", "self", ".", "_process", ".", "pid", ")", "try", ":", "while", "True", ":", "line", "=", "self", ".", "_process", ".", "stdout", ".", "readline", "(", ")", "if", "not", "line", ":", "# EOF", "break", "print", "line", ",", "self", ".", "_queue", ".", "put", "(", "RunProgramThread", ".", "PROGRESS", ",", "True", ")", "except", "IOError", ":", "pass", "# If we get here the process is done.", "gChildPIDs", ".", "remove", "(", "self", ".", "_process", ".", "pid", ")", "self", ".", "_queue", ".", "put", "(", "RunProgramThread", ".", "DONE", ")" ]
https://github.com/adobe/chromium/blob/cfe5bf0b51b1f6b9fe239c2a3c2f2364da9967d7/tools/code_coverage/coverage_posix.py#L233-L251
lhmRyan/deep-supervised-hashing-DSH
631901f82e2ab031fbac33f914a5b08ef8e21d57
scripts/cpp_lint.py
python
_CppLintState.SetVerboseLevel
(self, level)
return last_verbose_level
Sets the module's verbosity, and returns the previous setting.
Sets the module's verbosity, and returns the previous setting.
[ "Sets", "the", "module", "s", "verbosity", "and", "returns", "the", "previous", "setting", "." ]
def SetVerboseLevel(self, level): """Sets the module's verbosity, and returns the previous setting.""" last_verbose_level = self.verbose_level self.verbose_level = level return last_verbose_level
[ "def", "SetVerboseLevel", "(", "self", ",", "level", ")", ":", "last_verbose_level", "=", "self", ".", "verbose_level", "self", ".", "verbose_level", "=", "level", "return", "last_verbose_level" ]
https://github.com/lhmRyan/deep-supervised-hashing-DSH/blob/631901f82e2ab031fbac33f914a5b08ef8e21d57/scripts/cpp_lint.py#L707-L711
qt/qtbase
81b9ee66b8e40ed145185fe46b7c91929688cafd
util/locale_database/cldr.py
python
CldrAccess.fileLocales
(self)
Generator for locale IDs seen in file-names. All *.xml other than root.xml in common/main/ are assumed to identify locales.
Generator for locale IDs seen in file-names.
[ "Generator", "for", "locale", "IDs", "seen", "in", "file", "-", "names", "." ]
def fileLocales(self) -> Iterable[str]: """Generator for locale IDs seen in file-names. All *.xml other than root.xml in common/main/ are assumed to identify locales.""" for path in self.root.joinpath('common/main').glob('*.xml'): if path.stem != 'root': yield path.stem
[ "def", "fileLocales", "(", "self", ")", "->", "Iterable", "[", "str", "]", ":", "for", "path", "in", "self", ".", "root", ".", "joinpath", "(", "'common/main'", ")", ".", "glob", "(", "'*.xml'", ")", ":", "if", "path", ".", "stem", "!=", "'root'", ":", "yield", "path", ".", "stem" ]
https://github.com/qt/qtbase/blob/81b9ee66b8e40ed145185fe46b7c91929688cafd/util/locale_database/cldr.py#L283-L290
kamyu104/LeetCode-Solutions
77605708a927ea3b85aee5a479db733938c7c211
Python/maximum-score-from-performing-multiplication-operations.py
python
Solution.maximumScore
(self, nums, multipliers)
return dp[0]
:type nums: List[int] :type multipliers: List[int] :rtype: int
:type nums: List[int] :type multipliers: List[int] :rtype: int
[ ":", "type", "nums", ":", "List", "[", "int", "]", ":", "type", "multipliers", ":", "List", "[", "int", "]", ":", "rtype", ":", "int" ]
def maximumScore(self, nums, multipliers): """ :type nums: List[int] :type multipliers: List[int] :rtype: int """ dp = [0]*(len(multipliers)+1) for l, m in enumerate(reversed(multipliers), start=len(nums)-len(multipliers)): dp = [max(m*nums[i]+dp[i+1], m*nums[i+l]+dp[i]) for i in xrange(len(dp)-1)] return dp[0]
[ "def", "maximumScore", "(", "self", ",", "nums", ",", "multipliers", ")", ":", "dp", "=", "[", "0", "]", "*", "(", "len", "(", "multipliers", ")", "+", "1", ")", "for", "l", ",", "m", "in", "enumerate", "(", "reversed", "(", "multipliers", ")", ",", "start", "=", "len", "(", "nums", ")", "-", "len", "(", "multipliers", ")", ")", ":", "dp", "=", "[", "max", "(", "m", "*", "nums", "[", "i", "]", "+", "dp", "[", "i", "+", "1", "]", ",", "m", "*", "nums", "[", "i", "+", "l", "]", "+", "dp", "[", "i", "]", ")", "for", "i", "in", "xrange", "(", "len", "(", "dp", ")", "-", "1", ")", "]", "return", "dp", "[", "0", "]" ]
https://github.com/kamyu104/LeetCode-Solutions/blob/77605708a927ea3b85aee5a479db733938c7c211/Python/maximum-score-from-performing-multiplication-operations.py#L5-L14
lmb-freiburg/ogn
974f72ef4bf840d6f6693d22d1843a79223e77ce
python/caffe/io.py
python
Transformer.set_channel_swap
(self, in_, order)
Set the input channel order for e.g. RGB to BGR conversion as needed for the reference ImageNet model. N.B. this assumes the channels are the first dimension AFTER transpose. Parameters ---------- in_ : which input to assign this channel order order : the order to take the channels. (2,1,0) maps RGB to BGR for example.
Set the input channel order for e.g. RGB to BGR conversion as needed for the reference ImageNet model. N.B. this assumes the channels are the first dimension AFTER transpose.
[ "Set", "the", "input", "channel", "order", "for", "e", ".", "g", ".", "RGB", "to", "BGR", "conversion", "as", "needed", "for", "the", "reference", "ImageNet", "model", ".", "N", ".", "B", ".", "this", "assumes", "the", "channels", "are", "the", "first", "dimension", "AFTER", "transpose", "." ]
def set_channel_swap(self, in_, order): """ Set the input channel order for e.g. RGB to BGR conversion as needed for the reference ImageNet model. N.B. this assumes the channels are the first dimension AFTER transpose. Parameters ---------- in_ : which input to assign this channel order order : the order to take the channels. (2,1,0) maps RGB to BGR for example. """ self.__check_input(in_) if len(order) != self.inputs[in_][1]: raise Exception('Channel swap needs to have the same number of ' 'dimensions as the input channels.') self.channel_swap[in_] = order
[ "def", "set_channel_swap", "(", "self", ",", "in_", ",", "order", ")", ":", "self", ".", "__check_input", "(", "in_", ")", "if", "len", "(", "order", ")", "!=", "self", ".", "inputs", "[", "in_", "]", "[", "1", "]", ":", "raise", "Exception", "(", "'Channel swap needs to have the same number of '", "'dimensions as the input channels.'", ")", "self", ".", "channel_swap", "[", "in_", "]", "=", "order" ]
https://github.com/lmb-freiburg/ogn/blob/974f72ef4bf840d6f6693d22d1843a79223e77ce/python/caffe/io.py#L203-L219
cms-sw/cmssw
fd9de012d503d3405420bcbeec0ec879baa57cf2
Alignment/MuonAlignment/python/svgfig.py
python
Frame.__init__
(self, xmin, xmax, ymin, ymax, *d, **kwds)
Acts like Fig, but draws a coordinate frame around the data. You also need to supply plot ranges. Frame(xmin, xmax, ymin, ymax, obj, obj, obj..., keyword options...) xmin, xmax required minimum and maximum x values (in the objs' coordinates) ymin, ymax required minimum and maximum y values (in the objs' coordinates) obj optional list drawing primatives keyword options keyword list options defined below The following are keyword options, with their default values: x, y 20, 5 upper-left corner of the Frame in SVG coordinates width, height 75, 80 width and height of the Frame in SVG coordinates flipx, flipy False, True flip the sign of the coordinate axis minusInfinity -1000 if an axis is logarithmic and an object is plotted at 0 or a negative value, -1000 will be used as a stand-in for NaN xtitle None if a string, label the x axis xticks -10 request ticks according to the standard tick specification (see help(Ticks)) xminiticks True request miniticks according to the standard minitick specification xlabels True request tick labels according to the standard tick label specification xlogbase None if a number, the axis and transformation are logarithmic with ticks at the given base (10 being the most common) (same for y) text_attr {} a dictionary of attributes for label text axis_attr {} a dictionary of attributes for the axis lines
Acts like Fig, but draws a coordinate frame around the data. You also need to supply plot ranges.
[ "Acts", "like", "Fig", "but", "draws", "a", "coordinate", "frame", "around", "the", "data", ".", "You", "also", "need", "to", "supply", "plot", "ranges", "." ]
def __init__(self, xmin, xmax, ymin, ymax, *d, **kwds): """Acts like Fig, but draws a coordinate frame around the data. You also need to supply plot ranges. Frame(xmin, xmax, ymin, ymax, obj, obj, obj..., keyword options...) xmin, xmax required minimum and maximum x values (in the objs' coordinates) ymin, ymax required minimum and maximum y values (in the objs' coordinates) obj optional list drawing primatives keyword options keyword list options defined below The following are keyword options, with their default values: x, y 20, 5 upper-left corner of the Frame in SVG coordinates width, height 75, 80 width and height of the Frame in SVG coordinates flipx, flipy False, True flip the sign of the coordinate axis minusInfinity -1000 if an axis is logarithmic and an object is plotted at 0 or a negative value, -1000 will be used as a stand-in for NaN xtitle None if a string, label the x axis xticks -10 request ticks according to the standard tick specification (see help(Ticks)) xminiticks True request miniticks according to the standard minitick specification xlabels True request tick labels according to the standard tick label specification xlogbase None if a number, the axis and transformation are logarithmic with ticks at the given base (10 being the most common) (same for y) text_attr {} a dictionary of attributes for label text axis_attr {} a dictionary of attributes for the axis lines """ self.xmin, self.xmax, self.ymin, self.ymax = xmin, xmax, ymin, ymax self.d = list(d) defaults = {"x":20, "y":5, "width":75, "height":80, "flipx":False, "flipy":True, "minusInfinity":-1000, \ "xtitle":None, "xticks":-10, "xminiticks":True, "xlabels":True, "x2labels":None, "xlogbase":None, \ "ytitle":None, "yticks":-10, "yminiticks":True, "ylabels":True, "y2labels":None, "ylogbase":None, \ "text_attr":{}, "axis_attr":{}} defaults.update(kwds) kwds = defaults self.x = kwds["x"]; del kwds["x"] self.y = kwds["y"]; del kwds["y"] self.width = kwds["width"]; del kwds["width"] self.height = kwds["height"]; del kwds["height"] self.flipx = kwds["flipx"]; del kwds["flipx"] self.flipy = kwds["flipy"]; del kwds["flipy"] self.minusInfinity = kwds["minusInfinity"]; del kwds["minusInfinity"] self.xtitle = kwds["xtitle"]; del kwds["xtitle"] self.xticks = kwds["xticks"]; del kwds["xticks"] self.xminiticks = kwds["xminiticks"]; del kwds["xminiticks"] self.xlabels = kwds["xlabels"]; del kwds["xlabels"] self.x2labels = kwds["x2labels"]; del kwds["x2labels"] self.xlogbase = kwds["xlogbase"]; del kwds["xlogbase"] self.ytitle = kwds["ytitle"]; del kwds["ytitle"] self.yticks = kwds["yticks"]; del kwds["yticks"] self.yminiticks = kwds["yminiticks"]; del kwds["yminiticks"] self.ylabels = kwds["ylabels"]; del kwds["ylabels"] self.y2labels = kwds["y2labels"]; del kwds["y2labels"] self.ylogbase = kwds["ylogbase"]; del kwds["ylogbase"] self.text_attr = dict(self.text_defaults) self.text_attr.update(kwds["text_attr"]); del kwds["text_attr"] self.axis_attr = dict(self.axis_defaults) self.axis_attr.update(kwds["axis_attr"]); del kwds["axis_attr"] if len(kwds) != 0: raise TypeError("Frame() got unexpected keyword arguments %s" % kwds.keys())
[ "def", "__init__", "(", "self", ",", "xmin", ",", "xmax", ",", "ymin", ",", "ymax", ",", "*", "d", ",", "*", "*", "kwds", ")", ":", "self", ".", "xmin", ",", "self", ".", "xmax", ",", "self", ".", "ymin", ",", "self", ".", "ymax", "=", "xmin", ",", "xmax", ",", "ymin", ",", "ymax", "self", ".", "d", "=", "list", "(", "d", ")", "defaults", "=", "{", "\"x\"", ":", "20", ",", "\"y\"", ":", "5", ",", "\"width\"", ":", "75", ",", "\"height\"", ":", "80", ",", "\"flipx\"", ":", "False", ",", "\"flipy\"", ":", "True", ",", "\"minusInfinity\"", ":", "-", "1000", ",", "\"xtitle\"", ":", "None", ",", "\"xticks\"", ":", "-", "10", ",", "\"xminiticks\"", ":", "True", ",", "\"xlabels\"", ":", "True", ",", "\"x2labels\"", ":", "None", ",", "\"xlogbase\"", ":", "None", ",", "\"ytitle\"", ":", "None", ",", "\"yticks\"", ":", "-", "10", ",", "\"yminiticks\"", ":", "True", ",", "\"ylabels\"", ":", "True", ",", "\"y2labels\"", ":", "None", ",", "\"ylogbase\"", ":", "None", ",", "\"text_attr\"", ":", "{", "}", ",", "\"axis_attr\"", ":", "{", "}", "}", "defaults", ".", "update", "(", "kwds", ")", "kwds", "=", "defaults", "self", ".", "x", "=", "kwds", "[", "\"x\"", "]", "del", "kwds", "[", "\"x\"", "]", "self", ".", "y", "=", "kwds", "[", "\"y\"", "]", "del", "kwds", "[", "\"y\"", "]", "self", ".", "width", "=", "kwds", "[", "\"width\"", "]", "del", "kwds", "[", "\"width\"", "]", "self", ".", "height", "=", "kwds", "[", "\"height\"", "]", "del", "kwds", "[", "\"height\"", "]", "self", ".", "flipx", "=", "kwds", "[", "\"flipx\"", "]", "del", "kwds", "[", "\"flipx\"", "]", "self", ".", "flipy", "=", "kwds", "[", "\"flipy\"", "]", "del", "kwds", "[", "\"flipy\"", "]", "self", ".", "minusInfinity", "=", "kwds", "[", "\"minusInfinity\"", "]", "del", "kwds", "[", "\"minusInfinity\"", "]", "self", ".", "xtitle", "=", "kwds", "[", "\"xtitle\"", "]", "del", "kwds", "[", "\"xtitle\"", "]", "self", ".", "xticks", "=", "kwds", "[", "\"xticks\"", "]", "del", "kwds", "[", "\"xticks\"", "]", "self", ".", "xminiticks", "=", "kwds", "[", "\"xminiticks\"", "]", "del", "kwds", "[", "\"xminiticks\"", "]", "self", ".", "xlabels", "=", "kwds", "[", "\"xlabels\"", "]", "del", "kwds", "[", "\"xlabels\"", "]", "self", ".", "x2labels", "=", "kwds", "[", "\"x2labels\"", "]", "del", "kwds", "[", "\"x2labels\"", "]", "self", ".", "xlogbase", "=", "kwds", "[", "\"xlogbase\"", "]", "del", "kwds", "[", "\"xlogbase\"", "]", "self", ".", "ytitle", "=", "kwds", "[", "\"ytitle\"", "]", "del", "kwds", "[", "\"ytitle\"", "]", "self", ".", "yticks", "=", "kwds", "[", "\"yticks\"", "]", "del", "kwds", "[", "\"yticks\"", "]", "self", ".", "yminiticks", "=", "kwds", "[", "\"yminiticks\"", "]", "del", "kwds", "[", "\"yminiticks\"", "]", "self", ".", "ylabels", "=", "kwds", "[", "\"ylabels\"", "]", "del", "kwds", "[", "\"ylabels\"", "]", "self", ".", "y2labels", "=", "kwds", "[", "\"y2labels\"", "]", "del", "kwds", "[", "\"y2labels\"", "]", "self", ".", "ylogbase", "=", "kwds", "[", "\"ylogbase\"", "]", "del", "kwds", "[", "\"ylogbase\"", "]", "self", ".", "text_attr", "=", "dict", "(", "self", ".", "text_defaults", ")", "self", ".", "text_attr", ".", "update", "(", "kwds", "[", "\"text_attr\"", "]", ")", "del", "kwds", "[", "\"text_attr\"", "]", "self", ".", "axis_attr", "=", "dict", "(", "self", ".", "axis_defaults", ")", "self", ".", "axis_attr", ".", "update", "(", "kwds", "[", "\"axis_attr\"", "]", ")", "del", "kwds", "[", "\"axis_attr\"", "]", "if", "len", "(", "kwds", ")", "!=", "0", ":", "raise", "TypeError", "(", "\"Frame() got unexpected keyword arguments %s\"", "%", "kwds", ".", "keys", "(", ")", ")" ]
https://github.com/cms-sw/cmssw/blob/fd9de012d503d3405420bcbeec0ec879baa57cf2/Alignment/MuonAlignment/python/svgfig.py#L887-L954
stan-dev/math
5fd79f89933269a4ca4d8dd1fde2a36d53d4768c
lib/cpplint_1.4.5/cpplint.py
python
_CppLintState.ResetErrorCounts
(self)
Sets the module's error statistic back to zero.
Sets the module's error statistic back to zero.
[ "Sets", "the", "module", "s", "error", "statistic", "back", "to", "zero", "." ]
def ResetErrorCounts(self): """Sets the module's error statistic back to zero.""" self.error_count = 0 self.errors_by_category = {}
[ "def", "ResetErrorCounts", "(", "self", ")", ":", "self", ".", "error_count", "=", "0", "self", ".", "errors_by_category", "=", "{", "}" ]
https://github.com/stan-dev/math/blob/5fd79f89933269a4ca4d8dd1fde2a36d53d4768c/lib/cpplint_1.4.5/cpplint.py#L1087-L1090
hanpfei/chromium-net
392cc1fa3a8f92f42e4071ab6e674d8e0482f83f
tools/git/for-all-touched-files.py
python
FilenamesFromGit
(branch_name, extensions)
return filenames
Provides a list of all new and modified files listed by [git diff branch_name] where branch_name can be blank to get a diff of the workspace. Excludes deleted files. If extensions is not an empty list, include only files with one of the extensions on the list.
Provides a list of all new and modified files listed by [git diff branch_name] where branch_name can be blank to get a diff of the workspace.
[ "Provides", "a", "list", "of", "all", "new", "and", "modified", "files", "listed", "by", "[", "git", "diff", "branch_name", "]", "where", "branch_name", "can", "be", "blank", "to", "get", "a", "diff", "of", "the", "workspace", "." ]
def FilenamesFromGit(branch_name, extensions): """Provides a list of all new and modified files listed by [git diff branch_name] where branch_name can be blank to get a diff of the workspace. Excludes deleted files. If extensions is not an empty list, include only files with one of the extensions on the list. """ lines = GitShell('git diff --stat=600,500 %s' % branch_name) filenames = [] for line in lines: line = line.lstrip() # Avoid summary line, and files that have been deleted (no plus). if line.find('|') != -1 and line.find('+') != -1: filename = line.split()[0] if filename: filename = filename.rstrip() ext = filename.rsplit('.')[-1] if not extensions or ext in extensions: filenames.append(filename) return filenames
[ "def", "FilenamesFromGit", "(", "branch_name", ",", "extensions", ")", ":", "lines", "=", "GitShell", "(", "'git diff --stat=600,500 %s'", "%", "branch_name", ")", "filenames", "=", "[", "]", "for", "line", "in", "lines", ":", "line", "=", "line", ".", "lstrip", "(", ")", "# Avoid summary line, and files that have been deleted (no plus).", "if", "line", ".", "find", "(", "'|'", ")", "!=", "-", "1", "and", "line", ".", "find", "(", "'+'", ")", "!=", "-", "1", ":", "filename", "=", "line", ".", "split", "(", ")", "[", "0", "]", "if", "filename", ":", "filename", "=", "filename", ".", "rstrip", "(", ")", "ext", "=", "filename", ".", "rsplit", "(", "'.'", ")", "[", "-", "1", "]", "if", "not", "extensions", "or", "ext", "in", "extensions", ":", "filenames", ".", "append", "(", "filename", ")", "return", "filenames" ]
https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/tools/git/for-all-touched-files.py#L53-L75
eomahony/Numberjack
53fa9e994a36f881ffd320d8d04158097190aad8
Numberjack/__init__.py
python
VarArray.__ge__
(self, other)
return LeqLex(other, self)
Syntactic sugar for the lexicographic order constraint :class:`LeqLex` so it can be specified on two VarArray like so: `X >= Y` :param VarArray other: Another VarArray of the same length. :rtype: LeqLex
Syntactic sugar for the lexicographic order constraint :class:`LeqLex` so it can be specified on two VarArray like so: `X >= Y`
[ "Syntactic", "sugar", "for", "the", "lexicographic", "order", "constraint", ":", "class", ":", "LeqLex", "so", "it", "can", "be", "specified", "on", "two", "VarArray", "like", "so", ":", "X", ">", "=", "Y" ]
def __ge__(self, other): """ Syntactic sugar for the lexicographic order constraint :class:`LeqLex` so it can be specified on two VarArray like so: `X >= Y` :param VarArray other: Another VarArray of the same length. :rtype: LeqLex """ return LeqLex(other, self)
[ "def", "__ge__", "(", "self", ",", "other", ")", ":", "return", "LeqLex", "(", "other", ",", "self", ")" ]
https://github.com/eomahony/Numberjack/blob/53fa9e994a36f881ffd320d8d04158097190aad8/Numberjack/__init__.py#L1194-L1202
SoarGroup/Soar
a1c5e249499137a27da60533c72969eef3b8ab6b
scons/scons-local-4.1.0/SCons/Node/__init__.py
python
Node.get_binfo
(self)
return binfo
Fetch a node's build information. node - the node whose sources will be collected cache - alternate node to use for the signature cache returns - the build signature This no longer handles the recursive descent of the node's children's signatures. We expect that they're already built and updated by someone else, if that's what's wanted.
Fetch a node's build information.
[ "Fetch", "a", "node", "s", "build", "information", "." ]
def get_binfo(self): """ Fetch a node's build information. node - the node whose sources will be collected cache - alternate node to use for the signature cache returns - the build signature This no longer handles the recursive descent of the node's children's signatures. We expect that they're already built and updated by someone else, if that's what's wanted. """ try: return self.binfo except AttributeError: pass binfo = self.new_binfo() self.binfo = binfo executor = self.get_executor() ignore_set = self.ignore_set if self.has_builder(): binfo.bact = str(executor) binfo.bactsig = MD5signature(executor.get_contents()) if self._specific_sources: sources = [s for s in self.sources if s not in ignore_set] else: sources = executor.get_unignored_sources(self, self.ignore) seen = set() binfo.bsources = [s for s in sources if s not in seen and not seen.add(s)] binfo.bsourcesigs = [s.get_ninfo() for s in binfo.bsources] binfo.bdepends = [d for d in self.depends if d not in ignore_set] binfo.bdependsigs = [d.get_ninfo() for d in self.depends] # Because self.implicit is initialized to None (and not empty list []) # we have to handle this case if not self.implicit: binfo.bimplicit = [] binfo.bimplicitsigs = [] else: binfo.bimplicit = [i for i in self.implicit if i not in ignore_set] binfo.bimplicitsigs = [i.get_ninfo() for i in binfo.bimplicit] return binfo
[ "def", "get_binfo", "(", "self", ")", ":", "try", ":", "return", "self", ".", "binfo", "except", "AttributeError", ":", "pass", "binfo", "=", "self", ".", "new_binfo", "(", ")", "self", ".", "binfo", "=", "binfo", "executor", "=", "self", ".", "get_executor", "(", ")", "ignore_set", "=", "self", ".", "ignore_set", "if", "self", ".", "has_builder", "(", ")", ":", "binfo", ".", "bact", "=", "str", "(", "executor", ")", "binfo", ".", "bactsig", "=", "MD5signature", "(", "executor", ".", "get_contents", "(", ")", ")", "if", "self", ".", "_specific_sources", ":", "sources", "=", "[", "s", "for", "s", "in", "self", ".", "sources", "if", "s", "not", "in", "ignore_set", "]", "else", ":", "sources", "=", "executor", ".", "get_unignored_sources", "(", "self", ",", "self", ".", "ignore", ")", "seen", "=", "set", "(", ")", "binfo", ".", "bsources", "=", "[", "s", "for", "s", "in", "sources", "if", "s", "not", "in", "seen", "and", "not", "seen", ".", "add", "(", "s", ")", "]", "binfo", ".", "bsourcesigs", "=", "[", "s", ".", "get_ninfo", "(", ")", "for", "s", "in", "binfo", ".", "bsources", "]", "binfo", ".", "bdepends", "=", "[", "d", "for", "d", "in", "self", ".", "depends", "if", "d", "not", "in", "ignore_set", "]", "binfo", ".", "bdependsigs", "=", "[", "d", ".", "get_ninfo", "(", ")", "for", "d", "in", "self", ".", "depends", "]", "# Because self.implicit is initialized to None (and not empty list [])", "# we have to handle this case", "if", "not", "self", ".", "implicit", ":", "binfo", ".", "bimplicit", "=", "[", "]", "binfo", ".", "bimplicitsigs", "=", "[", "]", "else", ":", "binfo", ".", "bimplicit", "=", "[", "i", "for", "i", "in", "self", ".", "implicit", "if", "i", "not", "in", "ignore_set", "]", "binfo", ".", "bimplicitsigs", "=", "[", "i", ".", "get_ninfo", "(", ")", "for", "i", "in", "binfo", ".", "bimplicit", "]", "return", "binfo" ]
https://github.com/SoarGroup/Soar/blob/a1c5e249499137a27da60533c72969eef3b8ab6b/scons/scons-local-4.1.0/SCons/Node/__init__.py#L1144-L1194
wlanjie/AndroidFFmpeg
7baf9122f4b8e1c74e7baf4be5c422c7a5ba5aaf
tools/fdk-aac-build/armeabi/toolchain/lib/python2.7/htmllib.py
python
HTMLParser.__init__
(self, formatter, verbose=0)
Creates an instance of the HTMLParser class. The formatter parameter is the formatter instance associated with the parser.
Creates an instance of the HTMLParser class.
[ "Creates", "an", "instance", "of", "the", "HTMLParser", "class", "." ]
def __init__(self, formatter, verbose=0): """Creates an instance of the HTMLParser class. The formatter parameter is the formatter instance associated with the parser. """ sgmllib.SGMLParser.__init__(self, verbose) self.formatter = formatter
[ "def", "__init__", "(", "self", ",", "formatter", ",", "verbose", "=", "0", ")", ":", "sgmllib", ".", "SGMLParser", ".", "__init__", "(", "self", ",", "verbose", ")", "self", ".", "formatter", "=", "formatter" ]
https://github.com/wlanjie/AndroidFFmpeg/blob/7baf9122f4b8e1c74e7baf4be5c422c7a5ba5aaf/tools/fdk-aac-build/armeabi/toolchain/lib/python2.7/htmllib.py#L34-L42
pytorch/pytorch
7176c92687d3cc847cc046bf002269c6949a21c2
torch/package/package_importer.py
python
PackageImporter._add_file
(self, filename: str)
Assembles a Python module out of the given file. Will ignore files in the .data directory. Args: filename (str): the name of the file inside of the package archive to be added
Assembles a Python module out of the given file. Will ignore files in the .data directory.
[ "Assembles", "a", "Python", "module", "out", "of", "the", "given", "file", ".", "Will", "ignore", "files", "in", "the", ".", "data", "directory", "." ]
def _add_file(self, filename: str): """Assembles a Python module out of the given file. Will ignore files in the .data directory. Args: filename (str): the name of the file inside of the package archive to be added """ *prefix, last = filename.split("/") if len(prefix) > 1 and prefix[0] == ".data": return package = self._get_or_create_package(prefix) if isinstance(package, _ExternNode): raise ImportError( f"inconsistent module structure. package contains a module file {filename}" f" that is a subpackage of a module marked external." ) if last == "__init__.py": package.source_file = filename elif last.endswith(".py"): package_name = last[: -len(".py")] package.children[package_name] = _ModuleNode(filename)
[ "def", "_add_file", "(", "self", ",", "filename", ":", "str", ")", ":", "*", "prefix", ",", "last", "=", "filename", ".", "split", "(", "\"/\"", ")", "if", "len", "(", "prefix", ")", ">", "1", "and", "prefix", "[", "0", "]", "==", "\".data\"", ":", "return", "package", "=", "self", ".", "_get_or_create_package", "(", "prefix", ")", "if", "isinstance", "(", "package", ",", "_ExternNode", ")", ":", "raise", "ImportError", "(", "f\"inconsistent module structure. package contains a module file {filename}\"", "f\" that is a subpackage of a module marked external.\"", ")", "if", "last", "==", "\"__init__.py\"", ":", "package", ".", "source_file", "=", "filename", "elif", "last", ".", "endswith", "(", "\".py\"", ")", ":", "package_name", "=", "last", "[", ":", "-", "len", "(", "\".py\"", ")", "]", "package", ".", "children", "[", "package_name", "]", "=", "_ModuleNode", "(", "filename", ")" ]
https://github.com/pytorch/pytorch/blob/7176c92687d3cc847cc046bf002269c6949a21c2/torch/package/package_importer.py#L558-L577
miyosuda/TensorFlowAndroidDemo
35903e0221aa5f109ea2dbef27f20b52e317f42d
jni-build/jni/include/tensorflow/contrib/lookup/lookup_ops.py
python
LookupInterface.__init__
(self, key_dtype, value_dtype, name)
Construct a lookup table interface. Args: key_dtype: The table key type. value_dtype: The table value type. name: A name for the operation (optional).
Construct a lookup table interface.
[ "Construct", "a", "lookup", "table", "interface", "." ]
def __init__(self, key_dtype, value_dtype, name): """Construct a lookup table interface. Args: key_dtype: The table key type. value_dtype: The table value type. name: A name for the operation (optional). """ self._key_dtype = dtypes.as_dtype(key_dtype) self._value_dtype = dtypes.as_dtype(value_dtype) self._name = name
[ "def", "__init__", "(", "self", ",", "key_dtype", ",", "value_dtype", ",", "name", ")", ":", "self", ".", "_key_dtype", "=", "dtypes", ".", "as_dtype", "(", "key_dtype", ")", "self", ".", "_value_dtype", "=", "dtypes", ".", "as_dtype", "(", "value_dtype", ")", "self", ".", "_name", "=", "name" ]
https://github.com/miyosuda/TensorFlowAndroidDemo/blob/35903e0221aa5f109ea2dbef27f20b52e317f42d/jni-build/jni/include/tensorflow/contrib/lookup/lookup_ops.py#L32-L42
NVIDIAGameWorks/kaolin
e5148d05e9c1e2ce92a07881ce3593b1c5c3f166
kaolin/io/usd.py
python
add_pointcloud
(stage, points, scene_path, colors=None, time=None, points_type='point_instancer')
return stage
r"""Add a pointcloud to an existing USD stage. Create a pointcloud represented by point instances of a sphere centered at each point coordinate. The stage is modified but not saved to disk. Args: stage (Usd.Stage): Stage onto which to add the pointcloud. points (torch.FloatTensor): Pointcloud tensor containing ``N`` points of shape ``(N, 3)``. scene_path (str): Absolute path of pointcloud within the USD file scene. Must be a valid Sdf.Path. colors (torch.FloatTensor, optional): Color tensor corresponding each point in the pointcloud tensor of shape ``(N, 3)``. colors only works if points_type is 'usd_geom_points'. time (convertible to float, optional): Positive integer defining the time at which the supplied parameters correspond to. points_type (str): String that indicates whether to save pointcloud as UsdGeomPoints or PointInstancer. 'usd_geom_points' indicates UsdGeomPoints and 'point_instancer' indicates PointInstancer. Please refer here for UsdGeomPoints: https://graphics.pixar.com/usd/docs/api/class_usd_geom_points.html and here for PointInstancer https://graphics.pixar.com/usd/docs/api/class_usd_geom_point_instancer.html. Default: 'point_instancer'. Returns: (Usd.Stage) Example: >>> stage = create_stage('./new_stage.usd') >>> points = torch.rand(100, 3) >>> stage = add_pointcloud(stage, points, '/World/PointClouds/pointcloud_0') >>> stage.Save()
r"""Add a pointcloud to an existing USD stage.
[ "r", "Add", "a", "pointcloud", "to", "an", "existing", "USD", "stage", "." ]
def add_pointcloud(stage, points, scene_path, colors=None, time=None, points_type='point_instancer'): r"""Add a pointcloud to an existing USD stage. Create a pointcloud represented by point instances of a sphere centered at each point coordinate. The stage is modified but not saved to disk. Args: stage (Usd.Stage): Stage onto which to add the pointcloud. points (torch.FloatTensor): Pointcloud tensor containing ``N`` points of shape ``(N, 3)``. scene_path (str): Absolute path of pointcloud within the USD file scene. Must be a valid Sdf.Path. colors (torch.FloatTensor, optional): Color tensor corresponding each point in the pointcloud tensor of shape ``(N, 3)``. colors only works if points_type is 'usd_geom_points'. time (convertible to float, optional): Positive integer defining the time at which the supplied parameters correspond to. points_type (str): String that indicates whether to save pointcloud as UsdGeomPoints or PointInstancer. 'usd_geom_points' indicates UsdGeomPoints and 'point_instancer' indicates PointInstancer. Please refer here for UsdGeomPoints: https://graphics.pixar.com/usd/docs/api/class_usd_geom_points.html and here for PointInstancer https://graphics.pixar.com/usd/docs/api/class_usd_geom_point_instancer.html. Default: 'point_instancer'. Returns: (Usd.Stage) Example: >>> stage = create_stage('./new_stage.usd') >>> points = torch.rand(100, 3) >>> stage = add_pointcloud(stage, points, '/World/PointClouds/pointcloud_0') >>> stage.Save() """ scene_path = Sdf.Path(scene_path) if time is None: time = Usd.TimeCode.Default() if stage.GetPrimAtPath(scene_path): points_prim = stage.GetPrimAtPath(scene_path) else: if points_type == 'point_instancer': points_prim = stage.DefinePrim(scene_path, 'PointInstancer') elif points_type == 'usd_geom_points': points_prim = stage.DefinePrim(scene_path, 'Points') else: raise ValueError('Expected points_type to be "usd_geom_points" or "point_instancer", ' f'but got "{points_type}".') if points_type == 'point_instancer': geom_points = UsdGeom.PointInstancer(points_prim) sphere = UsdGeom.Sphere.Define(stage, f'{scene_path}/sphere') sphere.GetRadiusAttr().Set(0.5) geom_points.CreatePrototypesRel().SetTargets([sphere.GetPath()]) elif points_type == 'usd_geom_points': geom_points = UsdGeom.Points(points_prim) # Calculate default point scale bounds = points.max(dim=0)[0] - points.min(dim=0)[0] min_bound = min(bounds) scale = (min_bound / points.size(0) ** (1 / 3)).item() # Generate instancer parameters positions = points.detach().cpu().tolist() scales = np.asarray([scale, ] * points.size(0)) if points_type == 'point_instancer': indices = [0] * points.size(0) # Populate point instancer geom_points.GetProtoIndicesAttr().Set(indices, time=time) geom_points.GetPositionsAttr().Set(positions, time=time) scales = [(scale,) * 3] * points.size(0) geom_points.GetScalesAttr().Set(scales, time=time) elif points_type == 'usd_geom_points': # Populate UsdGeomPoints geom_points.GetPointsAttr().Set(points.numpy(), time=time) geom_points.GetWidthsAttr().Set(Vt.FloatArray.FromNumpy(scales), time=time) if colors is not None and points_type == 'usd_geom_points': assert colors.shape == points.shape, 'Colors and points must have the same shape.' geom_points.GetDisplayColorAttr().Set(colors.numpy(), time=time) return stage
[ "def", "add_pointcloud", "(", "stage", ",", "points", ",", "scene_path", ",", "colors", "=", "None", ",", "time", "=", "None", ",", "points_type", "=", "'point_instancer'", ")", ":", "scene_path", "=", "Sdf", ".", "Path", "(", "scene_path", ")", "if", "time", "is", "None", ":", "time", "=", "Usd", ".", "TimeCode", ".", "Default", "(", ")", "if", "stage", ".", "GetPrimAtPath", "(", "scene_path", ")", ":", "points_prim", "=", "stage", ".", "GetPrimAtPath", "(", "scene_path", ")", "else", ":", "if", "points_type", "==", "'point_instancer'", ":", "points_prim", "=", "stage", ".", "DefinePrim", "(", "scene_path", ",", "'PointInstancer'", ")", "elif", "points_type", "==", "'usd_geom_points'", ":", "points_prim", "=", "stage", ".", "DefinePrim", "(", "scene_path", ",", "'Points'", ")", "else", ":", "raise", "ValueError", "(", "'Expected points_type to be \"usd_geom_points\" or \"point_instancer\", '", "f'but got \"{points_type}\".'", ")", "if", "points_type", "==", "'point_instancer'", ":", "geom_points", "=", "UsdGeom", ".", "PointInstancer", "(", "points_prim", ")", "sphere", "=", "UsdGeom", ".", "Sphere", ".", "Define", "(", "stage", ",", "f'{scene_path}/sphere'", ")", "sphere", ".", "GetRadiusAttr", "(", ")", ".", "Set", "(", "0.5", ")", "geom_points", ".", "CreatePrototypesRel", "(", ")", ".", "SetTargets", "(", "[", "sphere", ".", "GetPath", "(", ")", "]", ")", "elif", "points_type", "==", "'usd_geom_points'", ":", "geom_points", "=", "UsdGeom", ".", "Points", "(", "points_prim", ")", "# Calculate default point scale", "bounds", "=", "points", ".", "max", "(", "dim", "=", "0", ")", "[", "0", "]", "-", "points", ".", "min", "(", "dim", "=", "0", ")", "[", "0", "]", "min_bound", "=", "min", "(", "bounds", ")", "scale", "=", "(", "min_bound", "/", "points", ".", "size", "(", "0", ")", "**", "(", "1", "/", "3", ")", ")", ".", "item", "(", ")", "# Generate instancer parameters", "positions", "=", "points", ".", "detach", "(", ")", ".", "cpu", "(", ")", ".", "tolist", "(", ")", "scales", "=", "np", ".", "asarray", "(", "[", "scale", ",", "]", "*", "points", ".", "size", "(", "0", ")", ")", "if", "points_type", "==", "'point_instancer'", ":", "indices", "=", "[", "0", "]", "*", "points", ".", "size", "(", "0", ")", "# Populate point instancer", "geom_points", ".", "GetProtoIndicesAttr", "(", ")", ".", "Set", "(", "indices", ",", "time", "=", "time", ")", "geom_points", ".", "GetPositionsAttr", "(", ")", ".", "Set", "(", "positions", ",", "time", "=", "time", ")", "scales", "=", "[", "(", "scale", ",", ")", "*", "3", "]", "*", "points", ".", "size", "(", "0", ")", "geom_points", ".", "GetScalesAttr", "(", ")", ".", "Set", "(", "scales", ",", "time", "=", "time", ")", "elif", "points_type", "==", "'usd_geom_points'", ":", "# Populate UsdGeomPoints", "geom_points", ".", "GetPointsAttr", "(", ")", ".", "Set", "(", "points", ".", "numpy", "(", ")", ",", "time", "=", "time", ")", "geom_points", ".", "GetWidthsAttr", "(", ")", ".", "Set", "(", "Vt", ".", "FloatArray", ".", "FromNumpy", "(", "scales", ")", ",", "time", "=", "time", ")", "if", "colors", "is", "not", "None", "and", "points_type", "==", "'usd_geom_points'", ":", "assert", "colors", ".", "shape", "==", "points", ".", "shape", ",", "'Colors and points must have the same shape.'", "geom_points", ".", "GetDisplayColorAttr", "(", ")", ".", "Set", "(", "colors", ".", "numpy", "(", ")", ",", "time", "=", "time", ")", "return", "stage" ]
https://github.com/NVIDIAGameWorks/kaolin/blob/e5148d05e9c1e2ce92a07881ce3593b1c5c3f166/kaolin/io/usd.py#L918-L994
OSGeo/gdal
3748fc4ba4fba727492774b2b908a2130c864a83
swig/python/osgeo/ogr.py
python
DataSource.CopyLayer
(self, *args, **kwargs)
return _ogr.DataSource_CopyLayer(self, *args, **kwargs)
r""" CopyLayer(DataSource self, Layer src_layer, char const * new_name, char ** options=None) -> Layer OGRLayerH OGR_DS_CopyLayer(OGRDataSourceH hDS, OGRLayerH hSrcLayer, const char *pszNewName, char **papszOptions) Duplicate an existing layer. This function creates a new layer, duplicate the field definitions of the source layer and then duplicate each features of the source layer. The papszOptions argument can be used to control driver specific creation options. These options are normally documented in the format specific documentation. The source layer may come from another dataset. Deprecated Use GDALDatasetCopyLayer() in GDAL 2.0 Parameters: ----------- hDS: handle to the data source where to create the new layer hSrcLayer: handle to the source layer. pszNewName: the name of the layer to create. papszOptions: a StringList of name=value options. Options are driver specific. a handle to the layer, or NULL if an error occurs.
r""" CopyLayer(DataSource self, Layer src_layer, char const * new_name, char ** options=None) -> Layer OGRLayerH OGR_DS_CopyLayer(OGRDataSourceH hDS, OGRLayerH hSrcLayer, const char *pszNewName, char **papszOptions)
[ "r", "CopyLayer", "(", "DataSource", "self", "Layer", "src_layer", "char", "const", "*", "new_name", "char", "**", "options", "=", "None", ")", "-", ">", "Layer", "OGRLayerH", "OGR_DS_CopyLayer", "(", "OGRDataSourceH", "hDS", "OGRLayerH", "hSrcLayer", "const", "char", "*", "pszNewName", "char", "**", "papszOptions", ")" ]
def CopyLayer(self, *args, **kwargs): r""" CopyLayer(DataSource self, Layer src_layer, char const * new_name, char ** options=None) -> Layer OGRLayerH OGR_DS_CopyLayer(OGRDataSourceH hDS, OGRLayerH hSrcLayer, const char *pszNewName, char **papszOptions) Duplicate an existing layer. This function creates a new layer, duplicate the field definitions of the source layer and then duplicate each features of the source layer. The papszOptions argument can be used to control driver specific creation options. These options are normally documented in the format specific documentation. The source layer may come from another dataset. Deprecated Use GDALDatasetCopyLayer() in GDAL 2.0 Parameters: ----------- hDS: handle to the data source where to create the new layer hSrcLayer: handle to the source layer. pszNewName: the name of the layer to create. papszOptions: a StringList of name=value options. Options are driver specific. a handle to the layer, or NULL if an error occurs. """ return _ogr.DataSource_CopyLayer(self, *args, **kwargs)
[ "def", "CopyLayer", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "_ogr", ".", "DataSource_CopyLayer", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")" ]
https://github.com/OSGeo/gdal/blob/3748fc4ba4fba727492774b2b908a2130c864a83/swig/python/osgeo/ogr.py#L739-L771
emscripten-core/emscripten
0d413d3c5af8b28349682496edc14656f5700c2f
third_party/ply/example/BASIC/basparse.py
python
p_command_def
(p)
command : DEF ID LPAREN ID RPAREN EQUALS expr
command : DEF ID LPAREN ID RPAREN EQUALS expr
[ "command", ":", "DEF", "ID", "LPAREN", "ID", "RPAREN", "EQUALS", "expr" ]
def p_command_def(p): '''command : DEF ID LPAREN ID RPAREN EQUALS expr''' p[0] = ('FUNC',p[2],p[4],p[7])
[ "def", "p_command_def", "(", "p", ")", ":", "p", "[", "0", "]", "=", "(", "'FUNC'", ",", "p", "[", "2", "]", ",", "p", "[", "4", "]", ",", "p", "[", "7", "]", ")" ]
https://github.com/emscripten-core/emscripten/blob/0d413d3c5af8b28349682496edc14656f5700c2f/third_party/ply/example/BASIC/basparse.py#L221-L223
windystrife/UnrealEngine_NVIDIAGameWorks
b50e6338a7c5b26374d66306ebc7807541ff815e
Engine/Extras/ThirdPartyNotUE/emsdk/Win64/python/2.7.5.3_64bit/Lib/lib-tk/turtle.py
python
RawTurtle._cc
(self, args)
return "#%02x%02x%02x" % (r, g, b)
Convert colortriples to hexstrings.
Convert colortriples to hexstrings.
[ "Convert", "colortriples", "to", "hexstrings", "." ]
def _cc(self, args): """Convert colortriples to hexstrings. """ if isinstance(args, str): return args try: r, g, b = args except: raise TurtleGraphicsError("bad color arguments: %s" % str(args)) if self.screen._colormode == 1.0: r, g, b = [round(255.0*x) for x in (r, g, b)] if not ((0 <= r <= 255) and (0 <= g <= 255) and (0 <= b <= 255)): raise TurtleGraphicsError("bad color sequence: %s" % str(args)) return "#%02x%02x%02x" % (r, g, b)
[ "def", "_cc", "(", "self", ",", "args", ")", ":", "if", "isinstance", "(", "args", ",", "str", ")", ":", "return", "args", "try", ":", "r", ",", "g", ",", "b", "=", "args", "except", ":", "raise", "TurtleGraphicsError", "(", "\"bad color arguments: %s\"", "%", "str", "(", "args", ")", ")", "if", "self", ".", "screen", ".", "_colormode", "==", "1.0", ":", "r", ",", "g", ",", "b", "=", "[", "round", "(", "255.0", "*", "x", ")", "for", "x", "in", "(", "r", ",", "g", ",", "b", ")", "]", "if", "not", "(", "(", "0", "<=", "r", "<=", "255", ")", "and", "(", "0", "<=", "g", "<=", "255", ")", "and", "(", "0", "<=", "b", "<=", "255", ")", ")", ":", "raise", "TurtleGraphicsError", "(", "\"bad color sequence: %s\"", "%", "str", "(", "args", ")", ")", "return", "\"#%02x%02x%02x\"", "%", "(", "r", ",", "g", ",", "b", ")" ]
https://github.com/windystrife/UnrealEngine_NVIDIAGameWorks/blob/b50e6338a7c5b26374d66306ebc7807541ff815e/Engine/Extras/ThirdPartyNotUE/emsdk/Win64/python/2.7.5.3_64bit/Lib/lib-tk/turtle.py#L2602-L2615
MythTV/mythtv
d282a209cb8be85d036f85a62a8ec971b67d45f4
mythtv/programs/scripts/internetcontent/nv_python_libs/pbs/pbs_api.py
python
Videos.getPBSConfig
(self)
return
Read the MNV PBS grabber "pbs_config.xml" configuration file return nothing
Read the MNV PBS grabber "pbs_config.xml" configuration file return nothing
[ "Read", "the", "MNV", "PBS", "grabber", "pbs_config", ".", "xml", "configuration", "file", "return", "nothing" ]
def getPBSConfig(self): ''' Read the MNV PBS grabber "pbs_config.xml" configuration file return nothing ''' # Read the grabber pbs_config.xml configuration file url = 'file://%s/nv_python_libs/configs/XML/pbs_config.xml' % (baseProcessingDir, ) if not os.path.isfile(url[7:]): raise PBSConfigFileError(self.error_messages['PBSConfigFileError'] % (url[7:], )) if self.config['debug_enabled']: print(url) print() try: self.pbs_config = etree.parse(url) except Exception as e: raise PBSUrlError(self.error_messages['PBSUrlError'] % (url, errormsg)) return
[ "def", "getPBSConfig", "(", "self", ")", ":", "# Read the grabber pbs_config.xml configuration file", "url", "=", "'file://%s/nv_python_libs/configs/XML/pbs_config.xml'", "%", "(", "baseProcessingDir", ",", ")", "if", "not", "os", ".", "path", ".", "isfile", "(", "url", "[", "7", ":", "]", ")", ":", "raise", "PBSConfigFileError", "(", "self", ".", "error_messages", "[", "'PBSConfigFileError'", "]", "%", "(", "url", "[", "7", ":", "]", ",", ")", ")", "if", "self", ".", "config", "[", "'debug_enabled'", "]", ":", "print", "(", "url", ")", "print", "(", ")", "try", ":", "self", ".", "pbs_config", "=", "etree", ".", "parse", "(", "url", ")", "except", "Exception", "as", "e", ":", "raise", "PBSUrlError", "(", "self", ".", "error_messages", "[", "'PBSUrlError'", "]", "%", "(", "url", ",", "errormsg", ")", ")", "return" ]
https://github.com/MythTV/mythtv/blob/d282a209cb8be85d036f85a62a8ec971b67d45f4/mythtv/programs/scripts/internetcontent/nv_python_libs/pbs/pbs_api.py#L200-L216
LiquidPlayer/LiquidCore
9405979363f2353ac9a71ad8ab59685dd7f919c9
deps/node-10.15.3/deps/npm/node_modules/node-gyp/gyp/pylib/gyp/generator/cmake.py
python
NormjoinPath
(base_path, rel_path)
return os.path.normpath(os.path.join(base_path, rel_path))
Resolves rel_path against base_path and returns the result. TODO: what is this really used for? If rel_path begins with '$' it is returned unchanged. Otherwise it is resolved against base_path if relative, then normalized.
Resolves rel_path against base_path and returns the result. TODO: what is this really used for? If rel_path begins with '$' it is returned unchanged. Otherwise it is resolved against base_path if relative, then normalized.
[ "Resolves", "rel_path", "against", "base_path", "and", "returns", "the", "result", ".", "TODO", ":", "what", "is", "this", "really", "used", "for?", "If", "rel_path", "begins", "with", "$", "it", "is", "returned", "unchanged", ".", "Otherwise", "it", "is", "resolved", "against", "base_path", "if", "relative", "then", "normalized", "." ]
def NormjoinPath(base_path, rel_path): """Resolves rel_path against base_path and returns the result. TODO: what is this really used for? If rel_path begins with '$' it is returned unchanged. Otherwise it is resolved against base_path if relative, then normalized. """ if rel_path.startswith('$') and not rel_path.startswith('${configuration}'): return rel_path return os.path.normpath(os.path.join(base_path, rel_path))
[ "def", "NormjoinPath", "(", "base_path", ",", "rel_path", ")", ":", "if", "rel_path", ".", "startswith", "(", "'$'", ")", "and", "not", "rel_path", ".", "startswith", "(", "'${configuration}'", ")", ":", "return", "rel_path", "return", "os", ".", "path", ".", "normpath", "(", "os", ".", "path", ".", "join", "(", "base_path", ",", "rel_path", ")", ")" ]
https://github.com/LiquidPlayer/LiquidCore/blob/9405979363f2353ac9a71ad8ab59685dd7f919c9/deps/node-10.15.3/deps/npm/node_modules/node-gyp/gyp/pylib/gyp/generator/cmake.py#L110-L118
miyosuda/TensorFlowAndroidMNIST
7b5a4603d2780a8a2834575706e9001977524007
jni-build/jni/include/tensorflow/contrib/distributions/python/ops/normal.py
python
Normal.log_prob
(self, x, name="log_prob")
Log prob of observations in `x` under these Normal distribution(s). Args: x: tensor of dtype `dtype`, must be broadcastable with `mu` and `sigma`. name: The name to give this op. Returns: log_prob: tensor of dtype `dtype`, the log-PDFs of `x`.
Log prob of observations in `x` under these Normal distribution(s).
[ "Log", "prob", "of", "observations", "in", "x", "under", "these", "Normal", "distribution", "(", "s", ")", "." ]
def log_prob(self, x, name="log_prob"): """Log prob of observations in `x` under these Normal distribution(s). Args: x: tensor of dtype `dtype`, must be broadcastable with `mu` and `sigma`. name: The name to give this op. Returns: log_prob: tensor of dtype `dtype`, the log-PDFs of `x`. """ with ops.name_scope(self.name): with ops.op_scope([self._mu, self._sigma, x], name): x = ops.convert_to_tensor(x) if x.dtype != self.dtype: raise TypeError("Input x dtype does not match dtype: %s vs. %s" % (x.dtype, self.dtype)) log_2_pi = constant_op.constant(math.log(2 * math.pi), dtype=self.dtype) return (-0.5*log_2_pi - math_ops.log(self._sigma) -0.5*math_ops.square((x - self._mu) / self._sigma))
[ "def", "log_prob", "(", "self", ",", "x", ",", "name", "=", "\"log_prob\"", ")", ":", "with", "ops", ".", "name_scope", "(", "self", ".", "name", ")", ":", "with", "ops", ".", "op_scope", "(", "[", "self", ".", "_mu", ",", "self", ".", "_sigma", ",", "x", "]", ",", "name", ")", ":", "x", "=", "ops", ".", "convert_to_tensor", "(", "x", ")", "if", "x", ".", "dtype", "!=", "self", ".", "dtype", ":", "raise", "TypeError", "(", "\"Input x dtype does not match dtype: %s vs. %s\"", "%", "(", "x", ".", "dtype", ",", "self", ".", "dtype", ")", ")", "log_2_pi", "=", "constant_op", ".", "constant", "(", "math", ".", "log", "(", "2", "*", "math", ".", "pi", ")", ",", "dtype", "=", "self", ".", "dtype", ")", "return", "(", "-", "0.5", "*", "log_2_pi", "-", "math_ops", ".", "log", "(", "self", ".", "_sigma", ")", "-", "0.5", "*", "math_ops", ".", "square", "(", "(", "x", "-", "self", ".", "_mu", ")", "/", "self", ".", "_sigma", ")", ")" ]
https://github.com/miyosuda/TensorFlowAndroidMNIST/blob/7b5a4603d2780a8a2834575706e9001977524007/jni-build/jni/include/tensorflow/contrib/distributions/python/ops/normal.py#L223-L241
aws/lumberyard
f85344403c1c2e77ec8c75deb2c116e97b713217
dev/Tools/Python/3.7.10/windows/Lib/tkinter/__init__.py
python
Button.invoke
(self)
return self.tk.call(self._w, 'invoke')
Invoke the command associated with the button. The return value is the return value from the command, or an empty string if there is no command associated with the button. This command is ignored if the button's state is disabled.
Invoke the command associated with the button.
[ "Invoke", "the", "command", "associated", "with", "the", "button", "." ]
def invoke(self): """Invoke the command associated with the button. The return value is the return value from the command, or an empty string if there is no command associated with the button. This command is ignored if the button's state is disabled. """ return self.tk.call(self._w, 'invoke')
[ "def", "invoke", "(", "self", ")", ":", "return", "self", ".", "tk", ".", "call", "(", "self", ".", "_w", ",", "'invoke'", ")" ]
https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/windows/Lib/tkinter/__init__.py#L2383-L2391
wxWidgets/wxPython-Classic
19571e1ae65f1ac445f5491474121998c97a1bf0
src/gtk/aui.py
python
AuiToolBarItem.GetBitmap
(*args, **kwargs)
return _aui.AuiToolBarItem_GetBitmap(*args, **kwargs)
GetBitmap(self) -> Bitmap
GetBitmap(self) -> Bitmap
[ "GetBitmap", "(", "self", ")", "-", ">", "Bitmap" ]
def GetBitmap(*args, **kwargs): """GetBitmap(self) -> Bitmap""" return _aui.AuiToolBarItem_GetBitmap(*args, **kwargs)
[ "def", "GetBitmap", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "_aui", ".", "AuiToolBarItem_GetBitmap", "(", "*", "args", ",", "*", "*", "kwargs", ")" ]
https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/gtk/aui.py#L1781-L1783
aws/lumberyard
f85344403c1c2e77ec8c75deb2c116e97b713217
dev/Tools/Python/3.7.10/mac/Python.framework/Versions/3.7/lib/python3.7/email/message.py
python
Message.get_content_charset
(self, failobj=None)
return charset.lower()
Return the charset parameter of the Content-Type header. The returned string is always coerced to lower case. If there is no Content-Type header, or if that header has no charset parameter, failobj is returned.
Return the charset parameter of the Content-Type header.
[ "Return", "the", "charset", "parameter", "of", "the", "Content", "-", "Type", "header", "." ]
def get_content_charset(self, failobj=None): """Return the charset parameter of the Content-Type header. The returned string is always coerced to lower case. If there is no Content-Type header, or if that header has no charset parameter, failobj is returned. """ missing = object() charset = self.get_param('charset', missing) if charset is missing: return failobj if isinstance(charset, tuple): # RFC 2231 encoded, so decode it, and it better end up as ascii. pcharset = charset[0] or 'us-ascii' try: # LookupError will be raised if the charset isn't known to # Python. UnicodeError will be raised if the encoded text # contains a character not in the charset. as_bytes = charset[2].encode('raw-unicode-escape') charset = str(as_bytes, pcharset) except (LookupError, UnicodeError): charset = charset[2] # charset characters must be in us-ascii range try: charset.encode('us-ascii') except UnicodeError: return failobj # RFC 2046, $4.1.2 says charsets are not case sensitive return charset.lower()
[ "def", "get_content_charset", "(", "self", ",", "failobj", "=", "None", ")", ":", "missing", "=", "object", "(", ")", "charset", "=", "self", ".", "get_param", "(", "'charset'", ",", "missing", ")", "if", "charset", "is", "missing", ":", "return", "failobj", "if", "isinstance", "(", "charset", ",", "tuple", ")", ":", "# RFC 2231 encoded, so decode it, and it better end up as ascii.", "pcharset", "=", "charset", "[", "0", "]", "or", "'us-ascii'", "try", ":", "# LookupError will be raised if the charset isn't known to", "# Python. UnicodeError will be raised if the encoded text", "# contains a character not in the charset.", "as_bytes", "=", "charset", "[", "2", "]", ".", "encode", "(", "'raw-unicode-escape'", ")", "charset", "=", "str", "(", "as_bytes", ",", "pcharset", ")", "except", "(", "LookupError", ",", "UnicodeError", ")", ":", "charset", "=", "charset", "[", "2", "]", "# charset characters must be in us-ascii range", "try", ":", "charset", ".", "encode", "(", "'us-ascii'", ")", "except", "UnicodeError", ":", "return", "failobj", "# RFC 2046, $4.1.2 says charsets are not case sensitive", "return", "charset", ".", "lower", "(", ")" ]
https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/mac/Python.framework/Versions/3.7/lib/python3.7/email/message.py#L881-L909
RobotLocomotion/drake
0e18a34604c45ed65bc9018a54f7610f91cdad5b
tools/workspace/drake_visualizer/_drake_visualizer_builtin_scripts/show_hydroelastic_contact.py
python
HydroelasticContactVisualizer.toggle_show_traction_vectors
(self, state)
Slot for dialog widget
Slot for dialog widget
[ "Slot", "for", "dialog", "widget" ]
def toggle_show_traction_vectors(self, state): """Slot for dialog widget""" self.show_traction_vectors = state self.update_visual_data_from_message()
[ "def", "toggle_show_traction_vectors", "(", "self", ",", "state", ")", ":", "self", ".", "show_traction_vectors", "=", "state", "self", ".", "update_visual_data_from_message", "(", ")" ]
https://github.com/RobotLocomotion/drake/blob/0e18a34604c45ed65bc9018a54f7610f91cdad5b/tools/workspace/drake_visualizer/_drake_visualizer_builtin_scripts/show_hydroelastic_contact.py#L1073-L1076
tensorflow/tensorflow
419e3a6b650ea4bd1b0cba23c4348f8a69f3272e
tensorflow/python/keras/layers/legacy_rnn/rnn_cell_impl.py
python
_zero_state_tensors
(state_size, batch_size, dtype)
return nest.map_structure(get_state_shape, state_size)
Create tensors of zeros based on state_size, batch_size, and dtype.
Create tensors of zeros based on state_size, batch_size, and dtype.
[ "Create", "tensors", "of", "zeros", "based", "on", "state_size", "batch_size", "and", "dtype", "." ]
def _zero_state_tensors(state_size, batch_size, dtype): """Create tensors of zeros based on state_size, batch_size, and dtype.""" def get_state_shape(s): """Combine s with batch_size to get a proper tensor shape.""" c = _concat(batch_size, s) size = array_ops.zeros(c, dtype=dtype) if not context.executing_eagerly(): c_static = _concat(batch_size, s, static=True) size.set_shape(c_static) return size return nest.map_structure(get_state_shape, state_size)
[ "def", "_zero_state_tensors", "(", "state_size", ",", "batch_size", ",", "dtype", ")", ":", "def", "get_state_shape", "(", "s", ")", ":", "\"\"\"Combine s with batch_size to get a proper tensor shape.\"\"\"", "c", "=", "_concat", "(", "batch_size", ",", "s", ")", "size", "=", "array_ops", ".", "zeros", "(", "c", ",", "dtype", "=", "dtype", ")", "if", "not", "context", ".", "executing_eagerly", "(", ")", ":", "c_static", "=", "_concat", "(", "batch_size", ",", "s", ",", "static", "=", "True", ")", "size", ".", "set_shape", "(", "c_static", ")", "return", "size", "return", "nest", ".", "map_structure", "(", "get_state_shape", ",", "state_size", ")" ]
https://github.com/tensorflow/tensorflow/blob/419e3a6b650ea4bd1b0cba23c4348f8a69f3272e/tensorflow/python/keras/layers/legacy_rnn/rnn_cell_impl.py#L168-L180
pmq20/node-packer
12c46c6e44fbc14d9ee645ebd17d5296b324f7e0
lts/deps/npm/node_modules/node-gyp/gyp/pylib/gyp/generator/make.py
python
MakefileWriter.WriteActions
(self, actions, extra_sources, extra_outputs, extra_mac_bundle_resources, part_of_all)
Write Makefile code for any 'actions' from the gyp input. extra_sources: a list that will be filled in with newly generated source files, if any extra_outputs: a list that will be filled in with any outputs of these actions (used to make other pieces dependent on these actions) part_of_all: flag indicating this target is part of 'all'
Write Makefile code for any 'actions' from the gyp input.
[ "Write", "Makefile", "code", "for", "any", "actions", "from", "the", "gyp", "input", "." ]
def WriteActions(self, actions, extra_sources, extra_outputs, extra_mac_bundle_resources, part_of_all): """Write Makefile code for any 'actions' from the gyp input. extra_sources: a list that will be filled in with newly generated source files, if any extra_outputs: a list that will be filled in with any outputs of these actions (used to make other pieces dependent on these actions) part_of_all: flag indicating this target is part of 'all' """ env = self.GetSortedXcodeEnv() for action in actions: name = StringToMakefileVariable('%s_%s' % (self.qualified_target, action['action_name'])) self.WriteLn('### Rules for action "%s":' % action['action_name']) inputs = action['inputs'] outputs = action['outputs'] # Build up a list of outputs. # Collect the output dirs we'll need. dirs = set() for out in outputs: dir = os.path.split(out)[0] if dir: dirs.add(dir) if int(action.get('process_outputs_as_sources', False)): extra_sources += outputs if int(action.get('process_outputs_as_mac_bundle_resources', False)): extra_mac_bundle_resources += outputs # Write the actual command. action_commands = action['action'] if self.flavor == 'mac': action_commands = [gyp.xcode_emulation.ExpandEnvVars(command, env) for command in action_commands] command = gyp.common.EncodePOSIXShellList(action_commands) if 'message' in action: self.WriteLn('quiet_cmd_%s = ACTION %s $@' % (name, action['message'])) else: self.WriteLn('quiet_cmd_%s = ACTION %s $@' % (name, name)) if len(dirs) > 0: command = 'mkdir -p %s' % ' '.join(dirs) + '; ' + command cd_action = 'cd %s; ' % Sourceify(self.path or '.') # command and cd_action get written to a toplevel variable called # cmd_foo. Toplevel variables can't handle things that change per # makefile like $(TARGET), so hardcode the target. command = command.replace('$(TARGET)', self.target) cd_action = cd_action.replace('$(TARGET)', self.target) # Set LD_LIBRARY_PATH in case the action runs an executable from this # build which links to shared libs from this build. # actions run on the host, so they should in theory only use host # libraries, but until everything is made cross-compile safe, also use # target libraries. # TODO(piman): when everything is cross-compile safe, remove lib.target self.WriteLn('cmd_%s = LD_LIBRARY_PATH=$(builddir)/lib.host:' '$(builddir)/lib.target:$$LD_LIBRARY_PATH; ' 'export LD_LIBRARY_PATH; ' '%s%s' % (name, cd_action, command)) self.WriteLn() outputs = [self.Absolutify(output) for output in outputs] # The makefile rules are all relative to the top dir, but the gyp actions # are defined relative to their containing dir. This replaces the obj # variable for the action rule with an absolute version so that the output # goes in the right place. # Only write the 'obj' and 'builddir' rules for the "primary" output (:1); # it's superfluous for the "extra outputs", and this avoids accidentally # writing duplicate dummy rules for those outputs. # Same for environment. self.WriteLn("%s: obj := $(abs_obj)" % QuoteSpaces(outputs[0])) self.WriteLn("%s: builddir := $(abs_builddir)" % QuoteSpaces(outputs[0])) self.WriteSortedXcodeEnv(outputs[0], self.GetSortedXcodeEnv()) for input in inputs: assert ' ' not in input, ( "Spaces in action input filenames not supported (%s)" % input) for output in outputs: assert ' ' not in output, ( "Spaces in action output filenames not supported (%s)" % output) # See the comment in WriteCopies about expanding env vars. outputs = [gyp.xcode_emulation.ExpandEnvVars(o, env) for o in outputs] inputs = [gyp.xcode_emulation.ExpandEnvVars(i, env) for i in inputs] self.WriteDoCmd(outputs, [Sourceify(self.Absolutify(i)) for i in inputs], part_of_all=part_of_all, command=name) # Stuff the outputs in a variable so we can refer to them later. outputs_variable = 'action_%s_outputs' % name self.WriteLn('%s := %s' % (outputs_variable, ' '.join(outputs))) extra_outputs.append('$(%s)' % outputs_variable) self.WriteLn() self.WriteLn()
[ "def", "WriteActions", "(", "self", ",", "actions", ",", "extra_sources", ",", "extra_outputs", ",", "extra_mac_bundle_resources", ",", "part_of_all", ")", ":", "env", "=", "self", ".", "GetSortedXcodeEnv", "(", ")", "for", "action", "in", "actions", ":", "name", "=", "StringToMakefileVariable", "(", "'%s_%s'", "%", "(", "self", ".", "qualified_target", ",", "action", "[", "'action_name'", "]", ")", ")", "self", ".", "WriteLn", "(", "'### Rules for action \"%s\":'", "%", "action", "[", "'action_name'", "]", ")", "inputs", "=", "action", "[", "'inputs'", "]", "outputs", "=", "action", "[", "'outputs'", "]", "# Build up a list of outputs.", "# Collect the output dirs we'll need.", "dirs", "=", "set", "(", ")", "for", "out", "in", "outputs", ":", "dir", "=", "os", ".", "path", ".", "split", "(", "out", ")", "[", "0", "]", "if", "dir", ":", "dirs", ".", "add", "(", "dir", ")", "if", "int", "(", "action", ".", "get", "(", "'process_outputs_as_sources'", ",", "False", ")", ")", ":", "extra_sources", "+=", "outputs", "if", "int", "(", "action", ".", "get", "(", "'process_outputs_as_mac_bundle_resources'", ",", "False", ")", ")", ":", "extra_mac_bundle_resources", "+=", "outputs", "# Write the actual command.", "action_commands", "=", "action", "[", "'action'", "]", "if", "self", ".", "flavor", "==", "'mac'", ":", "action_commands", "=", "[", "gyp", ".", "xcode_emulation", ".", "ExpandEnvVars", "(", "command", ",", "env", ")", "for", "command", "in", "action_commands", "]", "command", "=", "gyp", ".", "common", ".", "EncodePOSIXShellList", "(", "action_commands", ")", "if", "'message'", "in", "action", ":", "self", ".", "WriteLn", "(", "'quiet_cmd_%s = ACTION %s $@'", "%", "(", "name", ",", "action", "[", "'message'", "]", ")", ")", "else", ":", "self", ".", "WriteLn", "(", "'quiet_cmd_%s = ACTION %s $@'", "%", "(", "name", ",", "name", ")", ")", "if", "len", "(", "dirs", ")", ">", "0", ":", "command", "=", "'mkdir -p %s'", "%", "' '", ".", "join", "(", "dirs", ")", "+", "'; '", "+", "command", "cd_action", "=", "'cd %s; '", "%", "Sourceify", "(", "self", ".", "path", "or", "'.'", ")", "# command and cd_action get written to a toplevel variable called", "# cmd_foo. Toplevel variables can't handle things that change per", "# makefile like $(TARGET), so hardcode the target.", "command", "=", "command", ".", "replace", "(", "'$(TARGET)'", ",", "self", ".", "target", ")", "cd_action", "=", "cd_action", ".", "replace", "(", "'$(TARGET)'", ",", "self", ".", "target", ")", "# Set LD_LIBRARY_PATH in case the action runs an executable from this", "# build which links to shared libs from this build.", "# actions run on the host, so they should in theory only use host", "# libraries, but until everything is made cross-compile safe, also use", "# target libraries.", "# TODO(piman): when everything is cross-compile safe, remove lib.target", "self", ".", "WriteLn", "(", "'cmd_%s = LD_LIBRARY_PATH=$(builddir)/lib.host:'", "'$(builddir)/lib.target:$$LD_LIBRARY_PATH; '", "'export LD_LIBRARY_PATH; '", "'%s%s'", "%", "(", "name", ",", "cd_action", ",", "command", ")", ")", "self", ".", "WriteLn", "(", ")", "outputs", "=", "[", "self", ".", "Absolutify", "(", "output", ")", "for", "output", "in", "outputs", "]", "# The makefile rules are all relative to the top dir, but the gyp actions", "# are defined relative to their containing dir. This replaces the obj", "# variable for the action rule with an absolute version so that the output", "# goes in the right place.", "# Only write the 'obj' and 'builddir' rules for the \"primary\" output (:1);", "# it's superfluous for the \"extra outputs\", and this avoids accidentally", "# writing duplicate dummy rules for those outputs.", "# Same for environment.", "self", ".", "WriteLn", "(", "\"%s: obj := $(abs_obj)\"", "%", "QuoteSpaces", "(", "outputs", "[", "0", "]", ")", ")", "self", ".", "WriteLn", "(", "\"%s: builddir := $(abs_builddir)\"", "%", "QuoteSpaces", "(", "outputs", "[", "0", "]", ")", ")", "self", ".", "WriteSortedXcodeEnv", "(", "outputs", "[", "0", "]", ",", "self", ".", "GetSortedXcodeEnv", "(", ")", ")", "for", "input", "in", "inputs", ":", "assert", "' '", "not", "in", "input", ",", "(", "\"Spaces in action input filenames not supported (%s)\"", "%", "input", ")", "for", "output", "in", "outputs", ":", "assert", "' '", "not", "in", "output", ",", "(", "\"Spaces in action output filenames not supported (%s)\"", "%", "output", ")", "# See the comment in WriteCopies about expanding env vars.", "outputs", "=", "[", "gyp", ".", "xcode_emulation", ".", "ExpandEnvVars", "(", "o", ",", "env", ")", "for", "o", "in", "outputs", "]", "inputs", "=", "[", "gyp", ".", "xcode_emulation", ".", "ExpandEnvVars", "(", "i", ",", "env", ")", "for", "i", "in", "inputs", "]", "self", ".", "WriteDoCmd", "(", "outputs", ",", "[", "Sourceify", "(", "self", ".", "Absolutify", "(", "i", ")", ")", "for", "i", "in", "inputs", "]", ",", "part_of_all", "=", "part_of_all", ",", "command", "=", "name", ")", "# Stuff the outputs in a variable so we can refer to them later.", "outputs_variable", "=", "'action_%s_outputs'", "%", "name", "self", ".", "WriteLn", "(", "'%s := %s'", "%", "(", "outputs_variable", ",", "' '", ".", "join", "(", "outputs", ")", ")", ")", "extra_outputs", ".", "append", "(", "'$(%s)'", "%", "outputs_variable", ")", "self", ".", "WriteLn", "(", ")", "self", ".", "WriteLn", "(", ")" ]
https://github.com/pmq20/node-packer/blob/12c46c6e44fbc14d9ee645ebd17d5296b324f7e0/lts/deps/npm/node_modules/node-gyp/gyp/pylib/gyp/generator/make.py#L889-L986
OSGeo/gdal
3748fc4ba4fba727492774b2b908a2130c864a83
swig/python/osgeo/gdal.py
python
Band.GetColorTable
(self, *args)
return _gdal.Band_GetColorTable(self, *args)
r"""GetColorTable(Band self) -> ColorTable
r"""GetColorTable(Band self) -> ColorTable
[ "r", "GetColorTable", "(", "Band", "self", ")", "-", ">", "ColorTable" ]
def GetColorTable(self, *args): r"""GetColorTable(Band self) -> ColorTable""" return _gdal.Band_GetColorTable(self, *args)
[ "def", "GetColorTable", "(", "self", ",", "*", "args", ")", ":", "return", "_gdal", ".", "Band_GetColorTable", "(", "self", ",", "*", "args", ")" ]
https://github.com/OSGeo/gdal/blob/3748fc4ba4fba727492774b2b908a2130c864a83/swig/python/osgeo/gdal.py#L3512-L3514
PaddlePaddle/Paddle
1252f4bb3e574df80aa6d18c7ddae1b3a90bd81c
python/paddle/fluid/incubate/fleet/base/role_maker.py
python
MPISymetricRoleMaker._all_reduce
(self, input, output, mode="sum")
all reduce between trainers if current role is TRAINER, only support array of one dim. Args: input(list/numpy.array): array of one dim output(list/numpy.array): array of one dim mode(str): "sum" or "min" or "max"
all reduce between trainers if current role is TRAINER, only support array of one dim.
[ "all", "reduce", "between", "trainers", "if", "current", "role", "is", "TRAINER", "only", "support", "array", "of", "one", "dim", "." ]
def _all_reduce(self, input, output, mode="sum"): """ all reduce between trainers if current role is TRAINER, only support array of one dim. Args: input(list/numpy.array): array of one dim output(list/numpy.array): array of one dim mode(str): "sum" or "min" or "max" """ if not self._role_is_generated: self.generate_role() if mode == "sum": mode = self.MPI.SUM elif mode == "max": mode = self.MPI.MAX elif mode == "min": mode = self.MPI.MIN else: raise ValueError("unknown mode: %s" % mode) self._node_type_comm.Allreduce(input, output, op=mode)
[ "def", "_all_reduce", "(", "self", ",", "input", ",", "output", ",", "mode", "=", "\"sum\"", ")", ":", "if", "not", "self", ".", "_role_is_generated", ":", "self", ".", "generate_role", "(", ")", "if", "mode", "==", "\"sum\"", ":", "mode", "=", "self", ".", "MPI", ".", "SUM", "elif", "mode", "==", "\"max\"", ":", "mode", "=", "self", ".", "MPI", ".", "MAX", "elif", "mode", "==", "\"min\"", ":", "mode", "=", "self", ".", "MPI", ".", "MIN", "else", ":", "raise", "ValueError", "(", "\"unknown mode: %s\"", "%", "mode", ")", "self", ".", "_node_type_comm", ".", "Allreduce", "(", "input", ",", "output", ",", "op", "=", "mode", ")" ]
https://github.com/PaddlePaddle/Paddle/blob/1252f4bb3e574df80aa6d18c7ddae1b3a90bd81c/python/paddle/fluid/incubate/fleet/base/role_maker.py#L419-L439
wlanjie/AndroidFFmpeg
7baf9122f4b8e1c74e7baf4be5c422c7a5ba5aaf
tools/fdk-aac-build/armeabi/toolchain/lib/python2.7/lib2to3/pytree.py
python
Base.pre_order
(self)
Return a pre-order iterator for the tree. This must be implemented by the concrete subclass.
Return a pre-order iterator for the tree.
[ "Return", "a", "pre", "-", "order", "iterator", "for", "the", "tree", "." ]
def pre_order(self): """ Return a pre-order iterator for the tree. This must be implemented by the concrete subclass. """ raise NotImplementedError
[ "def", "pre_order", "(", "self", ")", ":", "raise", "NotImplementedError" ]
https://github.com/wlanjie/AndroidFFmpeg/blob/7baf9122f4b8e1c74e7baf4be5c422c7a5ba5aaf/tools/fdk-aac-build/armeabi/toolchain/lib/python2.7/lib2to3/pytree.py#L104-L110
Xilinx/Vitis-AI
fc74d404563d9951b57245443c73bef389f3657f
tools/Vitis-AI-Quantizer/vai_q_tensorflow1.x/tensorflow/python/ops/gradients_util.py
python
_GetGrad
(grads, t, unconnected_gradients)
return t_grad
Gets gradient for tensor "t".
Gets gradient for tensor "t".
[ "Gets", "gradient", "for", "tensor", "t", "." ]
def _GetGrad(grads, t, unconnected_gradients): """Gets gradient for tensor "t".""" op = t.op op_grads = grads.get(op) if not op_grads: if unconnected_gradients == UnconnectedGradients.ZERO: t_dtype = default_gradient.get_zeros_dtype(t) if t.dtype == dtypes.resource: return array_ops.zeros( resource_variable_ops.variable_shape(t), dtype=t_dtype) else: return array_ops.zeros_like(t, dtype=t_dtype) elif unconnected_gradients == UnconnectedGradients.NONE: return None else: raise ValueError( "Unknown value for unconnected_gradients: %r" % unconnected_gradients) t_grad = op_grads[t.value_index] assert not isinstance( t_grad, list), ("gradients list should have been aggregated by now.") return t_grad
[ "def", "_GetGrad", "(", "grads", ",", "t", ",", "unconnected_gradients", ")", ":", "op", "=", "t", ".", "op", "op_grads", "=", "grads", ".", "get", "(", "op", ")", "if", "not", "op_grads", ":", "if", "unconnected_gradients", "==", "UnconnectedGradients", ".", "ZERO", ":", "t_dtype", "=", "default_gradient", ".", "get_zeros_dtype", "(", "t", ")", "if", "t", ".", "dtype", "==", "dtypes", ".", "resource", ":", "return", "array_ops", ".", "zeros", "(", "resource_variable_ops", ".", "variable_shape", "(", "t", ")", ",", "dtype", "=", "t_dtype", ")", "else", ":", "return", "array_ops", ".", "zeros_like", "(", "t", ",", "dtype", "=", "t_dtype", ")", "elif", "unconnected_gradients", "==", "UnconnectedGradients", ".", "NONE", ":", "return", "None", "else", ":", "raise", "ValueError", "(", "\"Unknown value for unconnected_gradients: %r\"", "%", "unconnected_gradients", ")", "t_grad", "=", "op_grads", "[", "t", ".", "value_index", "]", "assert", "not", "isinstance", "(", "t_grad", ",", "list", ")", ",", "(", "\"gradients list should have been aggregated by now.\"", ")", "return", "t_grad" ]
https://github.com/Xilinx/Vitis-AI/blob/fc74d404563d9951b57245443c73bef389f3657f/tools/Vitis-AI-Quantizer/vai_q_tensorflow1.x/tensorflow/python/ops/gradients_util.py#L794-L815
wxWidgets/wxPython-Classic
19571e1ae65f1ac445f5491474121998c97a1bf0
src/gtk/combo.py
python
ComboPopup.DestroyPopup
(*args, **kwargs)
return _combo.ComboPopup_DestroyPopup(*args, **kwargs)
DestroyPopup(self)
DestroyPopup(self)
[ "DestroyPopup", "(", "self", ")" ]
def DestroyPopup(*args, **kwargs): """DestroyPopup(self)""" return _combo.ComboPopup_DestroyPopup(*args, **kwargs)
[ "def", "DestroyPopup", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "_combo", ".", "ComboPopup_DestroyPopup", "(", "*", "args", ",", "*", "*", "kwargs", ")" ]
https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/gtk/combo.py#L633-L635
google/llvm-propeller
45c226984fe8377ebfb2ad7713c680d652ba678d
compiler-rt/lib/sanitizer_common/scripts/cpplint.py
python
CheckCommaSpacing
(filename, clean_lines, linenum, error)
Checks for horizontal spacing near commas and semicolons. Args: filename: The name of the current file. clean_lines: A CleansedLines instance containing the file. linenum: The number of the line to check. error: The function to call with any errors found.
Checks for horizontal spacing near commas and semicolons.
[ "Checks", "for", "horizontal", "spacing", "near", "commas", "and", "semicolons", "." ]
def CheckCommaSpacing(filename, clean_lines, linenum, error): """Checks for horizontal spacing near commas and semicolons. Args: filename: The name of the current file. clean_lines: A CleansedLines instance containing the file. linenum: The number of the line to check. error: The function to call with any errors found. """ raw = clean_lines.lines_without_raw_strings line = clean_lines.elided[linenum] # You should always have a space after a comma (either as fn arg or operator) # # This does not apply when the non-space character following the # comma is another comma, since the only time when that happens is # for empty macro arguments. # # We run this check in two passes: first pass on elided lines to # verify that lines contain missing whitespaces, second pass on raw # lines to confirm that those missing whitespaces are not due to # elided comments. if (Search(r',[^,\s]', ReplaceAll(r'\boperator\s*,\s*\(', 'F(', line)) and Search(r',[^,\s]', raw[linenum])): error(filename, linenum, 'whitespace/comma', 3, 'Missing space after ,') # You should always have a space after a semicolon # except for few corner cases # TODO(unknown): clarify if 'if (1) { return 1;}' is requires one more # space after ; if Search(r';[^\s};\\)/]', line): error(filename, linenum, 'whitespace/semicolon', 3, 'Missing space after ;')
[ "def", "CheckCommaSpacing", "(", "filename", ",", "clean_lines", ",", "linenum", ",", "error", ")", ":", "raw", "=", "clean_lines", ".", "lines_without_raw_strings", "line", "=", "clean_lines", ".", "elided", "[", "linenum", "]", "# You should always have a space after a comma (either as fn arg or operator)", "#", "# This does not apply when the non-space character following the", "# comma is another comma, since the only time when that happens is", "# for empty macro arguments.", "#", "# We run this check in two passes: first pass on elided lines to", "# verify that lines contain missing whitespaces, second pass on raw", "# lines to confirm that those missing whitespaces are not due to", "# elided comments.", "if", "(", "Search", "(", "r',[^,\\s]'", ",", "ReplaceAll", "(", "r'\\boperator\\s*,\\s*\\('", ",", "'F('", ",", "line", ")", ")", "and", "Search", "(", "r',[^,\\s]'", ",", "raw", "[", "linenum", "]", ")", ")", ":", "error", "(", "filename", ",", "linenum", ",", "'whitespace/comma'", ",", "3", ",", "'Missing space after ,'", ")", "# You should always have a space after a semicolon", "# except for few corner cases", "# TODO(unknown): clarify if 'if (1) { return 1;}' is requires one more", "# space after ;", "if", "Search", "(", "r';[^\\s};\\\\)/]'", ",", "line", ")", ":", "error", "(", "filename", ",", "linenum", ",", "'whitespace/semicolon'", ",", "3", ",", "'Missing space after ;'", ")" ]
https://github.com/google/llvm-propeller/blob/45c226984fe8377ebfb2ad7713c680d652ba678d/compiler-rt/lib/sanitizer_common/scripts/cpplint.py#L3452-L3485
mindspore-ai/mindspore
fb8fd3338605bb34fa5cea054e535a8b1d753fab
mindspore/python/mindspore/ops/_op_impl/aicpu/uniform_real.py
python
_uniform_real_aicpu
()
return
RandomUniformReal AiCPU register
RandomUniformReal AiCPU register
[ "RandomUniformReal", "AiCPU", "register" ]
def _uniform_real_aicpu(): """RandomUniformReal AiCPU register""" return
[ "def", "_uniform_real_aicpu", "(", ")", ":", "return" ]
https://github.com/mindspore-ai/mindspore/blob/fb8fd3338605bb34fa5cea054e535a8b1d753fab/mindspore/python/mindspore/ops/_op_impl/aicpu/uniform_real.py#L31-L33
panda3d/panda3d
833ad89ebad58395d0af0b7ec08538e5e4308265
direct/src/controls/PhysicsWalker.py
python
PhysicsWalker.displayDebugInfo
(self)
For debug use.
For debug use.
[ "For", "debug", "use", "." ]
def displayDebugInfo(self): """ For debug use. """ onScreenDebug.add("w controls", "PhysicsWalker") if self.useLifter: onScreenDebug.add("w airborneHeight", self.lifter.getAirborneHeight()) onScreenDebug.add("w isOnGround", self.lifter.isOnGround()) #onScreenDebug.add("w gravity", self.lifter.getGravity()) onScreenDebug.add("w contact normal", self.lifter.getContactNormal().pPrintValues()) onScreenDebug.add("w impact", self.lifter.getImpactVelocity()) onScreenDebug.add("w velocity", self.lifter.getVelocity()) onScreenDebug.add("w hasContact", self.lifter.hasContact()) #onScreenDebug.add("w falling", self.falling) #onScreenDebug.add("w jumpForce", self.avatarControlJumpForce) #onScreenDebug.add("w mayJump", self.mayJump) onScreenDebug.add("w isAirborne", self.isAirborne)
[ "def", "displayDebugInfo", "(", "self", ")", ":", "onScreenDebug", ".", "add", "(", "\"w controls\"", ",", "\"PhysicsWalker\"", ")", "if", "self", ".", "useLifter", ":", "onScreenDebug", ".", "add", "(", "\"w airborneHeight\"", ",", "self", ".", "lifter", ".", "getAirborneHeight", "(", ")", ")", "onScreenDebug", ".", "add", "(", "\"w isOnGround\"", ",", "self", ".", "lifter", ".", "isOnGround", "(", ")", ")", "#onScreenDebug.add(\"w gravity\", self.lifter.getGravity())", "onScreenDebug", ".", "add", "(", "\"w contact normal\"", ",", "self", ".", "lifter", ".", "getContactNormal", "(", ")", ".", "pPrintValues", "(", ")", ")", "onScreenDebug", ".", "add", "(", "\"w impact\"", ",", "self", ".", "lifter", ".", "getImpactVelocity", "(", ")", ")", "onScreenDebug", ".", "add", "(", "\"w velocity\"", ",", "self", ".", "lifter", ".", "getVelocity", "(", ")", ")", "onScreenDebug", ".", "add", "(", "\"w hasContact\"", ",", "self", ".", "lifter", ".", "hasContact", "(", ")", ")", "#onScreenDebug.add(\"w falling\", self.falling)", "#onScreenDebug.add(\"w jumpForce\", self.avatarControlJumpForce)", "#onScreenDebug.add(\"w mayJump\", self.mayJump)", "onScreenDebug", ".", "add", "(", "\"w isAirborne\"", ",", "self", ".", "isAirborne", ")" ]
https://github.com/panda3d/panda3d/blob/833ad89ebad58395d0af0b7ec08538e5e4308265/direct/src/controls/PhysicsWalker.py#L373-L390
SequoiaDB/SequoiaDB
2894ed7e5bd6fe57330afc900cf76d0ff0df9f64
tools/server/php_linux/libxml2/lib/python2.4/site-packages/libxml2.py
python
xmlNs.setNs
(self, node)
Associate a namespace to a node, a posteriori.
Associate a namespace to a node, a posteriori.
[ "Associate", "a", "namespace", "to", "a", "node", "a", "posteriori", "." ]
def setNs(self, node): """Associate a namespace to a node, a posteriori. """ if node is None: node__o = None else: node__o = node._o libxml2mod.xmlSetNs(node__o, self._o)
[ "def", "setNs", "(", "self", ",", "node", ")", ":", "if", "node", "is", "None", ":", "node__o", "=", "None", "else", ":", "node__o", "=", "node", ".", "_o", "libxml2mod", ".", "xmlSetNs", "(", "node__o", ",", "self", ".", "_o", ")" ]
https://github.com/SequoiaDB/SequoiaDB/blob/2894ed7e5bd6fe57330afc900cf76d0ff0df9f64/tools/server/php_linux/libxml2/lib/python2.4/site-packages/libxml2.py#L5942-L5946
miyosuda/TensorFlowAndroidMNIST
7b5a4603d2780a8a2834575706e9001977524007
jni-build/jni/include/tensorflow/python/summary/impl/io_wrapper.py
python
ListDirectoryAbsolute
(directory)
Yields all files in the given directory. The paths are absolute.
Yields all files in the given directory. The paths are absolute.
[ "Yields", "all", "files", "in", "the", "given", "directory", ".", "The", "paths", "are", "absolute", "." ]
def ListDirectoryAbsolute(directory): """Yields all files in the given directory. The paths are absolute.""" if gcs.IsGCSPath(directory): return gcs.ListDirectory(directory) else: return (os.path.join(directory, path) for path in gfile.ListDirectory(directory))
[ "def", "ListDirectoryAbsolute", "(", "directory", ")", ":", "if", "gcs", ".", "IsGCSPath", "(", "directory", ")", ":", "return", "gcs", ".", "ListDirectory", "(", "directory", ")", "else", ":", "return", "(", "os", ".", "path", ".", "join", "(", "directory", ",", "path", ")", "for", "path", "in", "gfile", ".", "ListDirectory", "(", "directory", ")", ")" ]
https://github.com/miyosuda/TensorFlowAndroidMNIST/blob/7b5a4603d2780a8a2834575706e9001977524007/jni-build/jni/include/tensorflow/python/summary/impl/io_wrapper.py#L48-L54
aws/lumberyard
f85344403c1c2e77ec8c75deb2c116e97b713217
dev/Tools/Python/3.7.10/linux_x64/lib/python3.7/uuid.py
python
uuid3
(namespace, name)
return UUID(bytes=hash[:16], version=3)
Generate a UUID from the MD5 hash of a namespace UUID and a name.
Generate a UUID from the MD5 hash of a namespace UUID and a name.
[ "Generate", "a", "UUID", "from", "the", "MD5", "hash", "of", "a", "namespace", "UUID", "and", "a", "name", "." ]
def uuid3(namespace, name): """Generate a UUID from the MD5 hash of a namespace UUID and a name.""" from hashlib import md5 hash = md5(namespace.bytes + bytes(name, "utf-8")).digest() return UUID(bytes=hash[:16], version=3)
[ "def", "uuid3", "(", "namespace", ",", "name", ")", ":", "from", "hashlib", "import", "md5", "hash", "=", "md5", "(", "namespace", ".", "bytes", "+", "bytes", "(", "name", ",", "\"utf-8\"", ")", ")", ".", "digest", "(", ")", "return", "UUID", "(", "bytes", "=", "hash", "[", ":", "16", "]", ",", "version", "=", "3", ")" ]
https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/linux_x64/lib/python3.7/uuid.py#L753-L757
wxWidgets/wxPython-Classic
19571e1ae65f1ac445f5491474121998c97a1bf0
wx/lib/agw/toasterbox.py
python
ToasterBox.SetPopupScrollSpeed
(self, speed)
Sets the :class:`ToasterBox` scroll speed. :param `speed`: it is the pause time (in milliseconds) for every step in the `ScrollUp` method.
Sets the :class:`ToasterBox` scroll speed.
[ "Sets", "the", ":", "class", ":", "ToasterBox", "scroll", "speed", "." ]
def SetPopupScrollSpeed(self, speed): """ Sets the :class:`ToasterBox` scroll speed. :param `speed`: it is the pause time (in milliseconds) for every step in the `ScrollUp` method. """ self._sleeptime = speed
[ "def", "SetPopupScrollSpeed", "(", "self", ",", "speed", ")", ":", "self", ".", "_sleeptime", "=", "speed" ]
https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/wx/lib/agw/toasterbox.py#L473-L481
ceph/ceph
959663007321a369c83218414a29bd9dbc8bda3a
src/pybind/mgr/dashboard/controllers/_version.py
python
APIVersion.from_string
(cls, version_string: str)
return cls._make(int(s) for s in version_string.split('.'))
>>> APIVersion.from_string("1.0") APIVersion(major=1, minor=0)
>>> APIVersion.from_string("1.0") APIVersion(major=1, minor=0)
[ ">>>", "APIVersion", ".", "from_string", "(", "1", ".", "0", ")", "APIVersion", "(", "major", "=", "1", "minor", "=", "0", ")" ]
def from_string(cls, version_string: str) -> 'APIVersion': """ >>> APIVersion.from_string("1.0") APIVersion(major=1, minor=0) """ return cls._make(int(s) for s in version_string.split('.'))
[ "def", "from_string", "(", "cls", ",", "version_string", ":", "str", ")", "->", "'APIVersion'", ":", "return", "cls", ".", "_make", "(", "int", "(", "s", ")", "for", "s", "in", "version_string", ".", "split", "(", "'.'", ")", ")" ]
https://github.com/ceph/ceph/blob/959663007321a369c83218414a29bd9dbc8bda3a/src/pybind/mgr/dashboard/controllers/_version.py#L27-L32
wxWidgets/wxPython-Classic
19571e1ae65f1ac445f5491474121998c97a1bf0
src/osx_cocoa/_windows.py
python
Frame.DoMenuUpdates
(*args, **kwargs)
return _windows_.Frame_DoMenuUpdates(*args, **kwargs)
DoMenuUpdates(self, Menu menu=None)
DoMenuUpdates(self, Menu menu=None)
[ "DoMenuUpdates", "(", "self", "Menu", "menu", "=", "None", ")" ]
def DoMenuUpdates(*args, **kwargs): """DoMenuUpdates(self, Menu menu=None)""" return _windows_.Frame_DoMenuUpdates(*args, **kwargs)
[ "def", "DoMenuUpdates", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "_windows_", ".", "Frame_DoMenuUpdates", "(", "*", "args", ",", "*", "*", "kwargs", ")" ]
https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/osx_cocoa/_windows.py#L665-L667
domino-team/openwrt-cc
8b181297c34d14d3ca521cc9f31430d561dbc688
package/gli-pub/openwrt-node-packages-master/node/node-v6.9.1/deps/v8/tools/run_perf.py
python
AccumulateGenericResults
(graph_names, suite_units, iter_output)
return reduce(lambda r, t: r + t, traces.itervalues(), Results())
Iterates over the output of multiple benchmark reruns and accumulates generic results. Args: graph_names: List of names that configure the base path of the traces. E.g. ['v8', 'Octane']. suite_units: Measurement default units as defined by the benchmark suite. iter_output: Iterator over the standard output of each test run. Returns: A "Results" object.
Iterates over the output of multiple benchmark reruns and accumulates generic results.
[ "Iterates", "over", "the", "output", "of", "multiple", "benchmark", "reruns", "and", "accumulates", "generic", "results", "." ]
def AccumulateGenericResults(graph_names, suite_units, iter_output): """Iterates over the output of multiple benchmark reruns and accumulates generic results. Args: graph_names: List of names that configure the base path of the traces. E.g. ['v8', 'Octane']. suite_units: Measurement default units as defined by the benchmark suite. iter_output: Iterator over the standard output of each test run. Returns: A "Results" object. """ traces = OrderedDict() for stdout in iter_output(): if stdout is None: # The None value is used as a null object to simplify logic. continue for line in stdout.strip().splitlines(): match = GENERIC_RESULTS_RE.match(line) if match: stddev = "" graph = match.group(1) trace = match.group(2) body = match.group(3) units = match.group(4) match_stddev = RESULT_STDDEV_RE.match(body) match_list = RESULT_LIST_RE.match(body) errors = [] if match_stddev: result, stddev = map(str.strip, match_stddev.group(1).split(",")) results = [result] elif match_list: results = map(str.strip, match_list.group(1).split(",")) else: results = [body.strip()] try: results = map(lambda r: str(float(r)), results) except ValueError: results = [] errors = ["Found non-numeric in %s" % "/".join(graph_names + [graph, trace])] trace_result = traces.setdefault(trace, Results([{ "graphs": graph_names + [graph, trace], "units": (units or suite_units).strip(), "results": [], "stddev": "", }], errors)) trace_result.traces[0]["results"].extend(results) trace_result.traces[0]["stddev"] = stddev return reduce(lambda r, t: r + t, traces.itervalues(), Results())
[ "def", "AccumulateGenericResults", "(", "graph_names", ",", "suite_units", ",", "iter_output", ")", ":", "traces", "=", "OrderedDict", "(", ")", "for", "stdout", "in", "iter_output", "(", ")", ":", "if", "stdout", "is", "None", ":", "# The None value is used as a null object to simplify logic.", "continue", "for", "line", "in", "stdout", ".", "strip", "(", ")", ".", "splitlines", "(", ")", ":", "match", "=", "GENERIC_RESULTS_RE", ".", "match", "(", "line", ")", "if", "match", ":", "stddev", "=", "\"\"", "graph", "=", "match", ".", "group", "(", "1", ")", "trace", "=", "match", ".", "group", "(", "2", ")", "body", "=", "match", ".", "group", "(", "3", ")", "units", "=", "match", ".", "group", "(", "4", ")", "match_stddev", "=", "RESULT_STDDEV_RE", ".", "match", "(", "body", ")", "match_list", "=", "RESULT_LIST_RE", ".", "match", "(", "body", ")", "errors", "=", "[", "]", "if", "match_stddev", ":", "result", ",", "stddev", "=", "map", "(", "str", ".", "strip", ",", "match_stddev", ".", "group", "(", "1", ")", ".", "split", "(", "\",\"", ")", ")", "results", "=", "[", "result", "]", "elif", "match_list", ":", "results", "=", "map", "(", "str", ".", "strip", ",", "match_list", ".", "group", "(", "1", ")", ".", "split", "(", "\",\"", ")", ")", "else", ":", "results", "=", "[", "body", ".", "strip", "(", ")", "]", "try", ":", "results", "=", "map", "(", "lambda", "r", ":", "str", "(", "float", "(", "r", ")", ")", ",", "results", ")", "except", "ValueError", ":", "results", "=", "[", "]", "errors", "=", "[", "\"Found non-numeric in %s\"", "%", "\"/\"", ".", "join", "(", "graph_names", "+", "[", "graph", ",", "trace", "]", ")", "]", "trace_result", "=", "traces", ".", "setdefault", "(", "trace", ",", "Results", "(", "[", "{", "\"graphs\"", ":", "graph_names", "+", "[", "graph", ",", "trace", "]", ",", "\"units\"", ":", "(", "units", "or", "suite_units", ")", ".", "strip", "(", ")", ",", "\"results\"", ":", "[", "]", ",", "\"stddev\"", ":", "\"\"", ",", "}", "]", ",", "errors", ")", ")", "trace_result", ".", "traces", "[", "0", "]", "[", "\"results\"", "]", ".", "extend", "(", "results", ")", "trace_result", ".", "traces", "[", "0", "]", "[", "\"stddev\"", "]", "=", "stddev", "return", "reduce", "(", "lambda", "r", ",", "t", ":", "r", "+", "t", ",", "traces", ".", "itervalues", "(", ")", ",", "Results", "(", ")", ")" ]
https://github.com/domino-team/openwrt-cc/blob/8b181297c34d14d3ca521cc9f31430d561dbc688/package/gli-pub/openwrt-node-packages-master/node/node-v6.9.1/deps/v8/tools/run_perf.py#L288-L339
wlanjie/AndroidFFmpeg
7baf9122f4b8e1c74e7baf4be5c422c7a5ba5aaf
tools/fdk-aac-build/armeabi-v7a/toolchain/lib/python2.7/lib-tk/Tkinter.py
python
Wm.wm_aspect
(self, minNumer=None, minDenom=None, maxNumer=None, maxDenom=None)
return self._getints( self.tk.call('wm', 'aspect', self._w, minNumer, minDenom, maxNumer, maxDenom))
Instruct the window manager to set the aspect ratio (width/height) of this widget to be between MINNUMER/MINDENOM and MAXNUMER/MAXDENOM. Return a tuple of the actual values if no argument is given.
Instruct the window manager to set the aspect ratio (width/height) of this widget to be between MINNUMER/MINDENOM and MAXNUMER/MAXDENOM. Return a tuple of the actual values if no argument is given.
[ "Instruct", "the", "window", "manager", "to", "set", "the", "aspect", "ratio", "(", "width", "/", "height", ")", "of", "this", "widget", "to", "be", "between", "MINNUMER", "/", "MINDENOM", "and", "MAXNUMER", "/", "MAXDENOM", ".", "Return", "a", "tuple", "of", "the", "actual", "values", "if", "no", "argument", "is", "given", "." ]
def wm_aspect(self, minNumer=None, minDenom=None, maxNumer=None, maxDenom=None): """Instruct the window manager to set the aspect ratio (width/height) of this widget to be between MINNUMER/MINDENOM and MAXNUMER/MAXDENOM. Return a tuple of the actual values if no argument is given.""" return self._getints( self.tk.call('wm', 'aspect', self._w, minNumer, minDenom, maxNumer, maxDenom))
[ "def", "wm_aspect", "(", "self", ",", "minNumer", "=", "None", ",", "minDenom", "=", "None", ",", "maxNumer", "=", "None", ",", "maxDenom", "=", "None", ")", ":", "return", "self", ".", "_getints", "(", "self", ".", "tk", ".", "call", "(", "'wm'", ",", "'aspect'", ",", "self", ".", "_w", ",", "minNumer", ",", "minDenom", ",", "maxNumer", ",", "maxDenom", ")", ")" ]
https://github.com/wlanjie/AndroidFFmpeg/blob/7baf9122f4b8e1c74e7baf4be5c422c7a5ba5aaf/tools/fdk-aac-build/armeabi-v7a/toolchain/lib/python2.7/lib-tk/Tkinter.py#L1522-L1531
epiqc/ScaffCC
66a79944ee4cd116b27bc1a69137276885461db8
clang/bindings/python/clang/cindex.py
python
Type.get_named_type
(self)
return conf.lib.clang_Type_getNamedType(self)
Retrieve the type named by the qualified-id.
Retrieve the type named by the qualified-id.
[ "Retrieve", "the", "type", "named", "by", "the", "qualified", "-", "id", "." ]
def get_named_type(self): """ Retrieve the type named by the qualified-id. """ return conf.lib.clang_Type_getNamedType(self)
[ "def", "get_named_type", "(", "self", ")", ":", "return", "conf", ".", "lib", ".", "clang_Type_getNamedType", "(", "self", ")" ]
https://github.com/epiqc/ScaffCC/blob/66a79944ee4cd116b27bc1a69137276885461db8/clang/bindings/python/clang/cindex.py#L2371-L2375
avast/retdec
b9879088a5f0278508185ec645494e6c5c57a455
scripts/retdec-unpacker.py
python
Unpacker._unpack
(self, output)
Try to unpack the given file.
Try to unpack the given file.
[ "Try", "to", "unpack", "the", "given", "file", "." ]
def _unpack(self, output): """Try to unpack the given file. """ unpacker_params = [self.input, '-o', output] if self.args.max_memory: unpacker_params.extend(['--max-memory', self.args.max_memory]) elif self.args.max_memory_half_ram: unpacker_params.append('--max-memory-half-ram') self._print('\n##### Trying to unpack ' + self.input + ' into ' + output + ' by using generic unpacker...') out, unpacker_rc, _ = CmdRunner.run_cmd([UNPACKER] + unpacker_params, buffer_output=True, print_run_msg=True) self._print(out) if unpacker_rc == self.UNPACKER_EXIT_CODE_OK: self._print('##### Unpacking by using generic unpacker: successfully unpacked') return self.unpacker_output, self.RET_UNPACK_OK elif unpacker_rc == self.UNPACKER_EXIT_CODE_NOTHING_TO_DO: self._print('##### Unpacking by using generic unpacker: nothing to do') else: # Do not return -> try the next unpacker self._print('##### Unpacking by using generic unpacker: failed') if utils.tool_exists('upx'): # Do not return -> try the next unpacker # Try to unpack via UPX self._print('\n##### Trying to unpack ' + self.input + ' into ' + output + ' by using UPX...') out, upx_rc, _ = CmdRunner.run_cmd(['upx', '-d', self.input, '-o', output], buffer_output=True, discard_stdout=True, print_run_msg=True) self._print(out) if upx_rc == 0: self._print('##### Unpacking by using UPX: successfully unpacked') if self.args.extended_exit_codes: if unpacker_rc == self.UNPACKER_EXIT_CODE_NOTHING_TO_DO: return self.unpacker_output, self.RET_UNPACKER_NOTHING_TO_DO_OTHERS_OK elif unpacker_rc >= self.UNPACKER_EXIT_CODE_UNPACKING_FAILED: return self.unpacker_output, self.RET_UNPACKER_FAILED_OTHERS_OK else: return self.unpacker_output, self.RET_UNPACK_OK else: # We cannot distinguish whether upx failed or the input file was # not upx-packed self._print('##### Unpacking by using UPX: nothing to do') else: self._print('##### \'upx\' not available: nothing to do') if unpacker_rc >= self.UNPACKER_EXIT_CODE_UNPACKING_FAILED: return self.unpacker_output, self.RET_UNPACKER_FAILED else: return self.unpacker_output, self.RET_NOTHING_TO_DO
[ "def", "_unpack", "(", "self", ",", "output", ")", ":", "unpacker_params", "=", "[", "self", ".", "input", ",", "'-o'", ",", "output", "]", "if", "self", ".", "args", ".", "max_memory", ":", "unpacker_params", ".", "extend", "(", "[", "'--max-memory'", ",", "self", ".", "args", ".", "max_memory", "]", ")", "elif", "self", ".", "args", ".", "max_memory_half_ram", ":", "unpacker_params", ".", "append", "(", "'--max-memory-half-ram'", ")", "self", ".", "_print", "(", "'\\n##### Trying to unpack '", "+", "self", ".", "input", "+", "' into '", "+", "output", "+", "' by using generic unpacker...'", ")", "out", ",", "unpacker_rc", ",", "_", "=", "CmdRunner", ".", "run_cmd", "(", "[", "UNPACKER", "]", "+", "unpacker_params", ",", "buffer_output", "=", "True", ",", "print_run_msg", "=", "True", ")", "self", ".", "_print", "(", "out", ")", "if", "unpacker_rc", "==", "self", ".", "UNPACKER_EXIT_CODE_OK", ":", "self", ".", "_print", "(", "'##### Unpacking by using generic unpacker: successfully unpacked'", ")", "return", "self", ".", "unpacker_output", ",", "self", ".", "RET_UNPACK_OK", "elif", "unpacker_rc", "==", "self", ".", "UNPACKER_EXIT_CODE_NOTHING_TO_DO", ":", "self", ".", "_print", "(", "'##### Unpacking by using generic unpacker: nothing to do'", ")", "else", ":", "# Do not return -> try the next unpacker", "self", ".", "_print", "(", "'##### Unpacking by using generic unpacker: failed'", ")", "if", "utils", ".", "tool_exists", "(", "'upx'", ")", ":", "# Do not return -> try the next unpacker", "# Try to unpack via UPX", "self", ".", "_print", "(", "'\\n##### Trying to unpack '", "+", "self", ".", "input", "+", "' into '", "+", "output", "+", "' by using UPX...'", ")", "out", ",", "upx_rc", ",", "_", "=", "CmdRunner", ".", "run_cmd", "(", "[", "'upx'", ",", "'-d'", ",", "self", ".", "input", ",", "'-o'", ",", "output", "]", ",", "buffer_output", "=", "True", ",", "discard_stdout", "=", "True", ",", "print_run_msg", "=", "True", ")", "self", ".", "_print", "(", "out", ")", "if", "upx_rc", "==", "0", ":", "self", ".", "_print", "(", "'##### Unpacking by using UPX: successfully unpacked'", ")", "if", "self", ".", "args", ".", "extended_exit_codes", ":", "if", "unpacker_rc", "==", "self", ".", "UNPACKER_EXIT_CODE_NOTHING_TO_DO", ":", "return", "self", ".", "unpacker_output", ",", "self", ".", "RET_UNPACKER_NOTHING_TO_DO_OTHERS_OK", "elif", "unpacker_rc", ">=", "self", ".", "UNPACKER_EXIT_CODE_UNPACKING_FAILED", ":", "return", "self", ".", "unpacker_output", ",", "self", ".", "RET_UNPACKER_FAILED_OTHERS_OK", "else", ":", "return", "self", ".", "unpacker_output", ",", "self", ".", "RET_UNPACK_OK", "else", ":", "# We cannot distinguish whether upx failed or the input file was", "# not upx-packed", "self", ".", "_print", "(", "'##### Unpacking by using UPX: nothing to do'", ")", "else", ":", "self", ".", "_print", "(", "'##### \\'upx\\' not available: nothing to do'", ")", "if", "unpacker_rc", ">=", "self", ".", "UNPACKER_EXIT_CODE_UNPACKING_FAILED", ":", "return", "self", ".", "unpacker_output", ",", "self", ".", "RET_UNPACKER_FAILED", "else", ":", "return", "self", ".", "unpacker_output", ",", "self", ".", "RET_NOTHING_TO_DO" ]
https://github.com/avast/retdec/blob/b9879088a5f0278508185ec645494e6c5c57a455/scripts/retdec-unpacker.py#L129-L179
apache/incubator-mxnet
f03fb23f1d103fec9541b5ae59ee06b1734a51d9
python/mxnet/ndarray/ndarray.py
python
NDArray.mish
(self, *args, **kwargs)
return op.mish(self, *args, **kwargs)
Convenience fluent method for :py:func:`mish`. The arguments are the same as for :py:func:`mish`, with this array as data.
Convenience fluent method for :py:func:`mish`.
[ "Convenience", "fluent", "method", "for", ":", "py", ":", "func", ":", "mish", "." ]
def mish(self, *args, **kwargs): """Convenience fluent method for :py:func:`mish`. The arguments are the same as for :py:func:`mish`, with this array as data. """ return op.mish(self, *args, **kwargs)
[ "def", "mish", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "op", ".", "mish", "(", "self", ",", "*", "args", ",", "*", "*", "kwargs", ")" ]
https://github.com/apache/incubator-mxnet/blob/f03fb23f1d103fec9541b5ae59ee06b1734a51d9/python/mxnet/ndarray/ndarray.py#L2270-L2276
google/perfetto
fe68c7a7f7657aa71ced68efb126dcac4107c745
python/perfetto/batch_trace_processor/api.py
python
BatchTraceProcessor.query_and_flatten
(self, sql: str)
return self.execute_and_flatten(lambda tp: tp.query(sql). as_pandas_dataframe())
Executes the provided SQL statement and flattens the result. The execution happens in parallel across all the traces and the resulting Pandas dataframes are flattened into a single dataframe. Args: sql: The SQL statement to execute. Returns: A concatenated Pandas dataframe containing the result of executing the query across all the traces. If an URI or a trace resolver was passed to the constructor, the contents of the |metadata| dictionary emitted by the resolver will also be emitted as extra columns (key being column name, value being the value in the dataframe). For example: class CustomResolver(TraceResolver): def resolve(self): return [TraceResolver.Result(trace='/tmp/path', metadata={ 'path': '/tmp/path' 'foo': 'bar' })] with BatchTraceProcessor(CustomResolver()) as btp: df = btp.query_and_flatten('select count(1) as cnt from slice') Then df will look like this: cnt path foo 100 /tmp/path bar Raises: TraceProcessorException: An error occurred running the query.
Executes the provided SQL statement and flattens the result.
[ "Executes", "the", "provided", "SQL", "statement", "and", "flattens", "the", "result", "." ]
def query_and_flatten(self, sql: str): """Executes the provided SQL statement and flattens the result. The execution happens in parallel across all the traces and the resulting Pandas dataframes are flattened into a single dataframe. Args: sql: The SQL statement to execute. Returns: A concatenated Pandas dataframe containing the result of executing the query across all the traces. If an URI or a trace resolver was passed to the constructor, the contents of the |metadata| dictionary emitted by the resolver will also be emitted as extra columns (key being column name, value being the value in the dataframe). For example: class CustomResolver(TraceResolver): def resolve(self): return [TraceResolver.Result(trace='/tmp/path', metadata={ 'path': '/tmp/path' 'foo': 'bar' })] with BatchTraceProcessor(CustomResolver()) as btp: df = btp.query_and_flatten('select count(1) as cnt from slice') Then df will look like this: cnt path foo 100 /tmp/path bar Raises: TraceProcessorException: An error occurred running the query. """ return self.execute_and_flatten(lambda tp: tp.query(sql). as_pandas_dataframe())
[ "def", "query_and_flatten", "(", "self", ",", "sql", ":", "str", ")", ":", "return", "self", ".", "execute_and_flatten", "(", "lambda", "tp", ":", "tp", ".", "query", "(", "sql", ")", ".", "as_pandas_dataframe", "(", ")", ")" ]
https://github.com/google/perfetto/blob/fe68c7a7f7657aa71ced68efb126dcac4107c745/python/perfetto/batch_trace_processor/api.py#L152-L190
livecode/livecode
4606a10ea10b16d5071d0f9f263ccdd7ede8b31d
gyp/pylib/gyp/xcodeproj_file.py
python
XCObject.Name
(self)
Return the name corresponding to an object. Not all objects necessarily need to be nameable, and not all that do have a "name" property. Override as needed.
Return the name corresponding to an object.
[ "Return", "the", "name", "corresponding", "to", "an", "object", "." ]
def Name(self): """Return the name corresponding to an object. Not all objects necessarily need to be nameable, and not all that do have a "name" property. Override as needed. """ # If the schema indicates that "name" is required, try to access the # property even if it doesn't exist. This will result in a KeyError # being raised for the property that should be present, which seems more # appropriate than NotImplementedError in this case. if 'name' in self._properties or \ ('name' in self._schema and self._schema['name'][3]): return self._properties['name'] raise NotImplementedError(self.__class__.__name__ + ' must implement Name')
[ "def", "Name", "(", "self", ")", ":", "# If the schema indicates that \"name\" is required, try to access the", "# property even if it doesn't exist. This will result in a KeyError", "# being raised for the property that should be present, which seems more", "# appropriate than NotImplementedError in this case.", "if", "'name'", "in", "self", ".", "_properties", "or", "(", "'name'", "in", "self", ".", "_schema", "and", "self", ".", "_schema", "[", "'name'", "]", "[", "3", "]", ")", ":", "return", "self", ".", "_properties", "[", "'name'", "]", "raise", "NotImplementedError", "(", "self", ".", "__class__", ".", "__name__", "+", "' must implement Name'", ")" ]
https://github.com/livecode/livecode/blob/4606a10ea10b16d5071d0f9f263ccdd7ede8b31d/gyp/pylib/gyp/xcodeproj_file.py#L354-L369
wxWidgets/wxPython-Classic
19571e1ae65f1ac445f5491474121998c97a1bf0
src/msw/_gdi.py
python
NativeFontInfo.GetWeight
(*args, **kwargs)
return _gdi_.NativeFontInfo_GetWeight(*args, **kwargs)
GetWeight(self) -> int
GetWeight(self) -> int
[ "GetWeight", "(", "self", ")", "-", ">", "int" ]
def GetWeight(*args, **kwargs): """GetWeight(self) -> int""" return _gdi_.NativeFontInfo_GetWeight(*args, **kwargs)
[ "def", "GetWeight", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "_gdi_", ".", "NativeFontInfo_GetWeight", "(", "*", "args", ",", "*", "*", "kwargs", ")" ]
https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/msw/_gdi.py#L1978-L1980
eventql/eventql
7ca0dbb2e683b525620ea30dc40540a22d5eb227
deps/3rdparty/spidermonkey/mozjs/python/requests/requests/packages/urllib3/util/ssl_.py
python
create_urllib3_context
(ssl_version=None, cert_reqs=ssl.CERT_REQUIRED, options=None, ciphers=None)
return context
All arguments have the same meaning as ``ssl_wrap_socket``. By default, this function does a lot of the same work that ``ssl.create_default_context`` does on Python 3.4+. It: - Disables SSLv2, SSLv3, and compression - Sets a restricted set of server ciphers If you wish to enable SSLv3, you can do:: from urllib3.util import ssl_ context = ssl_.create_urllib3_context() context.options &= ~ssl_.OP_NO_SSLv3 You can do the same to enable compression (substituting ``COMPRESSION`` for ``SSLv3`` in the last line above). :param ssl_version: The desired protocol version to use. This will default to PROTOCOL_SSLv23 which will negotiate the highest protocol that both the server and your installation of OpenSSL support. :param cert_reqs: Whether to require the certificate verification. This defaults to ``ssl.CERT_REQUIRED``. :param options: Specific OpenSSL options. These default to ``ssl.OP_NO_SSLv2``, ``ssl.OP_NO_SSLv3``, ``ssl.OP_NO_COMPRESSION``. :param ciphers: Which cipher suites to allow the server to select. :returns: Constructed SSLContext object with specified options :rtype: SSLContext
All arguments have the same meaning as ``ssl_wrap_socket``.
[ "All", "arguments", "have", "the", "same", "meaning", "as", "ssl_wrap_socket", "." ]
def create_urllib3_context(ssl_version=None, cert_reqs=ssl.CERT_REQUIRED, options=None, ciphers=None): """All arguments have the same meaning as ``ssl_wrap_socket``. By default, this function does a lot of the same work that ``ssl.create_default_context`` does on Python 3.4+. It: - Disables SSLv2, SSLv3, and compression - Sets a restricted set of server ciphers If you wish to enable SSLv3, you can do:: from urllib3.util import ssl_ context = ssl_.create_urllib3_context() context.options &= ~ssl_.OP_NO_SSLv3 You can do the same to enable compression (substituting ``COMPRESSION`` for ``SSLv3`` in the last line above). :param ssl_version: The desired protocol version to use. This will default to PROTOCOL_SSLv23 which will negotiate the highest protocol that both the server and your installation of OpenSSL support. :param cert_reqs: Whether to require the certificate verification. This defaults to ``ssl.CERT_REQUIRED``. :param options: Specific OpenSSL options. These default to ``ssl.OP_NO_SSLv2``, ``ssl.OP_NO_SSLv3``, ``ssl.OP_NO_COMPRESSION``. :param ciphers: Which cipher suites to allow the server to select. :returns: Constructed SSLContext object with specified options :rtype: SSLContext """ context = SSLContext(ssl_version or ssl.PROTOCOL_SSLv23) if options is None: options = 0 # SSLv2 is easily broken and is considered harmful and dangerous options |= OP_NO_SSLv2 # SSLv3 has several problems and is now dangerous options |= OP_NO_SSLv3 # Disable compression to prevent CRIME attacks for OpenSSL 1.0+ # (issue #309) options |= OP_NO_COMPRESSION context.options |= options if getattr(context, 'supports_set_ciphers', True): # Platform-specific: Python 2.6 context.set_ciphers(ciphers or _DEFAULT_CIPHERS) context.verify_mode = cert_reqs if getattr(context, 'check_hostname', None) is not None: # Platform-specific: Python 3.2 context.check_hostname = (context.verify_mode == ssl.CERT_REQUIRED) return context
[ "def", "create_urllib3_context", "(", "ssl_version", "=", "None", ",", "cert_reqs", "=", "ssl", ".", "CERT_REQUIRED", ",", "options", "=", "None", ",", "ciphers", "=", "None", ")", ":", "context", "=", "SSLContext", "(", "ssl_version", "or", "ssl", ".", "PROTOCOL_SSLv23", ")", "if", "options", "is", "None", ":", "options", "=", "0", "# SSLv2 is easily broken and is considered harmful and dangerous", "options", "|=", "OP_NO_SSLv2", "# SSLv3 has several problems and is now dangerous", "options", "|=", "OP_NO_SSLv3", "# Disable compression to prevent CRIME attacks for OpenSSL 1.0+", "# (issue #309)", "options", "|=", "OP_NO_COMPRESSION", "context", ".", "options", "|=", "options", "if", "getattr", "(", "context", ",", "'supports_set_ciphers'", ",", "True", ")", ":", "# Platform-specific: Python 2.6", "context", ".", "set_ciphers", "(", "ciphers", "or", "_DEFAULT_CIPHERS", ")", "context", ".", "verify_mode", "=", "cert_reqs", "if", "getattr", "(", "context", ",", "'check_hostname'", ",", "None", ")", "is", "not", "None", ":", "# Platform-specific: Python 3.2", "context", ".", "check_hostname", "=", "(", "context", ".", "verify_mode", "==", "ssl", ".", "CERT_REQUIRED", ")", "return", "context" ]
https://github.com/eventql/eventql/blob/7ca0dbb2e683b525620ea30dc40540a22d5eb227/deps/3rdparty/spidermonkey/mozjs/python/requests/requests/packages/urllib3/util/ssl_.py#L160-L215
hanpfei/chromium-net
392cc1fa3a8f92f42e4071ab6e674d8e0482f83f
third_party/catapult/third_party/apiclient/googleapiclient/discovery.py
python
ResourceMethodParameters.__init__
(self, method_desc)
Constructor for ResourceMethodParameters. Sets default values and defers to set_parameters to populate. Args: method_desc: Dictionary with metadata describing an API method. Value comes from the dictionary of methods stored in the 'methods' key in the deserialized discovery document.
Constructor for ResourceMethodParameters.
[ "Constructor", "for", "ResourceMethodParameters", "." ]
def __init__(self, method_desc): """Constructor for ResourceMethodParameters. Sets default values and defers to set_parameters to populate. Args: method_desc: Dictionary with metadata describing an API method. Value comes from the dictionary of methods stored in the 'methods' key in the deserialized discovery document. """ self.argmap = {} self.required_params = [] self.repeated_params = [] self.pattern_params = {} self.query_params = [] # TODO(dhermes): Change path_params to a list if the extra URITEMPLATE # parsing is gotten rid of. self.path_params = set() self.param_types = {} self.enum_params = {} self.set_parameters(method_desc)
[ "def", "__init__", "(", "self", ",", "method_desc", ")", ":", "self", ".", "argmap", "=", "{", "}", "self", ".", "required_params", "=", "[", "]", "self", ".", "repeated_params", "=", "[", "]", "self", ".", "pattern_params", "=", "{", "}", "self", ".", "query_params", "=", "[", "]", "# TODO(dhermes): Change path_params to a list if the extra URITEMPLATE", "# parsing is gotten rid of.", "self", ".", "path_params", "=", "set", "(", ")", "self", ".", "param_types", "=", "{", "}", "self", ".", "enum_params", "=", "{", "}", "self", ".", "set_parameters", "(", "method_desc", ")" ]
https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/third_party/catapult/third_party/apiclient/googleapiclient/discovery.py#L540-L561
hanpfei/chromium-net
392cc1fa3a8f92f42e4071ab6e674d8e0482f83f
third_party/catapult/catapult_build/module_finder.py
python
FindModule
(name)
return imp.find_module(name)[1]
Gets the path of the named module. This is useful for cases where we want to use subprocess.call on a module we have imported, and safer than using __file__ since that can point to .pyc files. Args: name: the string name of a module (e.g. 'dev_appserver') Returns: The path to the module.
Gets the path of the named module.
[ "Gets", "the", "path", "of", "the", "named", "module", "." ]
def FindModule(name): """Gets the path of the named module. This is useful for cases where we want to use subprocess.call on a module we have imported, and safer than using __file__ since that can point to .pyc files. Args: name: the string name of a module (e.g. 'dev_appserver') Returns: The path to the module. """ return imp.find_module(name)[1]
[ "def", "FindModule", "(", "name", ")", ":", "return", "imp", ".", "find_module", "(", "name", ")", "[", "1", "]" ]
https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/third_party/catapult/catapult_build/module_finder.py#L7-L19
hfinkel/llvm-project-cxxjit
91084ef018240bbb8e24235ff5cd8c355a9c1a1e
lldb/examples/python/performance.py
python
MemoryMeasurement.__str__
(self)
return s
Dump the MemoryMeasurement current value
Dump the MemoryMeasurement current value
[ "Dump", "the", "MemoryMeasurement", "current", "value" ]
def __str__(self): '''Dump the MemoryMeasurement current value''' s = '' for key in self.value.keys(): if s: s += "\n" s += "%8s = %s" % (key, self.value[key]) return s
[ "def", "__str__", "(", "self", ")", ":", "s", "=", "''", "for", "key", "in", "self", ".", "value", ".", "keys", "(", ")", ":", "if", "s", ":", "s", "+=", "\"\\n\"", "s", "+=", "\"%8s = %s\"", "%", "(", "key", ",", "self", ".", "value", "[", "key", "]", ")", "return", "s" ]
https://github.com/hfinkel/llvm-project-cxxjit/blob/91084ef018240bbb8e24235ff5cd8c355a9c1a1e/lldb/examples/python/performance.py#L322-L329
hanpfei/chromium-net
392cc1fa3a8f92f42e4071ab6e674d8e0482f83f
third_party/catapult/third_party/coverage/coverage/python.py
python
PythonFileReporter.parser
(self)
return self._parser
Lazily create a :class:`PythonParser`.
Lazily create a :class:`PythonParser`.
[ "Lazily", "create", "a", ":", "class", ":", "PythonParser", "." ]
def parser(self): """Lazily create a :class:`PythonParser`.""" if self._parser is None: self._parser = PythonParser( filename=self.filename, exclude=self.coverage._exclude_regex('exclude'), ) return self._parser
[ "def", "parser", "(", "self", ")", ":", "if", "self", ".", "_parser", "is", "None", ":", "self", ".", "_parser", "=", "PythonParser", "(", "filename", "=", "self", ".", "filename", ",", "exclude", "=", "self", ".", "coverage", ".", "_exclude_regex", "(", "'exclude'", ")", ",", ")", "return", "self", ".", "_parser" ]
https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/third_party/catapult/third_party/coverage/coverage/python.py#L126-L133
generalized-intelligence/GAAS
29ab17d3e8a4ba18edef3a57c36d8db6329fac73
algorithms/src/LocalizationAndMapping/icp_lidar_localization/fast_gicp/thirdparty/pybind11/setup.py
python
TemporaryDirectory
()
Prepare a temporary directory, cleanup when done
Prepare a temporary directory, cleanup when done
[ "Prepare", "a", "temporary", "directory", "cleanup", "when", "done" ]
def TemporaryDirectory(): # noqa: N802 "Prepare a temporary directory, cleanup when done" try: tmpdir = tempfile.mkdtemp() yield tmpdir finally: shutil.rmtree(tmpdir)
[ "def", "TemporaryDirectory", "(", ")", ":", "# noqa: N802", "try", ":", "tmpdir", "=", "tempfile", ".", "mkdtemp", "(", ")", "yield", "tmpdir", "finally", ":", "shutil", ".", "rmtree", "(", "tmpdir", ")" ]
https://github.com/generalized-intelligence/GAAS/blob/29ab17d3e8a4ba18edef3a57c36d8db6329fac73/algorithms/src/LocalizationAndMapping/icp_lidar_localization/fast_gicp/thirdparty/pybind11/setup.py#L82-L88
hanpfei/chromium-net
392cc1fa3a8f92f42e4071ab6e674d8e0482f83f
third_party/catapult/third_party/gsutil/third_party/protorpc/protorpc/protojson.py
python
ProtoJson.decode_field
(self, field, value)
return value
Decode a JSON value to a python value. Args: field: A ProtoRPC field instance. value: A serialized JSON value. Return: A Python value compatible with field.
Decode a JSON value to a python value.
[ "Decode", "a", "JSON", "value", "to", "a", "python", "value", "." ]
def decode_field(self, field, value): """Decode a JSON value to a python value. Args: field: A ProtoRPC field instance. value: A serialized JSON value. Return: A Python value compatible with field. """ if isinstance(field, messages.EnumField): try: return field.type(value) except TypeError: raise messages.DecodeError('Invalid enum value "%s"' % (value or '')) elif isinstance(field, messages.BytesField): try: return base64.b64decode(value) except (binascii.Error, TypeError) as err: raise messages.DecodeError('Base64 decoding error: %s' % err) elif isinstance(field, message_types.DateTimeField): try: return util.decode_datetime(value) except ValueError as err: raise messages.DecodeError(err) elif (isinstance(field, messages.MessageField) and issubclass(field.type, messages.Message)): return self.__decode_dictionary(field.type, value) elif (isinstance(field, messages.FloatField) and isinstance(value, (six.integer_types, six.string_types))): try: return float(value) except: pass elif (isinstance(field, messages.IntegerField) and isinstance(value, six.string_types)): try: return int(value) except: pass return value
[ "def", "decode_field", "(", "self", ",", "field", ",", "value", ")", ":", "if", "isinstance", "(", "field", ",", "messages", ".", "EnumField", ")", ":", "try", ":", "return", "field", ".", "type", "(", "value", ")", "except", "TypeError", ":", "raise", "messages", ".", "DecodeError", "(", "'Invalid enum value \"%s\"'", "%", "(", "value", "or", "''", ")", ")", "elif", "isinstance", "(", "field", ",", "messages", ".", "BytesField", ")", ":", "try", ":", "return", "base64", ".", "b64decode", "(", "value", ")", "except", "(", "binascii", ".", "Error", ",", "TypeError", ")", "as", "err", ":", "raise", "messages", ".", "DecodeError", "(", "'Base64 decoding error: %s'", "%", "err", ")", "elif", "isinstance", "(", "field", ",", "message_types", ".", "DateTimeField", ")", ":", "try", ":", "return", "util", ".", "decode_datetime", "(", "value", ")", "except", "ValueError", "as", "err", ":", "raise", "messages", ".", "DecodeError", "(", "err", ")", "elif", "(", "isinstance", "(", "field", ",", "messages", ".", "MessageField", ")", "and", "issubclass", "(", "field", ".", "type", ",", "messages", ".", "Message", ")", ")", ":", "return", "self", ".", "__decode_dictionary", "(", "field", ".", "type", ",", "value", ")", "elif", "(", "isinstance", "(", "field", ",", "messages", ".", "FloatField", ")", "and", "isinstance", "(", "value", ",", "(", "six", ".", "integer_types", ",", "six", ".", "string_types", ")", ")", ")", ":", "try", ":", "return", "float", "(", "value", ")", "except", ":", "pass", "elif", "(", "isinstance", "(", "field", ",", "messages", ".", "IntegerField", ")", "and", "isinstance", "(", "value", ",", "six", ".", "string_types", ")", ")", ":", "try", ":", "return", "int", "(", "value", ")", "except", ":", "pass", "return", "value" ]
https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/third_party/catapult/third_party/gsutil/third_party/protorpc/protorpc/protojson.py#L292-L338
aws/lumberyard
f85344403c1c2e77ec8c75deb2c116e97b713217
dev/Tools/Python/3.7.10/linux_x64/lib/python3.7/tkinter/__init__.py
python
Misc.destroy
(self)
Internal function. Delete all Tcl commands created for this widget in the Tcl interpreter.
Internal function.
[ "Internal", "function", "." ]
def destroy(self): """Internal function. Delete all Tcl commands created for this widget in the Tcl interpreter.""" if self._tclCommands is not None: for name in self._tclCommands: #print '- Tkinter: deleted command', name self.tk.deletecommand(name) self._tclCommands = None
[ "def", "destroy", "(", "self", ")", ":", "if", "self", ".", "_tclCommands", "is", "not", "None", ":", "for", "name", "in", "self", ".", "_tclCommands", ":", "#print '- Tkinter: deleted command', name", "self", ".", "tk", ".", "deletecommand", "(", "name", ")", "self", ".", "_tclCommands", "=", "None" ]
https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/linux_x64/lib/python3.7/tkinter/__init__.py#L584-L593
OpenLightingProject/ola
d1433a1bed73276fbe55ce18c03b1c208237decc
python/ola/PidStore.py
python
PidStore._PidProtoToObject
(self, pid_pb)
return Pid(pid_pb.name, pid_pb.value, discovery_request, discovery_response, get_request, get_response, set_request, set_response, discovery_validators, get_validators, set_validators)
Convert the protobuf representation of a PID to a PID object. Args: pid_pb: The protobuf version of the pid Returns: A PIDStore.PID object.
Convert the protobuf representation of a PID to a PID object.
[ "Convert", "the", "protobuf", "representation", "of", "a", "PID", "to", "a", "PID", "object", "." ]
def _PidProtoToObject(self, pid_pb): """Convert the protobuf representation of a PID to a PID object. Args: pid_pb: The protobuf version of the pid Returns: A PIDStore.PID object. """ def BuildList(field_name): if not pid_pb.HasField(field_name): return None try: group = self._FrameFormatToGroup(getattr(pid_pb, field_name)) except PidStructureException as e: raise PidStructureException( "The structure for the %s in %s isn't valid: %s" % (field_name, pid_pb.name, e)) return group discovery_request = BuildList('discovery_request') discovery_response = BuildList('discovery_response') get_request = BuildList('get_request') get_response = BuildList('get_response') set_request = BuildList('set_request') set_response = BuildList('set_response') discovery_validators = [] if pid_pb.HasField('discovery_sub_device_range'): discovery_validators.append(self._SubDeviceRangeToValidator( pid_pb.discovery_sub_device_range)) get_validators = [] if pid_pb.HasField('get_sub_device_range'): get_validators.append(self._SubDeviceRangeToValidator( pid_pb.get_sub_device_range)) set_validators = [] if pid_pb.HasField('set_sub_device_range'): set_validators.append(self._SubDeviceRangeToValidator( pid_pb.set_sub_device_range)) return Pid(pid_pb.name, pid_pb.value, discovery_request, discovery_response, get_request, get_response, set_request, set_response, discovery_validators, get_validators, set_validators)
[ "def", "_PidProtoToObject", "(", "self", ",", "pid_pb", ")", ":", "def", "BuildList", "(", "field_name", ")", ":", "if", "not", "pid_pb", ".", "HasField", "(", "field_name", ")", ":", "return", "None", "try", ":", "group", "=", "self", ".", "_FrameFormatToGroup", "(", "getattr", "(", "pid_pb", ",", "field_name", ")", ")", "except", "PidStructureException", "as", "e", ":", "raise", "PidStructureException", "(", "\"The structure for the %s in %s isn't valid: %s\"", "%", "(", "field_name", ",", "pid_pb", ".", "name", ",", "e", ")", ")", "return", "group", "discovery_request", "=", "BuildList", "(", "'discovery_request'", ")", "discovery_response", "=", "BuildList", "(", "'discovery_response'", ")", "get_request", "=", "BuildList", "(", "'get_request'", ")", "get_response", "=", "BuildList", "(", "'get_response'", ")", "set_request", "=", "BuildList", "(", "'set_request'", ")", "set_response", "=", "BuildList", "(", "'set_response'", ")", "discovery_validators", "=", "[", "]", "if", "pid_pb", ".", "HasField", "(", "'discovery_sub_device_range'", ")", ":", "discovery_validators", ".", "append", "(", "self", ".", "_SubDeviceRangeToValidator", "(", "pid_pb", ".", "discovery_sub_device_range", ")", ")", "get_validators", "=", "[", "]", "if", "pid_pb", ".", "HasField", "(", "'get_sub_device_range'", ")", ":", "get_validators", ".", "append", "(", "self", ".", "_SubDeviceRangeToValidator", "(", "pid_pb", ".", "get_sub_device_range", ")", ")", "set_validators", "=", "[", "]", "if", "pid_pb", ".", "HasField", "(", "'set_sub_device_range'", ")", ":", "set_validators", ".", "append", "(", "self", ".", "_SubDeviceRangeToValidator", "(", "pid_pb", ".", "set_sub_device_range", ")", ")", "return", "Pid", "(", "pid_pb", ".", "name", ",", "pid_pb", ".", "value", ",", "discovery_request", ",", "discovery_response", ",", "get_request", ",", "get_response", ",", "set_request", ",", "set_response", ",", "discovery_validators", ",", "get_validators", ",", "set_validators", ")" ]
https://github.com/OpenLightingProject/ola/blob/d1433a1bed73276fbe55ce18c03b1c208237decc/python/ola/PidStore.py#L1170-L1221
trailofbits/llvm-sanitizer-tutorial
d29dfeec7f51fbf234fd0080f28f2b30cd0b6e99
llvm/examples/Kaleidoscope/MCJIT/cached/genk-timing.py
python
TimingScriptGenerator.writeTimingCall
(self, filename, numFuncs, funcsCalled, totalCalls)
Echo some comments and invoke both versions of toy
Echo some comments and invoke both versions of toy
[ "Echo", "some", "comments", "and", "invoke", "both", "versions", "of", "toy" ]
def writeTimingCall(self, filename, numFuncs, funcsCalled, totalCalls): """Echo some comments and invoke both versions of toy""" rootname = filename if '.' in filename: rootname = filename[:filename.rfind('.')] self.shfile.write("echo \"%s: Calls %d of %d functions, %d total\" >> %s\n" % (filename, funcsCalled, numFuncs, totalCalls, self.timeFile)) self.shfile.write("echo \"\" >> %s\n" % self.timeFile) self.shfile.write("echo \"With MCJIT\" >> %s\n" % self.timeFile) self.shfile.write("/usr/bin/time -f \"Command %C\\n\\tuser time: %U s\\n\\tsytem time: %S s\\n\\tmax set: %M kb\"") self.shfile.write(" -o %s -a " % self.timeFile) self.shfile.write("./toy-mcjit < %s > %s-mcjit.out 2> %s-mcjit.err\n" % (filename, rootname, rootname)) self.shfile.write("echo \"\" >> %s\n" % self.timeFile) self.shfile.write("echo \"With JIT\" >> %s\n" % self.timeFile) self.shfile.write("/usr/bin/time -f \"Command %C\\n\\tuser time: %U s\\n\\tsytem time: %S s\\n\\tmax set: %M kb\"") self.shfile.write(" -o %s -a " % self.timeFile) self.shfile.write("./toy-jit < %s > %s-jit.out 2> %s-jit.err\n" % (filename, rootname, rootname)) self.shfile.write("echo \"\" >> %s\n" % self.timeFile) self.shfile.write("echo \"\" >> %s\n" % self.timeFile)
[ "def", "writeTimingCall", "(", "self", ",", "filename", ",", "numFuncs", ",", "funcsCalled", ",", "totalCalls", ")", ":", "rootname", "=", "filename", "if", "'.'", "in", "filename", ":", "rootname", "=", "filename", "[", ":", "filename", ".", "rfind", "(", "'.'", ")", "]", "self", ".", "shfile", ".", "write", "(", "\"echo \\\"%s: Calls %d of %d functions, %d total\\\" >> %s\\n\"", "%", "(", "filename", ",", "funcsCalled", ",", "numFuncs", ",", "totalCalls", ",", "self", ".", "timeFile", ")", ")", "self", ".", "shfile", ".", "write", "(", "\"echo \\\"\\\" >> %s\\n\"", "%", "self", ".", "timeFile", ")", "self", ".", "shfile", ".", "write", "(", "\"echo \\\"With MCJIT\\\" >> %s\\n\"", "%", "self", ".", "timeFile", ")", "self", ".", "shfile", ".", "write", "(", "\"/usr/bin/time -f \\\"Command %C\\\\n\\\\tuser time: %U s\\\\n\\\\tsytem time: %S s\\\\n\\\\tmax set: %M kb\\\"\"", ")", "self", ".", "shfile", ".", "write", "(", "\" -o %s -a \"", "%", "self", ".", "timeFile", ")", "self", ".", "shfile", ".", "write", "(", "\"./toy-mcjit < %s > %s-mcjit.out 2> %s-mcjit.err\\n\"", "%", "(", "filename", ",", "rootname", ",", "rootname", ")", ")", "self", ".", "shfile", ".", "write", "(", "\"echo \\\"\\\" >> %s\\n\"", "%", "self", ".", "timeFile", ")", "self", ".", "shfile", ".", "write", "(", "\"echo \\\"With JIT\\\" >> %s\\n\"", "%", "self", ".", "timeFile", ")", "self", ".", "shfile", ".", "write", "(", "\"/usr/bin/time -f \\\"Command %C\\\\n\\\\tuser time: %U s\\\\n\\\\tsytem time: %S s\\\\n\\\\tmax set: %M kb\\\"\"", ")", "self", ".", "shfile", ".", "write", "(", "\" -o %s -a \"", "%", "self", ".", "timeFile", ")", "self", ".", "shfile", ".", "write", "(", "\"./toy-jit < %s > %s-jit.out 2> %s-jit.err\\n\"", "%", "(", "filename", ",", "rootname", ",", "rootname", ")", ")", "self", ".", "shfile", ".", "write", "(", "\"echo \\\"\\\" >> %s\\n\"", "%", "self", ".", "timeFile", ")", "self", ".", "shfile", ".", "write", "(", "\"echo \\\"\\\" >> %s\\n\"", "%", "self", ".", "timeFile", ")" ]
https://github.com/trailofbits/llvm-sanitizer-tutorial/blob/d29dfeec7f51fbf234fd0080f28f2b30cd0b6e99/llvm/examples/Kaleidoscope/MCJIT/cached/genk-timing.py#L15-L32
hanpfei/chromium-net
392cc1fa3a8f92f42e4071ab6e674d8e0482f83f
third_party/catapult/third_party/gsutil/third_party/protorpc/protorpc/remote.py
python
_RemoteMethodInfo.method
(self)
return self.__method
Original undecorated method.
Original undecorated method.
[ "Original", "undecorated", "method", "." ]
def method(self): """Original undecorated method.""" return self.__method
[ "def", "method", "(", "self", ")", ":", "return", "self", ".", "__method" ]
https://github.com/hanpfei/chromium-net/blob/392cc1fa3a8f92f42e4071ab6e674d8e0482f83f/third_party/catapult/third_party/gsutil/third_party/protorpc/protorpc/remote.py#L313-L315
ChromiumWebApps/chromium
c7361d39be8abd1574e6ce8957c8dbddd4c6ccf7
tools/telemetry/telemetry/page/actions/page_action.py
python
PageAction.BindMeasurementJavaScript
( self, tab, start_js, stop_js)
Let this action determine when measurements should start and stop. A measurement can call this method to provide the action with JavaScript code that starts and stops measurements. The action determines when to execute the provided JavaScript code, for more accurate timings. Args: tab: The tab to do everything on. start_js: JavaScript code that starts measurements. stop_js: JavaScript code that stops measurements.
Let this action determine when measurements should start and stop.
[ "Let", "this", "action", "determine", "when", "measurements", "should", "start", "and", "stop", "." ]
def BindMeasurementJavaScript( self, tab, start_js, stop_js): # pylint: disable=W0613 """Let this action determine when measurements should start and stop. A measurement can call this method to provide the action with JavaScript code that starts and stops measurements. The action determines when to execute the provided JavaScript code, for more accurate timings. Args: tab: The tab to do everything on. start_js: JavaScript code that starts measurements. stop_js: JavaScript code that stops measurements. """ raise Exception('This action cannot be bound.')
[ "def", "BindMeasurementJavaScript", "(", "self", ",", "tab", ",", "start_js", ",", "stop_js", ")", ":", "# pylint: disable=W0613", "raise", "Exception", "(", "'This action cannot be bound.'", ")" ]
https://github.com/ChromiumWebApps/chromium/blob/c7361d39be8abd1574e6ce8957c8dbddd4c6ccf7/tools/telemetry/telemetry/page/actions/page_action.py#L72-L86
apache/openoffice
97289b2620590d8b431bcc408f87252db6203818
main/pyuno/source/module/uno.py
python
getClass
( typeName )
return pyuno.getClass(typeName)
returns the class of a concrete uno exception, struct or interface
returns the class of a concrete uno exception, struct or interface
[ "returns", "the", "class", "of", "a", "concrete", "uno", "exception", "struct", "or", "interface" ]
def getClass( typeName ): """returns the class of a concrete uno exception, struct or interface """ return pyuno.getClass(typeName)
[ "def", "getClass", "(", "typeName", ")", ":", "return", "pyuno", ".", "getClass", "(", "typeName", ")" ]
https://github.com/apache/openoffice/blob/97289b2620590d8b431bcc408f87252db6203818/main/pyuno/source/module/uno.py#L73-L76
catboost/catboost
167f64f237114a4d10b2b4ee42adb4569137debe
contrib/python/scikit-learn/py2/sklearn/model_selection/_split.py
python
StratifiedShuffleSplit.split
(self, X, y, groups=None)
return super(StratifiedShuffleSplit, self).split(X, y, groups)
Generate indices to split data into training and test set. Parameters ---------- X : array-like, shape (n_samples, n_features) Training data, where n_samples is the number of samples and n_features is the number of features. Note that providing ``y`` is sufficient to generate the splits and hence ``np.zeros(n_samples)`` may be used as a placeholder for ``X`` instead of actual training data. y : array-like, shape (n_samples,) The target variable for supervised learning problems. Stratification is done based on the y labels. groups : object Always ignored, exists for compatibility. Returns ------- train : ndarray The training set indices for that split. test : ndarray The testing set indices for that split.
Generate indices to split data into training and test set.
[ "Generate", "indices", "to", "split", "data", "into", "training", "and", "test", "set", "." ]
def split(self, X, y, groups=None): """Generate indices to split data into training and test set. Parameters ---------- X : array-like, shape (n_samples, n_features) Training data, where n_samples is the number of samples and n_features is the number of features. Note that providing ``y`` is sufficient to generate the splits and hence ``np.zeros(n_samples)`` may be used as a placeholder for ``X`` instead of actual training data. y : array-like, shape (n_samples,) The target variable for supervised learning problems. Stratification is done based on the y labels. groups : object Always ignored, exists for compatibility. Returns ------- train : ndarray The training set indices for that split. test : ndarray The testing set indices for that split. """ y = check_array(y, ensure_2d=False, dtype=None) return super(StratifiedShuffleSplit, self).split(X, y, groups)
[ "def", "split", "(", "self", ",", "X", ",", "y", ",", "groups", "=", "None", ")", ":", "y", "=", "check_array", "(", "y", ",", "ensure_2d", "=", "False", ",", "dtype", "=", "None", ")", "return", "super", "(", "StratifiedShuffleSplit", ",", "self", ")", ".", "split", "(", "X", ",", "y", ",", "groups", ")" ]
https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/scikit-learn/py2/sklearn/model_selection/_split.py#L1296-L1325
catboost/catboost
167f64f237114a4d10b2b4ee42adb4569137debe
contrib/python/ipython/py2/IPython/core/completer.py
python
Completer.__init__
(self, namespace=None, global_namespace=None, **kwargs)
Create a new completer for the command line. Completer(namespace=ns, global_namespace=ns2) -> completer instance. If unspecified, the default namespace where completions are performed is __main__ (technically, __main__.__dict__). Namespaces should be given as dictionaries. An optional second namespace can be given. This allows the completer to handle cases where both the local and global scopes need to be distinguished. Completer instances should be used as the completion mechanism of readline via the set_completer() call: readline.set_completer(Completer(my_namespace).complete)
Create a new completer for the command line.
[ "Create", "a", "new", "completer", "for", "the", "command", "line", "." ]
def __init__(self, namespace=None, global_namespace=None, **kwargs): """Create a new completer for the command line. Completer(namespace=ns, global_namespace=ns2) -> completer instance. If unspecified, the default namespace where completions are performed is __main__ (technically, __main__.__dict__). Namespaces should be given as dictionaries. An optional second namespace can be given. This allows the completer to handle cases where both the local and global scopes need to be distinguished. Completer instances should be used as the completion mechanism of readline via the set_completer() call: readline.set_completer(Completer(my_namespace).complete) """ # Don't bind to namespace quite yet, but flag whether the user wants a # specific namespace or to use __main__.__dict__. This will allow us # to bind to __main__.__dict__ at completion time, not now. if namespace is None: self.use_main_ns = 1 else: self.use_main_ns = 0 self.namespace = namespace # The global namespace, if given, can be bound directly if global_namespace is None: self.global_namespace = {} else: self.global_namespace = global_namespace super(Completer, self).__init__(**kwargs)
[ "def", "__init__", "(", "self", ",", "namespace", "=", "None", ",", "global_namespace", "=", "None", ",", "*", "*", "kwargs", ")", ":", "# Don't bind to namespace quite yet, but flag whether the user wants a", "# specific namespace or to use __main__.__dict__. This will allow us", "# to bind to __main__.__dict__ at completion time, not now.", "if", "namespace", "is", "None", ":", "self", ".", "use_main_ns", "=", "1", "else", ":", "self", ".", "use_main_ns", "=", "0", "self", ".", "namespace", "=", "namespace", "# The global namespace, if given, can be bound directly", "if", "global_namespace", "is", "None", ":", "self", ".", "global_namespace", "=", "{", "}", "else", ":", "self", ".", "global_namespace", "=", "global_namespace", "super", "(", "Completer", ",", "self", ")", ".", "__init__", "(", "*", "*", "kwargs", ")" ]
https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/ipython/py2/IPython/core/completer.py#L251-L285
aws/lumberyard
f85344403c1c2e77ec8c75deb2c116e97b713217
dev/Tools/Python/3.7.10/linux_x64/lib/python3.7/site-packages/pip/_vendor/urllib3/util/url.py
python
split_first
(s, delims)
return s[:min_idx], s[min_idx + 1 :], min_delim
.. deprecated:: 1.25 Given a string and an iterable of delimiters, split on the first found delimiter. Return two split parts and the matched delimiter. If not found, then the first part is the full input string. Example:: >>> split_first('foo/bar?baz', '?/=') ('foo', 'bar?baz', '/') >>> split_first('foo/bar?baz', '123') ('foo/bar?baz', '', None) Scales linearly with number of delims. Not ideal for large number of delims.
[]
def split_first(s, delims): """ .. deprecated:: 1.25 Given a string and an iterable of delimiters, split on the first found delimiter. Return two split parts and the matched delimiter. If not found, then the first part is the full input string. Example:: >>> split_first('foo/bar?baz', '?/=') ('foo', 'bar?baz', '/') >>> split_first('foo/bar?baz', '123') ('foo/bar?baz', '', None) Scales linearly with number of delims. Not ideal for large number of delims. """ min_idx = None min_delim = None for d in delims: idx = s.find(d) if idx < 0: continue if min_idx is None or idx < min_idx: min_idx = idx min_delim = d if min_idx is None or min_idx < 0: return s, "", None return s[:min_idx], s[min_idx + 1 :], min_delim
[ "def", "split_first", "(", "s", ",", "delims", ")", ":", "min_idx", "=", "None", "min_delim", "=", "None", "for", "d", "in", "delims", ":", "idx", "=", "s", ".", "find", "(", "d", ")", "if", "idx", "<", "0", ":", "continue", "if", "min_idx", "is", "None", "or", "idx", "<", "min_idx", ":", "min_idx", "=", "idx", "min_delim", "=", "d", "if", "min_idx", "is", "None", "or", "min_idx", "<", "0", ":", "return", "s", ",", "\"\"", ",", "None", "return", "s", "[", ":", "min_idx", "]", ",", "s", "[", "min_idx", "+", "1", ":", "]", ",", "min_delim" ]
https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Tools/Python/3.7.10/linux_x64/lib/python3.7/site-packages/pip/_vendor/urllib3/util/url.py#L349-L413
arkenthera/electron-vibrancy
383153ef9ccb23a6c7517150d6bb0794dff3115e
scripts/cpplint.py
python
CheckVlogArguments
(filename, clean_lines, linenum, error)
Checks that VLOG() is only used for defining a logging level. For example, VLOG(2) is correct. VLOG(INFO), VLOG(WARNING), VLOG(ERROR), and VLOG(FATAL) are not. Args: filename: The name of the current file. clean_lines: A CleansedLines instance containing the file. linenum: The number of the line to check. error: The function to call with any errors found.
Checks that VLOG() is only used for defining a logging level.
[ "Checks", "that", "VLOG", "()", "is", "only", "used", "for", "defining", "a", "logging", "level", "." ]
def CheckVlogArguments(filename, clean_lines, linenum, error): """Checks that VLOG() is only used for defining a logging level. For example, VLOG(2) is correct. VLOG(INFO), VLOG(WARNING), VLOG(ERROR), and VLOG(FATAL) are not. Args: filename: The name of the current file. clean_lines: A CleansedLines instance containing the file. linenum: The number of the line to check. error: The function to call with any errors found. """ line = clean_lines.elided[linenum] if Search(r'\bVLOG\((INFO|ERROR|WARNING|DFATAL|FATAL)\)', line): error(filename, linenum, 'runtime/vlog', 5, 'VLOG() should be used with numeric verbosity level. ' 'Use LOG() if you want symbolic severity levels.')
[ "def", "CheckVlogArguments", "(", "filename", ",", "clean_lines", ",", "linenum", ",", "error", ")", ":", "line", "=", "clean_lines", ".", "elided", "[", "linenum", "]", "if", "Search", "(", "r'\\bVLOG\\((INFO|ERROR|WARNING|DFATAL|FATAL)\\)'", ",", "line", ")", ":", "error", "(", "filename", ",", "linenum", ",", "'runtime/vlog'", ",", "5", ",", "'VLOG() should be used with numeric verbosity level. '", "'Use LOG() if you want symbolic severity levels.'", ")" ]
https://github.com/arkenthera/electron-vibrancy/blob/383153ef9ccb23a6c7517150d6bb0794dff3115e/scripts/cpplint.py#L1726-L1742
gem5/gem5
141cc37c2d4b93959d4c249b8f7e6a8b2ef75338
ext/ply/example/ansic/cparse.py
python
p_function_definition_3
(t)
function_definition : declarator compound_statement
function_definition : declarator compound_statement
[ "function_definition", ":", "declarator", "compound_statement" ]
def p_function_definition_3(t): 'function_definition : declarator compound_statement' pass
[ "def", "p_function_definition_3", "(", "t", ")", ":", "pass" ]
https://github.com/gem5/gem5/blob/141cc37c2d4b93959d4c249b8f7e6a8b2ef75338/ext/ply/example/ansic/cparse.py#L44-L46
wlanjie/AndroidFFmpeg
7baf9122f4b8e1c74e7baf4be5c422c7a5ba5aaf
tools/fdk-aac-build/armeabi-v7a/toolchain/lib/python2.7/lib-tk/tkMessageBox.py
python
askretrycancel
(title=None, message=None, **options)
return s == RETRY
Ask if operation should be retried; return true if the answer is yes
Ask if operation should be retried; return true if the answer is yes
[ "Ask", "if", "operation", "should", "be", "retried", ";", "return", "true", "if", "the", "answer", "is", "yes" ]
def askretrycancel(title=None, message=None, **options): "Ask if operation should be retried; return true if the answer is yes" s = _show(title, message, WARNING, RETRYCANCEL, **options) return s == RETRY
[ "def", "askretrycancel", "(", "title", "=", "None", ",", "message", "=", "None", ",", "*", "*", "options", ")", ":", "s", "=", "_show", "(", "title", ",", "message", ",", "WARNING", ",", "RETRYCANCEL", ",", "*", "*", "options", ")", "return", "s", "==", "RETRY" ]
https://github.com/wlanjie/AndroidFFmpeg/blob/7baf9122f4b8e1c74e7baf4be5c422c7a5ba5aaf/tools/fdk-aac-build/armeabi-v7a/toolchain/lib/python2.7/lib-tk/tkMessageBox.py#L116-L119
gwaldron/osgearth
4c521857d59a69743e4a9cedba00afe570f984e8
src/third_party/tinygltf/deps/cpplint.py
python
CheckForFunctionLengths
(filename, clean_lines, linenum, function_state, error)
Reports for long function bodies. For an overview why this is done, see: http://google-styleguide.googlecode.com/svn/trunk/cppguide.xml#Write_Short_Functions Uses a simplistic algorithm assuming other style guidelines (especially spacing) are followed. Only checks unindented functions, so class members are unchecked. Trivial bodies are unchecked, so constructors with huge initializer lists may be missed. Blank/comment lines are not counted so as to avoid encouraging the removal of vertical space and comments just to get through a lint check. NOLINT *on the last line of a function* disables this check. Args: filename: The name of the current file. clean_lines: A CleansedLines instance containing the file. linenum: The number of the line to check. function_state: Current function name and lines in body so far. error: The function to call with any errors found.
Reports for long function bodies.
[ "Reports", "for", "long", "function", "bodies", "." ]
def CheckForFunctionLengths(filename, clean_lines, linenum, function_state, error): """Reports for long function bodies. For an overview why this is done, see: http://google-styleguide.googlecode.com/svn/trunk/cppguide.xml#Write_Short_Functions Uses a simplistic algorithm assuming other style guidelines (especially spacing) are followed. Only checks unindented functions, so class members are unchecked. Trivial bodies are unchecked, so constructors with huge initializer lists may be missed. Blank/comment lines are not counted so as to avoid encouraging the removal of vertical space and comments just to get through a lint check. NOLINT *on the last line of a function* disables this check. Args: filename: The name of the current file. clean_lines: A CleansedLines instance containing the file. linenum: The number of the line to check. function_state: Current function name and lines in body so far. error: The function to call with any errors found. """ lines = clean_lines.lines line = lines[linenum] joined_line = '' starting_func = False regexp = r'(\w(\w|::|\*|\&|\s)*)\(' # decls * & space::name( ... match_result = Match(regexp, line) if match_result: # If the name is all caps and underscores, figure it's a macro and # ignore it, unless it's TEST or TEST_F. function_name = match_result.group(1).split()[-1] if function_name == 'TEST' or function_name == 'TEST_F' or ( not Match(r'[A-Z_]+$', function_name)): starting_func = True if starting_func: body_found = False for start_linenum in xrange(linenum, clean_lines.NumLines()): start_line = lines[start_linenum] joined_line += ' ' + start_line.lstrip() if Search(r'(;|})', start_line): # Declarations and trivial functions body_found = True break # ... ignore elif Search(r'{', start_line): body_found = True function = Search(r'((\w|:)*)\(', line).group(1) if Match(r'TEST', function): # Handle TEST... macros parameter_regexp = Search(r'(\(.*\))', joined_line) if parameter_regexp: # Ignore bad syntax function += parameter_regexp.group(1) else: function += '()' function_state.Begin(function) break if not body_found: # No body for the function (or evidence of a non-function) was found. error(filename, linenum, 'readability/fn_size', 5, 'Lint failed to find start of function body.') elif Match(r'^\}\s*$', line): # function end function_state.Check(error, filename, linenum) function_state.End() elif not Match(r'^\s*$', line): function_state.Count()
[ "def", "CheckForFunctionLengths", "(", "filename", ",", "clean_lines", ",", "linenum", ",", "function_state", ",", "error", ")", ":", "lines", "=", "clean_lines", ".", "lines", "line", "=", "lines", "[", "linenum", "]", "joined_line", "=", "''", "starting_func", "=", "False", "regexp", "=", "r'(\\w(\\w|::|\\*|\\&|\\s)*)\\('", "# decls * & space::name( ...", "match_result", "=", "Match", "(", "regexp", ",", "line", ")", "if", "match_result", ":", "# If the name is all caps and underscores, figure it's a macro and", "# ignore it, unless it's TEST or TEST_F.", "function_name", "=", "match_result", ".", "group", "(", "1", ")", ".", "split", "(", ")", "[", "-", "1", "]", "if", "function_name", "==", "'TEST'", "or", "function_name", "==", "'TEST_F'", "or", "(", "not", "Match", "(", "r'[A-Z_]+$'", ",", "function_name", ")", ")", ":", "starting_func", "=", "True", "if", "starting_func", ":", "body_found", "=", "False", "for", "start_linenum", "in", "xrange", "(", "linenum", ",", "clean_lines", ".", "NumLines", "(", ")", ")", ":", "start_line", "=", "lines", "[", "start_linenum", "]", "joined_line", "+=", "' '", "+", "start_line", ".", "lstrip", "(", ")", "if", "Search", "(", "r'(;|})'", ",", "start_line", ")", ":", "# Declarations and trivial functions", "body_found", "=", "True", "break", "# ... ignore", "elif", "Search", "(", "r'{'", ",", "start_line", ")", ":", "body_found", "=", "True", "function", "=", "Search", "(", "r'((\\w|:)*)\\('", ",", "line", ")", ".", "group", "(", "1", ")", "if", "Match", "(", "r'TEST'", ",", "function", ")", ":", "# Handle TEST... macros", "parameter_regexp", "=", "Search", "(", "r'(\\(.*\\))'", ",", "joined_line", ")", "if", "parameter_regexp", ":", "# Ignore bad syntax", "function", "+=", "parameter_regexp", ".", "group", "(", "1", ")", "else", ":", "function", "+=", "'()'", "function_state", ".", "Begin", "(", "function", ")", "break", "if", "not", "body_found", ":", "# No body for the function (or evidence of a non-function) was found.", "error", "(", "filename", ",", "linenum", ",", "'readability/fn_size'", ",", "5", ",", "'Lint failed to find start of function body.'", ")", "elif", "Match", "(", "r'^\\}\\s*$'", ",", "line", ")", ":", "# function end", "function_state", ".", "Check", "(", "error", ",", "filename", ",", "linenum", ")", "function_state", ".", "End", "(", ")", "elif", "not", "Match", "(", "r'^\\s*$'", ",", "line", ")", ":", "function_state", ".", "Count", "(", ")" ]
https://github.com/gwaldron/osgearth/blob/4c521857d59a69743e4a9cedba00afe570f984e8/src/third_party/tinygltf/deps/cpplint.py#L2842-L2907
gromacs/gromacs
7dec3a3f99993cf5687a122de3e12de31c21c399
python_packaging/src/gmxapi/operation.py
python
SourceResource.data
(self)
Get the output data proxy.
Get the output data proxy.
[ "Get", "the", "output", "data", "proxy", "." ]
def data(self) -> _OutputDataProxyType: """Get the output data proxy.""" # Warning: this should probably be renamed, but "output_data_proxy" is already # a member in at least one derived class. ...
[ "def", "data", "(", "self", ")", "->", "_OutputDataProxyType", ":", "# Warning: this should probably be renamed, but \"output_data_proxy\" is already", "# a member in at least one derived class.", "..." ]
https://github.com/gromacs/gromacs/blob/7dec3a3f99993cf5687a122de3e12de31c21c399/python_packaging/src/gmxapi/operation.py#L768-L772
microsoft/CNTK
e9396480025b9ca457d26b6f33dd07c474c6aa04
bindings/python/cntk/contrib/deeprl/agent/agent.py
python
AgentBaseClass._choose_action
(self, state)
Choose an action according to the policy. Args: state (object): observation seen by agent, which can be different from what is provided by the environment. The difference comes from preprcessing. Returns: action (int): action choosen by agent. debug_info (str): auxiliary diagnostic information.
Choose an action according to the policy.
[ "Choose", "an", "action", "according", "to", "the", "policy", "." ]
def _choose_action(self, state): """ Choose an action according to the policy. Args: state (object): observation seen by agent, which can be different from what is provided by the environment. The difference comes from preprcessing. Returns: action (int): action choosen by agent. debug_info (str): auxiliary diagnostic information. """ pass
[ "def", "_choose_action", "(", "self", ",", "state", ")", ":", "pass" ]
https://github.com/microsoft/CNTK/blob/e9396480025b9ca457d26b6f33dd07c474c6aa04/bindings/python/cntk/contrib/deeprl/agent/agent.py#L160-L173
wxWidgets/wxPython-Classic
19571e1ae65f1ac445f5491474121998c97a1bf0
src/osx_carbon/stc.py
python
StyledTextCtrl.StyleResetDefault
(*args, **kwargs)
return _stc.StyledTextCtrl_StyleResetDefault(*args, **kwargs)
StyleResetDefault(self) Reset the default style to its state at startup
StyleResetDefault(self)
[ "StyleResetDefault", "(", "self", ")" ]
def StyleResetDefault(*args, **kwargs): """ StyleResetDefault(self) Reset the default style to its state at startup """ return _stc.StyledTextCtrl_StyleResetDefault(*args, **kwargs)
[ "def", "StyleResetDefault", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "_stc", ".", "StyledTextCtrl_StyleResetDefault", "(", "*", "args", ",", "*", "*", "kwargs", ")" ]
https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/osx_carbon/stc.py#L2570-L2576
qt/qt
0a2f2382541424726168804be2c90b91381608c6
src/3rdparty/webkit/Source/ThirdParty/gyp/pylib/gyp/generator/scons.py
python
EscapeShellArgument
(s)
return "'" + s.replace("'", "'\\''") + "'"
Quotes an argument so that it will be interpreted literally by a POSIX shell. Taken from http://stackoverflow.com/questions/35817/whats-the-best-way-to-escape-ossystem-calls-in-python
Quotes an argument so that it will be interpreted literally by a POSIX shell. Taken from http://stackoverflow.com/questions/35817/whats-the-best-way-to-escape-ossystem-calls-in-python
[ "Quotes", "an", "argument", "so", "that", "it", "will", "be", "interpreted", "literally", "by", "a", "POSIX", "shell", ".", "Taken", "from", "http", ":", "//", "stackoverflow", ".", "com", "/", "questions", "/", "35817", "/", "whats", "-", "the", "-", "best", "-", "way", "-", "to", "-", "escape", "-", "ossystem", "-", "calls", "-", "in", "-", "python" ]
def EscapeShellArgument(s): """Quotes an argument so that it will be interpreted literally by a POSIX shell. Taken from http://stackoverflow.com/questions/35817/whats-the-best-way-to-escape-ossystem-calls-in-python """ return "'" + s.replace("'", "'\\''") + "'"
[ "def", "EscapeShellArgument", "(", "s", ")", ":", "return", "\"'\"", "+", "s", ".", "replace", "(", "\"'\"", ",", "\"'\\\\''\"", ")", "+", "\"'\"" ]
https://github.com/qt/qt/blob/0a2f2382541424726168804be2c90b91381608c6/src/3rdparty/webkit/Source/ThirdParty/gyp/pylib/gyp/generator/scons.py#L170-L175
ApolloAuto/apollo-platform
86d9dc6743b496ead18d597748ebabd34a513289
ros/third_party/lib_x86_64/python2.7/dist-packages/numpy/oldnumeric/ma.py
python
right_shift
(a, n)
Right shift n bits
Right shift n bits
[ "Right", "shift", "n", "bits" ]
def right_shift (a, n): "Right shift n bits" m = getmask(a) if m is nomask: d = umath.right_shift(filled(a), n) return masked_array(d) else: d = umath.right_shift(filled(a, 0), n) return masked_array(d, m)
[ "def", "right_shift", "(", "a", ",", "n", ")", ":", "m", "=", "getmask", "(", "a", ")", "if", "m", "is", "nomask", ":", "d", "=", "umath", ".", "right_shift", "(", "filled", "(", "a", ")", ",", "n", ")", "return", "masked_array", "(", "d", ")", "else", ":", "d", "=", "umath", ".", "right_shift", "(", "filled", "(", "a", ",", "0", ")", ",", "n", ")", "return", "masked_array", "(", "d", ",", "m", ")" ]
https://github.com/ApolloAuto/apollo-platform/blob/86d9dc6743b496ead18d597748ebabd34a513289/ros/third_party/lib_x86_64/python2.7/dist-packages/numpy/oldnumeric/ma.py#L1522-L1530
baidu-research/tensorflow-allreduce
66d5b855e90b0949e9fa5cca5599fd729a70e874
tensorflow/python/summary/writer/writer.py
python
SummaryToEventTransformer.__init__
(self, event_writer, graph=None, graph_def=None)
Creates a `SummaryWriter` and an event file. On construction the summary writer creates a new event file in `logdir`. This event file will contain `Event` protocol buffers constructed when you call one of the following functions: `add_summary()`, `add_session_log()`, `add_event()`, or `add_graph()`. If you pass a `Graph` to the constructor it is added to the event file. (This is equivalent to calling `add_graph()` later). TensorBoard will pick the graph from the file and display it graphically so you can interactively explore the graph you built. You will usually pass the graph from the session in which you launched it: ```python ...create a graph... # Launch the graph in a session. sess = tf.Session() # Create a summary writer, add the 'graph' to the event file. writer = tf.summary.FileWriter(<some-directory>, sess.graph) ``` Args: event_writer: An EventWriter. Implements add_event and get_logdir. graph: A `Graph` object, such as `sess.graph`. graph_def: DEPRECATED: Use the `graph` argument instead.
Creates a `SummaryWriter` and an event file.
[ "Creates", "a", "SummaryWriter", "and", "an", "event", "file", "." ]
def __init__(self, event_writer, graph=None, graph_def=None): """Creates a `SummaryWriter` and an event file. On construction the summary writer creates a new event file in `logdir`. This event file will contain `Event` protocol buffers constructed when you call one of the following functions: `add_summary()`, `add_session_log()`, `add_event()`, or `add_graph()`. If you pass a `Graph` to the constructor it is added to the event file. (This is equivalent to calling `add_graph()` later). TensorBoard will pick the graph from the file and display it graphically so you can interactively explore the graph you built. You will usually pass the graph from the session in which you launched it: ```python ...create a graph... # Launch the graph in a session. sess = tf.Session() # Create a summary writer, add the 'graph' to the event file. writer = tf.summary.FileWriter(<some-directory>, sess.graph) ``` Args: event_writer: An EventWriter. Implements add_event and get_logdir. graph: A `Graph` object, such as `sess.graph`. graph_def: DEPRECATED: Use the `graph` argument instead. """ self.event_writer = event_writer # For storing used tags for session.run() outputs. self._session_run_tags = {} if graph is not None or graph_def is not None: # Calling it with both graph and graph_def for backward compatibility. self.add_graph(graph=graph, graph_def=graph_def) # Also export the meta_graph_def in this case. # graph may itself be a graph_def due to positional arguments maybe_graph_as_def = (graph.as_graph_def(add_shapes=True) if isinstance(graph, ops.Graph) else graph) self.add_meta_graph( meta_graph.create_meta_graph_def(graph_def=graph_def or maybe_graph_as_def)) # This set contains tags of Summary Values that have been encountered # already. The motivation here is that the SummaryWriter only keeps the # metadata property (which is a SummaryMetadata proto) of the first Summary # Value encountered for each tag. The SummaryWriter strips away the # SummaryMetadata for all subsequent Summary Values with tags seen # previously. This saves space. self._seen_summary_tags = set()
[ "def", "__init__", "(", "self", ",", "event_writer", ",", "graph", "=", "None", ",", "graph_def", "=", "None", ")", ":", "self", ".", "event_writer", "=", "event_writer", "# For storing used tags for session.run() outputs.", "self", ".", "_session_run_tags", "=", "{", "}", "if", "graph", "is", "not", "None", "or", "graph_def", "is", "not", "None", ":", "# Calling it with both graph and graph_def for backward compatibility.", "self", ".", "add_graph", "(", "graph", "=", "graph", ",", "graph_def", "=", "graph_def", ")", "# Also export the meta_graph_def in this case.", "# graph may itself be a graph_def due to positional arguments", "maybe_graph_as_def", "=", "(", "graph", ".", "as_graph_def", "(", "add_shapes", "=", "True", ")", "if", "isinstance", "(", "graph", ",", "ops", ".", "Graph", ")", "else", "graph", ")", "self", ".", "add_meta_graph", "(", "meta_graph", ".", "create_meta_graph_def", "(", "graph_def", "=", "graph_def", "or", "maybe_graph_as_def", ")", ")", "# This set contains tags of Summary Values that have been encountered", "# already. The motivation here is that the SummaryWriter only keeps the", "# metadata property (which is a SummaryMetadata proto) of the first Summary", "# Value encountered for each tag. The SummaryWriter strips away the", "# SummaryMetadata for all subsequent Summary Values with tags seen", "# previously. This saves space.", "self", ".", "_seen_summary_tags", "=", "set", "(", ")" ]
https://github.com/baidu-research/tensorflow-allreduce/blob/66d5b855e90b0949e9fa5cca5599fd729a70e874/tensorflow/python/summary/writer/writer.py#L46-L95
apple/turicreate
cce55aa5311300e3ce6af93cb45ba791fd1bdf49
deps/src/libxml2-2.9.1/python/libxml2class.py
python
xmlNode.xpathNodeEval
(self, str, ctx)
return xpathObjectRet(ret)
Evaluate the XPath Location Path in the given context. The node 'node' is set as the context node. The context node is not restored.
Evaluate the XPath Location Path in the given context. The node 'node' is set as the context node. The context node is not restored.
[ "Evaluate", "the", "XPath", "Location", "Path", "in", "the", "given", "context", ".", "The", "node", "node", "is", "set", "as", "the", "context", "node", ".", "The", "context", "node", "is", "not", "restored", "." ]
def xpathNodeEval(self, str, ctx): """Evaluate the XPath Location Path in the given context. The node 'node' is set as the context node. The context node is not restored. """ if ctx is None: ctx__o = None else: ctx__o = ctx._o ret = libxml2mod.xmlXPathNodeEval(self._o, str, ctx__o) if ret is None:raise xpathError('xmlXPathNodeEval() failed') return xpathObjectRet(ret)
[ "def", "xpathNodeEval", "(", "self", ",", "str", ",", "ctx", ")", ":", "if", "ctx", "is", "None", ":", "ctx__o", "=", "None", "else", ":", "ctx__o", "=", "ctx", ".", "_o", "ret", "=", "libxml2mod", ".", "xmlXPathNodeEval", "(", "self", ".", "_o", ",", "str", ",", "ctx__o", ")", "if", "ret", "is", "None", ":", "raise", "xpathError", "(", "'xmlXPathNodeEval() failed'", ")", "return", "xpathObjectRet", "(", "ret", ")" ]
https://github.com/apple/turicreate/blob/cce55aa5311300e3ce6af93cb45ba791fd1bdf49/deps/src/libxml2-2.9.1/python/libxml2class.py#L2935-L2943
aws/lumberyard
f85344403c1c2e77ec8c75deb2c116e97b713217
dev/Gems/CloudGemFramework/v1/ResourceManager/lib/Crypto/Cipher/_mode_eax.py
python
EaxMode.update
(self, assoc_data)
return self
Protect associated data If there is any associated data, the caller has to invoke this function one or more times, before using ``decrypt`` or ``encrypt``. By *associated data* it is meant any data (e.g. packet headers) that will not be encrypted and will be transmitted in the clear. However, the receiver is still able to detect any modification to it. If there is no associated data, this method must not be called. The caller may split associated data in segments of any size, and invoke this method multiple times, each time with the next segment. :Parameters: assoc_data : bytes/bytearray/memoryview A piece of associated data. There are no restrictions on its size.
Protect associated data
[ "Protect", "associated", "data" ]
def update(self, assoc_data): """Protect associated data If there is any associated data, the caller has to invoke this function one or more times, before using ``decrypt`` or ``encrypt``. By *associated data* it is meant any data (e.g. packet headers) that will not be encrypted and will be transmitted in the clear. However, the receiver is still able to detect any modification to it. If there is no associated data, this method must not be called. The caller may split associated data in segments of any size, and invoke this method multiple times, each time with the next segment. :Parameters: assoc_data : bytes/bytearray/memoryview A piece of associated data. There are no restrictions on its size. """ if self.update not in self._next: raise TypeError("update() can only be called" " immediately after initialization") self._next = [self.update, self.encrypt, self.decrypt, self.digest, self.verify] self._signer.update(assoc_data) return self
[ "def", "update", "(", "self", ",", "assoc_data", ")", ":", "if", "self", ".", "update", "not", "in", "self", ".", "_next", ":", "raise", "TypeError", "(", "\"update() can only be called\"", "\" immediately after initialization\"", ")", "self", ".", "_next", "=", "[", "self", ".", "update", ",", "self", ".", "encrypt", ",", "self", ".", "decrypt", ",", "self", ".", "digest", ",", "self", ".", "verify", "]", "self", ".", "_signer", ".", "update", "(", "assoc_data", ")", "return", "self" ]
https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Gems/CloudGemFramework/v1/ResourceManager/lib/Crypto/Cipher/_mode_eax.py#L127-L156
adobe/chromium
cfe5bf0b51b1f6b9fe239c2a3c2f2364da9967d7
tools/json_schema_compiler/cc_generator.py
python
CCGenerator._GenerateTypePopulateProperty
(self, prop, src, dst)
return c
Generate the code to populate a single property in a type. src: DictionaryValue* dst: Type*
Generate the code to populate a single property in a type.
[ "Generate", "the", "code", "to", "populate", "a", "single", "property", "in", "a", "type", "." ]
def _GenerateTypePopulateProperty(self, prop, src, dst): """Generate the code to populate a single property in a type. src: DictionaryValue* dst: Type* """ c = Code() value_var = prop.unix_name + '_value' c.Append('Value* %(value_var)s = NULL;') if prop.optional: (c.Sblock( 'if (%(src)s->GetWithoutPathExpansion("%(key)s", &%(value_var)s)) {' ) .Concat(self._GeneratePopulatePropertyFromValue( prop, value_var, dst, 'false')) .Eblock('}') ) else: (c.Append( 'if (!%(src)s->GetWithoutPathExpansion("%(key)s", &%(value_var)s))') .Append(' return false;') .Concat(self._GeneratePopulatePropertyFromValue( prop, value_var, dst, 'false')) ) c.Append() c.Substitute({'value_var': value_var, 'key': prop.name, 'src': src}) return c
[ "def", "_GenerateTypePopulateProperty", "(", "self", ",", "prop", ",", "src", ",", "dst", ")", ":", "c", "=", "Code", "(", ")", "value_var", "=", "prop", ".", "unix_name", "+", "'_value'", "c", ".", "Append", "(", "'Value* %(value_var)s = NULL;'", ")", "if", "prop", ".", "optional", ":", "(", "c", ".", "Sblock", "(", "'if (%(src)s->GetWithoutPathExpansion(\"%(key)s\", &%(value_var)s)) {'", ")", ".", "Concat", "(", "self", ".", "_GeneratePopulatePropertyFromValue", "(", "prop", ",", "value_var", ",", "dst", ",", "'false'", ")", ")", ".", "Eblock", "(", "'}'", ")", ")", "else", ":", "(", "c", ".", "Append", "(", "'if (!%(src)s->GetWithoutPathExpansion(\"%(key)s\", &%(value_var)s))'", ")", ".", "Append", "(", "' return false;'", ")", ".", "Concat", "(", "self", ".", "_GeneratePopulatePropertyFromValue", "(", "prop", ",", "value_var", ",", "dst", ",", "'false'", ")", ")", ")", "c", ".", "Append", "(", ")", "c", ".", "Substitute", "(", "{", "'value_var'", ":", "value_var", ",", "'key'", ":", "prop", ".", "name", ",", "'src'", ":", "src", "}", ")", "return", "c" ]
https://github.com/adobe/chromium/blob/cfe5bf0b51b1f6b9fe239c2a3c2f2364da9967d7/tools/json_schema_compiler/cc_generator.py#L209-L235
krishauser/Klampt
972cc83ea5befac3f653c1ba20f80155768ad519
Python/klampt/src/robotsim.py
python
Widget.hasFocus
(self)
return _robotsim.Widget_hasFocus(self)
r""" hasFocus(Widget self) -> bool
r""" hasFocus(Widget self) -> bool
[ "r", "hasFocus", "(", "Widget", "self", ")", "-", ">", "bool" ]
def hasFocus(self) -> "bool": r""" hasFocus(Widget self) -> bool """ return _robotsim.Widget_hasFocus(self)
[ "def", "hasFocus", "(", "self", ")", "->", "\"bool\"", ":", "return", "_robotsim", ".", "Widget_hasFocus", "(", "self", ")" ]
https://github.com/krishauser/Klampt/blob/972cc83ea5befac3f653c1ba20f80155768ad519/Python/klampt/src/robotsim.py#L3423-L3429
tensorflow/ngraph-bridge
ea6422491ec75504e78a63db029e7f74ec3479a5
examples/retrain_ngraph.py
python
ensure_dir_exists
(dir_name)
Makes sure the folder exists on disk. Args: dir_name: Path string to the folder we want to create.
Makes sure the folder exists on disk.
[ "Makes", "sure", "the", "folder", "exists", "on", "disk", "." ]
def ensure_dir_exists(dir_name): """Makes sure the folder exists on disk. Args: dir_name: Path string to the folder we want to create. """ if not os.path.exists(dir_name): os.makedirs(dir_name)
[ "def", "ensure_dir_exists", "(", "dir_name", ")", ":", "if", "not", "os", ".", "path", ".", "exists", "(", "dir_name", ")", ":", "os", ".", "makedirs", "(", "dir_name", ")" ]
https://github.com/tensorflow/ngraph-bridge/blob/ea6422491ec75504e78a63db029e7f74ec3479a5/examples/retrain_ngraph.py#L371-L378
Xilinx/Vitis-AI
fc74d404563d9951b57245443c73bef389f3657f
tools/Vitis-AI-Quantizer/vai_q_tensorflow1.x/tensorflow/python/keras/wrappers/scikit_learn.py
python
BaseWrapper.filter_sk_params
(self, fn, override=None)
return res
Filters `sk_params` and returns those in `fn`'s arguments. Arguments: fn : arbitrary function override: dictionary, values to override `sk_params` Returns: res : dictionary containing variables in both `sk_params` and `fn`'s arguments.
Filters `sk_params` and returns those in `fn`'s arguments.
[ "Filters", "sk_params", "and", "returns", "those", "in", "fn", "s", "arguments", "." ]
def filter_sk_params(self, fn, override=None): """Filters `sk_params` and returns those in `fn`'s arguments. Arguments: fn : arbitrary function override: dictionary, values to override `sk_params` Returns: res : dictionary containing variables in both `sk_params` and `fn`'s arguments. """ override = override or {} res = {} for name, value in self.sk_params.items(): if has_arg(fn, name): res.update({name: value}) res.update(override) return res
[ "def", "filter_sk_params", "(", "self", ",", "fn", ",", "override", "=", "None", ")", ":", "override", "=", "override", "or", "{", "}", "res", "=", "{", "}", "for", "name", ",", "value", "in", "self", ".", "sk_params", ".", "items", "(", ")", ":", "if", "has_arg", "(", "fn", ",", "name", ")", ":", "res", ".", "update", "(", "{", "name", ":", "value", "}", ")", "res", ".", "update", "(", "override", ")", "return", "res" ]
https://github.com/Xilinx/Vitis-AI/blob/fc74d404563d9951b57245443c73bef389f3657f/tools/Vitis-AI-Quantizer/vai_q_tensorflow1.x/tensorflow/python/keras/wrappers/scikit_learn.py#L170-L187
Cisco-Talos/moflow
ed71dfb0540d9e0d7a4c72f0881b58958d573728
BAP-0.7-moflow/libtracewrap/libtrace/protobuf/python/google/protobuf/service.py
python
RpcController.StartCancel
(self)
Initiate cancellation. Advises the RPC system that the caller desires that the RPC call be canceled. The RPC system may cancel it immediately, may wait awhile and then cancel it, or may not even cancel the call at all. If the call is canceled, the "done" callback will still be called and the RpcController will indicate that the call failed at that time.
Initiate cancellation.
[ "Initiate", "cancellation", "." ]
def StartCancel(self): """Initiate cancellation. Advises the RPC system that the caller desires that the RPC call be canceled. The RPC system may cancel it immediately, may wait awhile and then cancel it, or may not even cancel the call at all. If the call is canceled, the "done" callback will still be called and the RpcController will indicate that the call failed at that time. """ raise NotImplementedError
[ "def", "StartCancel", "(", "self", ")", ":", "raise", "NotImplementedError" ]
https://github.com/Cisco-Talos/moflow/blob/ed71dfb0540d9e0d7a4c72f0881b58958d573728/BAP-0.7-moflow/libtracewrap/libtrace/protobuf/python/google/protobuf/service.py#L154-L163
miyosuda/TensorFlowAndroidDemo
35903e0221aa5f109ea2dbef27f20b52e317f42d
jni-build/jni/include/tensorflow/contrib/losses/python/losses/loss_ops.py
python
_scale_losses
(losses, weight)
return math_ops.reduce_sum(reduced_losses)
Computes the scaled loss. Args: losses: A `Tensor` of size [batch_size, d1, ... dN]. weight: A `Tensor` of size [1], [batch_size] or [batch_size, d1, ... dN]. The `losses` are reduced (tf.reduce_sum) until its dimension matches that of `weight` at which point the reduced `losses` are element-wise multiplied by `weight` and a final reduce_sum is computed on the result. Conceptually, this operation is equivalent to broadcasting (tiling) `weight` to be the same size as `losses`, performing an element-wise multiplication, and summing the result. Returns: A scalar tf.float32 `Tensor` whose value represents the sum of the scaled `losses`.
Computes the scaled loss.
[ "Computes", "the", "scaled", "loss", "." ]
def _scale_losses(losses, weight): """Computes the scaled loss. Args: losses: A `Tensor` of size [batch_size, d1, ... dN]. weight: A `Tensor` of size [1], [batch_size] or [batch_size, d1, ... dN]. The `losses` are reduced (tf.reduce_sum) until its dimension matches that of `weight` at which point the reduced `losses` are element-wise multiplied by `weight` and a final reduce_sum is computed on the result. Conceptually, this operation is equivalent to broadcasting (tiling) `weight` to be the same size as `losses`, performing an element-wise multiplication, and summing the result. Returns: A scalar tf.float32 `Tensor` whose value represents the sum of the scaled `losses`. """ # First, compute the sum of the losses over all elements: start_index = max(0, weight.get_shape().ndims) reduction_indices = list(range(start_index, losses.get_shape().ndims)) reduced_losses = math_ops.reduce_sum(losses, reduction_indices=reduction_indices) reduced_losses = math_ops.mul(reduced_losses, weight) return math_ops.reduce_sum(reduced_losses)
[ "def", "_scale_losses", "(", "losses", ",", "weight", ")", ":", "# First, compute the sum of the losses over all elements:", "start_index", "=", "max", "(", "0", ",", "weight", ".", "get_shape", "(", ")", ".", "ndims", ")", "reduction_indices", "=", "list", "(", "range", "(", "start_index", ",", "losses", ".", "get_shape", "(", ")", ".", "ndims", ")", ")", "reduced_losses", "=", "math_ops", ".", "reduce_sum", "(", "losses", ",", "reduction_indices", "=", "reduction_indices", ")", "reduced_losses", "=", "math_ops", ".", "mul", "(", "reduced_losses", ",", "weight", ")", "return", "math_ops", ".", "reduce_sum", "(", "reduced_losses", ")" ]
https://github.com/miyosuda/TensorFlowAndroidDemo/blob/35903e0221aa5f109ea2dbef27f20b52e317f42d/jni-build/jni/include/tensorflow/contrib/losses/python/losses/loss_ops.py#L45-L68
Xilinx/Vitis-AI
fc74d404563d9951b57245443c73bef389f3657f
tools/Vitis-AI-Quantizer/vai_q_tensorflow1.x/tensorflow/python/ops/summary_ops_v2.py
python
create_file_writer_v2
(logdir, max_queue=None, flush_millis=None, filename_suffix=None, name=None)
Creates a summary file writer for the given log directory. Args: logdir: a string specifying the directory in which to write an event file. max_queue: the largest number of summaries to keep in a queue; will flush once the queue gets bigger than this. Defaults to 10. flush_millis: the largest interval between flushes. Defaults to 120,000. filename_suffix: optional suffix for the event file name. Defaults to `.v2`. name: a name for the op that creates the writer. Returns: A SummaryWriter object.
Creates a summary file writer for the given log directory.
[ "Creates", "a", "summary", "file", "writer", "for", "the", "given", "log", "directory", "." ]
def create_file_writer_v2(logdir, max_queue=None, flush_millis=None, filename_suffix=None, name=None): """Creates a summary file writer for the given log directory. Args: logdir: a string specifying the directory in which to write an event file. max_queue: the largest number of summaries to keep in a queue; will flush once the queue gets bigger than this. Defaults to 10. flush_millis: the largest interval between flushes. Defaults to 120,000. filename_suffix: optional suffix for the event file name. Defaults to `.v2`. name: a name for the op that creates the writer. Returns: A SummaryWriter object. """ if logdir is None: raise ValueError("logdir cannot be None") inside_function = ops.inside_function() with ops.name_scope(name, "create_file_writer") as scope, ops.device("cpu:0"): # Run init inside an init_scope() to hoist it out of tf.functions. with ops.init_scope(): if context.executing_eagerly(): _check_create_file_writer_args( inside_function, logdir=logdir, max_queue=max_queue, flush_millis=flush_millis, filename_suffix=filename_suffix) logdir = ops.convert_to_tensor(logdir, dtype=dtypes.string) if max_queue is None: max_queue = constant_op.constant(10) if flush_millis is None: flush_millis = constant_op.constant(2 * 60 * 1000) if filename_suffix is None: filename_suffix = constant_op.constant(".v2") # Prepend the PID and a process-local UID to the filename suffix to avoid # filename collisions within the machine (the filename already contains # the hostname to avoid cross-machine collisions). unique_prefix = constant_op.constant(".%s.%s" % (os.getpid(), ops.uid())) filename_suffix = unique_prefix + filename_suffix # Use a unique shared_name to prevent resource sharing. if context.executing_eagerly(): shared_name = context.shared_name() else: shared_name = ops.name_from_scope_name(scope) # pylint: disable=protected-access return ResourceSummaryWriter( shared_name=shared_name, init_op_fn=functools.partial( gen_summary_ops.create_summary_file_writer, logdir=logdir, max_queue=max_queue, flush_millis=flush_millis, filename_suffix=filename_suffix), name=name, v2=True)
[ "def", "create_file_writer_v2", "(", "logdir", ",", "max_queue", "=", "None", ",", "flush_millis", "=", "None", ",", "filename_suffix", "=", "None", ",", "name", "=", "None", ")", ":", "if", "logdir", "is", "None", ":", "raise", "ValueError", "(", "\"logdir cannot be None\"", ")", "inside_function", "=", "ops", ".", "inside_function", "(", ")", "with", "ops", ".", "name_scope", "(", "name", ",", "\"create_file_writer\"", ")", "as", "scope", ",", "ops", ".", "device", "(", "\"cpu:0\"", ")", ":", "# Run init inside an init_scope() to hoist it out of tf.functions.", "with", "ops", ".", "init_scope", "(", ")", ":", "if", "context", ".", "executing_eagerly", "(", ")", ":", "_check_create_file_writer_args", "(", "inside_function", ",", "logdir", "=", "logdir", ",", "max_queue", "=", "max_queue", ",", "flush_millis", "=", "flush_millis", ",", "filename_suffix", "=", "filename_suffix", ")", "logdir", "=", "ops", ".", "convert_to_tensor", "(", "logdir", ",", "dtype", "=", "dtypes", ".", "string", ")", "if", "max_queue", "is", "None", ":", "max_queue", "=", "constant_op", ".", "constant", "(", "10", ")", "if", "flush_millis", "is", "None", ":", "flush_millis", "=", "constant_op", ".", "constant", "(", "2", "*", "60", "*", "1000", ")", "if", "filename_suffix", "is", "None", ":", "filename_suffix", "=", "constant_op", ".", "constant", "(", "\".v2\"", ")", "# Prepend the PID and a process-local UID to the filename suffix to avoid", "# filename collisions within the machine (the filename already contains", "# the hostname to avoid cross-machine collisions).", "unique_prefix", "=", "constant_op", ".", "constant", "(", "\".%s.%s\"", "%", "(", "os", ".", "getpid", "(", ")", ",", "ops", ".", "uid", "(", ")", ")", ")", "filename_suffix", "=", "unique_prefix", "+", "filename_suffix", "# Use a unique shared_name to prevent resource sharing.", "if", "context", ".", "executing_eagerly", "(", ")", ":", "shared_name", "=", "context", ".", "shared_name", "(", ")", "else", ":", "shared_name", "=", "ops", ".", "name_from_scope_name", "(", "scope", ")", "# pylint: disable=protected-access", "return", "ResourceSummaryWriter", "(", "shared_name", "=", "shared_name", ",", "init_op_fn", "=", "functools", ".", "partial", "(", "gen_summary_ops", ".", "create_summary_file_writer", ",", "logdir", "=", "logdir", ",", "max_queue", "=", "max_queue", ",", "flush_millis", "=", "flush_millis", ",", "filename_suffix", "=", "filename_suffix", ")", ",", "name", "=", "name", ",", "v2", "=", "True", ")" ]
https://github.com/Xilinx/Vitis-AI/blob/fc74d404563d9951b57245443c73bef389f3657f/tools/Vitis-AI-Quantizer/vai_q_tensorflow1.x/tensorflow/python/ops/summary_ops_v2.py#L334-L391
aws/lumberyard
f85344403c1c2e77ec8c75deb2c116e97b713217
dev/Gems/CloudGemFramework/v1/AWS/common-code/lib/pycparser/plyparser.py
python
PLYParser._create_opt_rule
(self, rulename)
Given a rule name, creates an optional ply.yacc rule for it. The name of the optional rule is <rulename>_opt
Given a rule name, creates an optional ply.yacc rule for it. The name of the optional rule is <rulename>_opt
[ "Given", "a", "rule", "name", "creates", "an", "optional", "ply", ".", "yacc", "rule", "for", "it", ".", "The", "name", "of", "the", "optional", "rule", "is", "<rulename", ">", "_opt" ]
def _create_opt_rule(self, rulename): """ Given a rule name, creates an optional ply.yacc rule for it. The name of the optional rule is <rulename>_opt """ optname = rulename + '_opt' def optrule(self, p): p[0] = p[1] optrule.__doc__ = '%s : empty\n| %s' % (optname, rulename) optrule.__name__ = 'p_%s' % optname setattr(self.__class__, optrule.__name__, optrule)
[ "def", "_create_opt_rule", "(", "self", ",", "rulename", ")", ":", "optname", "=", "rulename", "+", "'_opt'", "def", "optrule", "(", "self", ",", "p", ")", ":", "p", "[", "0", "]", "=", "p", "[", "1", "]", "optrule", ".", "__doc__", "=", "'%s : empty\\n| %s'", "%", "(", "optname", ",", "rulename", ")", "optrule", ".", "__name__", "=", "'p_%s'", "%", "optname", "setattr", "(", "self", ".", "__class__", ",", "optrule", ".", "__name__", ",", "optrule", ")" ]
https://github.com/aws/lumberyard/blob/f85344403c1c2e77ec8c75deb2c116e97b713217/dev/Gems/CloudGemFramework/v1/AWS/common-code/lib/pycparser/plyparser.py#L35-L47
rapidsai/cudf
d5b2448fc69f17509304d594f029d0df56984962
python/cudf/cudf/core/column/string.py
python
StringMethods.upper
(self)
return self._return_or_inplace(libstrings.to_upper(self._column))
Convert each string to uppercase. This only applies to ASCII characters at this time. Equivalent to `str.upper() <https://docs.python.org/3/library/stdtypes.html#str.upper>`_. Returns ------- Series or Index of object See also -------- lower Converts all characters to lowercase. upper Converts all characters to uppercase. title Converts first character of each word to uppercase and remaining to lowercase. capitalize Converts first character to uppercase and remaining to lowercase. swapcase Converts uppercase to lowercase and lowercase to uppercase. Examples -------- >>> import cudf >>> data = ['lower', 'CAPITALS', 'this is a sentence', 'SwApCaSe'] >>> s = cudf.Series(data) >>> s 0 lower 1 CAPITALS 2 this is a sentence 3 SwApCaSe dtype: object >>> s.str.upper() 0 LOWER 1 CAPITALS 2 THIS IS A SENTENCE 3 SWAPCASE dtype: object
Convert each string to uppercase. This only applies to ASCII characters at this time.
[ "Convert", "each", "string", "to", "uppercase", ".", "This", "only", "applies", "to", "ASCII", "characters", "at", "this", "time", "." ]
def upper(self) -> SeriesOrIndex: """ Convert each string to uppercase. This only applies to ASCII characters at this time. Equivalent to `str.upper() <https://docs.python.org/3/library/stdtypes.html#str.upper>`_. Returns ------- Series or Index of object See also -------- lower Converts all characters to lowercase. upper Converts all characters to uppercase. title Converts first character of each word to uppercase and remaining to lowercase. capitalize Converts first character to uppercase and remaining to lowercase. swapcase Converts uppercase to lowercase and lowercase to uppercase. Examples -------- >>> import cudf >>> data = ['lower', 'CAPITALS', 'this is a sentence', 'SwApCaSe'] >>> s = cudf.Series(data) >>> s 0 lower 1 CAPITALS 2 this is a sentence 3 SwApCaSe dtype: object >>> s.str.upper() 0 LOWER 1 CAPITALS 2 THIS IS A SENTENCE 3 SWAPCASE dtype: object """ return self._return_or_inplace(libstrings.to_upper(self._column))
[ "def", "upper", "(", "self", ")", "->", "SeriesOrIndex", ":", "return", "self", ".", "_return_or_inplace", "(", "libstrings", ".", "to_upper", "(", "self", ".", "_column", ")", ")" ]
https://github.com/rapidsai/cudf/blob/d5b2448fc69f17509304d594f029d0df56984962/python/cudf/cudf/core/column/string.py#L1760-L1809
ChromiumWebApps/chromium
c7361d39be8abd1574e6ce8957c8dbddd4c6ccf7
tools/idl_parser/idl_ppapi_parser.py
python
IDLPPAPIParser.p_LabelCont
(self, p)
LabelCont : ',' LabelList |
LabelCont : ',' LabelList |
[ "LabelCont", ":", "LabelList", "|" ]
def p_LabelCont(self, p): """LabelCont : ',' LabelList |""" if len(p) > 1: p[0] = p[2]
[ "def", "p_LabelCont", "(", "self", ",", "p", ")", ":", "if", "len", "(", "p", ")", ">", "1", ":", "p", "[", "0", "]", "=", "p", "[", "2", "]" ]
https://github.com/ChromiumWebApps/chromium/blob/c7361d39be8abd1574e6ce8957c8dbddd4c6ccf7/tools/idl_parser/idl_ppapi_parser.py#L102-L106
digibyte/digibyte
0b8a04fb06d5470a15168e2f675aec57bcc24dac
contrib/devtools/optimize-pngs.py
python
content_hash
(filename)
return hashlib.sha256(data).hexdigest()
Return hash of RGBA contents of image
Return hash of RGBA contents of image
[ "Return", "hash", "of", "RGBA", "contents", "of", "image" ]
def content_hash(filename): '''Return hash of RGBA contents of image''' i = Image.open(filename) i = i.convert('RGBA') data = i.tobytes() return hashlib.sha256(data).hexdigest()
[ "def", "content_hash", "(", "filename", ")", ":", "i", "=", "Image", ".", "open", "(", "filename", ")", "i", "=", "i", ".", "convert", "(", "'RGBA'", ")", "data", "=", "i", ".", "tobytes", "(", ")", "return", "hashlib", ".", "sha256", "(", "data", ")", ".", "hexdigest", "(", ")" ]
https://github.com/digibyte/digibyte/blob/0b8a04fb06d5470a15168e2f675aec57bcc24dac/contrib/devtools/optimize-pngs.py#L20-L25
nasa/trick
7b85aa66329d62fe8816462627c09a353aac8299
share/trick/trickops/TrickWorkflow.py
python
SingleRun._sim_time
(self)
return 'Sim Time: {0:7.1f} sec'.format( self._tics.value / self._tics_per_sec.value)
Get a string displaying the sim time. Returns ------- str A string for displaying sim time.
Get a string displaying the sim time.
[ "Get", "a", "string", "displaying", "the", "sim", "time", "." ]
def _sim_time(self): """ Get a string displaying the sim time. Returns ------- str A string for displaying sim time. """ return 'Sim Time: {0:7.1f} sec'.format( self._tics.value / self._tics_per_sec.value)
[ "def", "_sim_time", "(", "self", ")", ":", "return", "'Sim Time: {0:7.1f} sec'", ".", "format", "(", "self", ".", "_tics", ".", "value", "/", "self", ".", "_tics_per_sec", ".", "value", ")" ]
https://github.com/nasa/trick/blob/7b85aa66329d62fe8816462627c09a353aac8299/share/trick/trickops/TrickWorkflow.py#L1440-L1450
HyeonwooNoh/caffe
d9e8494a2832d67b25dee37194c7bcb9d52d0e42
scripts/cpp_lint.py
python
CheckForCopyright
(filename, lines, error)
Logs an error if a Copyright message appears at the top of the file.
Logs an error if a Copyright message appears at the top of the file.
[ "Logs", "an", "error", "if", "a", "Copyright", "message", "appears", "at", "the", "top", "of", "the", "file", "." ]
def CheckForCopyright(filename, lines, error): """Logs an error if a Copyright message appears at the top of the file.""" # We'll check up to line 10. Don't forget there's a # dummy line at the front. for line in xrange(1, min(len(lines), 11)): if _RE_COPYRIGHT.search(lines[line], re.I): error(filename, 0, 'legal/copyright', 5, 'Copyright message found. ' 'You should not include a copyright line.')
[ "def", "CheckForCopyright", "(", "filename", ",", "lines", ",", "error", ")", ":", "# We'll check up to line 10. Don't forget there's a", "# dummy line at the front.", "for", "line", "in", "xrange", "(", "1", ",", "min", "(", "len", "(", "lines", ")", ",", "11", ")", ")", ":", "if", "_RE_COPYRIGHT", ".", "search", "(", "lines", "[", "line", "]", ",", "re", ".", "I", ")", ":", "error", "(", "filename", ",", "0", ",", "'legal/copyright'", ",", "5", ",", "'Copyright message found. '", "'You should not include a copyright line.'", ")" ]
https://github.com/HyeonwooNoh/caffe/blob/d9e8494a2832d67b25dee37194c7bcb9d52d0e42/scripts/cpp_lint.py#L1372-L1381
wxWidgets/wxPython-Classic
19571e1ae65f1ac445f5491474121998c97a1bf0
src/osx_carbon/grid.py
python
Grid.GetSelectionBlockBottomRight
(*args, **kwargs)
return _grid.Grid_GetSelectionBlockBottomRight(*args, **kwargs)
GetSelectionBlockBottomRight(self) -> wxGridCellCoordsArray
GetSelectionBlockBottomRight(self) -> wxGridCellCoordsArray
[ "GetSelectionBlockBottomRight", "(", "self", ")", "-", ">", "wxGridCellCoordsArray" ]
def GetSelectionBlockBottomRight(*args, **kwargs): """GetSelectionBlockBottomRight(self) -> wxGridCellCoordsArray""" return _grid.Grid_GetSelectionBlockBottomRight(*args, **kwargs)
[ "def", "GetSelectionBlockBottomRight", "(", "*", "args", ",", "*", "*", "kwargs", ")", ":", "return", "_grid", ".", "Grid_GetSelectionBlockBottomRight", "(", "*", "args", ",", "*", "*", "kwargs", ")" ]
https://github.com/wxWidgets/wxPython-Classic/blob/19571e1ae65f1ac445f5491474121998c97a1bf0/src/osx_carbon/grid.py#L2065-L2067
eclipse/sumo
7132a9b8b6eea734bdec38479026b4d8c4336d03
tools/traci/_polygon.py
python
PolygonDomain.getType
(self, polygonID)
return self._getUniversal(tc.VAR_TYPE, polygonID)
getType(string) -> string Returns the (abstract) type of the polygon.
getType(string) -> string
[ "getType", "(", "string", ")", "-", ">", "string" ]
def getType(self, polygonID): """getType(string) -> string Returns the (abstract) type of the polygon. """ return self._getUniversal(tc.VAR_TYPE, polygonID)
[ "def", "getType", "(", "self", ",", "polygonID", ")", ":", "return", "self", ".", "_getUniversal", "(", "tc", ".", "VAR_TYPE", ",", "polygonID", ")" ]
https://github.com/eclipse/sumo/blob/7132a9b8b6eea734bdec38479026b4d8c4336d03/tools/traci/_polygon.py#L33-L38
trailofbits/llvm-sanitizer-tutorial
d29dfeec7f51fbf234fd0080f28f2b30cd0b6e99
llvm/utils/lit/lit/llvm/subst.py
python
ToolSubst.__init__
(self, key, command=None, pre=r'.-^/\<', post='-.', verbatim=False, unresolved='warn', extra_args=None)
Construct a ToolSubst. key: The text which is to be substituted. command: The command to substitute when the key is matched. By default, this will treat `key` as a tool name and search for it. If it is a string, it is intereprted as an exact path. If it is an instance of FindTool, the specified tool name is searched for on disk. pre: If specified, the substitution will not find matches where the character immediately preceding the word-boundary that begins `key` is any of the characters in the string `pre`. post: If specified, the substitution will not find matches where the character immediately after the word-boundary that ends `key` is any of the characters specified in the string `post`. verbatim: If True, `key` is an exact regex that is passed to the underlying substitution unresolved: Action to take if the tool substitution cannot be resolved. Valid values: 'warn' - log a warning but add the substitution anyway. 'fatal' - Exit the test suite and log a fatal error. 'break' - Don't add any of the substitutions from the current group, and return a value indicating a failure. 'ignore' - Don't add the substitution, and don't log an error extra_args: If specified, represents a list of arguments that will be appended to the tool's substitution. explicit_path: If specified, the exact path will be used as a substitution. Otherwise, the tool will be searched for as if by calling which(tool)
Construct a ToolSubst.
[ "Construct", "a", "ToolSubst", "." ]
def __init__(self, key, command=None, pre=r'.-^/\<', post='-.', verbatim=False, unresolved='warn', extra_args=None): """Construct a ToolSubst. key: The text which is to be substituted. command: The command to substitute when the key is matched. By default, this will treat `key` as a tool name and search for it. If it is a string, it is intereprted as an exact path. If it is an instance of FindTool, the specified tool name is searched for on disk. pre: If specified, the substitution will not find matches where the character immediately preceding the word-boundary that begins `key` is any of the characters in the string `pre`. post: If specified, the substitution will not find matches where the character immediately after the word-boundary that ends `key` is any of the characters specified in the string `post`. verbatim: If True, `key` is an exact regex that is passed to the underlying substitution unresolved: Action to take if the tool substitution cannot be resolved. Valid values: 'warn' - log a warning but add the substitution anyway. 'fatal' - Exit the test suite and log a fatal error. 'break' - Don't add any of the substitutions from the current group, and return a value indicating a failure. 'ignore' - Don't add the substitution, and don't log an error extra_args: If specified, represents a list of arguments that will be appended to the tool's substitution. explicit_path: If specified, the exact path will be used as a substitution. Otherwise, the tool will be searched for as if by calling which(tool) """ self.unresolved = unresolved self.extra_args = extra_args self.key = key self.command = command if command is not None else FindTool(key) self.was_resolved = False if verbatim: self.regex = key return def not_in(chars, where=''): if not chars: return '' pattern_str = '|'.join(re.escape(x) for x in chars) return r'(?{}!({}))'.format(where, pattern_str) def wordify(word): match = wordifier.match(word) introducer = match.group(1) word = match.group(2) return introducer + r'\b' + word + r'\b' self.regex = not_in(pre, '<') + wordify(key) + not_in(post)
[ "def", "__init__", "(", "self", ",", "key", ",", "command", "=", "None", ",", "pre", "=", "r'.-^/\\<'", ",", "post", "=", "'-.'", ",", "verbatim", "=", "False", ",", "unresolved", "=", "'warn'", ",", "extra_args", "=", "None", ")", ":", "self", ".", "unresolved", "=", "unresolved", "self", ".", "extra_args", "=", "extra_args", "self", ".", "key", "=", "key", "self", ".", "command", "=", "command", "if", "command", "is", "not", "None", "else", "FindTool", "(", "key", ")", "self", ".", "was_resolved", "=", "False", "if", "verbatim", ":", "self", ".", "regex", "=", "key", "return", "def", "not_in", "(", "chars", ",", "where", "=", "''", ")", ":", "if", "not", "chars", ":", "return", "''", "pattern_str", "=", "'|'", ".", "join", "(", "re", ".", "escape", "(", "x", ")", "for", "x", "in", "chars", ")", "return", "r'(?{}!({}))'", ".", "format", "(", "where", ",", "pattern_str", ")", "def", "wordify", "(", "word", ")", ":", "match", "=", "wordifier", ".", "match", "(", "word", ")", "introducer", "=", "match", ".", "group", "(", "1", ")", "word", "=", "match", ".", "group", "(", "2", ")", "return", "introducer", "+", "r'\\b'", "+", "word", "+", "r'\\b'", "self", ".", "regex", "=", "not_in", "(", "pre", ",", "'<'", ")", "+", "wordify", "(", "key", ")", "+", "not_in", "(", "post", ")" ]
https://github.com/trailofbits/llvm-sanitizer-tutorial/blob/d29dfeec7f51fbf234fd0080f28f2b30cd0b6e99/llvm/utils/lit/lit/llvm/subst.py#L42-L100
BlzFans/wke
b0fa21158312e40c5fbd84682d643022b6c34a93
cygwin/lib/python2.6/trace.py
python
find_executable_linenos
(filename)
return find_lines(code, strs)
Return dict where keys are line numbers in the line number table.
Return dict where keys are line numbers in the line number table.
[ "Return", "dict", "where", "keys", "are", "line", "numbers", "in", "the", "line", "number", "table", "." ]
def find_executable_linenos(filename): """Return dict where keys are line numbers in the line number table.""" try: prog = open(filename, "rU").read() except IOError, err: print >> sys.stderr, ("Not printing coverage data for %r: %s" % (filename, err)) return {} code = compile(prog, filename, "exec") strs = find_strings(filename) return find_lines(code, strs)
[ "def", "find_executable_linenos", "(", "filename", ")", ":", "try", ":", "prog", "=", "open", "(", "filename", ",", "\"rU\"", ")", ".", "read", "(", ")", "except", "IOError", ",", "err", ":", "print", ">>", "sys", ".", "stderr", ",", "(", "\"Not printing coverage data for %r: %s\"", "%", "(", "filename", ",", "err", ")", ")", "return", "{", "}", "code", "=", "compile", "(", "prog", ",", "filename", ",", "\"exec\"", ")", "strs", "=", "find_strings", "(", "filename", ")", "return", "find_lines", "(", "code", ",", "strs", ")" ]
https://github.com/BlzFans/wke/blob/b0fa21158312e40c5fbd84682d643022b6c34a93/cygwin/lib/python2.6/trace.py#L427-L437
apiaryio/snowcrash
b5b39faa85f88ee17459edf39fdc6fe4fc70d2e3
tools/gyp/pylib/gyp/win_tool.py
python
WinTool.ExecActionWrapper
(self, arch, rspfile, *dir)
return subprocess.call(args, shell=True, env=env, cwd=dir)
Runs an action command line from a response file using the environment for |arch|. If |dir| is supplied, use that as the working directory.
Runs an action command line from a response file using the environment for |arch|. If |dir| is supplied, use that as the working directory.
[ "Runs", "an", "action", "command", "line", "from", "a", "response", "file", "using", "the", "environment", "for", "|arch|", ".", "If", "|dir|", "is", "supplied", "use", "that", "as", "the", "working", "directory", "." ]
def ExecActionWrapper(self, arch, rspfile, *dir): """Runs an action command line from a response file using the environment for |arch|. If |dir| is supplied, use that as the working directory.""" env = self._GetEnv(arch) # TODO(scottmg): This is a temporary hack to get some specific variables # through to actions that are set after gyp-time. http://crbug.com/333738. for k, v in os.environ.iteritems(): if k not in env: env[k] = v args = open(rspfile).read() dir = dir[0] if dir else None return subprocess.call(args, shell=True, env=env, cwd=dir)
[ "def", "ExecActionWrapper", "(", "self", ",", "arch", ",", "rspfile", ",", "*", "dir", ")", ":", "env", "=", "self", ".", "_GetEnv", "(", "arch", ")", "# TODO(scottmg): This is a temporary hack to get some specific variables", "# through to actions that are set after gyp-time. http://crbug.com/333738.", "for", "k", ",", "v", "in", "os", ".", "environ", ".", "iteritems", "(", ")", ":", "if", "k", "not", "in", "env", ":", "env", "[", "k", "]", "=", "v", "args", "=", "open", "(", "rspfile", ")", ".", "read", "(", ")", "dir", "=", "dir", "[", "0", "]", "if", "dir", "else", "None", "return", "subprocess", ".", "call", "(", "args", ",", "shell", "=", "True", ",", "env", "=", "env", ",", "cwd", "=", "dir", ")" ]
https://github.com/apiaryio/snowcrash/blob/b5b39faa85f88ee17459edf39fdc6fe4fc70d2e3/tools/gyp/pylib/gyp/win_tool.py#L289-L300
catboost/catboost
167f64f237114a4d10b2b4ee42adb4569137debe
contrib/python/pandas/py2/pandas/tseries/offsets.py
python
BusinessHourMixin.rollforward
(self, dt)
return dt
Roll provided date forward to next offset only if not on offset.
Roll provided date forward to next offset only if not on offset.
[ "Roll", "provided", "date", "forward", "to", "next", "offset", "only", "if", "not", "on", "offset", "." ]
def rollforward(self, dt): """ Roll provided date forward to next offset only if not on offset. """ if not self.onOffset(dt): if self.n >= 0: return self._next_opening_time(dt) else: return self._prev_opening_time(dt) return dt
[ "def", "rollforward", "(", "self", ",", "dt", ")", ":", "if", "not", "self", ".", "onOffset", "(", "dt", ")", ":", "if", "self", ".", "n", ">=", "0", ":", "return", "self", ".", "_next_opening_time", "(", "dt", ")", "else", ":", "return", "self", ".", "_prev_opening_time", "(", "dt", ")", "return", "dt" ]
https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/pandas/py2/pandas/tseries/offsets.py#L671-L680
catboost/catboost
167f64f237114a4d10b2b4ee42adb4569137debe
contrib/python/scipy/scipy/signal/ltisys.py
python
lti.output
(self, U, T, X0=None)
return lsim(self, U, T, X0=X0)
Return the response of a continuous-time system to input `U`. See `lsim` for details.
Return the response of a continuous-time system to input `U`. See `lsim` for details.
[ "Return", "the", "response", "of", "a", "continuous", "-", "time", "system", "to", "input", "U", ".", "See", "lsim", "for", "details", "." ]
def output(self, U, T, X0=None): """ Return the response of a continuous-time system to input `U`. See `lsim` for details. """ return lsim(self, U, T, X0=X0)
[ "def", "output", "(", "self", ",", "U", ",", "T", ",", "X0", "=", "None", ")", ":", "return", "lsim", "(", "self", ",", "U", ",", "T", ",", "X0", "=", "X0", ")" ]
https://github.com/catboost/catboost/blob/167f64f237114a4d10b2b4ee42adb4569137debe/contrib/python/scipy/scipy/signal/ltisys.py#L414-L419