org_text
stringlengths
761
968k
texts
sequence
scores
sequence
num_lines
int64
1
25.7k
avg_score
float64
0
0.31
#!/usr/bin/env python """ patrol_smach.py - Version 1.0 2013-04-12 Control a robot to patrol a square area using SMACH Created for the Pi Robot Project: http://www.pirobot.org Copyright (c) 2013 Patrick Goebel. All rights reserved. This program is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version.5 This program is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details at: http://www.gnu.org/licenses/gpl.htmlPoint """ import rospy from smach import State, StateMachine from smach_ros import SimpleActionState, IntrospectionServer from geometry_msgs.msg import Twist from rbx2_tasks.task_setup import * class Patrol(): def __init__(self): rospy.init_node('patrol_smach', anonymous=False) # Set the shutdown function (stop the robot) rospy.on_shutdown(self.shutdown) # Initialize a number of parameters and variables setup_task_environment(self) # Track success rate of getting to the goal locations self.n_succeeded = 0 self.n_aborted = 0 self.n_preempted = 0 # A list to hold then navigation waypoints nav_states = list() # Turn the waypoints into SMACH states for waypoint in self.waypoints: nav_goal = MoveBaseGoal() nav_goal.target_pose.header.frame_id = 'base_footprint' nav_goal.target_pose.pose = waypoint move_base_state = SimpleActionState('move_base', MoveBaseAction, goal=nav_goal, result_cb=self.move_base_result_cb, exec_timeout=rospy.Duration(10.0), server_wait_timeout=rospy.Duration(10.0)) nav_states.append(move_base_state) # Initialize the patrol state machine self.sm_patrol = StateMachine(outcomes=['succeeded','aborted','preempted']) # Add the states to the state machine with the appropriate transitions with self.sm_patrol: StateMachine.add('NAV_STATE_0', nav_states[0], transitions={'succeeded':'NAV_STATE_1','aborted':'NAV_STATE_1','preempted':'NAV_STATE_1'}) StateMachine.add('NAV_STATE_1', nav_states[1], transitions={'succeeded':'NAV_STATE_2','aborted':'NAV_STATE_2','preempted':'NAV_STATE_2'}) StateMachine.add('NAV_STATE_2', nav_states[2], transitions={'succeeded':'NAV_STATE_3','aborted':'NAV_STATE_3','preempted':'NAV_STATE_3'}) StateMachine.add('NAV_STATE_3', nav_states[3], transitions={'succeeded':'NAV_STATE_4','aborted':'NAV_STATE_4','preempted':'NAV_STATE_4'}) StateMachine.add('NAV_STATE_4', nav_states[0], transitions={'succeeded':'','aborted':'','preempted':''}) # Create and start the SMACH introspection server intro_server = IntrospectionServer('patrol', self.sm_patrol, '/SM_ROOT') intro_server.start() # Execute the state machine for the specified number of patrols while (self.n_patrols == -1 or self.patrol_count < self.n_patrols) and not rospy.is_shutdown(): sm_outcome = self.sm_patrol.execute() self.patrol_count += 1 rospy.loginfo("FINISHED PATROL LOOP: " + str(self.patrol_count)) rospy.loginfo('State Machine Outcome: ' + str(sm_outcome)) intro_server.stop() def move_base_result_cb(self, userdata, status, result): if status == actionlib.GoalStatus.SUCCEEDED: self.n_succeeded += 1 elif status == actionlib.GoalStatus.ABORTED: self.n_aborted += 1 elif status == actionlib.GoalStatus.PREEMPTED: self.n_preempted += 1 try: rospy.loginfo("Success rate: " + str(100.0 * self.n_succeeded / (self.n_succeeded + self.n_aborted + self.n_preempted))) except: pass def shutdown(self): rospy.loginfo("Stopping the robot...") self.sm_patrol.request_preempt() self.cmd_vel_pub.publish(Twist()) rospy.sleep(1) if __name__ == '__main__': try: Patrol() except rospy.ROSInterruptException: rospy.loginfo("SMACH test finished.")
[ "#!/usr/bin/env python\n", "\n", "\"\"\" patrol_smach.py - Version 1.0 2013-04-12\n", "\n", " Control a robot to patrol a square area using SMACH\n", "\n", " Created for the Pi Robot Project: http://www.pirobot.org\n", " Copyright (c) 2013 Patrick Goebel. All rights reserved.\n", "\n", " This program is free software; you can redistribute it and/or modify\n", " it under the terms of the GNU General Public License as published by\n", " the Free Software Foundation; either version 2 of the License, or\n", " (at your option) any later version.5\n", " \n", " This program is distributed in the hope that it will be useful,\n", " but WITHOUT ANY WARRANTY; without even the implied warranty of\n", " MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n", " GNU General Public License for more details at:\n", " \n", " http://www.gnu.org/licenses/gpl.htmlPoint\n", " \n", "\"\"\"\n", "\n", "import rospy\n", "from smach import State, StateMachine\n", "from smach_ros import SimpleActionState, IntrospectionServer\n", "from geometry_msgs.msg import Twist\n", "from rbx2_tasks.task_setup import *\n", "\n", "class Patrol():\n", " def __init__(self):\n", " rospy.init_node('patrol_smach', anonymous=False)\n", " \n", " # Set the shutdown function (stop the robot)\n", " rospy.on_shutdown(self.shutdown)\n", " \n", " # Initialize a number of parameters and variables\n", " setup_task_environment(self)\n", " \n", " # Track success rate of getting to the goal locations\n", " self.n_succeeded = 0\n", " self.n_aborted = 0\n", " self.n_preempted = 0\n", "\n", " # A list to hold then navigation waypoints\n", " nav_states = list()\n", " \n", " # Turn the waypoints into SMACH states\n", " for waypoint in self.waypoints: \n", " nav_goal = MoveBaseGoal()\n", " nav_goal.target_pose.header.frame_id = 'base_footprint'\n", " nav_goal.target_pose.pose = waypoint\n", " move_base_state = SimpleActionState('move_base', MoveBaseAction, goal=nav_goal, result_cb=self.move_base_result_cb,\n", " exec_timeout=rospy.Duration(10.0),\n", " server_wait_timeout=rospy.Duration(10.0))\n", " nav_states.append(move_base_state)\n", " \n", " # Initialize the patrol state machine\n", " self.sm_patrol = StateMachine(outcomes=['succeeded','aborted','preempted'])\n", "\n", " # Add the states to the state machine with the appropriate transitions\n", " with self.sm_patrol: \n", " StateMachine.add('NAV_STATE_0', nav_states[0], transitions={'succeeded':'NAV_STATE_1','aborted':'NAV_STATE_1','preempted':'NAV_STATE_1'})\n", " StateMachine.add('NAV_STATE_1', nav_states[1], transitions={'succeeded':'NAV_STATE_2','aborted':'NAV_STATE_2','preempted':'NAV_STATE_2'})\n", " StateMachine.add('NAV_STATE_2', nav_states[2], transitions={'succeeded':'NAV_STATE_3','aborted':'NAV_STATE_3','preempted':'NAV_STATE_3'})\n", " StateMachine.add('NAV_STATE_3', nav_states[3], transitions={'succeeded':'NAV_STATE_4','aborted':'NAV_STATE_4','preempted':'NAV_STATE_4'})\n", " StateMachine.add('NAV_STATE_4', nav_states[0], transitions={'succeeded':'','aborted':'','preempted':''})\n", " \n", " # Create and start the SMACH introspection server\n", " intro_server = IntrospectionServer('patrol', self.sm_patrol, '/SM_ROOT')\n", " intro_server.start()\n", " \n", " # Execute the state machine for the specified number of patrols\n", " while (self.n_patrols == -1 or self.patrol_count < self.n_patrols) and not rospy.is_shutdown():\n", " sm_outcome = self.sm_patrol.execute()\n", " self.patrol_count += 1\n", " rospy.loginfo(\"FINISHED PATROL LOOP: \" + str(self.patrol_count))\n", " \n", " rospy.loginfo('State Machine Outcome: ' + str(sm_outcome))\n", " \n", " intro_server.stop()\n", " \n", " def move_base_result_cb(self, userdata, status, result):\n", " if status == actionlib.GoalStatus.SUCCEEDED:\n", " self.n_succeeded += 1\n", " elif status == actionlib.GoalStatus.ABORTED:\n", " self.n_aborted += 1\n", " elif status == actionlib.GoalStatus.PREEMPTED:\n", " self.n_preempted += 1\n", "\n", " try:\n", " rospy.loginfo(\"Success rate: \" + str(100.0 * self.n_succeeded / (self.n_succeeded + self.n_aborted + self.n_preempted)))\n", " except:\n", " pass\n", "\n", " def shutdown(self):\n", " rospy.loginfo(\"Stopping the robot...\")\n", " \n", " self.sm_patrol.request_preempt()\n", " \n", " self.cmd_vel_pub.publish(Twist())\n", " \n", " rospy.sleep(1)\n", "\n", "if __name__ == '__main__':\n", " try:\n", " Patrol()\n", " except rospy.ROSInterruptException:\n", " rospy.loginfo(\"SMACH test finished.\")\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.2, 0, 0, 0, 0, 0.2, 0, 0.14285714285714285, 0, 0, 0, 0, 0, 0, 0, 0, 0.0625, 0, 0, 0.1111111111111111, 0, 0, 0.1111111111111111, 0, 0, 0.1111111111111111, 0, 0, 0, 0, 0, 0, 0, 0.1111111111111111, 0, 0.0196078431372549, 0, 0, 0, 0.0078125, 0.023809523809523808, 0.02197802197802198, 0, 0.1111111111111111, 0, 0.03571428571428571, 0, 0, 0.024390243902439025, 0.04, 0.04, 0.04, 0.04, 0.05128205128205128, 0.07692307692307693, 0, 0.012345679012345678, 0, 0.1111111111111111, 0, 0.009615384615384616, 0, 0, 0, 0.1111111111111111, 0, 0.058823529411764705, 0, 0.1111111111111111, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.014925373134328358, 0.0625, 0, 0, 0, 0, 0.1111111111111111, 0, 0.1111111111111111, 0, 0.1111111111111111, 0, 0, 0.037037037037037035, 0, 0, 0, 0 ]
109
0.022425
from __future__ import division, print_function, absolute_import import numpy as np # linear algebra import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv) import tensorflow as tf import tflearn from tflearn.data_utils import to_categorical import tflearn.data_utils as du from tflearn.layers.core import input_data, dropout, fully_connected from tflearn.layers.conv import conv_2d, max_pool_2d from tflearn.layers.normalization import local_response_normalization from tflearn.layers.estimator import regression import matplotlib.pyplot as plt import matplotlib as matplot import seaborn as sns import random trainx = pd.read_csv("/home/cheesecake/GAT/gat/scraping/ArabicTextExtractor/Arabic Handwritten Characters Dataset CSV/csvTrainImages 13440x1024.csv",header=None) trainy = pd.read_csv("/home/cheesecake/GAT/gat/scraping/ArabicTextExtractor/Arabic Handwritten Characters Dataset CSV/csvTrainLabel 13440x1.csv",header=None) testx = pd.read_csv("/home/cheesecake/GAT/gat/scraping/ArabicTextExtractor/Arabic Handwritten Characters Dataset CSV/csvTestImages 3360x1024.csv",header=None) testy = pd.read_csv("/home/cheesecake/GAT/gat/scraping/ArabicTextExtractor/Arabic Handwritten Characters Dataset CSV/csvTestLabel 3360x1.csv",header=None) # Split data into training set and validation set #training images trainx = trainx.values.astype('float32') #training labels trainy = trainy.values.astype('int32')-1 #testing images testx = testx.values.astype('float32') #testing labels testy = testy.values.astype('int32')-1 original_trainy = trainy #One Hot encoding of train labels. trainy = to_categorical(trainy[:,0],28) original_testy = testy #One Hot encoding of test labels. testy = to_categorical(testy[:,0],28) # reshape input images to 28x28x1 trainx = trainx.reshape([-1, 32, 32, 1]) testx = testx.reshape([-1, 32, 32, 1]) arabic_labels = ['alef', 'beh', 'teh', 'theh', 'jeem', 'hah', 'khah', 'dal', 'thal', 'reh', 'zain', 'seen', 'sheen', 'sad', 'dad', 'tah', 'zah', 'ain', 'ghain', 'feh', 'qaf', 'kaf', 'lam', 'meem', 'noon', 'heh', 'waw', 'yeh'] #size of images should be 1200 by 2300 #@return - trainy_new: List of tuples that represent the corners of the 32x32 character box in clockwise order starting from top left def get_image_batch(trainx, num_images): pad_left = random.randint(5,2265) pad_right = 2300-32-pad_left pad_top = random.randint(5,1165) pad_bottom = 1200-32-pad_top trainx_new = np.empty((num_images, pad_left + pad_right + 32, pad_top + pad_bottom + 32, 1)) for i in range(num_images): index = random.randint(0 ,len(trainx)-1) trainx_new[i] = np.pad(trainx[index], ((pad_left, pad_right), (pad_top, pad_bottom), (0, 0)), 'constant') trainy_new = [(pad_left,pad_top),(2300-pad_right,pad_top),(2300-pad_right,1200-pad_bottom),(pad_left,1200-pad_bottom)] return trainx_new, trainy_new for i in range(10): images, labels = get_image_batch(trainx,1) plt.imshow(images[0].squeeze().T) print(labels) plt.show() #x = random.randint(0, 13440) #plt.imshow(trainx_new[x].squeeze().T) #plt.title(arabic_labels[original_trainy[x][0]]) #plt.show() #Zero center every sample with specified mean. If not specified, the mean is evaluated over all samples. trainx, mean1 = du.featurewise_zero_center(trainx) testx, mean2 = du.featurewise_zero_center(testx) print(trainx.shape, trainy.shape, testx.shape, testy.shape) # Building convolutional network network = input_data(shape=[None, 32, 32, 1], name='input') network = conv_2d(network, 80, 3, activation='relu', regularizer="L2") network = max_pool_2d(network, 2) network = local_response_normalization(network) network = conv_2d(network, 64, 3, activation='relu', regularizer="L2") network = max_pool_2d(network, 2) network = local_response_normalization(network) network = fully_connected(network, 1024, activation='relu') network = dropout(network, 0.8) network = fully_connected(network, 512, activation='relu') network = dropout(network, 0.8) network = fully_connected(network, 28, activation='softmax') network = regression(network, optimizer='sgd', learning_rate=0.01, loss='categorical_crossentropy', name='target') #model complile model = tflearn.DNN(network, tensorboard_verbose=0) #model fitting scores = [] for i in range(100): model.fit({'input': trainx}, {'target': trainy}, n_epoch=1, validation_set=({'input': testx}, {'target': testy}), snapshot_step=100, show_metric=True, run_id='convnet_arabic_digits') score = model.evaluate(testx, testy) print('Test accuracy: %0.2f%%' % (score[0] * 100)) scores.append(score[0]*100) print(scores) x = list(range(len(scores))) y = [] for el in x: y.append(el + 1) plt.plot(y, scores, 'k-') plt.title("Accuracy vs Epochs Trained") plt.xlabel("Num Epochs") plt.ylabel("Accuracy on Testing Data") plt.grid() plt.show(block=False) plt.pause(.1) plt.savefig('AccuracyGraph.pdf') print(scores)
[ "from __future__ import division, print_function, absolute_import\n", "\n", "import numpy as np # linear algebra\n", "import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)\n", "import tensorflow as tf\n", "import tflearn\n", "from tflearn.data_utils import to_categorical\n", "import tflearn.data_utils as du\n", "from tflearn.layers.core import input_data, dropout, fully_connected\n", "from tflearn.layers.conv import conv_2d, max_pool_2d\n", "from tflearn.layers.normalization import local_response_normalization\n", "from tflearn.layers.estimator import regression\n", "import matplotlib.pyplot as plt\n", "import matplotlib as matplot\n", "import seaborn as sns\n", "import random\n", "\n", "\n", "trainx = pd.read_csv(\"/home/cheesecake/GAT/gat/scraping/ArabicTextExtractor/Arabic Handwritten Characters Dataset CSV/csvTrainImages 13440x1024.csv\",header=None)\n", "trainy = pd.read_csv(\"/home/cheesecake/GAT/gat/scraping/ArabicTextExtractor/Arabic Handwritten Characters Dataset CSV/csvTrainLabel 13440x1.csv\",header=None)\n", "\n", "testx = pd.read_csv(\"/home/cheesecake/GAT/gat/scraping/ArabicTextExtractor/Arabic Handwritten Characters Dataset CSV/csvTestImages 3360x1024.csv\",header=None)\n", "testy = pd.read_csv(\"/home/cheesecake/GAT/gat/scraping/ArabicTextExtractor/Arabic Handwritten Characters Dataset CSV/csvTestLabel 3360x1.csv\",header=None)\n", "# Split data into training set and validation set\n", "#training images\n", "trainx = trainx.values.astype('float32')\n", "#training labels\n", "trainy = trainy.values.astype('int32')-1\n", "#testing images\n", "testx = testx.values.astype('float32')\n", "#testing labels\n", "testy = testy.values.astype('int32')-1\n", "original_trainy = trainy\n", "#One Hot encoding of train labels.\n", "trainy = to_categorical(trainy[:,0],28)\n", "original_testy = testy\n", "#One Hot encoding of test labels.\n", "testy = to_categorical(testy[:,0],28)\n", "# reshape input images to 28x28x1\n", "trainx = trainx.reshape([-1, 32, 32, 1])\n", "testx = testx.reshape([-1, 32, 32, 1])\n", "\n", "arabic_labels = ['alef', 'beh', 'teh', 'theh', 'jeem', 'hah', 'khah', 'dal', 'thal',\n", " 'reh', 'zain', 'seen', 'sheen', 'sad', 'dad', 'tah', 'zah', 'ain',\n", " 'ghain', 'feh', 'qaf', 'kaf', 'lam', 'meem', 'noon', 'heh', 'waw', 'yeh']\n", "\n", "#size of images should be 1200 by 2300\n", "\n", "\n", "#@return - trainy_new: List of tuples that represent the corners of the 32x32 character box in clockwise order starting from top left\n", "def get_image_batch(trainx, num_images):\n", " pad_left = random.randint(5,2265)\n", " pad_right = 2300-32-pad_left\n", " pad_top = random.randint(5,1165)\n", " pad_bottom = 1200-32-pad_top\n", " trainx_new = np.empty((num_images, pad_left + pad_right + 32, pad_top + pad_bottom + 32, 1))\n", " for i in range(num_images):\n", " index = random.randint(0 ,len(trainx)-1)\n", " trainx_new[i] = np.pad(trainx[index], ((pad_left, pad_right), (pad_top, pad_bottom), (0, 0)), 'constant')\n", " trainy_new = [(pad_left,pad_top),(2300-pad_right,pad_top),(2300-pad_right,1200-pad_bottom),(pad_left,1200-pad_bottom)]\n", " return trainx_new, trainy_new\n", "\n", "\n", "\n", "for i in range(10):\n", " images, labels = get_image_batch(trainx,1)\n", " plt.imshow(images[0].squeeze().T)\n", " print(labels)\n", " plt.show()\n", "\n", "#x = random.randint(0, 13440)\n", "#plt.imshow(trainx_new[x].squeeze().T)\n", "#plt.title(arabic_labels[original_trainy[x][0]])\n", "#plt.show()\n", "\n", "#Zero center every sample with specified mean. If not specified, the mean is evaluated over all samples.\n", "trainx, mean1 = du.featurewise_zero_center(trainx)\n", "testx, mean2 = du.featurewise_zero_center(testx)\n", "\n", "print(trainx.shape, trainy.shape, testx.shape, testy.shape)\n", "\n", "# Building convolutional network\n", "network = input_data(shape=[None, 32, 32, 1], name='input')\n", "network = conv_2d(network, 80, 3, activation='relu', regularizer=\"L2\")\n", "network = max_pool_2d(network, 2)\n", "network = local_response_normalization(network)\n", "network = conv_2d(network, 64, 3, activation='relu', regularizer=\"L2\")\n", "network = max_pool_2d(network, 2)\n", "network = local_response_normalization(network)\n", "network = fully_connected(network, 1024, activation='relu')\n", "network = dropout(network, 0.8)\n", "network = fully_connected(network, 512, activation='relu')\n", "network = dropout(network, 0.8)\n", "network = fully_connected(network, 28, activation='softmax')\n", "network = regression(network, optimizer='sgd', learning_rate=0.01,\n", " loss='categorical_crossentropy', name='target')\n", "\n", "#model complile\n", "model = tflearn.DNN(network, tensorboard_verbose=0)\n", "#model fitting\n", "scores = []\n", "\n", "for i in range(100):\n", " model.fit({'input': trainx}, {'target': trainy}, n_epoch=1,\n", " validation_set=({'input': testx}, {'target': testy}),\n", " snapshot_step=100, show_metric=True, run_id='convnet_arabic_digits')\n", " score = model.evaluate(testx, testy)\n", " print('Test accuracy: %0.2f%%' % (score[0] * 100))\n", " scores.append(score[0]*100)\n", " print(scores)\n", " x = list(range(len(scores)))\n", " y = []\n", " for el in x:\n", " y.append(el + 1)\n", " plt.plot(y, scores, 'k-')\n", " plt.title(\"Accuracy vs Epochs Trained\")\n", " plt.xlabel(\"Num Epochs\")\n", " plt.ylabel(\"Accuracy on Testing Data\")\n", " plt.grid()\n", " plt.show(block=False)\n", " plt.pause(.1)\n", "\n", "plt.savefig('AccuracyGraph.pdf')\n", "\n", "print(scores)\n" ]
[ 0, 0, 0.027777777777777776, 0.014084507042253521, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.012345679012345678, 0.012658227848101266, 0, 0.012578616352201259, 0.012903225806451613, 0, 0.058823529411764705, 0, 0.058823529411764705, 0, 0.0625, 0, 0.0625, 0, 0, 0.02857142857142857, 0.05, 0, 0.029411764705882353, 0.05263157894736842, 0, 0, 0, 0, 0.011764705882352941, 0.024096385542168676, 0.022222222222222223, 0, 0.02564102564102564, 0, 0, 0.014925373134328358, 0, 0.02631578947368421, 0, 0.02702702702702703, 0, 0.010309278350515464, 0, 0.04081632653061224, 0.008771929824561403, 0.06504065040650407, 0, 0, 0, 0, 0.05, 0.02127659574468085, 0, 0, 0, 0, 0.03333333333333333, 0.02564102564102564, 0.02040816326530612, 0.08333333333333333, 0, 0.01904761904761905, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.0625, 0, 0.06666666666666667, 0, 0, 0, 0, 0.015384615384615385, 0.0125, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
125
0.009461
#!/usr/bin/env python3 # THIS FILE IS PART OF THE CYLC SUITE ENGINE. # Copyright (C) NIWA & British Crown (Met Office) & Contributors. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. """cylc [info] show [OPTIONS] ARGS Query a running workflow for: cylc show REG - workflow metadata cylc show REG TASK_NAME - task metadata cylc show REG TASK_GLOB - prerequisites and outputs of task instances Prerequisite and output status is indicated for current active tasks. """ import json import sys from ansimarkup import ansiprint from cylc.flow import ID_DELIM from cylc.flow.option_parsers import CylcOptionParser as COP from cylc.flow.network.client import SuiteRuntimeClient from cylc.flow.task_id import TaskID from cylc.flow.terminal import cli_function WORKFLOW_META_QUERY = ''' query ($wFlows: [ID]!) { workflows (ids: $wFlows, stripNull: false) { meta { title description URL userDefined } } } ''' TASK_META_QUERY = ''' query ($wFlows: [ID]!, $taskIds: [ID]) { tasks (workflows: $wFlows, ids: $taskIds, stripNull: false) { name meta { title description URL userDefined } } } ''' TASK_PREREQS_QUERY = ''' query ($wFlows: [ID]!, $taskIds: [ID]) { taskProxies (workflows: $wFlows, ids: $taskIds, stripNull: false) { name cyclePoint task { meta { title description URL userDefined } } prerequisites { expression conditions { exprAlias taskId reqState message satisfied } satisfied } outputs extras } } ''' def print_msg_state(msg, state): if state: ansiprint(f'<green> + {msg}</green>') else: ansiprint(f'<red> - {msg}</red>') def flatten_data(data, flat_data=None): if flat_data is None: flat_data = {} for key, value in data.items(): if isinstance(value, dict): flatten_data(value, flat_data) elif isinstance(value, list): for member in value: flatten_data(member, flat_data) else: flat_data[key] = value return flat_data def get_option_parser(): parser = COP( __doc__, comms=True, multitask=True, argdoc=[ ('REG', 'Suite name'), ('[TASK_NAME or TASK_GLOB ...]', 'Task names or match patterns')]) parser.add_option('--list-prereqs', action="store_true", default=False, help="Print a task's pre-requisites as a list.") parser.add_option('--json', action="store_true", default=False, help="Print output in JSON format.") return parser @cli_function(get_option_parser) def main(_, options, suite, *task_args): """Implement "cylc show" CLI.""" pclient = SuiteRuntimeClient(suite, timeout=options.comms_timeout) json_filter = {} if not task_args: query = WORKFLOW_META_QUERY query_kwargs = { 'request_string': query, 'variables': {'wFlows': [suite]} } # Print suite info. results = pclient('graphql', query_kwargs) for workflow in results['workflows']: flat_data = flatten_data(workflow) if options.json: json_filter.update(flat_data) else: for key, value in sorted(flat_data.items(), reverse=True): ansiprint( f'<bold>{key}:</bold> {value or "<m>(not given)</m>"}') task_names = [arg for arg in task_args if TaskID.is_valid_name(arg)] task_ids = [arg for arg in task_args if TaskID.is_valid_id_2(arg)] if task_names: tasks_query = TASK_META_QUERY tasks_kwargs = { 'request_string': tasks_query, 'variables': {'wFlows': [suite], 'taskIds': task_names} } # Print suite info. results = pclient('graphql', tasks_kwargs) multi = len(results['tasks']) > 1 for task in results['tasks']: flat_data = flatten_data(task['meta']) if options.json: json_filter.update({task['name']: flat_data}) else: if multi: print(f'----\nTASK NAME: {task["name"]}') for key, value in sorted(flat_data.items(), reverse=True): ansiprint( f'<bold>{key}:</bold> {value or "<m>(not given)</m>"}') if task_ids: tp_query = TASK_PREREQS_QUERY tp_kwargs = { 'request_string': tp_query, 'variables': { 'wFlows': [suite], 'taskIds': [ f'{c}{ID_DELIM}{n}' for n, c in [ TaskID.split(t_id) for t_id in task_ids if TaskID.is_valid_id(t_id) ] ] + [ f'{c}{ID_DELIM}{n}' for c, n in [ t_id.rsplit(TaskID.DELIM2, 1) for t_id in task_ids if not TaskID.is_valid_id(t_id) ] ] } } results = pclient('graphql', tp_kwargs) multi = len(results['taskProxies']) > 1 for t_proxy in results['taskProxies']: task_id = TaskID.get(t_proxy['name'], t_proxy['cyclePoint']) if options.json: json_filter.update({task_id: t_proxy}) else: if multi: print(f'----\nTASK ID: {task_id}') prereqs = [] for item in t_proxy['prerequisites']: prefix = '' multi_cond = len(item['conditions']) > 1 if multi_cond: prereqs.append([ True, '', item['expression'].replace('c', ''), item['satisfied'] ]) for cond in item['conditions']: if multi_cond and not options.list_prereqs: prefix = f'\t{cond["exprAlias"].strip("c")} = ' _, _, point, name = cond['taskId'].split(ID_DELIM) cond_id = TaskID.get(name, point) prereqs.append([ False, prefix, f'{cond_id} {cond["reqState"]}', cond['satisfied'] ]) if options.list_prereqs: for composite, _, msg, _ in prereqs: if not composite: print(msg) else: flat_meta = flatten_data(t_proxy['task']['meta']) for key, value in sorted(flat_meta.items(), reverse=True): ansiprint( f'<bold>{key}:</bold>' f' {value or "<m>(not given)</m>"}') ansiprint( '\n<bold>prerequisites</bold>' ' (<red>- => not satisfied</red>):') if not prereqs: print(' (None)') for _, prefix, msg, state in prereqs: print_msg_state(f'{prefix}{msg}', state) ansiprint( '\n<bold>outputs</bold>' ' (<red>- => not completed</red>):') if not t_proxy['outputs']: print(' (None)') for key, val in t_proxy['outputs'].items(): print_msg_state(f'{task_id} {key}', val) if t_proxy['extras']: print('\nother:') for key, value in t_proxy['extras'].items(): print(' o %s ... %s' % (key, value)) if not results['taskProxies']: ansiprint( f"<red>No matching tasks found: {task_ids}", file=sys.stderr) sys.exit(1) if options.json: print(json.dumps(json_filter, indent=4)) if __name__ == "__main__": main()
[ "#!/usr/bin/env python3\n", "\n", "# THIS FILE IS PART OF THE CYLC SUITE ENGINE.\n", "# Copyright (C) NIWA & British Crown (Met Office) & Contributors.\n", "#\n", "# This program is free software: you can redistribute it and/or modify\n", "# it under the terms of the GNU General Public License as published by\n", "# the Free Software Foundation, either version 3 of the License, or\n", "# (at your option) any later version.\n", "#\n", "# This program is distributed in the hope that it will be useful,\n", "# but WITHOUT ANY WARRANTY; without even the implied warranty of\n", "# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n", "# GNU General Public License for more details.\n", "#\n", "# You should have received a copy of the GNU General Public License\n", "# along with this program. If not, see <http://www.gnu.org/licenses/>.\n", "\n", "\"\"\"cylc [info] show [OPTIONS] ARGS\n", "\n", "Query a running workflow for:\n", " cylc show REG - workflow metadata\n", " cylc show REG TASK_NAME - task metadata\n", " cylc show REG TASK_GLOB - prerequisites and outputs of task instances\n", "\n", "Prerequisite and output status is indicated for current active tasks.\n", "\"\"\"\n", "\n", "import json\n", "import sys\n", "\n", "from ansimarkup import ansiprint\n", "\n", "from cylc.flow import ID_DELIM\n", "from cylc.flow.option_parsers import CylcOptionParser as COP\n", "from cylc.flow.network.client import SuiteRuntimeClient\n", "from cylc.flow.task_id import TaskID\n", "from cylc.flow.terminal import cli_function\n", "\n", "\n", "WORKFLOW_META_QUERY = '''\n", "query ($wFlows: [ID]!) {\n", " workflows (ids: $wFlows, stripNull: false) {\n", " meta {\n", " title\n", " description\n", " URL\n", " userDefined\n", " }\n", " }\n", "}\n", "'''\n", "\n", "TASK_META_QUERY = '''\n", "query ($wFlows: [ID]!, $taskIds: [ID]) {\n", " tasks (workflows: $wFlows, ids: $taskIds, stripNull: false) {\n", " name\n", " meta {\n", " title\n", " description\n", " URL\n", " userDefined\n", " }\n", " }\n", "}\n", "'''\n", "\n", "TASK_PREREQS_QUERY = '''\n", "query ($wFlows: [ID]!, $taskIds: [ID]) {\n", " taskProxies (workflows: $wFlows, ids: $taskIds, stripNull: false) {\n", " name\n", " cyclePoint\n", " task {\n", " meta {\n", " title\n", " description\n", " URL\n", " userDefined\n", " }\n", " }\n", " prerequisites {\n", " expression\n", " conditions {\n", " exprAlias\n", " taskId\n", " reqState\n", " message\n", " satisfied\n", " }\n", " satisfied\n", " }\n", " outputs\n", " extras\n", " }\n", "}\n", "'''\n", "\n", "\n", "def print_msg_state(msg, state):\n", " if state:\n", " ansiprint(f'<green> + {msg}</green>')\n", " else:\n", " ansiprint(f'<red> - {msg}</red>')\n", "\n", "\n", "def flatten_data(data, flat_data=None):\n", " if flat_data is None:\n", " flat_data = {}\n", " for key, value in data.items():\n", " if isinstance(value, dict):\n", " flatten_data(value, flat_data)\n", " elif isinstance(value, list):\n", " for member in value:\n", " flatten_data(member, flat_data)\n", " else:\n", " flat_data[key] = value\n", " return flat_data\n", "\n", "\n", "def get_option_parser():\n", " parser = COP(\n", " __doc__, comms=True, multitask=True,\n", " argdoc=[\n", " ('REG', 'Suite name'),\n", " ('[TASK_NAME or TASK_GLOB ...]', 'Task names or match patterns')])\n", "\n", " parser.add_option('--list-prereqs', action=\"store_true\", default=False,\n", " help=\"Print a task's pre-requisites as a list.\")\n", "\n", " parser.add_option('--json', action=\"store_true\", default=False,\n", " help=\"Print output in JSON format.\")\n", "\n", " return parser\n", "\n", "\n", "@cli_function(get_option_parser)\n", "def main(_, options, suite, *task_args):\n", " \"\"\"Implement \"cylc show\" CLI.\"\"\"\n", " pclient = SuiteRuntimeClient(suite, timeout=options.comms_timeout)\n", " json_filter = {}\n", "\n", " if not task_args:\n", " query = WORKFLOW_META_QUERY\n", " query_kwargs = {\n", " 'request_string': query,\n", " 'variables': {'wFlows': [suite]}\n", " }\n", " # Print suite info.\n", " results = pclient('graphql', query_kwargs)\n", " for workflow in results['workflows']:\n", " flat_data = flatten_data(workflow)\n", " if options.json:\n", " json_filter.update(flat_data)\n", " else:\n", " for key, value in sorted(flat_data.items(), reverse=True):\n", " ansiprint(\n", " f'<bold>{key}:</bold> {value or \"<m>(not given)</m>\"}')\n", "\n", " task_names = [arg for arg in task_args if TaskID.is_valid_name(arg)]\n", " task_ids = [arg for arg in task_args if TaskID.is_valid_id_2(arg)]\n", "\n", " if task_names:\n", " tasks_query = TASK_META_QUERY\n", " tasks_kwargs = {\n", " 'request_string': tasks_query,\n", " 'variables': {'wFlows': [suite], 'taskIds': task_names}\n", " }\n", " # Print suite info.\n", " results = pclient('graphql', tasks_kwargs)\n", " multi = len(results['tasks']) > 1\n", " for task in results['tasks']:\n", " flat_data = flatten_data(task['meta'])\n", " if options.json:\n", " json_filter.update({task['name']: flat_data})\n", " else:\n", " if multi:\n", " print(f'----\\nTASK NAME: {task[\"name\"]}')\n", " for key, value in sorted(flat_data.items(), reverse=True):\n", " ansiprint(\n", " f'<bold>{key}:</bold> {value or \"<m>(not given)</m>\"}')\n", "\n", " if task_ids:\n", " tp_query = TASK_PREREQS_QUERY\n", " tp_kwargs = {\n", " 'request_string': tp_query,\n", " 'variables': {\n", " 'wFlows': [suite],\n", " 'taskIds': [\n", " f'{c}{ID_DELIM}{n}'\n", " for n, c in [\n", " TaskID.split(t_id)\n", " for t_id in task_ids\n", " if TaskID.is_valid_id(t_id)\n", " ]\n", " ] + [\n", " f'{c}{ID_DELIM}{n}'\n", " for c, n in [\n", " t_id.rsplit(TaskID.DELIM2, 1)\n", " for t_id in task_ids\n", " if not TaskID.is_valid_id(t_id)\n", " ]\n", " ]\n", " }\n", " }\n", " results = pclient('graphql', tp_kwargs)\n", " multi = len(results['taskProxies']) > 1\n", " for t_proxy in results['taskProxies']:\n", " task_id = TaskID.get(t_proxy['name'], t_proxy['cyclePoint'])\n", " if options.json:\n", " json_filter.update({task_id: t_proxy})\n", " else:\n", " if multi:\n", " print(f'----\\nTASK ID: {task_id}')\n", " prereqs = []\n", " for item in t_proxy['prerequisites']:\n", " prefix = ''\n", " multi_cond = len(item['conditions']) > 1\n", " if multi_cond:\n", " prereqs.append([\n", " True,\n", " '',\n", " item['expression'].replace('c', ''),\n", " item['satisfied']\n", " ])\n", " for cond in item['conditions']:\n", " if multi_cond and not options.list_prereqs:\n", " prefix = f'\\t{cond[\"exprAlias\"].strip(\"c\")} = '\n", " _, _, point, name = cond['taskId'].split(ID_DELIM)\n", " cond_id = TaskID.get(name, point)\n", " prereqs.append([\n", " False,\n", " prefix,\n", " f'{cond_id} {cond[\"reqState\"]}',\n", " cond['satisfied']\n", " ])\n", " if options.list_prereqs:\n", " for composite, _, msg, _ in prereqs:\n", " if not composite:\n", " print(msg)\n", " else:\n", " flat_meta = flatten_data(t_proxy['task']['meta'])\n", " for key, value in sorted(flat_meta.items(), reverse=True):\n", " ansiprint(\n", " f'<bold>{key}:</bold>'\n", " f' {value or \"<m>(not given)</m>\"}')\n", " ansiprint(\n", " '\\n<bold>prerequisites</bold>'\n", " ' (<red>- => not satisfied</red>):')\n", " if not prereqs:\n", " print(' (None)')\n", " for _, prefix, msg, state in prereqs:\n", " print_msg_state(f'{prefix}{msg}', state)\n", "\n", " ansiprint(\n", " '\\n<bold>outputs</bold>'\n", " ' (<red>- => not completed</red>):')\n", " if not t_proxy['outputs']:\n", " print(' (None)')\n", " for key, val in t_proxy['outputs'].items():\n", " print_msg_state(f'{task_id} {key}', val)\n", " if t_proxy['extras']:\n", " print('\\nother:')\n", " for key, value in t_proxy['extras'].items():\n", " print(' o %s ... %s' % (key, value))\n", " if not results['taskProxies']:\n", " ansiprint(\n", " f\"<red>No matching tasks found: {task_ids}\",\n", " file=sys.stderr)\n", " sys.exit(1)\n", "\n", " if options.json:\n", " print(json.dumps(json_filter, indent=4))\n", "\n", "\n", "if __name__ == \"__main__\":\n", " main()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
276
0
#!/usr/bin/env python # # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. import logging import os import re import subprocess # Matches the output lines from the 'ldd' tool. For example: # libcrypto.so.10 => /path/to/usr/lib64/libcrypto.so.10 (0x00007fb0cb0a5000) # # Note: The following pattern will not match the following two types of # dependencies and so they will not be included in the output from this module: # # 1. The dynamic linker: # /lib64/ld-linux-x86-64.so.2 (0x00007f6f7ab79000) # 2. Linux virtual dynamic shared objects: # linux-vdso.so.1 (0x00007ffc06cfb000) # LDD_RE = re.compile(r'^\s+.+? => (\S+) \(0x.+\)') class DependencyExtractor(object): """ This class extracts native library dependencies from the given executable. """ def __init__(self): self.deps_cache = {} self.lib_allowed_filter = lambda path: True self.enable_expand_symlinks = False def set_library_filter(self, lib_allowed_filter): """ Specify a filter predicate that should return True iff the specified library path should be included in the result from extract_deps(). By default, all libraries are included in the result. """ self.lib_allowed_filter = lib_allowed_filter def set_expand_symlinks(self, expand): """ Specify whether symlinks should be expanded in the output from extract_deps(). By default, symlinks are not expanded. See expand_symlinks(). """ self.enable_expand_symlinks = expand def expand_symlinks(self, deps): """ ldd will often point to symlinks. Return a list including any symlink in the specified dependency list as well as whatever it's pointing to, recursively. """ expanded = [] for path in deps: expanded.append(path) while os.path.islink(path): # TODO(mpercy): os.readlink() can return an absolute path. Should we more carefully handle # the path concatenation here? path = os.path.join(os.path.dirname(path), os.readlink(path)) expanded.append(path) return expanded def extract_deps(self, exe): """ Runs 'ldd' on the provided 'exe' path, returning a list of any libraries it depends on. Blacklisted libraries are removed from this list. If the provided 'exe' is not a binary executable, returns an empty list. """ if (exe.endswith(".jar") or exe.endswith(".pl") or exe.endswith(".py") or exe.endswith(".sh") or exe.endswith(".txt") or os.path.isdir(exe)): return [] if exe not in self.deps_cache: p = subprocess.Popen(["ldd", exe], stdout=subprocess.PIPE) out, err = p.communicate() self.deps_cache[exe] = (out, err, p.returncode) out, err, rc = self.deps_cache[exe] if rc != 0: logging.warning("failed to run ldd on %s", exe) return [] deps = [] for line in out.splitlines(): match = LDD_RE.match(line) if not match: continue dep = match.group(1) # Apply the provided predicate. if not self.lib_allowed_filter(dep): continue deps.append(dep) if self.enable_expand_symlinks: deps = self.expand_symlinks(deps) return deps
[ "#!/usr/bin/env python\n", "#\n", "# Licensed to the Apache Software Foundation (ASF) under one\n", "# or more contributor license agreements. See the NOTICE file\n", "# distributed with this work for additional information\n", "# regarding copyright ownership. The ASF licenses this file\n", "# to you under the Apache License, Version 2.0 (the\n", "# \"License\"); you may not use this file except in compliance\n", "# with the License. You may obtain a copy of the License at\n", "#\n", "# http://www.apache.org/licenses/LICENSE-2.0\n", "#\n", "# Unless required by applicable law or agreed to in writing,\n", "# software distributed under the License is distributed on an\n", "# \"AS IS\" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY\n", "# KIND, either express or implied. See the License for the\n", "# specific language governing permissions and limitations\n", "# under the License.\n", "\n", "import logging\n", "import os\n", "import re\n", "import subprocess\n", "\n", "# Matches the output lines from the 'ldd' tool. For example:\n", "# libcrypto.so.10 => /path/to/usr/lib64/libcrypto.so.10 (0x00007fb0cb0a5000)\n", "#\n", "# Note: The following pattern will not match the following two types of\n", "# dependencies and so they will not be included in the output from this module:\n", "#\n", "# 1. The dynamic linker:\n", "# /lib64/ld-linux-x86-64.so.2 (0x00007f6f7ab79000)\n", "# 2. Linux virtual dynamic shared objects:\n", "# linux-vdso.so.1 (0x00007ffc06cfb000)\n", "#\n", "LDD_RE = re.compile(r'^\\s+.+? => (\\S+) \\(0x.+\\)')\n", "\n", "class DependencyExtractor(object):\n", " \"\"\"\n", " This class extracts native library dependencies from the given executable.\n", " \"\"\"\n", " def __init__(self):\n", " self.deps_cache = {}\n", " self.lib_allowed_filter = lambda path: True\n", " self.enable_expand_symlinks = False\n", "\n", " def set_library_filter(self, lib_allowed_filter):\n", " \"\"\"\n", " Specify a filter predicate that should return True iff the specified\n", " library path should be included in the result from extract_deps().\n", " By default, all libraries are included in the result.\n", " \"\"\"\n", " self.lib_allowed_filter = lib_allowed_filter\n", "\n", " def set_expand_symlinks(self, expand):\n", " \"\"\"\n", " Specify whether symlinks should be expanded in the output from\n", " extract_deps(). By default, symlinks are not expanded. See\n", " expand_symlinks().\n", " \"\"\"\n", " self.enable_expand_symlinks = expand\n", "\n", " def expand_symlinks(self, deps):\n", " \"\"\"\n", " ldd will often point to symlinks. Return a list including any symlink in\n", " the specified dependency list as well as whatever it's pointing to,\n", " recursively.\n", " \"\"\"\n", " expanded = []\n", " for path in deps:\n", " expanded.append(path)\n", " while os.path.islink(path):\n", " # TODO(mpercy): os.readlink() can return an absolute path. Should we more carefully handle\n", " # the path concatenation here?\n", " path = os.path.join(os.path.dirname(path), os.readlink(path))\n", " expanded.append(path)\n", " return expanded\n", "\n", " def extract_deps(self, exe):\n", " \"\"\"\n", " Runs 'ldd' on the provided 'exe' path, returning a list of\n", " any libraries it depends on. Blacklisted libraries are\n", " removed from this list.\n", "\n", " If the provided 'exe' is not a binary executable, returns\n", " an empty list.\n", " \"\"\"\n", " if (exe.endswith(\".jar\") or\n", " exe.endswith(\".pl\") or\n", " exe.endswith(\".py\") or\n", " exe.endswith(\".sh\") or\n", " exe.endswith(\".txt\") or\n", " os.path.isdir(exe)):\n", " return []\n", "\n", " if exe not in self.deps_cache:\n", " p = subprocess.Popen([\"ldd\", exe], stdout=subprocess.PIPE)\n", " out, err = p.communicate()\n", " self.deps_cache[exe] = (out, err, p.returncode)\n", "\n", " out, err, rc = self.deps_cache[exe]\n", " if rc != 0:\n", " logging.warning(\"failed to run ldd on %s\", exe)\n", " return []\n", "\n", " deps = []\n", " for line in out.splitlines():\n", " match = LDD_RE.match(line)\n", " if not match:\n", " continue\n", " dep = match.group(1)\n", " # Apply the provided predicate.\n", " if not self.lib_allowed_filter(dep):\n", " continue\n", " deps.append(dep)\n", "\n", " if self.enable_expand_symlinks:\n", " deps = self.expand_symlinks(deps)\n", " return deps\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.02857142857142857, 0.16666666666666666, 0, 0, 0.045454545454545456, 0, 0, 0, 0, 0.019230769230769232, 0, 0, 0, 0, 0, 0, 0, 0.024390243902439025, 0, 0, 0, 0, 0, 0, 0, 0.02857142857142857, 0, 0, 0, 0, 0, 0, 0, 0.03571428571428571, 0.029411764705882353, 0.010101010101010102, 0, 0, 0, 0, 0, 0.03225806451612903, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.034482758620689655, 0.0625, 0, 0, 0.015384615384615385, 0.030303030303030304, 0.018518518518518517, 0, 0, 0, 0.018518518518518517, 0.0625, 0, 0, 0, 0.030303030303030304, 0.05, 0, 0.037037037037037035, 0.02631578947368421, 0.023255813953488372, 0, 0.043478260869565216, 0, 0, 0.025, 0 ]
119
0.007546
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import absolute_import, division, print_function __metaclass__ = type # Copyright 2019 Palo Alto Networks, Inc # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ANSIBLE_METADATA = {'metadata_version': '1.1', 'status': ['preview'], 'supported_by': 'community'} DOCUMENTATION = ''' --- module: panos_log_forwarding_profile short_description: Manage log forwarding profiles. description: - Manages log forwarding profiles. author: "Garfield Lee Freeman (@shinmog)" version_added: "2.8" requirements: - pan-python - pandevice >= 0.11.1 notes: - Panorama is supported. - Check mode is supported. extends_documentation_fragment: - panos.transitional_provider - panos.vsys_shared - panos.device_group options: name: description: - Name of the profile. required: true description: description: - Profile description enhanced_logging: description: - Valid for PAN-OS 8.1+ - Enabling enhanced application logging. type: 'bool' ''' EXAMPLES = ''' # Create a profile - name: Create log forwarding profile panos_log_forwarding_profile: provider: '{{ provider }}' name: 'my-profile' enhanced_logging: true ''' RETURN = ''' # Default return values ''' from ansible.module_utils.basic import AnsibleModule from ansible.module_utils.network.panos.panos import get_connection try: from pandevice.objects import LogForwardingProfile from pandevice.errors import PanDeviceError except ImportError: pass def main(): helper = get_connection( vsys_shared=True, device_group=True, with_state=True, with_classic_provider_spec=True, min_pandevice_version=(0, 11, 1), min_panos_version=(8, 0, 0), argument_spec=dict( name=dict(required=True), description=dict(), enhanced_logging=dict(type='bool'), ), ) module = AnsibleModule( argument_spec=helper.argument_spec, supports_check_mode=True, required_one_of=helper.required_one_of, ) # Verify imports, build pandevice object tree. parent = helper.get_pandevice_parent(module) try: listing = LogForwardingProfile.refreshall(parent) except PanDeviceError as e: module.fail_json(msg='Failed refresh: {0}'.format(e)) spec = { 'name': module.params['name'], 'description': module.params['description'], 'enhanced_logging': module.params['enhanced_logging'], } obj = LogForwardingProfile(**spec) parent.add(obj) changed = helper.apply_state(obj, listing, module) module.exit_json(changed=changed, msg='Done') if __name__ == '__main__': main()
[ "#!/usr/bin/env python\n", "# -*- coding: utf-8 -*-\n", "\n", "from __future__ import absolute_import, division, print_function\n", "__metaclass__ = type\n", "\n", "# Copyright 2019 Palo Alto Networks, Inc\n", "#\n", "# Licensed under the Apache License, Version 2.0 (the \"License\");\n", "# you may not use this file except in compliance with the License.\n", "# You may obtain a copy of the License at\n", "#\n", "# http://www.apache.org/licenses/LICENSE-2.0\n", "#\n", "# Unless required by applicable law or agreed to in writing, software\n", "# distributed under the License is distributed on an \"AS IS\" BASIS,\n", "# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n", "# See the License for the specific language governing permissions and\n", "# limitations under the License.\n", "\n", "ANSIBLE_METADATA = {'metadata_version': '1.1',\n", " 'status': ['preview'],\n", " 'supported_by': 'community'}\n", "\n", "DOCUMENTATION = '''\n", "---\n", "module: panos_log_forwarding_profile\n", "short_description: Manage log forwarding profiles.\n", "description:\n", " - Manages log forwarding profiles.\n", "author: \"Garfield Lee Freeman (@shinmog)\"\n", "version_added: \"2.8\"\n", "requirements:\n", " - pan-python\n", " - pandevice >= 0.11.1\n", "notes:\n", " - Panorama is supported.\n", " - Check mode is supported.\n", "extends_documentation_fragment:\n", " - panos.transitional_provider\n", " - panos.vsys_shared\n", " - panos.device_group\n", "options:\n", " name:\n", " description:\n", " - Name of the profile.\n", " required: true\n", " description:\n", " description:\n", " - Profile description\n", " enhanced_logging:\n", " description:\n", " - Valid for PAN-OS 8.1+\n", " - Enabling enhanced application logging.\n", " type: 'bool'\n", "'''\n", "\n", "EXAMPLES = '''\n", "# Create a profile\n", "- name: Create log forwarding profile\n", " panos_log_forwarding_profile:\n", " provider: '{{ provider }}'\n", " name: 'my-profile'\n", " enhanced_logging: true\n", "'''\n", "\n", "RETURN = '''\n", "# Default return values\n", "'''\n", "\n", "from ansible.module_utils.basic import AnsibleModule\n", "from ansible.module_utils.network.panos.panos import get_connection\n", "\n", "\n", "try:\n", " from pandevice.objects import LogForwardingProfile\n", " from pandevice.errors import PanDeviceError\n", "except ImportError:\n", " pass\n", "\n", "\n", "def main():\n", " helper = get_connection(\n", " vsys_shared=True,\n", " device_group=True,\n", " with_state=True,\n", " with_classic_provider_spec=True,\n", " min_pandevice_version=(0, 11, 1),\n", " min_panos_version=(8, 0, 0),\n", " argument_spec=dict(\n", " name=dict(required=True),\n", " description=dict(),\n", " enhanced_logging=dict(type='bool'),\n", " ),\n", " )\n", " module = AnsibleModule(\n", " argument_spec=helper.argument_spec,\n", " supports_check_mode=True,\n", " required_one_of=helper.required_one_of,\n", " )\n", "\n", " # Verify imports, build pandevice object tree.\n", " parent = helper.get_pandevice_parent(module)\n", "\n", " try:\n", " listing = LogForwardingProfile.refreshall(parent)\n", " except PanDeviceError as e:\n", " module.fail_json(msg='Failed refresh: {0}'.format(e))\n", "\n", " spec = {\n", " 'name': module.params['name'],\n", " 'description': module.params['description'],\n", " 'enhanced_logging': module.params['enhanced_logging'],\n", " }\n", " obj = LogForwardingProfile(**spec)\n", " parent.add(obj)\n", "\n", " changed = helper.apply_state(obj, listing, module)\n", " module.exit_json(changed=changed, msg='Done')\n", "\n", "\n", "if __name__ == '__main__':\n", " main()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.018867924528301886, 0.014705882352941176, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
123
0.000273
# Copyright 2015, Yahoo Inc. # Licensed under the terms of the Apache License, Version 2.0. See the LICENSE file associated with the project for terms. from pyspark.context import SparkContext import numpy as np import base64 import cPickle as pkl from tempfile import NamedTemporaryFile import os import subprocess from operator import add def default_data_loading(sc, data_path, sampling_ratio, seed): """ This function loads training data from a text file, sampling it by the provided ratio and random seed, and interprets each line as a tab-separated (id, data) pair where 'data' is assumed to be a base64-encoded pickled numpy array. The ids are discarded. The data is returned as an RDD of numpy arrays. """ # Compute the number of cores in our cluster - used below to heuristically set the number of partitions # TypeError: int() argument must be a string or a number, not 'NoneType' ? print sc._conf.toDebugString() #print sc._conf.getAll() try: nb_instances = int(sc._conf.get('spark.executor.instances')) except Exception as inst: print "[default_data_loading: error] {}. Setting nb_instances to 2.".format(inst) nb_instances = 2 try: nb_executor_cores = int(sc._conf.get('spark.executor.cores')) except Exception as inst: print "[default_data_loading: error] {}. Setting nb_executor_cores to 2.".format(inst) nb_executor_cores = 2 total_cores = nb_instances * nb_executor_cores # Load and sample down the dataset d = sc.textFile(data_path, total_cores * 3).sample(False, sampling_ratio, seed) # The data is (id, vector) tab-delimited pairs where each vector is # a base64-encoded pickled numpy array deserialize_vec = lambda s: pkl.loads(base64.decodestring(s.split('\t')[1])) vecs = d.map(deserialize_vec) return vecs def main(sc, args, data_load_fn=default_data_loading): def seqOp(a, b): a += np.outer(b, b) return a def combOp(a, b): a += b return a # Load data d = data_load_fn(sc, args.data, args.sampling_ratio, args.seed) d.cache() # Determine the data dimension D = len(d.first()) print "d.first: {}, D: {}".format(d.first(),D) # Count data points count = d.count() mu = d.aggregate(np.zeros(D), add, add) mu = mu / float(count) # Compute covariance estimator summed_covar = d.treeAggregate(np.zeros((D, D)), seqOp, combOp, depth=args.agg_depth) A = summed_covar / (count - 1) - np.outer(mu, mu) E, P = np.linalg.eigh(A) params = { 'mu': mu, # mean 'P': P, # PCA matrix 'E': E, # eigenvalues 'A': A, # covariance matrix 'c': count # sample size } save_hdfs_pickle(params, args.output) def save_hdfs_pickle(m, pkl_path): """ Given a python object and a path on hdfs, save the object as a pickle file locally and copy the file to the hdfs path. """ print 'Saving pickle to temp file...' f = NamedTemporaryFile(delete=False) pkl.dump(m, f, -1) f.close() print 'Copying pickle file to hdfs...' copy_to_hdfs(f, pkl_path) os.remove(f.name) def copy_to_hdfs(f, hdfs_path): subprocess.call(['hadoop', 'fs', '-copyFromLocal', f.name, hdfs_path]) if __name__ == '__main__': from argparse import ArgumentParser parser = ArgumentParser() # Data handling parameters parser.add_argument('--data', dest='data', type=str, required=True, help='hdfs path to input data') parser.add_argument('--data_udf', dest='data_udf', type=str, default=None, help='module name from which to load a data loading UDF') parser.add_argument('--seed', dest='seed', type=int, default=None, help='optional random seed') parser.add_argument('--sampling_ratio', dest='sampling_ratio', type=float, default=1.0, help='proportion of data to sample for training') parser.add_argument('--agg_depth', dest='agg_depth', type=int, default=4, help='depth of tree aggregation to compute covariance estimator') parser.add_argument('--output', dest='output', type=str, default=None, help='hdfs path to output pickle file of parameters') args = parser.parse_args() sc = SparkContext(appName='PCA') # Load UDF module if provided if args.data_udf: sc.addPyFile('hdfs://memex/user/skaraman/build-lopq-index/lopq/spark/memex_udf.py') sc.addPyFile('hdfs://memex/user/skaraman/build-lopq-index/lopq/spark/deepsentibanktf_udf.py') udf_module = __import__(args.data_udf, fromlist=['udf']) load_udf = udf_module.udf main(sc, args, data_load_fn=load_udf) else: main(sc, args) sc.stop()
[ "# Copyright 2015, Yahoo Inc.\n", "# Licensed under the terms of the Apache License, Version 2.0. See the LICENSE file associated with the project for terms.\n", "from pyspark.context import SparkContext\n", "\n", "import numpy as np\n", "import base64\n", "import cPickle as pkl\n", "from tempfile import NamedTemporaryFile\n", "import os\n", "import subprocess\n", "from operator import add\n", "\n", "\n", "def default_data_loading(sc, data_path, sampling_ratio, seed):\n", " \"\"\"\n", " This function loads training data from a text file, sampling it by the provided\n", " ratio and random seed, and interprets each line as a tab-separated (id, data) pair\n", " where 'data' is assumed to be a base64-encoded pickled numpy array. The ids are discarded.\n", " The data is returned as an RDD of numpy arrays.\n", " \"\"\"\n", " # Compute the number of cores in our cluster - used below to heuristically set the number of partitions\n", " # TypeError: int() argument must be a string or a number, not 'NoneType' ?\n", " print sc._conf.toDebugString()\n", " #print sc._conf.getAll()\n", " try:\n", " nb_instances = int(sc._conf.get('spark.executor.instances'))\n", " except Exception as inst:\n", " print \"[default_data_loading: error] {}. Setting nb_instances to 2.\".format(inst)\n", " nb_instances = 2\n", " try:\n", " nb_executor_cores = int(sc._conf.get('spark.executor.cores'))\n", " except Exception as inst:\n", " print \"[default_data_loading: error] {}. Setting nb_executor_cores to 2.\".format(inst)\n", " nb_executor_cores = 2\n", "\n", "\n", " total_cores = nb_instances * nb_executor_cores\n", "\n", " # Load and sample down the dataset\n", " d = sc.textFile(data_path, total_cores * 3).sample(False, sampling_ratio, seed)\n", "\n", " # The data is (id, vector) tab-delimited pairs where each vector is\n", " # a base64-encoded pickled numpy array\n", " deserialize_vec = lambda s: pkl.loads(base64.decodestring(s.split('\\t')[1]))\n", " vecs = d.map(deserialize_vec)\n", "\n", " return vecs\n", "\n", "\n", "def main(sc, args, data_load_fn=default_data_loading):\n", "\n", " def seqOp(a, b):\n", " a += np.outer(b, b)\n", " return a\n", "\n", " def combOp(a, b):\n", " a += b\n", " return a\n", "\n", " # Load data\n", " d = data_load_fn(sc, args.data, args.sampling_ratio, args.seed)\n", " d.cache()\n", "\n", " # Determine the data dimension\n", " D = len(d.first())\n", " print \"d.first: {}, D: {}\".format(d.first(),D)\n", "\n", " # Count data points\n", " count = d.count()\n", " mu = d.aggregate(np.zeros(D), add, add)\n", " mu = mu / float(count)\n", "\n", " # Compute covariance estimator\n", " summed_covar = d.treeAggregate(np.zeros((D, D)), seqOp, combOp, depth=args.agg_depth)\n", "\n", " A = summed_covar / (count - 1) - np.outer(mu, mu)\n", " E, P = np.linalg.eigh(A)\n", "\n", " params = {\n", " 'mu': mu, # mean\n", " 'P': P, # PCA matrix\n", " 'E': E, # eigenvalues\n", " 'A': A, # covariance matrix\n", " 'c': count # sample size\n", " }\n", "\n", " save_hdfs_pickle(params, args.output)\n", "\n", "\n", "def save_hdfs_pickle(m, pkl_path):\n", " \"\"\"\n", " Given a python object and a path on hdfs, save the object as a pickle file locally and copy the file\n", " to the hdfs path.\n", " \"\"\"\n", " print 'Saving pickle to temp file...'\n", " f = NamedTemporaryFile(delete=False)\n", " pkl.dump(m, f, -1)\n", " f.close()\n", "\n", " print 'Copying pickle file to hdfs...'\n", " copy_to_hdfs(f, pkl_path)\n", " os.remove(f.name)\n", "\n", "\n", "def copy_to_hdfs(f, hdfs_path):\n", " subprocess.call(['hadoop', 'fs', '-copyFromLocal', f.name, hdfs_path])\n", "\n", "\n", "if __name__ == '__main__':\n", " from argparse import ArgumentParser\n", " parser = ArgumentParser()\n", "\n", " # Data handling parameters\n", " parser.add_argument('--data', dest='data', type=str, required=True, help='hdfs path to input data')\n", " parser.add_argument('--data_udf', dest='data_udf', type=str, default=None, help='module name from which to load a data loading UDF')\n", " parser.add_argument('--seed', dest='seed', type=int, default=None, help='optional random seed')\n", " parser.add_argument('--sampling_ratio', dest='sampling_ratio', type=float, default=1.0, help='proportion of data to sample for training')\n", " parser.add_argument('--agg_depth', dest='agg_depth', type=int, default=4, help='depth of tree aggregation to compute covariance estimator')\n", "\n", " parser.add_argument('--output', dest='output', type=str, default=None, help='hdfs path to output pickle file of parameters')\n", "\n", " args = parser.parse_args()\n", "\n", " sc = SparkContext(appName='PCA')\n", "\n", " # Load UDF module if provided\n", " if args.data_udf:\n", " sc.addPyFile('hdfs://memex/user/skaraman/build-lopq-index/lopq/spark/memex_udf.py')\n", " sc.addPyFile('hdfs://memex/user/skaraman/build-lopq-index/lopq/spark/deepsentibanktf_udf.py')\n", " udf_module = __import__(args.data_udf, fromlist=['udf'])\n", " load_udf = udf_module.udf\n", " main(sc, args, data_load_fn=load_udf)\n", " else:\n", " main(sc, args)\n", "\n", " sc.stop()\n" ]
[ 0, 0.008130081300813009, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.011904761904761904, 0.011494252873563218, 0.010526315789473684, 0, 0, 0.009259259259259259, 0, 0, 0.034482758620689655, 0, 0.014285714285714285, 0, 0.011111111111111112, 0, 0, 0, 0, 0.010526315789473684, 0, 0, 0, 0.0196078431372549, 0, 0, 0.011904761904761904, 0, 0, 0, 0.024691358024691357, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.0196078431372549, 0, 0, 0, 0, 0, 0, 0, 0.011111111111111112, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.009523809523809525, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.009615384615384616, 0.0072992700729927005, 0.01, 0.007042253521126761, 0.006944444444444444, 0, 0.007751937984496124, 0, 0, 0, 0, 0, 0, 0, 0.010869565217391304, 0.00980392156862745, 0, 0, 0, 0, 0, 0, 0 ]
136
0.002114
import requests from requests.adapters import HTTPAdapter from requests.packages.urllib3.util.retry import Retry import logging import os from datetime import datetime try: import defusedxml.minidom as xml except ImportError: import xml.dom.minidom as xml class instagram: def __init__(self, cookie): """This sets up this class to communicate with Instagram. Args: cookie: A dictionary object with the required cookie values (ds_user_id, sessionid, csrftoken). """ self.userid = cookie["ds_user_id"] self.sessionid = cookie["sessionid"] self.csrftoken = cookie["csrftoken"] self.mid = cookie["mid"] self.headers = { "accept" : "*/*", "accept-encoding" : "gzip, deflate", "accept-language" : "en-US", "content_type" : "application/x-www-form-urlencoded; charset=UTF-8", "cache-control" : "no-cache", "cookie" : "ds_user_id=" + self.userid + "; sessionid=" + self.sessionid + "; csrftoken=" + self.csrftoken + "; mid=" + self.mid, "dnt" : "1", # "pragma" : "no-cache", # "referer" : "https://www.instagram.com/", "user-agent" : "Instagram 10.26.0 (iPhone7,2; iOS 10_1_1; en_US; en-US; scale=2.00; gamut=normal; 750x1334) AppleWebKit/420+", "x-ig-capabilities": "36oD", # "x-ig-connection-type" : "WIFI", # "x-ig-fb-http-engine" : "Liger" } self.session = requests.Session() max_tries = 3 backoff_factor = 0.2 status_forcelist = (500, 502, 503, 504) retry = Retry(total=max_tries, read=max_tries, connect=max_tries, backoff_factor=backoff_factor, status_forcelist=status_forcelist) adapter = HTTPAdapter(max_retries=retry) self.session.mount('http://', adapter) self.session.mount('https://', adapter) self.session.headers = self.headers def getReelTray(self): """Get reel tray from API. Returns: Response object with reel tray API response """ endpoint = "https://i.instagram.com/api/v1/feed/reels_tray/" response = self.session.get(endpoint, timeout=60) if response.status_code != requests.codes.ok: logging.error("Status Code Error." + str(response.status_code)) response.raise_for_status() return response def getReelMedia(self, user): """Get reel media of a user from API. Args: user: User ID Returns: Response object with reel media API response """ endpoint = "https://i.instagram.com/api/v1/feed/user/" + str(user) + "/reel_media/" response = self.session.get(endpoint, timeout=60) if response.status_code != requests.codes.ok: logging.error("Status Code Error." + str(response.status_code)) response.raise_for_status() return response def getStories(self): return self.getReelTray() def getUserStories(self, user): return self.getReelMedia(user) def getUserIDs(self, json: dict) -> list: """Extract user IDs from reel tray JSON. Args: json: Reel tray response from IG Returns: List of user IDs """ users = [] for user in json['tray']: users.append(user['user']['pk']) return users def getFile(self, url: str, dest: str): """Download file and save to destination Args: url: URL of item to download dest: File system destination to save item to Returns: None """ logging.debug("URL: %s", url) logging.debug("Dest: %s", dest) try: if os.path.getsize(dest) == 0: logging.info("Empty file exists. Removing.") os.remove(dest) except FileNotFoundError: pass try: dirpath = os.path.dirname(dest) os.makedirs(dirpath, exist_ok=True) with open(dest, "xb") as handle: response = self.session.get(url, stream=True, timeout=60) if response.status_code != requests.codes.ok: logging.error("Status Code Error." + str(response.status_code)) response.raise_for_status() for data in response.iter_content(chunk_size=4194304): handle.write(data) handle.close() except FileExistsError: logging.info("File already exists.") if os.path.getsize(dest) == 0: logging.info("Error downloading. Removing.") os.remove(dest) def formatPath(self, user: str, pk: int, timestamp: int, postid: str, mediatype: int) -> str: """Format download path to a specific format/template Args: user: User name pk: User ID timestamp: UTC Unix timestamp postid: Post ID mediatype: Media type as defined by IG Returns: None """ dirpath = os.path.dirname(__file__) utcdatetime = datetime.utcfromtimestamp(timestamp).strftime("%Y-%m-%d-%H-%M-%S") if mediatype == 1: ext = ".jpg" type = "stories" elif mediatype == 2: ext = ".mp4" type = "stories" elif mediatype == 3: ext = ".mp4" type = "livestories" else: ext = "" type = "other" path = os.path.join(dirpath, "downloads", user + "_" + str(pk), type, utcdatetime + "_" + str(timestamp) + "_" + postid + ext) return path def downloadReel(self, resp): """Download stories of a followed user's tray. Download the stories of a followed user. Args: resp: JSON dictionary of reel from IG API Returns: None """ try: for index, item in enumerate(resp['items']): logging.debug(' ' + str(index)) username = item['user']['username'] userpk = item['user']['pk'] timestamp = item['taken_at'] postid = item['id'] mediatype = item['media_type'] if mediatype == 2: # Video largest = 0 for versionindex, video in enumerate(item['video_versions']): itemsize = video['width'] * video['height'] largestsize = item['video_versions'][largest]['width'] * \ item['video_versions'][largest]['height'] if itemsize > largestsize: largest = versionindex logging.debug(' V' + str(largest)) url = item['video_versions'][largest]['url'] logging.debug(' ' + url) elif mediatype == 1: # Image largest = 0 for versionindex, image in enumerate(item['image_versions2']['candidates']): itemsize = image['width'] * image['height'] largestsize = item['image_versions2']['candidates'][largest]['width'] * \ item['image_versions2']['candidates'][largest]['height'] if itemsize > largestsize: largest = versionindex logging.debug(' I' + str(largest)) url = item['image_versions2']['candidates'][largest]['url'] logging.debug(' ' + url) else: # Unknown logging.debug(' E') url = None pass path = self.formatPath(username, userpk, timestamp, postid, mediatype) self.getFile(url, path) except KeyError: # JSON 'item' key does not exist for later items in tray as of 6/2/2017 pass def downloadTray(self, resp): """Download stories of logged in user's tray. Download the stories as available in the tray. The tray contains a list of reels, a collection of the stories posted by a followed user. The tray only contains a small set of reels of the first few users. To download the rest, a reel must be obtained for each user in the tray. Args: resp: JSON dictionary of tray from IG API Returns: None """ for reel in resp['tray']: self.downloadReel(reel) def downloadStoryLive(self, resp): """Download post-live stories of a followed user's tray. Download the post-live stories of a followed user. Args: resp: JSON dictionary of reel from IG API Returns: None """ try: for index,item in enumerate(resp["post_live"]["post_live_items"]): logging.debug(' ' + str(index)) username = item["user"]["username"] userpk = item["user"]["pk"] for bindex,broadcast in enumerate(item["broadcasts"]): logging.debug(' ' + str(bindex)) timestamp = broadcast["published_time"] postid = broadcast["media_id"] dash = broadcast["dash_manifest"] dashxml = xml.parseString(dash) elements = dashxml.getElementsByTagName("BaseURL") for eindex,element in enumerate(elements): for node in element.childNodes: if node.nodeType == node.TEXT_NODE: url = node.data mediatype = 3 path = self.formatPath(username, userpk, timestamp, postid + "_" + str(eindex), mediatype) self.getFile(url, path) except KeyError: # No "post_live" key logging.debug(' ' + 'No live stories.') def close(self): """Close seesion to IG Returns: None """ self.session.close()
[ "import requests\n", "from requests.adapters import HTTPAdapter\n", "from requests.packages.urllib3.util.retry import Retry\n", "import logging\n", "import os\n", "from datetime import datetime\n", "try:\n", " import defusedxml.minidom as xml\n", "except ImportError:\n", " import xml.dom.minidom as xml\n", "\n", "class instagram:\n", " def __init__(self, cookie):\n", " \"\"\"This sets up this class to communicate with Instagram.\n", "\n", " Args:\n", " cookie: A dictionary object with the required cookie values (ds_user_id, sessionid, csrftoken).\n", " \"\"\"\n", " self.userid = cookie[\"ds_user_id\"]\n", " self.sessionid = cookie[\"sessionid\"]\n", " self.csrftoken = cookie[\"csrftoken\"]\n", " self.mid = cookie[\"mid\"]\n", " self.headers = {\n", " \"accept\" : \"*/*\",\n", " \"accept-encoding\" : \"gzip, deflate\",\n", " \"accept-language\" : \"en-US\",\n", " \"content_type\" : \"application/x-www-form-urlencoded; charset=UTF-8\",\n", " \"cache-control\" : \"no-cache\",\n", " \"cookie\" : \"ds_user_id=\" + self.userid + \"; sessionid=\" + self.sessionid + \"; csrftoken=\" + self.csrftoken + \"; mid=\" + self.mid,\n", " \"dnt\" : \"1\",\n", " # \"pragma\" : \"no-cache\",\n", " # \"referer\" : \"https://www.instagram.com/\",\n", " \"user-agent\" : \"Instagram 10.26.0 (iPhone7,2; iOS 10_1_1; en_US; en-US; scale=2.00; gamut=normal; 750x1334) AppleWebKit/420+\",\n", " \"x-ig-capabilities\": \"36oD\",\n", " # \"x-ig-connection-type\" : \"WIFI\",\n", " # \"x-ig-fb-http-engine\" : \"Liger\"\n", " }\n", " self.session = requests.Session()\n", " max_tries = 3\n", " backoff_factor = 0.2\n", " status_forcelist = (500, 502, 503, 504)\n", " retry = Retry(total=max_tries, read=max_tries, connect=max_tries, backoff_factor=backoff_factor, status_forcelist=status_forcelist)\n", " adapter = HTTPAdapter(max_retries=retry)\n", " self.session.mount('http://', adapter)\n", " self.session.mount('https://', adapter)\n", " self.session.headers = self.headers\n", "\n", " def getReelTray(self):\n", " \"\"\"Get reel tray from API.\n", "\n", " Returns:\n", " Response object with reel tray API response\n", " \"\"\"\n", " endpoint = \"https://i.instagram.com/api/v1/feed/reels_tray/\"\n", " response = self.session.get(endpoint, timeout=60)\n", " if response.status_code != requests.codes.ok:\n", " logging.error(\"Status Code Error.\" + str(response.status_code))\n", " response.raise_for_status()\n", " return response\n", "\n", " def getReelMedia(self, user):\n", " \"\"\"Get reel media of a user from API.\n", "\n", " Args:\n", " user: User ID\n", "\n", " Returns:\n", " Response object with reel media API response\n", " \"\"\"\n", " endpoint = \"https://i.instagram.com/api/v1/feed/user/\" + str(user) + \"/reel_media/\"\n", " response = self.session.get(endpoint, timeout=60)\n", " if response.status_code != requests.codes.ok:\n", " logging.error(\"Status Code Error.\" + str(response.status_code))\n", " response.raise_for_status()\n", " return response\n", "\n", " def getStories(self):\n", " return self.getReelTray()\n", "\n", " def getUserStories(self, user):\n", " return self.getReelMedia(user)\n", "\n", " def getUserIDs(self, json: dict) -> list:\n", " \"\"\"Extract user IDs from reel tray JSON.\n", "\n", " Args:\n", " json: Reel tray response from IG\n", "\n", " Returns:\n", " List of user IDs\n", " \"\"\"\n", " users = []\n", " for user in json['tray']:\n", " users.append(user['user']['pk'])\n", " return users\n", "\n", " def getFile(self, url: str, dest: str):\n", " \"\"\"Download file and save to destination\n", "\n", " Args:\n", " url: URL of item to download\n", " dest: File system destination to save item to\n", "\n", " Returns:\n", " None\n", " \"\"\"\n", " logging.debug(\"URL: %s\", url)\n", " logging.debug(\"Dest: %s\", dest)\n", " try:\n", " if os.path.getsize(dest) == 0:\n", " logging.info(\"Empty file exists. Removing.\")\n", " os.remove(dest)\n", " except FileNotFoundError:\n", " pass\n", "\n", " try:\n", " dirpath = os.path.dirname(dest)\n", " os.makedirs(dirpath, exist_ok=True)\n", " with open(dest, \"xb\") as handle:\n", " response = self.session.get(url, stream=True, timeout=60)\n", " if response.status_code != requests.codes.ok:\n", " logging.error(\"Status Code Error.\" + str(response.status_code))\n", " response.raise_for_status()\n", " for data in response.iter_content(chunk_size=4194304):\n", " handle.write(data)\n", " handle.close()\n", " except FileExistsError:\n", " logging.info(\"File already exists.\")\n", "\n", " if os.path.getsize(dest) == 0:\n", " logging.info(\"Error downloading. Removing.\")\n", " os.remove(dest)\n", "\n", " def formatPath(self, user: str, pk: int, timestamp: int, postid: str, mediatype: int) -> str:\n", " \"\"\"Format download path to a specific format/template\n", "\n", " Args:\n", " user: User name\n", " pk: User ID\n", " timestamp: UTC Unix timestamp\n", " postid: Post ID\n", " mediatype: Media type as defined by IG\n", "\n", " Returns:\n", " None\n", " \"\"\"\n", " dirpath = os.path.dirname(__file__)\n", " utcdatetime = datetime.utcfromtimestamp(timestamp).strftime(\"%Y-%m-%d-%H-%M-%S\")\n", " if mediatype == 1:\n", " ext = \".jpg\"\n", " type = \"stories\"\n", " elif mediatype == 2:\n", " ext = \".mp4\"\n", " type = \"stories\"\n", " elif mediatype == 3:\n", " ext = \".mp4\"\n", " type = \"livestories\"\n", " else:\n", " ext = \"\"\n", " type = \"other\"\n", " path = os.path.join(dirpath, \"downloads\", user + \"_\" + str(pk), type, utcdatetime + \"_\" + str(timestamp) + \"_\" + postid + ext)\n", " return path\n", "\n", " def downloadReel(self, resp):\n", " \"\"\"Download stories of a followed user's tray.\n", "\n", " Download the stories of a followed user.\n", "\n", " Args:\n", " resp: JSON dictionary of reel from IG API\n", "\n", " Returns:\n", " None\n", " \"\"\"\n", " try:\n", " for index, item in enumerate(resp['items']):\n", " logging.debug(' ' + str(index))\n", " username = item['user']['username']\n", " userpk = item['user']['pk']\n", " timestamp = item['taken_at']\n", " postid = item['id']\n", " mediatype = item['media_type']\n", " if mediatype == 2: # Video\n", " largest = 0\n", " for versionindex, video in enumerate(item['video_versions']):\n", " itemsize = video['width'] * video['height']\n", " largestsize = item['video_versions'][largest]['width'] * \\\n", " item['video_versions'][largest]['height']\n", " if itemsize > largestsize:\n", " largest = versionindex\n", " logging.debug(' V' + str(largest))\n", " url = item['video_versions'][largest]['url']\n", " logging.debug(' ' + url)\n", " elif mediatype == 1: # Image\n", " largest = 0\n", " for versionindex, image in enumerate(item['image_versions2']['candidates']):\n", " itemsize = image['width'] * image['height']\n", " largestsize = item['image_versions2']['candidates'][largest]['width'] * \\\n", " item['image_versions2']['candidates'][largest]['height']\n", " if itemsize > largestsize:\n", " largest = versionindex\n", " logging.debug(' I' + str(largest))\n", " url = item['image_versions2']['candidates'][largest]['url']\n", " logging.debug(' ' + url)\n", " else: # Unknown\n", " logging.debug(' E')\n", " url = None\n", " pass\n", "\n", " path = self.formatPath(username, userpk, timestamp, postid, mediatype)\n", " self.getFile(url, path)\n", " except KeyError: # JSON 'item' key does not exist for later items in tray as of 6/2/2017\n", " pass\n", "\n", " def downloadTray(self, resp):\n", " \"\"\"Download stories of logged in user's tray.\n", "\n", " Download the stories as available in the tray. The tray contains a list of\n", " reels, a collection of the stories posted by a followed user.\n", "\n", " The tray only contains a small set of reels of the first few users. To download\n", " the rest, a reel must be obtained for each user in the tray.\n", "\n", " Args:\n", " resp: JSON dictionary of tray from IG API\n", "\n", " Returns:\n", " None\n", " \"\"\"\n", " for reel in resp['tray']:\n", " self.downloadReel(reel)\n", "\n", " def downloadStoryLive(self, resp):\n", " \"\"\"Download post-live stories of a followed user's tray.\n", "\n", " Download the post-live stories of a followed user.\n", "\n", " Args:\n", " resp: JSON dictionary of reel from IG API\n", "\n", " Returns:\n", " None\n", " \"\"\"\n", " try:\n", " for index,item in enumerate(resp[\"post_live\"][\"post_live_items\"]):\n", " logging.debug(' ' + str(index))\n", " username = item[\"user\"][\"username\"]\n", " userpk = item[\"user\"][\"pk\"]\n", " for bindex,broadcast in enumerate(item[\"broadcasts\"]):\n", " logging.debug(' ' + str(bindex))\n", " timestamp = broadcast[\"published_time\"]\n", " postid = broadcast[\"media_id\"]\n", " dash = broadcast[\"dash_manifest\"]\n", " dashxml = xml.parseString(dash)\n", " elements = dashxml.getElementsByTagName(\"BaseURL\")\n", " for eindex,element in enumerate(elements):\n", " for node in element.childNodes:\n", " if node.nodeType == node.TEXT_NODE:\n", " url = node.data\n", " mediatype = 3\n", " path = self.formatPath(username, userpk, timestamp, postid + \"_\" + str(eindex), mediatype)\n", " self.getFile(url, path)\n", " except KeyError: # No \"post_live\" key\n", " logging.debug(' ' + 'No live stories.')\n", "\n", " def close(self):\n", " \"\"\"Close seesion to IG\n", "\n", " Returns:\n", " None\n", " \"\"\"\n", " self.session.close()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.058823529411764705, 0, 0, 0, 0, 0.009259259259259259, 0, 0, 0, 0, 0, 0, 0.025, 0.02, 0.023809523809523808, 0.023529411764705882, 0.022222222222222223, 0.013157894736842105, 0.02631578947368421, 0, 0, 0.013793103448275862, 0, 0, 0, 0, 0, 0, 0, 0, 0.007142857142857143, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.010869565217391304, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.011904761904761904, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.01020408163265306, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.011235955056179775, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.007407407407407408, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.023255813953488372, 0, 0.012195121951219513, 0, 0.012048192771084338, 0.0125, 0, 0, 0, 0, 0, 0.022222222222222223, 0, 0.010309278350515464, 0, 0.01020408163265306, 0.021052631578947368, 0, 0, 0, 0, 0, 0.03125, 0, 0, 0, 0, 0.011494252873563218, 0, 0.020618556701030927, 0, 0, 0, 0, 0, 0.012048192771084338, 0, 0, 0.011363636363636364, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.012658227848101266, 0, 0, 0, 0.014084507042253521, 0, 0, 0, 0, 0, 0, 0.015873015873015872, 0, 0, 0, 0, 0.008130081300813009, 0, 0.021739130434782608, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
272
0.002124
from airflow.hooks.base_hook import BaseHook from airflow import configuration try: snakebite_imported = True from snakebite.client import Client, HAClient, Namenode except ImportError: snakebite_imported = False from airflow.utils import AirflowException class HDFSHookException(AirflowException): pass class HDFSHook(BaseHook): ''' Interact with HDFS. This class is a wrapper around the snakebite library. ''' def __init__(self, hdfs_conn_id='hdfs_default', proxy_user=None): if not snakebite_imported: raise ImportError( 'This HDFSHook implementation requires snakebite, but ' 'snakebite is not compatible with Python 3 ' '(as of August 2015). Please use Python 2 if you require ' 'this hook -- or help by submitting a PR!') self.hdfs_conn_id = hdfs_conn_id self.proxy_user = proxy_user def get_conn(self): ''' Returns a snakebite HDFSClient object. ''' use_sasl = False if configuration.get('core', 'security') == 'kerberos': use_sasl = True connections = self.get_connections(self.hdfs_conn_id) client = None ''' When using HAClient, proxy_user must be the same, so is ok to always take the first ''' effective_user = self.proxy_user or connections[0].login if len(connections) == 1: client = Client(connections[0].host, connections[0].port, use_sasl=use_sasl, effective_user=effective_user) elif len(connections) > 1: nn = [Namenode(conn.host, conn.port) for conn in connections] client = HAClient(nn, use_sasl=use_sasl, effective_user=effective_user) else: raise HDFSHookException("conn_id doesn't exist in the repository") return client
[ "from airflow.hooks.base_hook import BaseHook\n", "from airflow import configuration\n", "\n", "try:\n", " snakebite_imported = True\n", " from snakebite.client import Client, HAClient, Namenode\n", "except ImportError:\n", " snakebite_imported = False\n", "\n", "from airflow.utils import AirflowException\n", "\n", "\n", "class HDFSHookException(AirflowException):\n", " pass\n", "\n", "\n", "class HDFSHook(BaseHook):\n", " '''\n", " Interact with HDFS. This class is a wrapper around the snakebite library.\n", " '''\n", " def __init__(self, hdfs_conn_id='hdfs_default', proxy_user=None):\n", " if not snakebite_imported:\n", " raise ImportError(\n", " 'This HDFSHook implementation requires snakebite, but '\n", " 'snakebite is not compatible with Python 3 '\n", " '(as of August 2015). Please use Python 2 if you require '\n", " 'this hook -- or help by submitting a PR!')\n", " self.hdfs_conn_id = hdfs_conn_id\n", " self.proxy_user = proxy_user\n", "\n", " def get_conn(self):\n", " '''\n", " Returns a snakebite HDFSClient object.\n", " '''\n", " use_sasl = False\n", " if configuration.get('core', 'security') == 'kerberos':\n", " use_sasl = True\n", "\n", " connections = self.get_connections(self.hdfs_conn_id)\n", " client = None\n", "\t''' When using HAClient, proxy_user must be the same, so is ok to always take the first '''\n", "\teffective_user = self.proxy_user or connections[0].login\n", " if len(connections) == 1:\n", " client = Client(connections[0].host, connections[0].port, use_sasl=use_sasl, effective_user=effective_user)\n", " elif len(connections) > 1:\n", " nn = [Namenode(conn.host, conn.port) for conn in connections]\n", " client = HAClient(nn, use_sasl=use_sasl, effective_user=effective_user)\n", " else:\n", " raise HDFSHookException(\"conn_id doesn't exist in the repository\")\n", " return client\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.03225806451612903, 0.017241379310344827, 0.029411764705882353, 0.008333333333333333, 0, 0, 0.011904761904761904, 0, 0, 0 ]
50
0.001983
import json import csv """ def write_to_text(data, filename): f = open(filename, 'w', encoding = 'utf-8') for line in data: f.write(line + '\n') f.close() """ def read_text_file(filename): ignore = [] ignore.append('[') ignore.append(']') ignore.append('') f = open(filename, encoding='utf-8') table_original = [] for line in f: line = line.rstrip('\n') if line not in ignore: if 'Getting snapshot pages' not in line: table_original.append(line.rstrip(',')) return table_original def write_to_csv(data, filename): with open(filename, 'w') as csvfile: w = csv.writer(csvfile, delimiter = ',') for row in data: try: w.writerow(row) except: print(row, 'not written to file.') print(filename, 'created.') def extract_url_database(filename): u = read_text_file(filename) url_list = [['url', 'timestamp', 'id']] for article in u: url_object = [] url_object.append(json.loads(article)['file_url']) url_object.append(json.loads(article)['timestamp']) url_object.append(json.loads(article)['file_id']) url_list.append(url_object) return(url_list) # write list to text file, one URL per line write_to_csv(extract_url_database('whg.txt'), 'whg_url_data.csv')
[ "import json\n", "import csv\n", "\n", "\"\"\"\n", "def write_to_text(data, filename):\n", " f = open(filename, 'w', encoding = 'utf-8')\n", " for line in data:\n", " f.write(line + '\\n')\n", " f.close()\n", "\"\"\"\n", "\n", "def read_text_file(filename):\n", " ignore = []\n", " ignore.append('[')\n", " ignore.append(']')\n", " ignore.append('')\n", " f = open(filename, encoding='utf-8')\n", " table_original = []\n", " for line in f:\n", " line = line.rstrip('\\n')\n", " if line not in ignore:\n", " if 'Getting snapshot pages' not in line:\n", " table_original.append(line.rstrip(','))\n", " return table_original\n", "\n", "def write_to_csv(data, filename):\n", " with open(filename, 'w') as csvfile:\n", " w = csv.writer(csvfile, delimiter = ',')\n", " for row in data:\n", " try:\n", " w.writerow(row)\n", " except:\n", " print(row, 'not written to file.')\n", " print(filename, 'created.')\n", "\n", "def extract_url_database(filename):\n", " u = read_text_file(filename)\n", " url_list = [['url', 'timestamp', 'id']]\n", " for article in u:\n", " url_object = []\n", " url_object.append(json.loads(article)['file_url'])\n", " url_object.append(json.loads(article)['timestamp'])\n", " url_object.append(json.loads(article)['file_id'])\n", " url_list.append(url_object)\n", " return(url_list)\n", "\n", "# write list to text file, one URL per line\n", "write_to_csv(extract_url_database('whg.txt'), 'whg_url_data.csv')\n", "\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.03333333333333333, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.029411764705882353, 0, 0.04081632653061224, 0, 0, 0, 0.05, 0, 0, 0, 0.027777777777777776, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.015151515151515152, 1 ]
49
0.024418
#!/usr/bin/python -u from optparse import OptionParser, OptionGroup from sys import argv as sys_argv, exit from swiftclient import Client def upload(options): values = { 'username': options.os_username, 'password': options.os_password, 'auth_url': options.os_auth_url, 'project_name': options.os_project_name, 'user_domain_name': options.os_user_domain_name, 'project_domain_name': options.os_project_domain_name, 'storage_url': options.os_storage_url, } container = options.container path = options.upload_path cli = Client(values) cli.upload(container, path) def download(options): values = { 'username': options.os_username, 'password': options.os_password, 'auth_url': options.os_auth_url, 'project_name': options.os_project_name, 'user_domain_name': options.os_user_domain_name, 'project_domain_name': options.os_project_domain_name, 'storage_url': options.os_storage_url, } container = options.container objectname = options.object download_path = options.download_path cli = Client(values) cli.download(container, objectname, download_path) def main(arguments=None): if arguments: argv = arguments else: argv = sys_argv version = '0.0.1' parser = OptionParser(version='%%prog %s' % version, usage=''' Command-line interface to the OpenStack Swift API. usage: %%prog [--version] [--help] Mandatory Switch: [--os-username <auth-user-name>] [--os-password <auth-password>] [--os-project-name <auth-project-name>] [--os-auth-url <auth-url>] [--os-user-domain-name <auth-user-domain-name>] [--os-project-domain-name <auth-project-domain-name>] [--os-storage-url <storage-url>] [--operation-type <operation-type>] [--container <container-name>] Command Specific Switch: For Upload (Uploads files or directories to the given container from the upload path.): [--upload-path <upload-path>] For Download (Downloads files from the given container in the download path.): [--object <object-name>] [--download-path <download-path>] '''.strip('\n') % globals()) parser.add_option('--insecure', action="store_true", dest="insecure", default=True, help='Allow swiftclient to access servers without ' 'having to verify the SSL certificate. ' 'Defaults to env[SWIFTCLIENT_INSECURE] ' '(set to \'true\' to enable).') os_grp = OptionGroup(parser, "OpenStack authentication options") os_grp.add_option('--os-username', metavar='<auth-user-name>', help='OpenStack username required to authenticate with OpenStack swift. ') os_grp.add_option('--os_username', help='OpenStack username required to authenticate with OpenStack swift. ') os_grp.add_option('--os-password', metavar='<auth-password>', help='OpenStack password required to authenticate with OpenStack swift.') os_grp.add_option('--os-user-domain-name', metavar='<user-domain-name>', help='OpenStack user domain name required to connect with OpenStack swift.') os_grp.add_option('--os-project-name', metavar='<project-name>', help='OpenStack project name required to connect with OpenStack swift.') os_grp.add_option('--os-project-domain-name', metavar='<project-domain-name>', help='OpenStack project domain name required to connect with OpenStack swift.') os_grp.add_option('--os-auth-url', metavar='<auth-url>', help='OpenStack auth URL required to authenticate with OpenStack Identity to get the ' 'authentication token.') os_grp.add_option('--os-storage-url', metavar='<storage-url>', help='OpenStack storage URL required to connect with the OpenStack Swift.') os_grp.add_option('--operation-type', metavar='<operation-type>', help='Specified OpenStack swift related operation which can be upload or download.') os_grp.add_option('--container', metavar='<container-name>', help='Specified container name to upload/download object.') os_grp.add_option('--object', metavar='<object-name>', help='Specified object name to be downloaded in the downloaded path.') os_grp.add_option('--upload-path', metavar='<upload-path>', help='Upload path of the file or directory.') os_grp.add_option('--download-path', metavar='<download-path>', help='Download path to download the object.') (options, args) = parser.parse_args(argv[1:]) try: if(options.operation_type == 'upload'): if(options.upload_path is None): parser.print_help() exit() else: upload(options) elif(options.operation_type == 'download'): if (options.object is None and options.download_path is None): parser.print_help() exit() else: download(options) else: parser.print_help() exit() except Exception as err: print(str(err)) if __name__ == '__main__': main()
[ "#!/usr/bin/python -u\n", "\n", "from optparse import OptionParser, OptionGroup\n", "from sys import argv as sys_argv, exit\n", "\n", "from swiftclient import Client\n", "\n", "def upload(options):\n", "\n", " values = {\n", " 'username': options.os_username,\n", " 'password': options.os_password,\n", " 'auth_url': options.os_auth_url,\n", " 'project_name': options.os_project_name,\n", " 'user_domain_name': options.os_user_domain_name,\n", " 'project_domain_name': options.os_project_domain_name,\n", " 'storage_url': options.os_storage_url,\n", " }\n", "\n", " container = options.container\n", " path = options.upload_path\n", "\n", " cli = Client(values)\n", " cli.upload(container, path)\n", "\n", "def download(options):\n", "\n", " values = {\n", " 'username': options.os_username,\n", " 'password': options.os_password,\n", " 'auth_url': options.os_auth_url,\n", " 'project_name': options.os_project_name,\n", " 'user_domain_name': options.os_user_domain_name,\n", " 'project_domain_name': options.os_project_domain_name,\n", " 'storage_url': options.os_storage_url,\n", " }\n", "\n", " container = options.container\n", " objectname = options.object\n", " download_path = options.download_path\n", "\n", " cli = Client(values)\n", " cli.download(container, objectname, download_path)\n", "\n", "def main(arguments=None):\n", " if arguments:\n", " argv = arguments\n", " else:\n", " argv = sys_argv\n", "\n", " version = '0.0.1'\n", "\n", " parser = OptionParser(version='%%prog %s' % version,\n", " usage='''\n", "Command-line interface to the OpenStack Swift API.\n", "\n", "usage: %%prog [--version] [--help]\n", "\n", "Mandatory Switch: \n", " [--os-username <auth-user-name>]\n", " [--os-password <auth-password>]\n", " [--os-project-name <auth-project-name>]\n", " [--os-auth-url <auth-url>]\n", " [--os-user-domain-name <auth-user-domain-name>]\n", " [--os-project-domain-name <auth-project-domain-name>]\n", " [--os-storage-url <storage-url>]\n", " [--operation-type <operation-type>]\n", " [--container <container-name>]\n", "\n", "Command Specific Switch:\n", "\n", "For Upload (Uploads files or directories to the given container from the upload path.):\n", " [--upload-path <upload-path>]\n", "\n", "For Download (Downloads files from the given container in the download path.):\n", " [--object <object-name>]\n", " [--download-path <download-path>] \n", "'''.strip('\\n') % globals())\n", " parser.add_option('--insecure',\n", " action=\"store_true\", dest=\"insecure\",\n", " default=True,\n", " help='Allow swiftclient to access servers without '\n", " 'having to verify the SSL certificate. '\n", " 'Defaults to env[SWIFTCLIENT_INSECURE] '\n", " '(set to \\'true\\' to enable).')\n", "\n", " os_grp = OptionGroup(parser, \"OpenStack authentication options\")\n", "\n", " os_grp.add_option('--os-username',\n", " metavar='<auth-user-name>',\n", " help='OpenStack username required to authenticate with OpenStack swift. ')\n", " os_grp.add_option('--os_username',\n", " help='OpenStack username required to authenticate with OpenStack swift. ')\n", " os_grp.add_option('--os-password',\n", " metavar='<auth-password>',\n", " help='OpenStack password required to authenticate with OpenStack swift.')\n", " os_grp.add_option('--os-user-domain-name',\n", " metavar='<user-domain-name>',\n", " help='OpenStack user domain name required to connect with OpenStack swift.')\n", " os_grp.add_option('--os-project-name',\n", " metavar='<project-name>',\n", " help='OpenStack project name required to connect with OpenStack swift.')\n", " os_grp.add_option('--os-project-domain-name',\n", " metavar='<project-domain-name>',\n", " help='OpenStack project domain name required to connect with OpenStack swift.')\n", " os_grp.add_option('--os-auth-url',\n", " metavar='<auth-url>',\n", " help='OpenStack auth URL required to authenticate with OpenStack Identity to get the '\n", " 'authentication token.')\n", " os_grp.add_option('--os-storage-url',\n", " metavar='<storage-url>',\n", " help='OpenStack storage URL required to connect with the OpenStack Swift.')\n", " os_grp.add_option('--operation-type',\n", " metavar='<operation-type>',\n", " help='Specified OpenStack swift related operation which can be upload or download.')\n", " os_grp.add_option('--container',\n", " metavar='<container-name>',\n", " help='Specified container name to upload/download object.')\n", " os_grp.add_option('--object',\n", " metavar='<object-name>',\n", " help='Specified object name to be downloaded in the downloaded path.')\n", " os_grp.add_option('--upload-path',\n", " metavar='<upload-path>',\n", " help='Upload path of the file or directory.')\n", " os_grp.add_option('--download-path',\n", " metavar='<download-path>',\n", " help='Download path to download the object.')\n", "\n", " (options, args) = parser.parse_args(argv[1:])\n", "\n", " try:\n", " if(options.operation_type == 'upload'):\n", " if(options.upload_path is None):\n", " parser.print_help()\n", " exit()\n", " else:\n", " upload(options)\n", " elif(options.operation_type == 'download'):\n", " if (options.object is None and options.download_path is None):\n", " parser.print_help()\n", " exit()\n", " else:\n", " download(options)\n", " else:\n", " parser.print_help()\n", " exit()\n", "\n", " except Exception as err:\n", " print(str(err))\n", "\n", "if __name__ == '__main__':\n", " main()\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0.047619047619047616, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.043478260869565216, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.038461538461538464, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.05263157894736842, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.011363636363636364, 0, 0, 0, 0, 0.017543859649122806, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.010309278350515464, 0, 0.010309278350515464, 0, 0, 0.010416666666666666, 0, 0, 0.010101010101010102, 0, 0, 0.010526315789473684, 0, 0, 0.00980392156862745, 0, 0, 0.009174311926605505, 0, 0, 0, 0.020202020202020204, 0, 0, 0.009345794392523364, 0, 0, 0.012195121951219513, 0, 0, 0.010752688172043012, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.037037037037037035, 0 ]
152
0.002443
# Copyright (c) 2015 Huawei, Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. # import paste.urlmap from congress.api import application from congress.api import versions def wsgi_app(): mapper = paste.urlmap.URLMap() mapper['/'] = versions.Versions() api_resource_mgr = application.ResourceManager() api_resource_mgr.register_handler(versions.VersionV1Handler(r'/v1[/]?')) app = application.ApiApplication(api_resource_mgr) mapper['/v1'] = app return mapper
[ "# Copyright (c) 2015 Huawei, Inc. All rights reserved.\n", "#\n", "# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n", "# not use this file except in compliance with the License. You may obtain\n", "# a copy of the License at\n", "#\n", "# http://www.apache.org/licenses/LICENSE-2.0\n", "#\n", "# Unless required by applicable law or agreed to in writing, software\n", "# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n", "# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n", "# License for the specific language governing permissions and limitations\n", "# under the License.\n", "#\n", "\n", "import paste.urlmap\n", "\n", "from congress.api import application\n", "from congress.api import versions\n", "\n", "\n", "def wsgi_app():\n", "\n", " mapper = paste.urlmap.URLMap()\n", " mapper['/'] = versions.Versions()\n", "\n", " api_resource_mgr = application.ResourceManager()\n", " api_resource_mgr.register_handler(versions.VersionV1Handler(r'/v1[/]?'))\n", " app = application.ApiApplication(api_resource_mgr)\n", " mapper['/v1'] = app\n", "\n", " return mapper\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
32
0
from __future__ import print_function from ase.db.table import dict2forces, hill from ase.data import atomic_masses, chemical_symbols from ase.db.core import float_to_time_string, now import numpy as np class Summary: def __init__(self, dct, subscript=None): self.dct = dct self.cell = [['{0:.3f}'.format(a) for a in axis] for axis in dct.cell] forces = dict2forces(dct) if forces is None: fmax = None self.forces = None else: fmax = (forces**2).sum(1).max()**0.5 N = len(forces) self.forces = [] for n, f in enumerate(forces): if n < 5 or n >= N - 5: f = tuple('{0:10.3f}'.format(x) for x in f) symbol = chemical_symbols[dct.numbers[n]] self.forces.append((n, symbol) + f) elif n == 5: self.forces.append((' ...', '', ' ...', ' ...', ' ...')) self.stress = dct.get('stress') if self.stress is not None: self.stress = ', '.join('{0:.3f}'.format(s) for s in self.stress) if 'masses' in dct: mass = dct.masses.sum() else: mass = atomic_masses[dct.numbers].sum() formula = hill(dct.numbers) if subscript: formula = subscript.sub(r'<sub>\1</sub>', formula) table = [ ('id', dct.id), ('age', float_to_time_string(now() - dct.ctime, True)), ('formula', formula), ('user', dct.user), ('calculator', dct.get('calculator')), ('energy [eV]', dct.get('energy')), ('fmax [eV/Ang]', fmax), ('charge [|e|]', dct.get('charge')), ('mass [au]', mass), ('unique id', dct.unique_id), ('volume [Ang^3]', abs(np.linalg.det(dct.cell)))] self.table = [(name, value) for name, value in table if value is not None] if 'key_value_pairs' in dct: self.key_value_pairs = sorted(dct.key_value_pairs.items()) else: self.key_value_pairs = None if 'keywords' in dct: self.keywords = ', '.join(sorted(dct.keywords)) else: self.keywords = None self.dipole = dct.get('dipole') if self.dipole is not None: self.dipole = ', '.join('{0:.3f}'.format(d) for d in self.dipole) self.data = dct.get('data') if self.data: self.data = ', '.join(self.data.keys()) self.constraints = dct.get('constraints') if self.constraints: self.constraints = ', '.join(d['name'] for d in self.constraints) def write(self): dct = self.dct width = max(len(name) for name, value in self.table) for name, value in self.table: print('{0:{width}}|{1}'.format(name, value, width=width)) print('\nUnit cell in Ang:') print('axis|periodic| x| y| z') c = 1 for p, axis in zip(dct.pbc, self.cell): print(' {0}| {1}|{2[0]:>11}|{2[1]:>11}|{2[2]:>11}'.format( c, [' no', 'yes'][p], axis)) c += 1 if self.key_value_pairs: print('\nKey-value pairs:') width = max(len(key) for key, value in self.key_value_pairs) for key, value in self.key_value_pairs: print('{0:{width}}|{1}'.format(key, value, width=width)) if self.keywords: print('\nKeywords:', self.keywords) if self.forces: print('\nForces in ev/Ang:') for f in self.forces: print('{0:4}|{1:2}|{2}|{3}|{4}'.format(*f)) if self.stress: print('\nStress tensor (xx, yy, zz, zy, zx, yx) in eV/Ang^3:') print(' ', self.stress) if self.dipole: print('\nDipole moment in e*Ang: ({0})'.format(self.dipole)) if self.constraints: print('\nConstraints:', self.constraints) if self.data: print('\nData:', self.data)
[ "from __future__ import print_function\n", "\n", "from ase.db.table import dict2forces, hill\n", "from ase.data import atomic_masses, chemical_symbols\n", "from ase.db.core import float_to_time_string, now\n", "\n", "import numpy as np\n", "\n", "\n", "class Summary:\n", " def __init__(self, dct, subscript=None):\n", " self.dct = dct\n", " \n", " self.cell = [['{0:.3f}'.format(a) for a in axis] for axis in dct.cell]\n", " \n", " forces = dict2forces(dct)\n", " if forces is None:\n", " fmax = None\n", " self.forces = None\n", " else:\n", " fmax = (forces**2).sum(1).max()**0.5\n", " N = len(forces)\n", " self.forces = []\n", " for n, f in enumerate(forces):\n", " if n < 5 or n >= N - 5:\n", " f = tuple('{0:10.3f}'.format(x) for x in f)\n", " symbol = chemical_symbols[dct.numbers[n]]\n", " self.forces.append((n, symbol) + f)\n", " elif n == 5:\n", " self.forces.append((' ...', '',\n", " ' ...',\n", " ' ...',\n", " ' ...'))\n", " \n", " self.stress = dct.get('stress')\n", " if self.stress is not None:\n", " self.stress = ', '.join('{0:.3f}'.format(s) for s in self.stress)\n", " \n", " if 'masses' in dct:\n", " mass = dct.masses.sum()\n", " else:\n", " mass = atomic_masses[dct.numbers].sum()\n", " \n", " formula = hill(dct.numbers)\n", " if subscript:\n", " formula = subscript.sub(r'<sub>\\1</sub>', formula)\n", " \n", " table = [\n", " ('id', dct.id),\n", " ('age', float_to_time_string(now() - dct.ctime, True)),\n", " ('formula', formula),\n", " ('user', dct.user),\n", " ('calculator', dct.get('calculator')),\n", " ('energy [eV]', dct.get('energy')),\n", " ('fmax [eV/Ang]', fmax),\n", " ('charge [|e|]', dct.get('charge')),\n", " ('mass [au]', mass),\n", " ('unique id', dct.unique_id),\n", " ('volume [Ang^3]', abs(np.linalg.det(dct.cell)))]\n", " self.table = [(name, value) for name, value in table\n", " if value is not None]\n", "\n", " if 'key_value_pairs' in dct:\n", " self.key_value_pairs = sorted(dct.key_value_pairs.items())\n", " else:\n", " self.key_value_pairs = None\n", "\n", " if 'keywords' in dct:\n", " self.keywords = ', '.join(sorted(dct.keywords))\n", " else:\n", " self.keywords = None\n", " \n", " self.dipole = dct.get('dipole')\n", " if self.dipole is not None:\n", " self.dipole = ', '.join('{0:.3f}'.format(d) for d in self.dipole)\n", " \n", " self.data = dct.get('data')\n", " if self.data:\n", " self.data = ', '.join(self.data.keys())\n", " \n", " self.constraints = dct.get('constraints')\n", " if self.constraints:\n", " self.constraints = ', '.join(d['name'] for d in self.constraints)\n", " \n", " def write(self):\n", " dct = self.dct\n", " \n", " width = max(len(name) for name, value in self.table)\n", " for name, value in self.table:\n", " print('{0:{width}}|{1}'.format(name, value, width=width))\n", "\n", " print('\\nUnit cell in Ang:')\n", " print('axis|periodic| x| y| z')\n", " c = 1\n", " for p, axis in zip(dct.pbc, self.cell):\n", " print(' {0}| {1}|{2[0]:>11}|{2[1]:>11}|{2[2]:>11}'.format(\n", " c, [' no', 'yes'][p], axis))\n", " c += 1\n", " \n", " if self.key_value_pairs:\n", " print('\\nKey-value pairs:')\n", " width = max(len(key) for key, value in self.key_value_pairs)\n", " for key, value in self.key_value_pairs:\n", " print('{0:{width}}|{1}'.format(key, value, width=width))\n", " \n", " if self.keywords:\n", " print('\\nKeywords:', self.keywords)\n", " \n", " if self.forces:\n", " print('\\nForces in ev/Ang:')\n", " for f in self.forces:\n", " print('{0:4}|{1:2}|{2}|{3}|{4}'.format(*f))\n", "\n", " if self.stress:\n", " print('\\nStress tensor (xx, yy, zz, zy, zx, yx) in eV/Ang^3:')\n", " print(' ', self.stress)\n", "\n", " if self.dipole:\n", " print('\\nDipole moment in e*Ang: ({0})'.format(self.dipole))\n", " \n", " if self.constraints:\n", " print('\\nConstraints:', self.constraints)\n", " \n", " if self.data:\n", " print('\\nData:', self.data)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.1111111111111111, 0, 0.1111111111111111, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.047619047619047616, 0, 0, 0, 0.07692307692307693, 0, 0, 0, 0, 0.07692307692307693, 0, 0, 0, 0.07692307692307693, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.07692307692307693, 0, 0, 0, 0.1111111111111111, 0, 0, 0, 0.07692307692307693, 0, 0, 0, 0.1111111111111111, 0, 0, 0.1111111111111111, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.07692307692307693, 0, 0, 0, 0, 0, 0.058823529411764705, 0, 0, 0.058823529411764705, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.1111111111111111, 0, 0, 0.07692307692307693, 0, 0 ]
125
0.010963
import sys from metatlas import metatlas_objects as metob import metatlas.helpers.metatlas_get_data_helper_fun as ma_data import os from IPython.display import display import matplotlib.pyplot as plt try: import ipywidgets as widgets except ImportError: from IPython.html import widgets try: import traitlets except ImportError: from IPython.utils import traitlets from ipywidgets import interact, interactive, fixed, FloatSlider import copy data = [] groups = [] file_names = [] compound_names = [] compound_objects = [] files_idx = dict() compound_idx = dict() groups_idx = dict() # one select for the compound wcompounds = widgets.Select( description="compounds", options=[] ) # have a multiple select for the files wfiles = widgets.SelectMultiple( description="files", options=[] ) wfname = widgets.Text( description='Atlas Name', value='myAtlas', ) all_files = widgets.Checkbox( description='Select All Files', value=False, ) plot_button = widgets.Button(description='Plot me') create_atlas_btn = widgets.Button(description="Create Atlas") rtmin_widget = FloatSlider() rtpeak_widget = FloatSlider() rtmax_widget = FloatSlider() ########################################################################### ### def plot_intensity(cval, fvals, rt_min, rt_max, rt_peak): for i in range(len(fvals)): d = data[files_idx[fvals[i]]][compound_idx[cval]] if len(d['data']['eic']['rt']) > 0: x = d['data']['eic']['rt'] y = d['data']['eic']['intensity'] plt.plot(x, y, 'k-', ms=1, mew=0, mfc='b', alpha=1.0) plt.axvline(rt_min, color='b', linewidth=2.0) plt.axvline(rt_max, color='g', linewidth=2.0) plt.axvline(rt_peak, color='r', linewidth=2.0) ########################################################################### ### def create_atlas(b): identifications = list() file_names = wfiles.value compound_name = wcompounds.value idx2 = compound_idx[compound_name] atlas = metob.Atlas() atlas.name = wfname.value # create an empty rt reference rt_ref = metob.RtReference() rt_ref.rt_min = rtmin_widget.value rt_ref.rt_max = rtmax_widget.value rt_ref.rt_peak = rtpeak_widget.value rt_ref.rt_units = data[0][idx2]['identification'].rt_references[0].rt_units # create an empty mz_reference mz_ref = metob.MzReference() mz_ref.mz = data[0][idx2]['identification'].mz_references[0].mz mz_ref.mz_tolerance = data[0][idx2]['identification'].mz_references[0].mz_tolerance mz_ref.mz_tolerance_units = data[0][idx2]['identification'].mz_references[0].mz_tolerance_units mz_ref.detected_polarity = data[0][idx2]['identification'].mz_references[0].detected_polarity identification = metob.CompoundIdentification() identification.compoud = compound_name identification.name = compound_name identification.rt_references = [rt_ref] identification.mz_references = [mz_ref] identifications.append(identification) atlas.compound_identifications = identifications #metob.store(atlas) def select_files(b): all_files.value = not all_files.value ########################################################################### ## def plot_button_clicked(b): plt.cla() plt.clf() plt.close() fvals = list(wfiles.value) cval = wcompounds.value global rtmin_widget, rtmax_widget, rtpeak_widget min_x = list() max_x = list() if len(fvals) == 1 and fvals[0] == 'all': fvals = file_names elif len(fvals) > 1 and 'all' in fvals: fvals.remove('all') #if all_files.value == True: # fvals = file_names #else: # fvals = wfiles.value for i in range(len(fvals)): d = data[files_idx[fvals[i]]][compound_idx[cval]] rt_min = d['identification'].rt_references[0].rt_min rt_max = d['identification'].rt_references[0].rt_max rt_peak = d['identification'].rt_references[0].rt_peak if len(d['data']['eic']['rt']) > 0: x = d['data']['eic']['rt'] y = d['data']['eic']['intensity'] min_x.append(min(x)) max_x.append(max(x)) plt.plot(x, y, 'k-', ms=1, mew=0, mfc='b', alpha=1.0) plt.axvline(rt_min, color='b', linewidth=2.0) plt.axvline(rt_max, color='g', linewidth=2.0) plt.axvline(rt_peak, color='r', linewidth=2.0) rtmin_widget.close() rtpeak_widget.close() rtmax_widget.close() rtmin_widget = FloatSlider(min=min(min_x), max=max(max_x), step=0.01, value=rt_min, color='blue') rtpeak_widget = FloatSlider(min=min(min_x), max=max(max_x), step=0.01, value=rt_peak, color='red') rtmax_widget = FloatSlider(min=min(min_x), max=max(max_x), step=0.01, value=rt_max, color='green') interact(plot_intensity, cval=fixed(cval), fvals=fixed(fvals), rt_min=rtmin_widget, rt_peak=rtpeak_widget, rt_max=rtmax_widget) def dill2atlas(fname): global data, groups, file_names, compound_names, compound_objects, files_idx, compound_idx, groups_idx data = ma_data.get_dill_data(fname) groups = ma_data.get_group_names(data) file_names = ma_data.get_file_names(data) (compound_names, compound_objects) = ma_data.get_compound_names(data) files_idx = dict() for f_idx, f_name in enumerate(file_names): files_idx[f_name] = f_idx compound_idx = dict() for cpd_idx, cpd_name in enumerate(compound_names): compound_idx[cpd_name] = cpd_idx groups_idx = dict() for grp_idx, grp_name in enumerate(groups): groups_idx[grp_name] = grp_idx wcompounds.options=compound_names wfiles.options= ['all'] + file_names display(widgets.HBox((wfname, create_atlas_btn))) display(widgets.HBox((wcompounds, wfiles))) display(plot_button) plot_button.on_click(plot_button_clicked) create_atlas_btn.on_click(create_atlas) all_files.observe(select_files)
[ "import sys\n", "\n", "from metatlas import metatlas_objects as metob\n", "import metatlas.helpers.metatlas_get_data_helper_fun as ma_data\n", "import os\n", "\n", "\n", "from IPython.display import display\n", "import matplotlib.pyplot as plt\n", "try:\n", " import ipywidgets as widgets\n", "except ImportError:\n", " from IPython.html import widgets\n", "try:\n", " import traitlets\n", "except ImportError:\n", " from IPython.utils import traitlets\n", "\n", "from ipywidgets import interact, interactive, fixed, FloatSlider\n", "\n", "\n", "import copy\n", "\n", "\n", "\n", "data = []\n", "groups = []\n", "file_names = []\n", "compound_names = []\n", "compound_objects = []\n", "\n", "files_idx = dict()\n", "compound_idx = dict()\n", "groups_idx = dict()\n", "\n", "\n", "\n", "# one select for the compound\n", "wcompounds = widgets.Select(\n", " description=\"compounds\",\n", " options=[]\n", ")\n", "\n", "# have a multiple select for the files\n", "wfiles = widgets.SelectMultiple(\n", " description=\"files\",\n", " options=[]\n", ")\n", "\n", "wfname = widgets.Text(\n", " description='Atlas Name',\n", " value='myAtlas',\n", ")\n", "\n", "all_files = widgets.Checkbox(\n", " description='Select All Files',\n", " value=False,\n", ")\n", "\n", "plot_button = widgets.Button(description='Plot me')\n", "create_atlas_btn = widgets.Button(description=\"Create Atlas\")\n", "\n", "rtmin_widget = FloatSlider()\n", "rtpeak_widget = FloatSlider()\n", "rtmax_widget = FloatSlider()\n", "\n", "\n", "###########################################################################\n", "###\n", "def plot_intensity(cval, fvals, rt_min, rt_max, rt_peak):\n", " for i in range(len(fvals)):\n", " d = data[files_idx[fvals[i]]][compound_idx[cval]]\n", "\n", " if len(d['data']['eic']['rt']) > 0:\n", " x = d['data']['eic']['rt']\n", " y = d['data']['eic']['intensity']\n", " plt.plot(x, y, 'k-', ms=1, mew=0, mfc='b', alpha=1.0)\n", "\n", " plt.axvline(rt_min, color='b', linewidth=2.0)\n", " plt.axvline(rt_max, color='g', linewidth=2.0)\n", " plt.axvline(rt_peak, color='r', linewidth=2.0)\n", "\n", "\n", "###########################################################################\n", "###\n", "def create_atlas(b):\n", " identifications = list()\n", " file_names = wfiles.value\n", " compound_name = wcompounds.value\n", " idx2 = compound_idx[compound_name]\n", "\n", " atlas = metob.Atlas()\n", " atlas.name = wfname.value\n", "\n", " # create an empty rt reference\n", " rt_ref = metob.RtReference()\n", "\n", " rt_ref.rt_min = rtmin_widget.value\n", " rt_ref.rt_max = rtmax_widget.value\n", " rt_ref.rt_peak = rtpeak_widget.value\n", " rt_ref.rt_units = data[0][idx2]['identification'].rt_references[0].rt_units\n", "\n", " # create an empty mz_reference\n", " mz_ref = metob.MzReference()\n", "\n", " mz_ref.mz = data[0][idx2]['identification'].mz_references[0].mz\n", " mz_ref.mz_tolerance = data[0][idx2]['identification'].mz_references[0].mz_tolerance\n", " mz_ref.mz_tolerance_units = data[0][idx2]['identification'].mz_references[0].mz_tolerance_units\n", " mz_ref.detected_polarity = data[0][idx2]['identification'].mz_references[0].detected_polarity\n", "\n", " identification = metob.CompoundIdentification()\n", " identification.compoud = compound_name\n", " identification.name = compound_name\n", " identification.rt_references = [rt_ref]\n", " identification.mz_references = [mz_ref]\n", "\n", " identifications.append(identification)\n", "\n", " atlas.compound_identifications = identifications\n", " #metob.store(atlas)\n", "\n", "\n", "def select_files(b):\n", " all_files.value = not all_files.value\n", "\n", "\n", "###########################################################################\n", "##\n", "def plot_button_clicked(b):\n", " plt.cla()\n", " plt.clf()\n", " plt.close()\n", "\n", " fvals = list(wfiles.value)\n", " cval = wcompounds.value\n", " global rtmin_widget, rtmax_widget, rtpeak_widget\n", "\n", " min_x = list()\n", " max_x = list()\n", "\n", " if len(fvals) == 1 and fvals[0] == 'all':\n", " fvals = file_names\n", " elif len(fvals) > 1 and 'all' in fvals:\n", " fvals.remove('all')\n", "\n", "\n", "\n", "\n", "\n", " #if all_files.value == True:\n", " # fvals = file_names\n", " #else:\n", " # fvals = wfiles.value\n", "\n", " for i in range(len(fvals)):\n", " d = data[files_idx[fvals[i]]][compound_idx[cval]]\n", " rt_min = d['identification'].rt_references[0].rt_min\n", " rt_max = d['identification'].rt_references[0].rt_max\n", " rt_peak = d['identification'].rt_references[0].rt_peak\n", "\n", " if len(d['data']['eic']['rt']) > 0:\n", " x = d['data']['eic']['rt']\n", " y = d['data']['eic']['intensity']\n", " min_x.append(min(x))\n", " max_x.append(max(x))\n", " plt.plot(x, y, 'k-', ms=1, mew=0, mfc='b', alpha=1.0)\n", "\n", " plt.axvline(rt_min, color='b', linewidth=2.0)\n", " plt.axvline(rt_max, color='g', linewidth=2.0)\n", " plt.axvline(rt_peak, color='r', linewidth=2.0)\n", "\n", " rtmin_widget.close()\n", " rtpeak_widget.close()\n", " rtmax_widget.close()\n", " rtmin_widget = FloatSlider(min=min(min_x), max=max(max_x), step=0.01, value=rt_min, color='blue')\n", " rtpeak_widget = FloatSlider(min=min(min_x), max=max(max_x), step=0.01, value=rt_peak, color='red')\n", " rtmax_widget = FloatSlider(min=min(min_x), max=max(max_x), step=0.01, value=rt_max, color='green')\n", " interact(plot_intensity,\n", " cval=fixed(cval),\n", " fvals=fixed(fvals),\n", " rt_min=rtmin_widget,\n", " rt_peak=rtpeak_widget,\n", " rt_max=rtmax_widget)\n", "\n", "\n", "def dill2atlas(fname):\n", " global data, groups, file_names, compound_names, compound_objects, files_idx, compound_idx, groups_idx\n", "\n", " data = ma_data.get_dill_data(fname)\n", " groups = ma_data.get_group_names(data)\n", " file_names = ma_data.get_file_names(data)\n", " (compound_names, compound_objects) = ma_data.get_compound_names(data)\n", "\n", " files_idx = dict()\n", " for f_idx, f_name in enumerate(file_names):\n", " files_idx[f_name] = f_idx\n", "\n", " compound_idx = dict()\n", " for cpd_idx, cpd_name in enumerate(compound_names):\n", " compound_idx[cpd_name] = cpd_idx\n", "\n", " groups_idx = dict()\n", " for grp_idx, grp_name in enumerate(groups):\n", " groups_idx[grp_name] = grp_idx\n", "\n", "\n", " wcompounds.options=compound_names\n", "\n", " wfiles.options= ['all'] + file_names\n", "\n", "\n", "\n", " display(widgets.HBox((wfname, create_atlas_btn)))\n", " display(widgets.HBox((wcompounds, wfiles)))\n", " display(plot_button)\n", "\n", "\n", " plot_button.on_click(plot_button_clicked)\n", " create_atlas_btn.on_click(create_atlas)\n", " all_files.observe(select_files)\n", "\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.03333333333333333, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.011363636363636364, 0.01, 0.01020408163265306, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.041666666666666664, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.06060606060606061, 0, 0.09090909090909091, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.00980392156862745, 0.009708737864077669, 0.009708737864077669, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.009259259259259259, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.05263157894736842, 0, 0.024390243902439025, 0, 0, 0, 0.018518518518518517, 0, 0, 0, 0, 0.021739130434782608, 0, 0, 1 ]
221
0.00685
#!/usr/bin/env python ''' Plots the gravitational potential of the earth and moon assumes the earth is at 0,0 and moon is on x-axis. Inspired by http://www.wired.com/2014/07/contour-plots-with-python-and-plotly/?mbid=social_twitter and: https://gist.github.com/rhettallain/1aa12b44d59562ce08fc Practicing plotly ''' __author__ = 'julenka' import numpy as np import plotly.plotly as py import math from plotly.graph_objs import * massEarth = 5.9729e24 massMoon = 7.3477e22 distanceFromEarthToMoon = 384400 # units is km G = 6.67384e-11 earthOffset = 100000 earthLocation = [earthOffset,earthOffset] moonLocation = [earthOffset + distanceFromEarthToMoon/2, earthOffset + math.sqrt(math.pow(distanceFromEarthToMoon,2) - math.pow(distanceFromEarthToMoon/2, 2))] # make mesh step = 10000 x = np.arange(1, 2 * distanceFromEarthToMoon, step) y = np.arange(1, 2 * distanceFromEarthToMoon, step) X,Y = np.meshgrid(x,y) # gravitational potential values V = X * 0 Vmax = 4e10 for r in range(len(X)): for c in range(len(Y)): currentLocation = np.array([X[r,c], Y[r,c]]) distanceToEarth = np.linalg.norm(np.subtract(currentLocation, earthLocation)) gpFromEarth = G * massEarth / distanceToEarth #V(r) = Gm/r distanceToMoon = np.linalg.norm(np.subtract(currentLocation, moonLocation)) gpFromMoon = G * massEarth / distanceToMoon #V(r) = Gm/r totalGp = max(0, min(Vmax, gpFromEarth + gpFromMoon)) V[r,c] = totalGp data=[{'x':x, 'y':y, 'z':V, 'type':'contour'}] plot_url = py.plot(data, filename='earth-moon gravitational potential')
[ "#!/usr/bin/env python\n", "'''\n", "Plots the gravitational potential of the earth and moon\n", "assumes the earth is at 0,0 and moon is on x-axis.\n", "\n", "Inspired by http://www.wired.com/2014/07/contour-plots-with-python-and-plotly/?mbid=social_twitter\n", "and: https://gist.github.com/rhettallain/1aa12b44d59562ce08fc\n", "\n", "Practicing plotly\n", "'''\n", "\n", "__author__ = 'julenka'\n", "\n", "import numpy as np\n", "import plotly.plotly as py\n", "import math\n", "from plotly.graph_objs import *\n", "\n", "massEarth = 5.9729e24\n", "massMoon = 7.3477e22\n", "distanceFromEarthToMoon = 384400 # units is km\n", "G = 6.67384e-11\n", "\n", "earthOffset = 100000\n", "earthLocation = [earthOffset,earthOffset]\n", "moonLocation = [earthOffset + distanceFromEarthToMoon/2,\n", " earthOffset + math.sqrt(math.pow(distanceFromEarthToMoon,2) - math.pow(distanceFromEarthToMoon/2, 2))]\n", "\n", "# make mesh\n", "step = 10000\n", "x = np.arange(1, 2 * distanceFromEarthToMoon, step)\n", "y = np.arange(1, 2 * distanceFromEarthToMoon, step)\n", "X,Y = np.meshgrid(x,y)\n", "\n", "\n", "# gravitational potential values\n", "V = X * 0\n", "Vmax = 4e10\n", "\n", "for r in range(len(X)):\n", " for c in range(len(Y)):\n", " currentLocation = np.array([X[r,c], Y[r,c]])\n", "\n", " distanceToEarth = np.linalg.norm(np.subtract(currentLocation, earthLocation))\n", " gpFromEarth = G * massEarth / distanceToEarth #V(r) = Gm/r\n", "\n", " distanceToMoon = np.linalg.norm(np.subtract(currentLocation, moonLocation))\n", " gpFromMoon = G * massEarth / distanceToMoon #V(r) = Gm/r\n", "\n", " totalGp = max(0, min(Vmax, gpFromEarth + gpFromMoon))\n", " V[r,c] = totalGp\n", "\n", "data=[{'x':x, 'y':y, 'z':V, 'type':'contour'}]\n", "plot_url = py.plot(data, filename='earth-moon gravitational potential')" ]
[ 0, 0, 0, 0, 0, 0.010101010101010102, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.02127659574468085, 0, 0, 0, 0.023809523809523808, 0, 0.01680672268907563, 0, 0, 0, 0, 0, 0.08695652173913043, 0, 0, 0, 0, 0, 0, 0, 0, 0.03773584905660377, 0, 0.011627906976744186, 0.029850746268656716, 0, 0.011904761904761904, 0.03076923076923077, 0, 0, 0.04, 0, 0.10638297872340426, 0.014084507042253521 ]
54
0.008172
""" EULA APP This module provides additional functionality to the EUAL app. Classes: EULAAcceptedMixin Functions: n/a Created on 23 Oct 2013 @author: michael """ from django.core.exceptions import PermissionDenied from django.core.urlresolvers import reverse_lazy from django.contrib.auth.decorators import login_required from django.contrib.contenttypes.models import ContentType from django.shortcuts import get_object_or_404 from django.utils.decorators import method_decorator from tunobase.eula import models, utils class EULAAcceptedMixin(object): """Render EUAL to users.""" eula_url = 'eula_sign' raise_exception = False @method_decorator(login_required) def dispatch(self, request, *args, **kwargs): """ Confirm that the EULA has been agreed to by the user and if not render EULA. """ self.eula = get_object_or_404(models.EULA).latest_version() # If the user has accepted the latest EULA if not models.UserEULA.objects\ .filter(user=request.user, eula=self.eula)\ .exists(): if self.raise_exception: raise PermissionDenied else: if hasattr(self, 'get_object'): self.object = self.get_object() eula_url_kwargs = { 'content_type_id': ContentType.objects\ .get_for_model(self.object).id, 'object_pk': self.object.pk } eula_url = reverse_lazy( self.eula_url, kwargs=eula_url_kwargs ) else: eula_url = reverse_lazy(self.eula_url) return utils.redirect_to_eula( request.get_full_path(), eula_url ) return super(EULAAcceptedMixin, self).dispatch( request, *args, **kwargs )
[ "\"\"\"\n", "EULA APP\n", "\n", "This module provides additional functionality to the EUAL app.\n", "\n", "Classes:\n", " EULAAcceptedMixin\n", "\n", "Functions:\n", " n/a\n", "\n", "Created on 23 Oct 2013\n", "\n", "@author: michael\n", "\n", "\"\"\"\n", "from django.core.exceptions import PermissionDenied\n", "from django.core.urlresolvers import reverse_lazy\n", "from django.contrib.auth.decorators import login_required\n", "from django.contrib.contenttypes.models import ContentType\n", "from django.shortcuts import get_object_or_404\n", "from django.utils.decorators import method_decorator\n", "\n", "from tunobase.eula import models, utils\n", "\n", "class EULAAcceptedMixin(object):\n", " \"\"\"Render EUAL to users.\"\"\"\n", "\n", " eula_url = 'eula_sign'\n", " raise_exception = False\n", "\n", " @method_decorator(login_required)\n", " def dispatch(self, request, *args, **kwargs):\n", " \"\"\"\n", " Confirm that the EULA has been agreed to by the user\n", " and if not render EULA.\n", "\n", " \"\"\"\n", " self.eula = get_object_or_404(models.EULA).latest_version()\n", "\n", " # If the user has accepted the latest EULA\n", " if not models.UserEULA.objects\\\n", " .filter(user=request.user, eula=self.eula)\\\n", " .exists():\n", " if self.raise_exception:\n", " raise PermissionDenied\n", " else:\n", " if hasattr(self, 'get_object'):\n", " self.object = self.get_object()\n", " eula_url_kwargs = {\n", " 'content_type_id': ContentType.objects\\\n", " .get_for_model(self.object).id,\n", " 'object_pk': self.object.pk\n", " }\n", " eula_url = reverse_lazy(\n", " self.eula_url, kwargs=eula_url_kwargs\n", " )\n", " else:\n", " eula_url = reverse_lazy(self.eula_url)\n", "\n", " return utils.redirect_to_eula(\n", " request.get_full_path(),\n", " eula_url\n", " )\n", "\n", " return super(EULAAcceptedMixin, self).dispatch(\n", " request,\n", " *args,\n", " **kwargs\n", " )\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.030303030303030304, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.015625, 0.015625, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
70
0.000879
"""Fixtures for tests in this directory.""" import multiprocessing import multiprocessing.queues import sys import httpretty import pytest from sphinx import build_main def run_build_main(docs_dir, html_dir, overflow): """Run build_main(). :param str docs_dir: Path to input docs directory. :param str html_dir: Path to output html directory. :param iter overflow: Append these args to sphinx-build call. :return: Value from build_main(). :rtype: int """ argv = ('sphinx-build', str(docs_dir), str(html_dir)) if overflow: argv += overflow result = build_main(argv) return result def run_build_main_post_multiprocessing(docs_dir, html_dir, cached_responses, queue, overflow): """Run Sphinx's build_main after setting up httpretty mock responses. Called by multiprocess.Process. Need to use this instead of httpretty pytest fixtures since forking doesn't exist in Windows and multiprocess runs in "spawn" mode. This means that everything setup by pytest is lost since subprocesses are generated from scratch on Windows. :raise: RuntimeError on Sphinx non-zero exit. This causes multiprocessing.Process().exitcode to be != 0. :param str docs_dir: Path to input docs directory. :param str html_dir: Path to output html directory. :param dict cached_responses: URL keys and serialized JSON values. :param multiprocessing.queues.Queue queue: Queue to transmit stdout/err back to parent process. :param iter overflow: Append these args to sphinx-build call. """ # Capture stdout/stderr after forking/spawning. capture = __import__('_pytest').capture try: capsys = capture.CaptureFixture(capture.SysCapture) except TypeError: capsys = capture.CaptureFixture(capture.SysCapture, None) getattr(capsys, '_start')() # Re-run httpretty on Windows (due to lack of forking). if sys.platform == 'win32': httpretty.enable() if cached_responses: for url, body in cached_responses.items(): httpretty.register_uri(httpretty.GET, url, body=body) # Run. result = run_build_main(docs_dir, html_dir, overflow) stdout, stderr = capsys.readouterr() queue.put((stdout, stderr)) if result != 0: raise RuntimeError(result, stdout, stderr) def pytest_namespace(): """Add objects to the pytest namespace. E.g. Returning {'func': lambda: True} allows import pytest; assert pytest.func() is True. :return: Namespace names and objects. :rtype: dict """ def add_page(root, name, append=''): """Add a page to the sample Sphinx docs. :param py.path.local root: Path to docs root dir. :param str name: Page name. :param str append: Append text to RST document body. :return: Path to new page RST file. :rtype: py.path.local """ root.join('contents.rst').write(' {}\n'.format(name), mode='a') page = root.join('{}.rst'.format(name)) page.write('.. _{}:\n\n{}\n{}\n\n{}'.format(name, name.capitalize(), '=' * len(name), append)) return page def build_isolated(docs_dir, html_dir, cached_responses, overflow=None): """Run build_main() through multiprocessing.Process. :param str docs_dir: Path to input docs directory. :param str html_dir: Path to output html directory. :param dict cached_responses: URL keys and serialized JSON values. :param iter overflow: Append these args to sphinx-build call. :return: Exit code of subprocess, stdout, and stderr. :rtype: tuple """ queue = multiprocessing.Queue() args = docs_dir, html_dir, cached_responses, queue, overflow child = multiprocessing.Process(target=run_build_main_post_multiprocessing, args=args) child.start() child.join() result = child.exitcode try: stdout, stderr = queue.get(False) except multiprocessing.queues.Empty: stdout, stderr = '', '' return result, stdout, stderr return dict(add_page=add_page, build_isolated=build_isolated) @pytest.fixture def docs(tmpdir): """Create sample docs used in this test module. :param tmpdir: pytest fixture. :return: Path to docs root. :rtype: py.path """ root = tmpdir.ensure_dir('docs') # Create Sphinx config. root.join('conf.py').write("extensions = ['sphinxcontrib.imgur']\nimgur_client_id = 'a0b1c2d3e4f56789'\n") # Create Sphinx docs. root.join('contents.rst').write( 'Test\n' '====\n' '\n' 'Sample documentation.\n' '\n' '.. toctree::\n' ' ignore\n' ) root.join('ignore.rst').write('.. _ignore:\n\nIgnore\n======\n\nHello World.\n') return root
[ "\"\"\"Fixtures for tests in this directory.\"\"\"\n", "\n", "import multiprocessing\n", "import multiprocessing.queues\n", "import sys\n", "\n", "import httpretty\n", "import pytest\n", "from sphinx import build_main\n", "\n", "\n", "def run_build_main(docs_dir, html_dir, overflow):\n", " \"\"\"Run build_main().\n", "\n", " :param str docs_dir: Path to input docs directory.\n", " :param str html_dir: Path to output html directory.\n", " :param iter overflow: Append these args to sphinx-build call.\n", "\n", " :return: Value from build_main().\n", " :rtype: int\n", " \"\"\"\n", " argv = ('sphinx-build', str(docs_dir), str(html_dir))\n", " if overflow:\n", " argv += overflow\n", " result = build_main(argv)\n", " return result\n", "\n", "\n", "def run_build_main_post_multiprocessing(docs_dir, html_dir, cached_responses, queue, overflow):\n", " \"\"\"Run Sphinx's build_main after setting up httpretty mock responses. Called by multiprocess.Process.\n", "\n", " Need to use this instead of httpretty pytest fixtures since forking doesn't exist in Windows and multiprocess runs\n", " in \"spawn\" mode. This means that everything setup by pytest is lost since subprocesses are generated from scratch on\n", " Windows.\n", "\n", " :raise: RuntimeError on Sphinx non-zero exit. This causes multiprocessing.Process().exitcode to be != 0.\n", "\n", " :param str docs_dir: Path to input docs directory.\n", " :param str html_dir: Path to output html directory.\n", " :param dict cached_responses: URL keys and serialized JSON values.\n", " :param multiprocessing.queues.Queue queue: Queue to transmit stdout/err back to parent process.\n", " :param iter overflow: Append these args to sphinx-build call.\n", " \"\"\"\n", " # Capture stdout/stderr after forking/spawning.\n", " capture = __import__('_pytest').capture\n", " try:\n", " capsys = capture.CaptureFixture(capture.SysCapture)\n", " except TypeError:\n", " capsys = capture.CaptureFixture(capture.SysCapture, None)\n", " getattr(capsys, '_start')()\n", "\n", " # Re-run httpretty on Windows (due to lack of forking).\n", " if sys.platform == 'win32':\n", " httpretty.enable()\n", " if cached_responses:\n", " for url, body in cached_responses.items():\n", " httpretty.register_uri(httpretty.GET, url, body=body)\n", "\n", " # Run.\n", " result = run_build_main(docs_dir, html_dir, overflow)\n", " stdout, stderr = capsys.readouterr()\n", " queue.put((stdout, stderr))\n", " if result != 0:\n", " raise RuntimeError(result, stdout, stderr)\n", "\n", "\n", "def pytest_namespace():\n", " \"\"\"Add objects to the pytest namespace.\n", "\n", " E.g. Returning {'func': lambda: True} allows import pytest; assert pytest.func() is True.\n", "\n", " :return: Namespace names and objects.\n", " :rtype: dict\n", " \"\"\"\n", " def add_page(root, name, append=''):\n", " \"\"\"Add a page to the sample Sphinx docs.\n", "\n", " :param py.path.local root: Path to docs root dir.\n", " :param str name: Page name.\n", " :param str append: Append text to RST document body.\n", "\n", " :return: Path to new page RST file.\n", " :rtype: py.path.local\n", " \"\"\"\n", " root.join('contents.rst').write(' {}\\n'.format(name), mode='a')\n", " page = root.join('{}.rst'.format(name))\n", " page.write('.. _{}:\\n\\n{}\\n{}\\n\\n{}'.format(name, name.capitalize(), '=' * len(name), append))\n", " return page\n", "\n", " def build_isolated(docs_dir, html_dir, cached_responses, overflow=None):\n", " \"\"\"Run build_main() through multiprocessing.Process.\n", "\n", " :param str docs_dir: Path to input docs directory.\n", " :param str html_dir: Path to output html directory.\n", " :param dict cached_responses: URL keys and serialized JSON values.\n", " :param iter overflow: Append these args to sphinx-build call.\n", "\n", " :return: Exit code of subprocess, stdout, and stderr.\n", " :rtype: tuple\n", " \"\"\"\n", " queue = multiprocessing.Queue()\n", " args = docs_dir, html_dir, cached_responses, queue, overflow\n", " child = multiprocessing.Process(target=run_build_main_post_multiprocessing, args=args)\n", " child.start()\n", " child.join()\n", " result = child.exitcode\n", " try:\n", " stdout, stderr = queue.get(False)\n", " except multiprocessing.queues.Empty:\n", " stdout, stderr = '', ''\n", " return result, stdout, stderr\n", "\n", " return dict(add_page=add_page, build_isolated=build_isolated)\n", "\n", "\n", "@pytest.fixture\n", "def docs(tmpdir):\n", " \"\"\"Create sample docs used in this test module.\n", "\n", " :param tmpdir: pytest fixture.\n", "\n", " :return: Path to docs root.\n", " :rtype: py.path\n", " \"\"\"\n", " root = tmpdir.ensure_dir('docs')\n", "\n", " # Create Sphinx config.\n", " root.join('conf.py').write(\"extensions = ['sphinxcontrib.imgur']\\nimgur_client_id = 'a0b1c2d3e4f56789'\\n\")\n", "\n", " # Create Sphinx docs.\n", " root.join('contents.rst').write(\n", " 'Test\\n'\n", " '====\\n'\n", " '\\n'\n", " 'Sample documentation.\\n'\n", " '\\n'\n", " '.. toctree::\\n'\n", " ' ignore\\n'\n", " )\n", " root.join('ignore.rst').write('.. _ignore:\\n\\nIgnore\\n======\\n\\nHello World.\\n')\n", "\n", " return root\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.010416666666666666, 0.009433962264150943, 0, 0.008403361344537815, 0.008264462809917356, 0, 0, 0.009174311926605505, 0, 0, 0, 0, 0.01, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.010638297872340425, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.009708737864077669, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.010526315789473684, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.009009009009009009, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0.011764705882352941, 0, 0 ]
142
0.000756
# Copyright 2017 Cisco Systems # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import logging LOG = logging.getLogger(__name__) # A list of features and their supported microversions. Note that these are # explicit functioning versions, not a range. # There should be a minimum of two versions per feature. The first entry in # this list should always be the lowest possible API microversion for a # feature i.e. the version at which that feature was introduced. The second # entry should be the current service version when the feature was added to # horizon. # Further documentation can be found at # https://docs.openstack.org/horizon/latest/contributor/topics/ # microversion_support.html MICROVERSION_FEATURES = { "nova": { "locked_attribute": ["2.9", "2.42"], "instance_description": ["2.19", "2.42"], "remote_console_mks": ["2.8", "2.53"], "servergroup_soft_policies": ["2.15", "2.60"], "servergroup_user_info": ["2.13", "2.60"], "multiattach": ["2.60"], "auto_allocated_network": ["2.37", "2.42"], "key_types": ["2.2", "2.9"], "key_type_list": ["2.9"], }, "cinder": { "groups": ["3.27", "3.43", "3.48", "3.58"], "consistency_groups": ["2.0", "3.10"], "message_list": ["3.5", "3.29"], "limits_project_id_query": ["3.43", "3.50", "3.55"], } } class MicroVersionNotFound(Exception): def __init__(self, features): self.features = features def __str__(self): return "Insufficient microversion for %s" % self.features def get_requested_versions(service, features): if not features: return None # Convert a single feature string into a list for backward compatibility. if isinstance(features, str): features = [features] try: service_features = MICROVERSION_FEATURES[service] except KeyError: LOG.debug("'%s' could not be found in the MICROVERSION_FEATURES dict", service) return None feature_versions = set(service_features[features[0]]) for feature in features[1:]: feature_versions &= set(service_features[feature]) if not feature_versions: return None # Sort version candidates from larger versins feature_versions = sorted(feature_versions, reverse=True, key=lambda v: [int(i) for i in v.split('.')]) return feature_versions # NOTE(robcresswell): Since each client implements their own wrapper class for # API objects, we'll need to allow that to be passed in. In the future this # should be replaced by some common handling in Oslo. def get_microversion_for_features(service, features, wrapper_class, min_ver, max_ver): """Retrieves that highest known functional microversion for features""" feature_versions = get_requested_versions(service, features) if not feature_versions: return None for version in feature_versions: microversion = wrapper_class(version) if microversion.matches(min_ver, max_ver): return microversion return None
[ "# Copyright 2017 Cisco Systems\n", "#\n", "# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n", "# not use this file except in compliance with the License. You may obtain\n", "# a copy of the License at\n", "#\n", "# http://www.apache.org/licenses/LICENSE-2.0\n", "#\n", "# Unless required by applicable law or agreed to in writing, software\n", "# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n", "# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n", "# License for the specific language governing permissions and limitations\n", "# under the License.\n", "\n", "import logging\n", "\n", "LOG = logging.getLogger(__name__)\n", "\n", "# A list of features and their supported microversions. Note that these are\n", "# explicit functioning versions, not a range.\n", "# There should be a minimum of two versions per feature. The first entry in\n", "# this list should always be the lowest possible API microversion for a\n", "# feature i.e. the version at which that feature was introduced. The second\n", "# entry should be the current service version when the feature was added to\n", "# horizon.\n", "# Further documentation can be found at\n", "# https://docs.openstack.org/horizon/latest/contributor/topics/\n", "# microversion_support.html\n", "MICROVERSION_FEATURES = {\n", " \"nova\": {\n", " \"locked_attribute\": [\"2.9\", \"2.42\"],\n", " \"instance_description\": [\"2.19\", \"2.42\"],\n", " \"remote_console_mks\": [\"2.8\", \"2.53\"],\n", " \"servergroup_soft_policies\": [\"2.15\", \"2.60\"],\n", " \"servergroup_user_info\": [\"2.13\", \"2.60\"],\n", " \"multiattach\": [\"2.60\"],\n", " \"auto_allocated_network\": [\"2.37\", \"2.42\"],\n", " \"key_types\": [\"2.2\", \"2.9\"],\n", " \"key_type_list\": [\"2.9\"],\n", " },\n", " \"cinder\": {\n", " \"groups\": [\"3.27\", \"3.43\", \"3.48\", \"3.58\"],\n", " \"consistency_groups\": [\"2.0\", \"3.10\"],\n", " \"message_list\": [\"3.5\", \"3.29\"],\n", " \"limits_project_id_query\": [\"3.43\", \"3.50\", \"3.55\"],\n", " }\n", "}\n", "\n", "\n", "class MicroVersionNotFound(Exception):\n", " def __init__(self, features):\n", " self.features = features\n", "\n", " def __str__(self):\n", " return \"Insufficient microversion for %s\" % self.features\n", "\n", "\n", "def get_requested_versions(service, features):\n", " if not features:\n", " return None\n", " # Convert a single feature string into a list for backward compatibility.\n", " if isinstance(features, str):\n", " features = [features]\n", " try:\n", " service_features = MICROVERSION_FEATURES[service]\n", " except KeyError:\n", " LOG.debug(\"'%s' could not be found in the MICROVERSION_FEATURES dict\",\n", " service)\n", " return None\n", "\n", " feature_versions = set(service_features[features[0]])\n", " for feature in features[1:]:\n", " feature_versions &= set(service_features[feature])\n", " if not feature_versions:\n", " return None\n", " # Sort version candidates from larger versins\n", " feature_versions = sorted(feature_versions, reverse=True,\n", " key=lambda v: [int(i) for i in v.split('.')])\n", " return feature_versions\n", "\n", "\n", "# NOTE(robcresswell): Since each client implements their own wrapper class for\n", "# API objects, we'll need to allow that to be passed in. In the future this\n", "# should be replaced by some common handling in Oslo.\n", "def get_microversion_for_features(service, features, wrapper_class,\n", " min_ver, max_ver):\n", " \"\"\"Retrieves that highest known functional microversion for features\"\"\"\n", " feature_versions = get_requested_versions(service, features)\n", " if not feature_versions:\n", " return None\n", "\n", " for version in feature_versions:\n", " microversion = wrapper_class(version)\n", " if microversion.matches(min_ver, max_ver):\n", " return microversion\n", " return None\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
96
0
import numpy from chainer.backends import cuda from chainer.functions.loss import black_out from chainer import link from chainer.utils import walker_alias from chainer import variable class BlackOut(link.Link): """BlackOut loss layer. .. seealso:: :func:`~chainer.functions.black_out` for more detail. Args: in_size (int): Dimension of input vectors. counts (int list): Number of each identifiers. sample_size (int): Number of negative samples. Attributes: W (~chainer.Parameter): Weight parameter matrix. """ sample_data = None def __init__(self, in_size, counts, sample_size): super(BlackOut, self).__init__() vocab_size = len(counts) p = numpy.array(counts, dtype=numpy.float32) self.sampler = walker_alias.WalkerAlias(p) self.sample_size = sample_size with self.init_scope(): self.W = variable.Parameter(shape=(vocab_size, in_size)) def to_cpu(self): super(BlackOut, self).to_cpu() self.sampler.to_cpu() def to_gpu(self, device=None): with cuda._get_device(device): super(BlackOut, self).to_gpu() self.sampler.to_gpu() def forward(self, x, t): """Computes the loss value for given input and ground truth labels. Args: x (~chainer.Variable): Input of the weight matrix multiplication. t (~chainer.Variable): Batch of ground truth labels. Returns: ~chainer.Variable: Loss value. """ batch_size = x.shape[0] if self.sample_data is not None: # for test sample_data = self.sample_data else: shape = (batch_size, self.sample_size) sample_data = self.sampler.sample(shape) samples = variable.Variable(sample_data) return black_out.black_out(x, t, self.W, samples)
[ "import numpy\n", "\n", "from chainer.backends import cuda\n", "from chainer.functions.loss import black_out\n", "from chainer import link\n", "from chainer.utils import walker_alias\n", "from chainer import variable\n", "\n", "\n", "class BlackOut(link.Link):\n", "\n", " \"\"\"BlackOut loss layer.\n", "\n", " .. seealso:: :func:`~chainer.functions.black_out` for more detail.\n", "\n", " Args:\n", " in_size (int): Dimension of input vectors.\n", " counts (int list): Number of each identifiers.\n", " sample_size (int): Number of negative samples.\n", "\n", " Attributes:\n", " W (~chainer.Parameter): Weight parameter matrix.\n", "\n", " \"\"\"\n", "\n", " sample_data = None\n", "\n", " def __init__(self, in_size, counts, sample_size):\n", " super(BlackOut, self).__init__()\n", " vocab_size = len(counts)\n", " p = numpy.array(counts, dtype=numpy.float32)\n", " self.sampler = walker_alias.WalkerAlias(p)\n", " self.sample_size = sample_size\n", "\n", " with self.init_scope():\n", " self.W = variable.Parameter(shape=(vocab_size, in_size))\n", "\n", " def to_cpu(self):\n", " super(BlackOut, self).to_cpu()\n", " self.sampler.to_cpu()\n", "\n", " def to_gpu(self, device=None):\n", " with cuda._get_device(device):\n", " super(BlackOut, self).to_gpu()\n", " self.sampler.to_gpu()\n", "\n", " def forward(self, x, t):\n", " \"\"\"Computes the loss value for given input and ground truth labels.\n", "\n", " Args:\n", " x (~chainer.Variable): Input of the weight matrix multiplication.\n", " t (~chainer.Variable): Batch of ground truth labels.\n", "\n", " Returns:\n", " ~chainer.Variable: Loss value.\n", "\n", " \"\"\"\n", "\n", " batch_size = x.shape[0]\n", " if self.sample_data is not None:\n", " # for test\n", " sample_data = self.sample_data\n", " else:\n", " shape = (batch_size, self.sample_size)\n", " sample_data = self.sampler.sample(shape)\n", " samples = variable.Variable(sample_data)\n", " return black_out.black_out(x, t, self.W, samples)\n" ]
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
67
0
README.md exists but content is empty. Use the Edit dataset card button to edit it.
Downloads last month
0
Edit dataset card

Models trained or fine-tuned on kejian/codeparrot-train-more-filter-3.3b-cleaned