repo_name
stringlengths
5
100
path
stringlengths
4
375
copies
stringlengths
1
5
size
stringlengths
4
7
content
stringlengths
666
1M
license
stringclasses
15 values
blitzmann/Pyfa
eos/saveddata/fighterAbility.py
1
5887
# =============================================================================== # Copyright (C) 2010 Diego Duclos # # This file is part of eos. # # eos is free software: you can redistribute it and/or modify # it under the terms of the GNU Lesser General Public License as published by # the Free Software Foundation, either version 2 of the License, or # (at your option) any later version. # # eos is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public License # along with eos. If not, see <http://www.gnu.org/licenses/>. # =============================================================================== from logbook import Logger from sqlalchemy.orm import reconstructor from eos.utils.stats import DmgTypes pyfalog = Logger(__name__) class FighterAbility(object): # We aren't able to get data on the charges that can be stored with fighters. So we hardcode that data here, keyed # with the fighter squadron role NUM_SHOTS_MAPPING = { 1: 0, # Superiority fighter / Attack 2: 12, # Light fighter / Attack 4: 6, # Heavy fighter / Heavy attack 5: 3, # Heavy fighter / Long range attack } # Same as above REARM_TIME_MAPPING = { 1: 0, # Superiority fighter / Attack 2: 4000, # Light fighter / Attack 4: 6000, # Heavy fighter / Heavy attack 5: 20000, # Heavy fighter / Long range attack } def __init__(self, effect): """Initialize from the program""" self.__effect = effect self.effectID = effect.ID if effect is not None else None self.active = False self.build() @reconstructor def init(self): """Initialize from the database""" self.__effect = None if self.effectID: self.__effect = next((x for x in self.fighter.item.effects.values() if x.ID == self.effectID), None) if self.__effect is None: pyfalog.error("Effect (id: {0}) does not exist", self.effectID) return self.build() def build(self): pass @property def effect(self): return self.__effect @property def name(self): return self.__effect.getattr('displayName') or self.__effect.name @property def attrPrefix(self): return self.__effect.getattr('prefix') @property def dealsDamage(self): attr = "{}DamageMultiplier".format(self.attrPrefix) return attr in self.fighter.itemModifiedAttributes or self.fighter.charge is not None @property def grouped(self): # is the ability applied per fighter (webs, returns False), or as a group (MWD, returned True) return self.__effect.getattr('grouped') @property def hasCharges(self): return self.__effect.getattr('hasCharges') @property def reloadTime(self): rearm_time = (self.REARM_TIME_MAPPING[self.fighter.getModifiedItemAttr("fighterSquadronRole")] or 0 if self.hasCharges else 0) return self.fighter.getModifiedItemAttr("fighterRefuelingTime") + rearm_time * self.numShots @property def numShots(self): return self.NUM_SHOTS_MAPPING[self.fighter.getModifiedItemAttr("fighterSquadronRole")] or 0 if self.hasCharges else 0 @property def cycleTime(self): speed = self.fighter.getModifiedItemAttr("{}Duration".format(self.attrPrefix)) # Factor in reload ''' reload = self.reloadTime if self.fighter.owner.factorReload: numShots = self.numShots # Speed here already takes into consideration reactivation time speed = (speed * numShots + reload) / numShots if numShots > 0 else speed ''' return speed def getVolley(self, targetResists=None): if not self.dealsDamage or not self.active: return DmgTypes(0, 0, 0, 0) if self.attrPrefix == "fighterAbilityLaunchBomb": em = self.fighter.getModifiedChargeAttr("emDamage", 0) therm = self.fighter.getModifiedChargeAttr("thermalDamage", 0) kin = self.fighter.getModifiedChargeAttr("kineticDamage", 0) exp = self.fighter.getModifiedChargeAttr("explosiveDamage", 0) else: em = self.fighter.getModifiedItemAttr("{}DamageEM".format(self.attrPrefix), 0) therm = self.fighter.getModifiedItemAttr("{}DamageTherm".format(self.attrPrefix), 0) kin = self.fighter.getModifiedItemAttr("{}DamageKin".format(self.attrPrefix), 0) exp = self.fighter.getModifiedItemAttr("{}DamageExp".format(self.attrPrefix), 0) dmgMult = self.fighter.amountActive * self.fighter.getModifiedItemAttr("{}DamageMultiplier".format(self.attrPrefix), 1) volley = DmgTypes( em=em * dmgMult * (1 - getattr(targetResists, "emAmount", 0)), thermal=therm * dmgMult * (1 - getattr(targetResists, "thermalAmount", 0)), kinetic=kin * dmgMult * (1 - getattr(targetResists, "kineticAmount", 0)), explosive=exp * dmgMult * (1 - getattr(targetResists, "explosiveAmount", 0))) return volley def getDps(self, targetResists=None): volley = self.getVolley(targetResists=targetResists) if not volley: return DmgTypes(0, 0, 0, 0) dpsFactor = 1 / (self.cycleTime / 1000) dps = DmgTypes( em=volley.em * dpsFactor, thermal=volley.thermal * dpsFactor, kinetic=volley.kinetic * dpsFactor, explosive=volley.explosive * dpsFactor) return dps def clear(self): self.__dps = None self.__volley = None
gpl-3.0
davidwaroquiers/pymatgen
pymatgen/io/abinit/tests/test_inputs.py
5
10923
# coding: utf-8 # Copyright (c) Pymatgen Development Team. # Distributed under the terms of the MIT License. import os import tempfile import unittest import numpy as np from pymatgen.core.structure import Structure from pymatgen.io.abinit.inputs import ( BasicAbinitInput, BasicMultiDataset, ShiftMode, calc_shiftk, ebands_input, gs_input, ion_ioncell_relax_input, num_valence_electrons, ) from pymatgen.util.testing import PymatgenTest _test_dir = os.path.join(os.path.dirname(__file__), "..", "..", "..", "..", "test_files", "abinit") def abiref_file(filename): """Return absolute path to filename in ~pymatgen/test_files/abinit""" return os.path.join(_test_dir, filename) def abiref_files(*filenames): """Return list of absolute paths to filenames in ~pymatgen/test_files/abinit""" return [os.path.join(_test_dir, f) for f in filenames] class AbinitInputTestCase(PymatgenTest): """Unit tests for BasicAbinitInput.""" def test_api(self): """Testing BasicAbinitInput API.""" # Build simple input with structure and pseudos unit_cell = { "acell": 3 * [10.217], "rprim": [[0.0, 0.5, 0.5], [0.5, 0.0, 0.5], [0.5, 0.5, 0.0]], "ntypat": 1, "znucl": [14], "natom": 2, "typat": [1, 1], "xred": [[0.0, 0.0, 0.0], [0.25, 0.25, 0.25]], } inp = BasicAbinitInput(structure=unit_cell, pseudos=abiref_file("14si.pspnc")) shiftk = [[0.5, 0.5, 0.5], [0.5, 0.0, 0.0], [0.0, 0.5, 0.0], [0.0, 0.0, 0.5]] self.assertArrayEqual(calc_shiftk(inp.structure), shiftk) assert num_valence_electrons(inp.structure, inp.pseudos) == 8 repr(inp), str(inp) assert len(inp) == 0 and not inp assert inp.get("foo", "bar") == "bar" and inp.pop("foo", "bar") == "bar" assert inp.comment is None inp.set_comment("This is a comment") assert inp.comment == "This is a comment" assert inp.isnc and not inp.ispaw inp["ecut"] = 1 assert inp.get("ecut") == 1 and len(inp) == 1 and "ecut" in inp.keys() and "foo" not in inp # Test to_string assert inp.to_string(with_structure=True, with_pseudos=True) assert inp.to_string(with_structure=False, with_pseudos=False) inp.set_vars(ecut=5, toldfe=1e-6) assert inp["ecut"] == 5 inp.set_vars_ifnotin(ecut=-10) assert inp["ecut"] == 5 _, tmpname = tempfile.mkstemp(text=True) inp.write(filepath=tmpname) # Cannot change structure variables directly. with self.assertRaises(inp.Error): inp.set_vars(unit_cell) with self.assertRaises(TypeError): inp.add_abiobjects({}) with self.assertRaises(KeyError): inp.remove_vars("foo", strict=True) assert not inp.remove_vars("foo", strict=False) # Test deepcopy and remove_vars. inp["bdgw"] = [1, 2] inp_copy = inp.deepcopy() inp_copy["bdgw"][1] = 3 assert inp["bdgw"] == [1, 2] assert inp.remove_vars("bdgw") and "bdgw" not in inp removed = inp.pop_tolerances() assert len(removed) == 1 and removed["toldfe"] == 1e-6 # Test set_spin_mode old_vars = inp.set_spin_mode("polarized") assert "nsppol" in inp and inp["nspden"] == 2 and inp["nspinor"] == 1 inp.set_vars(old_vars) # Test set_structure new_structure = inp.structure.copy() new_structure.perturb(distance=0.1) inp.set_structure(new_structure) assert inp.structure == new_structure # Compatible with Pickle and MSONable? self.serialize_with_pickle(inp, test_eq=False) def test_input_errors(self): """Testing typical BasicAbinitInput Error""" si_structure = Structure.from_file(abiref_file("si.cif")) # Ambiguous list of pseudos. with self.assertRaises(BasicAbinitInput.Error): BasicAbinitInput(si_structure, pseudos=abiref_files("14si.pspnc", "14si.4.hgh")) # Pseudos do not match structure. with self.assertRaises(BasicAbinitInput.Error): BasicAbinitInput(si_structure, pseudos=abiref_file("H-wdr.oncvpsp")) si1_negative_volume = dict( ntypat=1, natom=1, typat=[1], znucl=14, acell=3 * [7.60], rprim=[[0.0, 0.5, 0.5], [-0.5, -0.0, -0.5], [0.5, 0.5, 0.0]], xred=[[0.0, 0.0, 0.0]], ) # Negative triple product. with self.assertRaises(BasicAbinitInput.Error): BasicAbinitInput(si1_negative_volume, pseudos=abiref_files("14si.pspnc")) def test_helper_functions(self): """Testing BasicAbinitInput helper functions.""" inp = BasicAbinitInput(structure=abiref_file("si.cif"), pseudos="14si.pspnc", pseudo_dir=_test_dir) inp.set_kmesh(ngkpt=(1, 2, 3), shiftk=(1, 2, 3, 4, 5, 6)) assert inp["kptopt"] == 1 and inp["nshiftk"] == 2 inp.set_gamma_sampling() assert inp["kptopt"] == 1 and inp["nshiftk"] == 1 assert np.all(inp["shiftk"] == 0) inp.set_kpath(ndivsm=3, kptbounds=None) assert inp["ndivsm"] == 3 and inp["iscf"] == -2 and len(inp["kptbounds"]) == 12 class TestMultiDataset(PymatgenTest): """Unit tests for BasicMultiDataset.""" def test_api(self): """Testing BasicMultiDataset API.""" structure = Structure.from_file(abiref_file("si.cif")) pseudo = abiref_file("14si.pspnc") pseudo_dir = os.path.dirname(pseudo) multi = BasicMultiDataset(structure=structure, pseudos=pseudo) with self.assertRaises(ValueError): BasicMultiDataset(structure=structure, pseudos=pseudo, ndtset=-1) multi = BasicMultiDataset(structure=structure, pseudos=pseudo, pseudo_dir=pseudo_dir) assert len(multi) == 1 and multi.ndtset == 1 assert multi.isnc for i, inp in enumerate(multi): assert list(inp.keys()) == list(multi[i].keys()) multi.addnew_from(0) assert multi.ndtset == 2 and multi[0] is not multi[1] assert multi[0].structure == multi[1].structure assert multi[0].structure is not multi[1].structure multi.set_vars(ecut=2) assert all(inp["ecut"] == 2 for inp in multi) self.assertEqual(multi.get("ecut"), [2, 2]) multi[1].set_vars(ecut=1) assert multi[0]["ecut"] == 2 and multi[1]["ecut"] == 1 self.assertEqual(multi.get("ecut"), [2, 1]) self.assertEqual(multi.get("foo", "default"), ["default", "default"]) multi[1].set_vars(paral_kgb=1) assert "paral_kgb" not in multi[0] self.assertEqual(multi.get("paral_kgb"), [None, 1]) pert_structure = structure.copy() pert_structure.perturb(distance=0.1) assert structure != pert_structure assert multi.set_structure(structure) == multi.ndtset * [structure] assert all(s == structure for s in multi.structure) assert multi.has_same_structures multi[1].set_structure(pert_structure) assert multi[0].structure != multi[1].structure and multi[1].structure == pert_structure assert not multi.has_same_structures split = multi.split_datasets() assert len(split) == 2 and all(split[i] == multi[i] for i in range(multi.ndtset)) repr(multi) str(multi) assert multi.to_string(with_pseudos=False) tmpdir = tempfile.mkdtemp() filepath = os.path.join(tmpdir, "run.abi") inp.write(filepath=filepath) multi.write(filepath=filepath) new_multi = BasicMultiDataset.from_inputs([inp for inp in multi]) assert new_multi.ndtset == multi.ndtset assert new_multi.structure == multi.structure for old_inp, new_inp in zip(multi, new_multi): assert old_inp is not new_inp self.assertDictEqual(old_inp.as_dict(), new_inp.as_dict()) ref_input = multi[0] new_multi = BasicMultiDataset.replicate_input(input=ref_input, ndtset=4) assert new_multi.ndtset == 4 for inp in new_multi: assert ref_input is not inp self.assertDictEqual(ref_input.as_dict(), inp.as_dict()) # Compatible with Pickle and MSONable? self.serialize_with_pickle(multi, test_eq=False) class ShiftModeTest(PymatgenTest): def test_shiftmode(self): """Testing shiftmode""" gamma = ShiftMode.GammaCentered assert ShiftMode.from_object("G") == gamma assert ShiftMode.from_object(gamma) == gamma with self.assertRaises(TypeError): ShiftMode.from_object({}) class FactoryTest(PymatgenTest): def setUp(self): # Si ebands self.si_structure = Structure.from_file(abiref_file("si.cif")) self.si_pseudo = abiref_file("14si.pspnc") def test_gs_input(self): """Testing gs_input factory.""" inp = gs_input(self.si_structure, self.si_pseudo, kppa=10, ecut=10, spin_mode="polarized") str(inp) assert inp["nsppol"] == 2 assert inp["nband"] == 14 self.assertArrayEqual(inp["ngkpt"], [2, 2, 2]) def test_ebands_input(self): """Testing ebands_input factory.""" multi = ebands_input(self.si_structure, self.si_pseudo, kppa=10, ecut=2) str(multi) scf_inp, nscf_inp = multi.split_datasets() # Test dos_kppa and other options. multi_dos = ebands_input( self.si_structure, self.si_pseudo, nscf_nband=10, kppa=10, ecut=2, spin_mode="unpolarized", smearing=None, charge=2.0, dos_kppa=50, ) assert len(multi_dos) == 3 assert all(i["charge"] == 2 for i in multi_dos) self.assertEqual(multi_dos.get("nsppol"), [1, 1, 1]) self.assertEqual(multi_dos.get("iscf"), [None, -2, -2]) multi_dos = ebands_input( self.si_structure, self.si_pseudo, nscf_nband=10, kppa=10, ecut=2, spin_mode="unpolarized", smearing=None, charge=2.0, dos_kppa=[50, 100], ) assert len(multi_dos) == 4 self.assertEqual(multi_dos.get("iscf"), [None, -2, -2, -2]) str(multi_dos) def test_ion_ioncell_relax_input(self): """Testing ion_ioncell_relax_input factory.""" multi = ion_ioncell_relax_input(self.si_structure, self.si_pseudo, kppa=10, ecut=2) str(multi) ion_inp, ioncell_inp = multi.split_datasets() assert ion_inp["chksymbreak"] == 0 assert ion_inp["ionmov"] == 3 and ion_inp["optcell"] == 0 assert ioncell_inp["ionmov"] == 3 and ioncell_inp["optcell"] == 2
mit
yasoob/youtube-dl-GUI
youtube_dl/extractor/glide.py
64
1543
# coding: utf-8 from __future__ import unicode_literals from .common import InfoExtractor class GlideIE(InfoExtractor): IE_DESC = 'Glide mobile video messages (glide.me)' _VALID_URL = r'https?://share\.glide\.me/(?P<id>[A-Za-z0-9\-=_+]+)' _TEST = { 'url': 'http://share.glide.me/UZF8zlmuQbe4mr+7dCiQ0w==', 'md5': '4466372687352851af2d131cfaa8a4c7', 'info_dict': { 'id': 'UZF8zlmuQbe4mr+7dCiQ0w==', 'ext': 'mp4', 'title': "Damon's Glide message", 'thumbnail': r're:^https?://.*?\.cloudfront\.net/.*\.jpg$', } } def _real_extract(self, url): video_id = self._match_id(url) webpage = self._download_webpage(url, video_id) title = self._html_search_regex( r'<title>(.+?)</title>', webpage, 'title', default=None) or self._og_search_title(webpage) video_url = self._proto_relative_url(self._search_regex( r'<source[^>]+src=(["\'])(?P<url>.+?)\1', webpage, 'video URL', default=None, group='url')) or self._og_search_video_url(webpage) thumbnail = self._proto_relative_url(self._search_regex( r'<img[^>]+id=["\']video-thumbnail["\'][^>]+src=(["\'])(?P<url>.+?)\1', webpage, 'thumbnail url', default=None, group='url')) or self._og_search_thumbnail(webpage) return { 'id': video_id, 'title': title, 'url': video_url, 'thumbnail': thumbnail, }
mit
acshi/osf.io
website/util/rubeus.py
7
11153
# -*- coding: utf-8 -*- """Contains helper functions for generating correctly formatted hgrid list/folders. """ import logging import hurry.filesize from django.utils import timezone from framework import sentry from framework.auth.decorators import Auth from django.apps import apps from website import settings from website.util import paths from website.util import sanitize from website.settings import DISK_SAVING_MODE logger = logging.getLogger(__name__) FOLDER = 'folder' FILE = 'file' KIND = 'kind' # TODO: Validate the JSON schema, esp. for addons DEFAULT_PERMISSIONS = { 'view': True, 'edit': False, } def format_filesize(size): return hurry.filesize.size(size, system=hurry.filesize.alternative) def default_urls(node_api, short_name): return { 'fetch': u'{node_api}{addonshort}/hgrid/'.format(node_api=node_api, addonshort=short_name), 'upload': u'{node_api}{addonshort}/'.format(node_api=node_api, addonshort=short_name), } def to_hgrid(node, auth, **data): """Converts a node into a rubeus grid format :param Node node: the node to be parsed :param Auth auth: the user authorization object :returns: rubeus-formatted dict """ return NodeFileCollector(node, auth, **data).to_hgrid() def build_addon_root(node_settings, name, permissions=None, urls=None, extra=None, buttons=None, user=None, private_key=None, **kwargs): """Builds the root or "dummy" folder for an addon. :param addonNodeSettingsBase node_settings: Addon settings :param String name: Additional information for the folder title eg. Repo name for Github or bucket name for S3 :param dict or Auth permissions: Dictionary of permissions for the addon's content or Auth for use in node.can_X methods :param dict urls: Hgrid related urls :param String extra: Html to be appened to the addon folder name eg. Branch switcher for github :param list of dicts buttons: List of buttons to appear in HGrid row. Each dict must have 'text', a string that will appear on the button, and 'action', the name of a function in :param bool private_key: Used to check if information should be stripped from anonymous links :param dict kwargs: Any additional information to add to the root folder :return dict: Hgrid formatted dictionary for the addon root folder """ from website.util import check_private_key_for_anonymized_link permissions = permissions or DEFAULT_PERMISSIONS if name and not check_private_key_for_anonymized_link(private_key): name = u'{0}: {1}'.format(node_settings.config.full_name, name) else: name = node_settings.config.full_name if hasattr(node_settings.config, 'urls') and node_settings.config.urls: urls = node_settings.config.urls if urls is None: urls = default_urls(node_settings.owner.api_url, node_settings.config.short_name) forbid_edit = DISK_SAVING_MODE if node_settings.config.short_name == 'osfstorage' else False if isinstance(permissions, Auth): auth = permissions permissions = { 'view': node_settings.owner.can_view(auth), 'edit': (node_settings.owner.can_edit(auth) and not node_settings.owner.is_registration and not forbid_edit), } max_size = node_settings.config.max_file_size if user and 'high_upload_limit' in user.system_tags: max_size = node_settings.config.high_max_file_size ret = { 'provider': node_settings.config.short_name, 'addonFullname': node_settings.config.full_name, 'name': name, 'iconUrl': node_settings.config.icon_url, KIND: FOLDER, 'extra': extra, 'buttons': buttons, 'isAddonRoot': True, 'permissions': permissions, 'accept': { 'maxSize': max_size, 'acceptedFiles': node_settings.config.accept_extensions, }, 'urls': urls, 'isPointer': False, 'nodeId': node_settings.owner._id, 'nodeUrl': node_settings.owner.url, 'nodeApiUrl': node_settings.owner.api_url, } ret.update(kwargs) return ret def build_addon_button(text, action, title=''): """Builds am action button to be rendered in HGrid :param str text: A string or html to appear on the button itself :param str action: The name of the HGrid action for the button to call. The callback for the HGrid action must be defined as a member of HGrid.Actions :return dict: Hgrid formatted dictionary for custom buttons """ button = { 'text': text, 'action': action, } if title: button['attributes'] = 'title="{title}" data-toggle="tooltip" data-placement="right" '.format(title=title) return button def sort_by_name(hgrid_data): return_value = hgrid_data if hgrid_data is not None: return_value = sorted(hgrid_data, key=lambda item: item['name'].lower()) return return_value class NodeFileCollector(object): """A utility class for creating rubeus formatted node data""" def __init__(self, node, auth, **kwargs): NodeRelation = apps.get_model('osf.NodeRelation') self.node = node.child if isinstance(node, NodeRelation) else node self.auth = auth self.extra = kwargs self.can_view = self.node.can_view(auth) self.can_edit = self.node.can_edit(auth) and not self.node.is_registration def to_hgrid(self): """Return the Rubeus.JS representation of the node's file data, including addons and components """ root = self._serialize_node(self.node) return [root] def _collect_components(self, node, visited): rv = [] if not node.can_view(self.auth): return rv for child in node.get_nodes(is_deleted=False): if not child.can_view(self.auth): if child.primary: for desc in child.find_readable_descendants(self.auth): visited.append(desc.resolve()._id) rv.append(self._serialize_node(desc, visited=visited, parent=node)) elif child.resolve()._id not in visited: visited.append(child.resolve()._id) rv.append(self._serialize_node(child, visited=visited, parent=node)) return rv def _get_node_name(self, node): """Input node object, return the project name to be display. """ NodeRelation = apps.get_model('osf.NodeRelation') is_node_relation = isinstance(node, NodeRelation) node = node.child if is_node_relation else node can_view = node.can_view(auth=self.auth) if can_view: node_name = sanitize.unescape_entities(node.title) elif node.is_registration: node_name = u'Private Registration' elif node.is_fork: node_name = u'Private Fork' elif is_node_relation: node_name = u'Private Link' else: node_name = u'Private Component' return node_name def _serialize_node(self, node, visited=None, parent=None): """Returns the rubeus representation of a node folder. """ visited = visited or [] visited.append(node.resolve()._id) can_view = node.can_view(auth=self.auth) if can_view: children = self._collect_addons(node) + self._collect_components(node, visited) else: children = [] is_pointer = parent and parent.has_node_link_to(node) return { # TODO: Remove safe_unescape_html when mako html safe comes in 'name': self._get_node_name(node), 'category': node.category, 'kind': FOLDER, 'permissions': { 'edit': node.can_edit(self.auth) and not node.is_registration, 'view': can_view, }, 'urls': { 'upload': None, 'fetch': None, }, 'children': children, 'isPointer': is_pointer, 'isSmartFolder': False, 'nodeType': node.project_or_component, 'nodeID': node.resolve()._id, } def _collect_addons(self, node): rv = [] for addon in node.get_addons(): if addon.config.has_hgrid_files: # WARNING: get_hgrid_data can return None if the addon is added but has no credentials. try: temp = addon.config.get_hgrid_data(addon, self.auth, **self.extra) except Exception as e: logger.warn( getattr( e, 'data', 'Unexpected error when fetching file contents for {0}.'.format(addon.config.full_name) ) ) sentry.log_exception() rv.append({ KIND: FOLDER, 'unavailable': True, 'iconUrl': addon.config.icon_url, 'provider': addon.config.short_name, 'addonFullname': addon.config.full_name, 'permissions': {'view': False, 'edit': False}, 'name': '{} is currently unavailable'.format(addon.config.full_name), }) continue rv.extend(sort_by_name(temp) or []) return rv # TODO: these might belong in addons module def collect_addon_assets(node): """Return a dictionary containing lists of JS and CSS assets for a node's addons. :rtype: {'tree_js': <list of JS scripts>, 'tree_css': <list of CSS files>} """ return { 'tree_js': list(collect_addon_js(node)), 'tree_css': list(collect_addon_css(node)), } # TODO: Abstract static collectors def collect_addon_js(node, visited=None, filename='files.js', config_entry='files'): """Collect JavaScript includes for all add-ons implementing HGrid views. :return list: List of JavaScript include paths """ js = [] for addon_config in settings.ADDONS_AVAILABLE_DICT.values(): # JS modules configured in each addon's __init__ file js.extend(addon_config.include_js.get(config_entry, [])) # Webpack bundle js_path = paths.resolve_addon_path(addon_config, filename) if js_path: js.append(js_path) return js def collect_addon_css(node, visited=None): """Collect CSS includes for all addons-ons implementing Hgrid views. :return: List of CSS include paths :rtype: list """ css = [] for addon_config in settings.ADDONS_AVAILABLE_DICT.values(): # CSS modules configured in each addon's __init__ file css.extend(addon_config.include_css.get('files', [])) return css def delta_date(d): diff = d - timezone.now() s = diff.total_seconds() return s
apache-2.0
daafgo/Server_LRS
lrs/util/retrieve_statement.py
1
11615
import bencode import hashlib import json from datetime import datetime from itertools import chain from django.core.cache import cache from django.conf import settings from django.core.paginator import Paginator from django.db.models import Q from util import convert_to_utc, convert_to_dict from ..models import Statement, Agent from ..objects.AgentManager import AgentManager from ..exceptions import NotFound, IDNotFoundError MORE_ENDPOINT = '/xapi/statements/more/' def complex_get(param_dict, limit, language, format, attachments): # Tests if value is True or "true" voidQ = Q(voided=False) # keep track if a filter other than time or sequence is used reffilter = False sinceQ = Q() if 'since' in param_dict: sinceQ = Q(stored__gt=convert_to_utc(param_dict['since'])) untilQ = Q() if 'until' in param_dict: untilQ = Q(stored__lte=convert_to_utc(param_dict['until'])) # For statements/read/mine oauth scope authQ = Q() if 'auth' in param_dict and (param_dict['auth'] and 'statements_mine_only' in param_dict['auth']): q_auth = param_dict['auth']['authority'] # If oauth - set authority to look for as the user if q_auth.oauth_identifier: authQ = Q(authority=q_auth) | Q(authority=q_auth.get_user_from_oauth_group()) # Chain all of user's oauth clients as well else: oauth_clients = Agent.objects.filter(member__in=[q_auth]) authQ = Q(authority=q_auth) for client in oauth_clients: authQ = authQ | Q(authority=client.get_user_from_oauth_group()) agentQ = Q() if 'agent' in param_dict: reffilter = True agent = None data = param_dict['agent'] related = 'related_agents' in param_dict and param_dict['related_agents'] if not type(data) is dict: data = convert_to_dict(data) try: agent = AgentManager(data).Agent if agent.objectType == "Group": groups = [] else: groups = agent.member.all() agentQ = Q(actor=agent) for g in groups: agentQ = agentQ | Q(actor=g) if related: me = chain([agent], groups) for a in me: agentQ = agentQ | Q(object_agent=a) | Q(authority=a) \ | Q(context_instructor=a) | Q(context_team=a) \ | Q(object_substatement__actor=a) \ | Q(object_substatement__object_agent=a) \ | Q(object_substatement__context_instructor=a) \ | Q(object_substatement__context_team=a) except IDNotFoundError: return[] verbQ = Q() if 'verb' in param_dict: reffilter = True verbQ = Q(verb__verb_id=param_dict['verb']) # activity activityQ = Q() if 'activity' in param_dict: reffilter = True activityQ = Q(object_activity__activity_id=param_dict['activity']) if 'related_activities' in param_dict and param_dict['related_activities']: activityQ = activityQ | Q(statementcontextactivity__context_activity__activity_id=param_dict['activity']) \ | Q(object_substatement__object_activity__activity_id=param_dict['activity']) \ | Q(object_substatement__substatementcontextactivity__context_activity__activity_id=param_dict['activity']) registrationQ = Q() if 'registration' in param_dict: reffilter = True registrationQ = Q(context_registration=param_dict['registration']) # If want ordered by ascending stored_param = '-stored' if 'ascending' in param_dict and param_dict['ascending']: stored_param = 'stored' stmtset = Statement.objects.filter(voidQ & untilQ & sinceQ & authQ & agentQ & verbQ & activityQ & registrationQ) # only find references when a filter other than # since, until, or limit was used if reffilter: stmtset = findstmtrefs(stmtset.distinct(), sinceQ, untilQ) # Calculate limit of stmts to return return_limit = set_limit(limit) # If there are more stmts than the limit, need to break it up and return more id if stmtset.count() > return_limit: return initial_cache_return(stmtset, stored_param, return_limit, language, format, attachments) else: return create_stmt_result(stmtset, stored_param, language, format) def create_stmt_result(stmt_set, stored, language, format): stmt_result = {} # blows up if the idlist is empty... so i gotta check for that idlist = stmt_set.values_list('id', flat=True) if idlist > 0: if format == 'exact': stmt_result = '{"statements": [%s], "more": ""}' % ",".join([json.dumps(stmt.full_statement) for stmt in \ Statement.objects.filter(id__in=idlist).order_by(stored)]) else: stmt_result['statements'] = [stmt.to_dict(language, format) for stmt in \ Statement.objects.filter(id__in=idlist).order_by(stored)] stmt_result['more'] = "" else: stmt_result['statements'] = [] stmt_result['more'] = "" return stmt_result def findstmtrefs(stmtset, sinceQ, untilQ): if stmtset.count() == 0: return stmtset q = Q() for s in stmtset: q = q | Q(object_statementref__ref_id=s.statement_id) if sinceQ and untilQ: q = q & Q(sinceQ, untilQ) elif sinceQ: q = q & sinceQ elif untilQ: q = q & untilQ # finally weed out voided statements in this lookup q = q & Q(voided=False) return findstmtrefs(Statement.objects.filter(q).distinct(), sinceQ, untilQ) | stmtset def create_cache_key(stmt_list): # Create unique hash data to use for the cache key hash_data = [] hash_data.append(str(datetime.now())) hash_data.append(str(stmt_list)) # Create cache key from hashed data (always 32 digits) key = hashlib.md5(bencode.bencode(hash_data)).hexdigest() return key def initial_cache_return(stmt_list, stored, limit, language, format, attachments): # First time someone queries POST/GET result = {} cache_list = [] cache_list.append([s for s in stmt_list.order_by(stored).values_list('id', flat=True)]) stmt_pager = Paginator(cache_list[0], limit) # Always start on first page current_page = 1 total_pages = stmt_pager.num_pages # Create cache key from hashed data (always 32 digits) cache_key = create_cache_key(cache_list[0]) # Add data to cache cache_list.append(current_page) cache_list.append(total_pages) cache_list.append(limit) cache_list.append(attachments) cache_list.append(language) cache_list.append(format) cache_list.append(stored) # Encode data encoded_info = json.dumps(cache_list) # Save encoded_dict in cache cache.set(cache_key,encoded_info) # Return first page of results if format == 'exact': result = '{"statements": [%s], "more": "%s"}' % (",".join([json.dumps(stmt.full_statement) for stmt in \ Statement.objects.filter(id__in=stmt_pager.page(1).object_list).order_by(stored)]), MORE_ENDPOINT + cache_key) else: result['statements'] = [stmt.to_dict(language, format) for stmt in \ Statement.objects.filter(id__in=stmt_pager.page(1).object_list).order_by(stored)] result['more'] = MORE_ENDPOINT + cache_key return result def set_limit(req_limit): if not req_limit or req_limit > settings.SERVER_STMT_LIMIT: req_limit = settings.SERVER_STMT_LIMIT return req_limit def get_more_statement_request(req_id): # Retrieve encoded info for statements encoded_info = cache.get(req_id) # Could have expired or never existed if not encoded_info: raise NotFound("List does not exist - may have expired after 24 hours") # Decode info decoded_info = json.loads(encoded_info) # Info is always cached as [stmt_list, start_page, total_pages, limit, attachments, language, format] stmt_list = decoded_info[0] start_page = decoded_info[1] total_pages = decoded_info[2] limit = decoded_info[3] attachments = decoded_info[4] language = decoded_info[5] format = decoded_info[6] stored = decoded_info[7] # Build statementResult stmt_result = build_statement_result(stmt_list, start_page, total_pages, limit, attachments, language, format, stored, req_id) return stmt_result, attachments # Gets called from req_process after complex_get with list of django objects and also gets called from get_more_statement_request when # more_id is used so list will be serialized def build_statement_result(stmt_list, start_page, total_pages, limit, attachments, language, format, stored, more_id): result = {} current_page = start_page + 1 # If that was the last page to display then just return the remaining stmts if current_page == total_pages: stmt_pager = Paginator(stmt_list, limit) # Return first page of results if format == 'exact': result = '{"statements": [%s], "more": ""}' % ",".join([json.dumps(stmt.to_dict(language, format)) for stmt in \ Statement.objects.filter(id__in=stmt_pager.page(current_page).object_list).order_by(stored)]) else: result['statements'] = [stmt.to_dict(language, format) for stmt in \ Statement.objects.filter(id__in=stmt_pager.page(current_page).object_list).order_by(stored)] result['more'] = "" # Set current page back for when someone hits the URL again current_page -= 1 # Retrieve list stored in cache encoded_list = cache.get(more_id) # Decode info to set the current page back then encode again decoded_list = json.loads(encoded_list) decoded_list[1] = current_page encoded_list = json.dumps(decoded_list) cache.set(more_id, encoded_list) # There are more pages to display else: stmt_pager = Paginator(stmt_list, limit) # Create cache key from hashed data (always 32 digits) cache_key = create_cache_key(stmt_list) # Return first page of results if format == 'exact': result = '{"statements": [%s], "more": "%s"}' % (",".join([json.dumps(stmt.to_dict(language, format)) for stmt in \ Statement.objects.filter(id__in=stmt_pager.page(current_page).object_list).order_by(stored)]), MORE_ENDPOINT + cache_key) else: # Set result to have selected page of stmts and more endpoint result['statements'] = [stmt.to_dict(language, format) for stmt in \ Statement.objects.filter(id__in=stmt_pager.page(current_page).object_list).order_by(stored)] result['more'] = MORE_ENDPOINT + cache_key more_cache_list = [] # Increment next page start_page = current_page more_cache_list.append(stmt_list) more_cache_list.append(start_page) more_cache_list.append(total_pages) more_cache_list.append(limit) more_cache_list.append(attachments) more_cache_list.append(language) more_cache_list.append(format) more_cache_list.append(stored) # Encode info encoded_list = json.dumps(more_cache_list) cache.set(cache_key, encoded_list) return result
apache-2.0
RAtechntukan/CouchPotatoServer
libs/subliminal/services/tvsubtitles.py
106
6240
# -*- coding: utf-8 -*- # Copyright 2012 Nicolas Wack <wackou@gmail.com> # # This file is part of subliminal. # # subliminal is free software; you can redistribute it and/or modify it under # the terms of the GNU Lesser General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # subliminal is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public License # along with subliminal. If not, see <http://www.gnu.org/licenses/>. from . import ServiceBase from ..cache import cachedmethod from ..language import language_set, Language from ..subtitles import get_subtitle_path, ResultSubtitle from ..utils import get_keywords from ..videos import Episode from bs4 import BeautifulSoup import logging import re logger = logging.getLogger(__name__) def match(pattern, string): try: return re.search(pattern, string).group(1) except AttributeError: logger.debug(u'Could not match %r on %r' % (pattern, string)) return None class TvSubtitles(ServiceBase): server_url = 'http://www.tvsubtitles.net' api_based = False languages = language_set(['ar', 'bg', 'cs', 'da', 'de', 'el', 'en', 'es', 'fi', 'fr', 'hu', 'it', 'ja', 'ko', 'nl', 'pl', 'pt', 'ro', 'ru', 'sv', 'tr', 'uk', 'zh', 'pt-br']) #TODO: Find more exceptions language_map = {'gr': Language('gre'), 'cz': Language('cze'), 'ua': Language('ukr'), 'cn': Language('chi')} videos = [Episode] require_video = False required_features = ['permissive'] @cachedmethod def get_likely_series_id(self, name): r = self.session.post('%s/search.php' % self.server_url, data={'q': name}) soup = BeautifulSoup(r.content, self.required_features) maindiv = soup.find('div', 'left') results = [] for elem in maindiv.find_all('li'): sid = int(match('tvshow-([0-9]+)\.html', elem.a['href'])) show_name = match('(.*) \(', elem.a.text) results.append((show_name, sid)) #TODO: pick up the best one in a smart way result = results[0] return result[1] @cachedmethod def get_episode_id(self, series_id, season, number): """Get the TvSubtitles id for the given episode. Raises KeyError if none could be found.""" # download the page of the season, contains ids for all episodes episode_id = None r = self.session.get('%s/tvshow-%d-%d.html' % (self.server_url, series_id, season)) soup = BeautifulSoup(r.content, self.required_features) table = soup.find('table', id='table5') for row in table.find_all('tr'): cells = row.find_all('td') if not cells: continue episode_number = match('x([0-9]+)', cells[0].text) if not episode_number: continue episode_number = int(episode_number) episode_id = int(match('episode-([0-9]+)', cells[1].a['href'])) # we could just return the id of the queried episode, but as we # already downloaded the whole page we might as well fill in the # information for all the episodes of the season self.cache_for(self.get_episode_id, args=(series_id, season, episode_number), result=episode_id) # raises KeyError if not found return self.cached_value(self.get_episode_id, args=(series_id, season, number)) # Do not cache this method in order to always check for the most recent # subtitles def get_sub_ids(self, episode_id): subids = [] r = self.session.get('%s/episode-%d.html' % (self.server_url, episode_id)) epsoup = BeautifulSoup(r.content, self.required_features) for subdiv in epsoup.find_all('a'): if 'href' not in subdiv.attrs or not subdiv['href'].startswith('/subtitle'): continue subid = int(match('([0-9]+)', subdiv['href'])) lang = self.get_language(match('flags/(.*).gif', subdiv.img['src'])) result = {'subid': subid, 'language': lang} for p in subdiv.find_all('p'): if 'alt' in p.attrs and p['alt'] == 'rip': result['rip'] = p.text.strip() if 'alt' in p.attrs and p['alt'] == 'release': result['release'] = p.text.strip() subids.append(result) return subids def list_checked(self, video, languages): return self.query(video.path or video.release, languages, get_keywords(video.guess), video.series, video.season, video.episode) def query(self, filepath, languages, keywords, series, season, episode): logger.debug(u'Getting subtitles for %s season %d episode %d with languages %r' % (series, season, episode, languages)) self.init_cache() sid = self.get_likely_series_id(series.lower()) try: ep_id = self.get_episode_id(sid, season, episode) except KeyError: logger.debug(u'Could not find episode id for %s season %d episode %d' % (series, season, episode)) return [] subids = self.get_sub_ids(ep_id) # filter the subtitles with our queried languages subtitles = [] for subid in subids: language = subid['language'] if language not in languages: continue path = get_subtitle_path(filepath, language, self.config.multi) subtitle = ResultSubtitle(path, language, self.__class__.__name__.lower(), '%s/download-%d.html' % (self.server_url, subid['subid']), keywords=[subid['rip'], subid['release']]) subtitles.append(subtitle) return subtitles def download(self, subtitle): self.download_zip_file(subtitle.link, subtitle.path) return subtitle Service = TvSubtitles
gpl-3.0
ds-hwang/chromium-crosswalk
remoting/tools/zip2msi.py
89
8691
#!/usr/bin/env python # Copyright (c) 2012 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Generates .msi from a .zip archive or an unpacked directory. The structure of the input archive or directory should look like this: +- archive.zip +- archive +- parameters.json The name of the archive and the top level directory in the archive must match. When an unpacked directory is used as the input "archive.zip/archive" should be passed via the command line. 'parameters.json' specifies the parameters to be passed to candle/light and must have the following structure: { "defines": { "name": "value" }, "extensions": [ "WixFirewallExtension.dll" ], "switches": [ '-nologo' ], "source": "chromoting.wxs", "bind_path": "files", "sign": [ ... ], "candle": { ... }, "light": { ... } } "source" specifies the name of the input .wxs relative to "archive.zip/archive". "bind_path" specifies the path where to look for binary files referenced by .wxs relative to "archive.zip/archive". This script is used for both building Chromoting Host installation during Chromuim build and for signing Chromoting Host installation later. There are two copies of this script because of that: - one in Chromium tree at src/remoting/tools/zip2msi.py. - another one next to the signing scripts. The copies of the script can be out of sync so make sure that a newer version is compatible with the older ones when updating the script. """ import copy import json from optparse import OptionParser import os import re import subprocess import sys import zipfile def UnpackZip(target, source): """Unpacks |source| archive to |target| directory.""" target = os.path.normpath(target) archive = zipfile.ZipFile(source, 'r') for f in archive.namelist(): target_file = os.path.normpath(os.path.join(target, f)) # Sanity check to make sure .zip uses relative paths. if os.path.commonprefix([target_file, target]) != target: print "Failed to unpack '%s': '%s' is not under '%s'" % ( source, target_file, target) return 1 # Create intermediate directories. target_dir = os.path.dirname(target_file) if not os.path.exists(target_dir): os.makedirs(target_dir) archive.extract(f, target) return 0 def Merge(left, right): """Merges two values. Raises: TypeError: |left| and |right| cannot be merged. Returns: - if both |left| and |right| are dictionaries, they are merged recursively. - if both |left| and |right| are lists, the result is a list containing elements from both lists. - if both |left| and |right| are simple value, |right| is returned. - |TypeError| exception is raised if a dictionary or a list are merged with a non-dictionary or non-list correspondingly. """ if isinstance(left, dict): if isinstance(right, dict): retval = copy.copy(left) for key, value in right.iteritems(): if key in retval: retval[key] = Merge(retval[key], value) else: retval[key] = value return retval else: raise TypeError('Error: merging a dictionary and non-dictionary value') elif isinstance(left, list): if isinstance(right, list): return left + right else: raise TypeError('Error: merging a list and non-list value') else: if isinstance(right, dict): raise TypeError('Error: merging a dictionary and non-dictionary value') elif isinstance(right, list): raise TypeError('Error: merging a dictionary and non-dictionary value') else: return right quote_matcher_regex = re.compile(r'\s|"') quote_replacer_regex = re.compile(r'(\\*)"') def QuoteArgument(arg): """Escapes a Windows command-line argument. So that the Win32 CommandLineToArgv function will turn the escaped result back into the original string. See http://msdn.microsoft.com/en-us/library/17w5ykft.aspx ("Parsing C++ Command-Line Arguments") to understand why we have to do this. Args: arg: the string to be escaped. Returns: the escaped string. """ def _Replace(match): # For a literal quote, CommandLineToArgv requires an odd number of # backslashes preceding it, and it produces half as many literal backslashes # (rounded down). So we need to produce 2n+1 backslashes. return 2 * match.group(1) + '\\"' if re.search(quote_matcher_regex, arg): # Escape all quotes so that they are interpreted literally. arg = quote_replacer_regex.sub(_Replace, arg) # Now add unescaped quotes so that any whitespace is interpreted literally. return '"' + arg + '"' else: return arg def GenerateCommandLine(tool, source, dest, parameters): """Generates the command line for |tool|.""" # Merge/apply tool-specific parameters params = copy.copy(parameters) if tool in parameters: params = Merge(params, params[tool]) wix_path = os.path.normpath(params.get('wix_path', '')) switches = [os.path.join(wix_path, tool), '-nologo'] # Append the list of defines and extensions to the command line switches. for name, value in params.get('defines', {}).iteritems(): switches.append('-d%s=%s' % (name, value)) for ext in params.get('extensions', []): switches += ('-ext', os.path.join(wix_path, ext)) # Append raw switches switches += params.get('switches', []) # Append the input and output files switches += ('-out', dest, source) # Generate the actual command line #return ' '.join(map(QuoteArgument, switches)) return switches def Run(args): """Runs a command interpreting the passed |args| as a command line.""" command = ' '.join(map(QuoteArgument, args)) popen = subprocess.Popen( command, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) out, _ = popen.communicate() if popen.returncode: print command for line in out.splitlines(): print line print '%s returned %d' % (args[0], popen.returncode) return popen.returncode def GenerateMsi(target, source, parameters): """Generates .msi from the installation files prepared by Chromium build.""" parameters['basename'] = os.path.splitext(os.path.basename(source))[0] # The script can handle both forms of input a directory with unpacked files or # a ZIP archive with the same files. In the latter case the archive should be # unpacked to the intermediate directory. source_dir = None if os.path.isdir(source): # Just use unpacked files from the supplied directory. source_dir = source else: # Unpack .zip rc = UnpackZip(parameters['intermediate_dir'], source) if rc != 0: return rc source_dir = '%(intermediate_dir)s\\%(basename)s' % parameters # Read parameters from 'parameters.json'. f = open(os.path.join(source_dir, 'parameters.json')) parameters = Merge(json.load(f), parameters) f.close() if 'source' not in parameters: print 'The source .wxs is not specified' return 1 if 'bind_path' not in parameters: print 'The binding path is not specified' return 1 wxs = os.path.join(source_dir, parameters['source']) # Add the binding path to the light-specific parameters. bind_path = os.path.join(source_dir, parameters['bind_path']) parameters = Merge(parameters, {'light': {'switches': ['-b', bind_path]}}) target_arch = parameters['target_arch'] if target_arch == 'ia32': arch_param = 'x86' elif target_arch == 'x64': arch_param = 'x64' else: print 'Invalid target_arch parameter value' return 1 # Add the architecture to candle-specific parameters. parameters = Merge( parameters, {'candle': {'switches': ['-arch', arch_param]}}) # Run candle and light to generate the installation. wixobj = '%(intermediate_dir)s\\%(basename)s.wixobj' % parameters args = GenerateCommandLine('candle', wxs, wixobj, parameters) rc = Run(args) if rc: return rc args = GenerateCommandLine('light', wixobj, target, parameters) rc = Run(args) if rc: return rc return 0 def main(): usage = 'Usage: zip2msi [options] <input.zip> <output.msi>' parser = OptionParser(usage=usage) parser.add_option('--intermediate_dir', dest='intermediate_dir', default='.') parser.add_option('--wix_path', dest='wix_path', default='.') parser.add_option('--target_arch', dest='target_arch', default='x86') options, args = parser.parse_args() if len(args) != 2: parser.error('two positional arguments expected') return GenerateMsi(args[1], args[0], dict(options.__dict__)) if __name__ == '__main__': sys.exit(main())
bsd-3-clause
rolobio/DictORM
dictorm/pg.py
1
10974
""" This module is used to build queries for Postgresql. You shouldn't really need to import anything from this file because they can all be built using dictorm.Table. Sqlite queries are slightly different, but use these methods as their base. """ from copy import copy from typing import Union global sort_keys sort_keys = False __all__ = [ 'And', 'Column', 'Comparison', 'Delete', 'Insert', 'Null', 'Operator', 'Or', 'Select', 'set_sort_keys', 'Update', ] def set_sort_keys(val): """Used only for testing""" global sort_keys sort_keys = val class Select(object): query = 'SELECT * FROM "{table}"' def __init__(self, table, operators_or_comp=None, returning=None): self.table = table self.operators_or_comp = operators_or_comp or [] self.returning = returning self._order_by = None self._limit = None self._offset = None def __repr__(self): # pragma: no cover return 'Select({0}, {1}, ret:{2}, order:{3}, limit:{4}, offset:{5}'.format( self.table, repr(self.operators_or_comp), self.returning, self._order_by, self._limit, self._offset) def _copy(self): try: ooc = self.operators_or_comp[:] except TypeError: ooc = self.operators_or_comp._copy() new = type(self)(self.table, ooc, copy(self.returning)) new._order_by = copy(self._order_by) new._limit = copy(self._limit) new._offset = copy(self._offset) return new def __str__(self): parts = [] formats = {'table': self.table, } ooc = self.operators_or_comp if (isinstance(ooc, Operator) and ooc.operators_or_comp) or ( isinstance(ooc, Comparison) ): parts.append(' WHERE {comp}') formats['comp'] = str(ooc) if self._order_by: parts.append(' ORDER BY {0}'.format(str(self._order_by))) if self.returning == '*': parts.append(' RETURNING *') elif self.returning: parts.append(' RETURNING "{0}"'.format(str(self.returning))) if self._limit: parts.append(' LIMIT {0}'.format(str(self._limit))) if self._offset: parts.append(' OFFSET {0}'.format(str(self._offset))) sql = self.query + ''.join(parts) return sql.format(**formats) def values(self): return list(self.operators_or_comp or []) def build(self): return (str(self), self.values()) def order_by(self, order_by): self._order_by = order_by return self def limit(self, limit): self._limit = limit return self def offset(self, offset): self._offset = offset return self def __add__(self, item): self.operators_or_comp += item return self class Insert(object): query = 'INSERT INTO "{table}" {cvp}' cvp = '({0}) VALUES ({1})' interpolation_str = '%s' append_returning = None last_row = 'SELECT {0} FROM "{1}" WHERE "rowid" = last_insert_rowid()' def __init__(self, table, **values): self.table = table self._values = values self._returning = None self._ordered_keys = values.keys() if sort_keys: self._ordered_keys = sorted(self._ordered_keys) def _build_cvp(self): return (', '.join(['"{}"'.format(i) for i in self._ordered_keys]), ', '.join([self.interpolation_str, ] * len(self._values))) def __str__(self): sql = self.query if self._returning == '*': sql += ' RETURNING *' elif self._returning: sql += ' RETURNING "{0}"'.format(self._returning) if not self._values: return sql.format(table=self.table, cvp='DEFAULT VALUES') return sql.format(table=self.table, cvp=self.cvp.format(*self._build_cvp())) def values(self): return [self._values[k] for k in self._ordered_keys] def build(self): sql, values = str(self), self.values() if self.append_returning: ret = [(sql, values), ] ret.append((self.last_row.format( self.append_returning, self.table), [])) return ret return (sql, values) def returning(self, returning): self._returning = returning return self class Update(Insert): query = 'UPDATE "{table}" SET {cvp}' interpolation_str = '%s' def __init__(self, table, **values): self.operators_or_comp = None super(Update, self).__init__(table, **values) def _build_cvp(self): return ', '.join(('"{0}"={1}'.format(k, self.interpolation_str) \ for k in self._ordered_keys)) def __str__(self): parts = [] formats = {'table': self.table, 'cvp': self._build_cvp()} if self.operators_or_comp: parts.append(' WHERE {comps}') formats['comps'] = str(self.operators_or_comp) if self._returning == '*': parts.append(' RETURNING *') elif self._returning: parts.append(' RETURNING "{0}"'.format(self._returning)) sql = self.query + ''.join(parts) return sql.format(**formats) def values(self): values = super(Update, self).values() if self.operators_or_comp: values.extend(list(self.operators_or_comp)) return values def where(self, operators_or_comp): self.operators_or_comp = operators_or_comp return self class Delete(Update): query = 'DELETE FROM "{table}"' QueryHint = Union[Select, Insert, Update, Delete] class Comparison(object): interpolation_str = '%s' many = False def __init__(self, column1, column2, kind): self.column1 = column1 self.column2 = column2 self.kind = kind self._substratum = None self._aggregate = False self._array_exp = False def __repr__(self): # pragma: no cover if isinstance(self.column2, Null): ret = 'Comparison({0}{1})'.format(self.column1, self.kind) ret = 'Comparison{0}({1}{2}{3})'.format('Many' if self.many else '', self.column1, str(self.kind), self.column2) if self._substratum: ret += '.substratum({0})'.format(self._substratum) return ret def __str__(self): c1 = self.column1.column if self._null_kind(): return '"{0}"{1}'.format(c1, self.kind) # Surround the expression with parentheses if self._array_exp: return '"{0}"{1}({2})'.format(c1, self.kind, self.interpolation_str) return '"{0}"{1}{2}'.format(c1, self.kind, self.interpolation_str) def _copy(self): new = type(self)(self.column1, self.column2, self.kind) new._substratum = self._substratum new._aggregate = self._aggregate return new def value(self): return self.column2 def __iter__(self): if self._null_kind(): return iter([]) return iter([self.column2, ]) def substratum(self, column): comp = Comparison(self.column1, self.column2, self.kind) comp._substratum = column comp.many = self.many return comp def aggregate(self, column): comp = self.substratum(column) comp._aggregate = True return comp def _null_kind(self): return isinstance(self.column2, Null) def Or(self, comp2): return Or(self, comp2) def And(self, comp2): return And(self, comp2) class Null(): pass class Column(object): comparison = Comparison def __init__(self, table, column): self.table = table self.column = column def __repr__(self): # pragma: no cover return '{0}.{1}'.format(self.table, self.column) def many(self, column): c = self.comparison(self, column, '=') c.many = True return c def __eq__(self, column): return self.comparison(self, column, '=') def __gt__(self, column): return self.comparison(self, column, '>') def __ge__(self, column): return self.comparison(self, column, '>=') def __lt__(self, column): return self.comparison(self, column, '<') def __le__(self, column): return self.comparison(self, column, '<=') def __ne__(self, column): return self.comparison(self, column, '!=') def Is(self, column): return self.comparison(self, column, ' IS ') def IsNot(self, column): return self.comparison(self, column, ' IS NOT ') def IsDistinct(self, column): return self.comparison(self, column, ' IS DISTINCT FROM ') def IsNotDistinct(self, column): return self.comparison(self, column, ' IS NOT DISTINCT FROM ') def IsNull(self): return self.comparison(self, Null(), ' IS NULL') def IsNotNull(self): return self.comparison(self, Null(), ' IS NOT NULL') def In(self, tup): if isinstance(tup, list): tup = tuple(tup) return self.comparison(self, tup, ' IN ') def Like(self, column): return self.comparison(self, column, ' LIKE ') def Ilike(self, column): return self.comparison(self, column, ' ILIKE ') def Any(self, column): comp = self.comparison(self, column, ' = ANY ') comp._array_exp = True return comp def wrap_ooc(ooc): if isinstance(ooc, Comparison): return '%s' % str(ooc) return '(%s)' % str(ooc) class Operator(object): def __init__(self, kind, operators_or_comp): self.kind = kind self.operators_or_comp = operators_or_comp def __repr__(self): # pragma: no cover return '{0}{1}'.format(self.kind, repr(self.operators_or_comp)) def __str__(self): kind = ' {0} '.format(self.kind) return kind.join(map(wrap_ooc, self.operators_or_comp)) def __iter__(self): i = [] for comp in self.operators_or_comp: if isinstance(comp, Operator): i.extend(comp) elif isinstance(comp, Comparison) and not comp._null_kind(): i.append(comp.value()) return iter(i) def __add__(self, i): if isinstance(i, tuple): self.operators_or_comp += i else: self.operators_or_comp += (i,) return self def _copy(self): new = type(self)() new.operators_or_comp = tuple(i._copy() for i in self.operators_or_comp) return new class Or(Operator): def __init__(self, *operators_or_comp): super(Or, self).__init__('OR', operators_or_comp) class And(Operator): def __init__(self, *operators_or_comp): super(And, self).__init__('AND', operators_or_comp)
apache-2.0
megraf/asuswrt-merlin
release/src/router/libxml2/python/tests/xpath.py
87
1188
#!/usr/bin/python -u # # this test exercise the XPath basic engine, parser, etc, and # allows to detect memory leaks # import sys import libxml2 # Memory debug specific libxml2.debugMemory(1) doc = libxml2.parseFile("tst.xml") if doc.name != "tst.xml": print "doc.name error" sys.exit(1); ctxt = doc.xpathNewContext() res = ctxt.xpathEval("//*") if len(res) != 2: print "xpath query: wrong node set size" sys.exit(1) if res[0].name != "doc" or res[1].name != "foo": print "xpath query: wrong node set value" sys.exit(1) ctxt.setContextNode(res[0]) res = ctxt.xpathEval("foo") if len(res) != 1: print "xpath query: wrong node set size" sys.exit(1) if res[0].name != "foo": print "xpath query: wrong node set value" sys.exit(1) doc.freeDoc() ctxt.xpathFreeContext() i = 1000 while i > 0: doc = libxml2.parseFile("tst.xml") ctxt = doc.xpathNewContext() res = ctxt.xpathEval("//*") doc.freeDoc() ctxt.xpathFreeContext() i = i -1 del ctxt # Memory debug specific libxml2.cleanupParser() if libxml2.debugMemory(1) == 0: print "OK" else: print "Memory leak %d bytes" % (libxml2.debugMemory(1)) libxml2.dumpMemory()
gpl-2.0
mmatyas/skia
tools/misc_utils.py
68
7711
# Copyright 2014 Google Inc. # # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Module to host the VerboseSubprocess, ChangeDir, and ReSearch classes. """ import os import re import subprocess def print_subprocess_args(prefix, *args, **kwargs): """Print out args in a human-readable manner.""" def quote_and_escape(string): """Quote and escape a string if necessary.""" if ' ' in string or '\n' in string: string = '"%s"' % string.replace('"', '\\"') return string if 'cwd' in kwargs: print '%scd %s' % (prefix, kwargs['cwd']) print prefix + ' '.join(quote_and_escape(arg) for arg in args[0]) if 'cwd' in kwargs: print '%scd -' % prefix class VerboseSubprocess(object): """Call subprocess methods, but print out command before executing. Attributes: verbose: (boolean) should we print out the command or not. If not, this is the same as calling the subprocess method quiet: (boolean) suppress stdout on check_call and call. prefix: (string) When verbose, what to print before each command. """ def __init__(self, verbose): self.verbose = verbose self.quiet = not verbose self.prefix = '~~$ ' def check_call(self, *args, **kwargs): """Wrapper for subprocess.check_call(). Args: *args: to be passed to subprocess.check_call() **kwargs: to be passed to subprocess.check_call() Returns: Whatever subprocess.check_call() returns. Raises: OSError or subprocess.CalledProcessError: raised by check_call. """ if self.verbose: print_subprocess_args(self.prefix, *args, **kwargs) if self.quiet: with open(os.devnull, 'w') as devnull: return subprocess.check_call(*args, stdout=devnull, **kwargs) else: return subprocess.check_call(*args, **kwargs) def call(self, *args, **kwargs): """Wrapper for subprocess.check(). Args: *args: to be passed to subprocess.check_call() **kwargs: to be passed to subprocess.check_call() Returns: Whatever subprocess.call() returns. Raises: OSError or subprocess.CalledProcessError: raised by call. """ if self.verbose: print_subprocess_args(self.prefix, *args, **kwargs) if self.quiet: with open(os.devnull, 'w') as devnull: return subprocess.call(*args, stdout=devnull, **kwargs) else: return subprocess.call(*args, **kwargs) def check_output(self, *args, **kwargs): """Wrapper for subprocess.check_output(). Args: *args: to be passed to subprocess.check_output() **kwargs: to be passed to subprocess.check_output() Returns: Whatever subprocess.check_output() returns. Raises: OSError or subprocess.CalledProcessError: raised by check_output. """ if self.verbose: print_subprocess_args(self.prefix, *args, **kwargs) return subprocess.check_output(*args, **kwargs) def strip_output(self, *args, **kwargs): """Wrap subprocess.check_output and str.strip(). Pass the given arguments into subprocess.check_output() and return the results, after stripping any excess whitespace. Args: *args: to be passed to subprocess.check_output() **kwargs: to be passed to subprocess.check_output() Returns: The output of the process as a string without leading or trailing whitespace. Raises: OSError or subprocess.CalledProcessError: raised by check_output. """ if self.verbose: print_subprocess_args(self.prefix, *args, **kwargs) return str(subprocess.check_output(*args, **kwargs)).strip() def popen(self, *args, **kwargs): """Wrapper for subprocess.Popen(). Args: *args: to be passed to subprocess.Popen() **kwargs: to be passed to subprocess.Popen() Returns: The output of subprocess.Popen() Raises: OSError or subprocess.CalledProcessError: raised by Popen. """ if self.verbose: print_subprocess_args(self.prefix, *args, **kwargs) return subprocess.Popen(*args, **kwargs) class ChangeDir(object): """Use with a with-statement to temporarily change directories.""" # pylint: disable=I0011,R0903 def __init__(self, directory, verbose=False): self._directory = directory self._verbose = verbose def __enter__(self): if self._directory != os.curdir: if self._verbose: print '~~$ cd %s' % self._directory cwd = os.getcwd() os.chdir(self._directory) self._directory = cwd def __exit__(self, etype, value, traceback): if self._directory != os.curdir: if self._verbose: print '~~$ cd %s' % self._directory os.chdir(self._directory) class ReSearch(object): """A collection of static methods for regexing things.""" @staticmethod def search_within_stream(input_stream, pattern, default=None): """Search for regular expression in a file-like object. Opens a file for reading and searches line by line for a match to the regex and returns the parenthesized group named return for the first match. Does not search across newlines. For example: pattern = '^root(:[^:]*){4}:(?P<return>[^:]*)' with open('/etc/passwd', 'r') as stream: return search_within_file(stream, pattern) should return root's home directory (/root on my system). Args: input_stream: file-like object to be read pattern: (string) to be passed to re.compile default: what to return if no match Returns: A string or whatever default is """ pattern_object = re.compile(pattern) for line in input_stream: match = pattern_object.search(line) if match: return match.group('return') return default @staticmethod def search_within_string(input_string, pattern, default=None): """Search for regular expression in a string. Args: input_string: (string) to be searched pattern: (string) to be passed to re.compile default: what to return if no match Returns: A string or whatever default is """ match = re.search(pattern, input_string) return match.group('return') if match else default @staticmethod def search_within_output(verbose, pattern, default, *args, **kwargs): """Search for regular expression in a process output. Does not search across newlines. Args: verbose: (boolean) shoule we call print_subprocess_args? pattern: (string) to be passed to re.compile default: what to return if no match *args: to be passed to subprocess.Popen() **kwargs: to be passed to subprocess.Popen() Returns: A string or whatever default is """ if verbose: print_subprocess_args('~~$ ', *args, **kwargs) proc = subprocess.Popen(*args, stdout=subprocess.PIPE, **kwargs) return ReSearch.search_within_stream(proc.stdout, pattern, default)
bsd-3-clause
mbauskar/helpdesk-frappe
frappe/patches/v5_2/change_checks_to_not_null.py
58
1267
from __future__ import unicode_literals import frappe from frappe.utils import cint from frappe.model import default_fields def execute(): for table in frappe.db.get_tables(): doctype = table[3:] if frappe.db.exists("DocType", doctype): fieldnames = [df["fieldname"] for df in frappe.get_all("DocField", fields=["fieldname"], filters={"parent": doctype})] custom_fieldnames = [df["fieldname"] for df in frappe.get_all("Custom Field", fields=["fieldname"], filters={"dt": doctype})] else: fieldnames = custom_fieldnames = [] for column in frappe.db.sql("""desc `{0}`""".format(table), as_dict=True): if column["Type"]=="int(1)": fieldname = column["Field"] # only change for defined fields, ignore old fields that don't exist in meta if not (fieldname in default_fields or fieldname in fieldnames or fieldname in custom_fieldnames): continue # set 0 frappe.db.sql("""update `{table}` set `{column}`=0 where `{column}` is null"""\ .format(table=table, column=fieldname)) frappe.db.commit() # change definition frappe.db.sql_ddl("""alter table `{table}` modify `{column}` int(1) not null default {default}"""\ .format(table=table, column=fieldname, default=cint(column["Default"])))
mit
wiltonlazary/arangodb
3rdParty/V8/V8-5.0.71.39/tools/swarming_client/tests/lru_test.py
6
6265
#!/usr/bin/env python # Copyright 2013 The Swarming Authors. All rights reserved. # Use of this source code is governed under the Apache License, Version 2.0 that # can be found in the LICENSE file. import json import logging import os import sys import tempfile import unittest ROOT_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) sys.path.insert(0, ROOT_DIR) from utils import lru class LRUDictTest(unittest.TestCase): @staticmethod def prepare_lru_dict(keys): """Returns new LRUDict with given |keys| added one by one.""" lru_dict = lru.LRUDict() for key in keys: lru_dict.add(key, None) return lru_dict def assert_order(self, lru_dict, expected_keys): """Asserts order of keys in |lru_dict| is |expected_keys|. expected_keys[0] is supposedly oldest key, expected_keys[-1] is newest. Destroys |lru_dict| state in the process. """ # Check keys iteration works. self.assertEqual(lru_dict.keys_set(), set(expected_keys)) # Check pop_oldest returns keys in expected order. actual_keys = [] while lru_dict: oldest_key, _ = lru_dict.pop_oldest() actual_keys.append(oldest_key) self.assertEqual(actual_keys, expected_keys) def assert_same_data(self, lru_dict, regular_dict): """Asserts that given |lru_dict| contains same data as |regular_dict|.""" self.assertEqual(lru_dict.keys_set(), set(regular_dict.keys())) self.assertEqual(set(lru_dict.itervalues()), set(regular_dict.values())) for k, v in regular_dict.items(): self.assertEqual(lru_dict.get(k), v) def test_basic_dict_funcs(self): lru_dict = lru.LRUDict() # Add a bunch. data = {1: 'one', 2: 'two', 3: 'three'} for k, v in data.items(): lru_dict.add(k, v) # Check its there. self.assert_same_data(lru_dict, data) # Replace value. lru_dict.add(1, 'one!!!') data[1] = 'one!!!' self.assert_same_data(lru_dict, data) # Check pop works. self.assertEqual(lru_dict.pop(2), 'two') data.pop(2) self.assert_same_data(lru_dict, data) # Pop missing key. with self.assertRaises(KeyError): lru_dict.pop(2) # Touch has no effect on set of keys and values. lru_dict.touch(1) self.assert_same_data(lru_dict, data) # Touch fails on missing key. with self.assertRaises(KeyError): lru_dict.touch(22) def test_magic_methods(self): # Check __nonzero__, __len__ and __contains__ for empty dict. lru_dict = lru.LRUDict() self.assertFalse(lru_dict) self.assertEqual(len(lru_dict), 0) self.assertFalse(1 in lru_dict) # Dict with one item. lru_dict.add(1, 'one') self.assertTrue(lru_dict) self.assertEqual(len(lru_dict), 1) self.assertTrue(1 in lru_dict) self.assertFalse(2 in lru_dict) def test_order(self): data = [1, 2, 3] # Edge cases. self.assert_order(self.prepare_lru_dict([]), []) self.assert_order(self.prepare_lru_dict([1]), [1]) # No touches. self.assert_order(self.prepare_lru_dict(data), data) # Touching newest item is noop. lru_dict = self.prepare_lru_dict(data) lru_dict.touch(3) self.assert_order(lru_dict, data) # Touch to move to newest. lru_dict = self.prepare_lru_dict(data) lru_dict.touch(2) self.assert_order(lru_dict, [1, 3, 2]) # Pop newest. lru_dict = self.prepare_lru_dict(data) lru_dict.pop(1) self.assert_order(lru_dict, [2, 3]) # Pop in the middle. lru_dict = self.prepare_lru_dict(data) lru_dict.pop(2) self.assert_order(lru_dict, [1, 3]) # Pop oldest. lru_dict = self.prepare_lru_dict(data) lru_dict.pop(3) self.assert_order(lru_dict, [1, 2]) # Add oldest. lru_dict = self.prepare_lru_dict(data) lru_dict.batch_insert_oldest([(4, 4), (5, 5)]) self.assert_order(lru_dict, [4, 5] + data) # Add newest. lru_dict = self.prepare_lru_dict(data) lru_dict.add(4, 4) self.assert_order(lru_dict, data + [4]) def test_load_save(self): def save_and_load(lru_dict): handle, tmp_name = tempfile.mkstemp(prefix=u'lru_test') os.close(handle) try: lru_dict.save(tmp_name) return lru.LRUDict.load(tmp_name) finally: try: os.unlink(tmp_name) except OSError: pass data = [1, 2, 3] # Edge case. empty = save_and_load(lru.LRUDict()) self.assertFalse(empty) # Normal flow. lru_dict = save_and_load(self.prepare_lru_dict(data)) self.assert_order(lru_dict, data) # After touches. lru_dict = self.prepare_lru_dict(data) lru_dict.touch(2) lru_dict = save_and_load(lru_dict) self.assert_order(lru_dict, [1, 3, 2]) # After pop. lru_dict = self.prepare_lru_dict(data) lru_dict.pop(2) lru_dict = save_and_load(lru_dict) self.assert_order(lru_dict, [1, 3]) # After add. lru_dict = self.prepare_lru_dict(data) lru_dict.add(4, 4) lru_dict.batch_insert_oldest([(5, 5), (6, 6)]) lru_dict = save_and_load(lru_dict) self.assert_order(lru_dict, [5, 6] + data + [4]) def test_corrupted_state_file(self): def load_from_state(state_text): handle, tmp_name = tempfile.mkstemp(prefix=u'lru_test') os.close(handle) try: with open(tmp_name, 'w') as f: f.write(state_text) return lru.LRUDict.load(tmp_name) finally: os.unlink(tmp_name) # Loads correct state just fine. self.assertIsNotNone(load_from_state(json.dumps([ ['key1', 'value1'], ['key2', 'value2'], ]))) # Not a json. with self.assertRaises(ValueError): load_from_state('garbage, not a state') # Not a list. with self.assertRaises(ValueError): load_from_state('{}') # Not a list of pairs. with self.assertRaises(ValueError): load_from_state(json.dumps([ ['key', 'value', 'and whats this?'], ])) # Duplicate keys. with self.assertRaises(ValueError): load_from_state(json.dumps([ ['key', 'value'], ['key', 'another_value'], ])) if __name__ == '__main__': VERBOSE = '-v' in sys.argv logging.basicConfig(level=logging.DEBUG if VERBOSE else logging.ERROR) unittest.main()
apache-2.0
kkampardi/Plinth
plinth/modules/matrixsynapse/views.py
2
2670
# # This file is part of Plinth. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # """ Views for the Matrix Synapse module. """ from django.shortcuts import redirect from django.urls import reverse_lazy from django.views.generic import FormView from plinth import actions from plinth import views from plinth.modules import matrixsynapse from plinth.forms import DomainSelectionForm from plinth.utils import get_domain_names class SetupView(FormView): """Show matrix-synapse setup page.""" template_name = 'matrix-synapse-pre-setup.html' form_class = DomainSelectionForm success_url = reverse_lazy('matrixsynapse:index') def form_valid(self, form): """Handle valid form submission.""" domain_name = form.cleaned_data['domain_name'] actions.superuser_run('matrixsynapse', ['setup', '--domain-name', domain_name]) return super().form_valid(form) def get_context_data(self, *args, **kwargs): """Provide context data to the template.""" context = super().get_context_data(**kwargs) context['title'] = matrixsynapse.name context['description'] = matrixsynapse.description context['domain_names'] = get_domain_names() return context class ServiceView(views.ServiceView): """Show matrix-synapse service page.""" service_id = matrixsynapse.managed_services[0] template_name = 'matrix-synapse.html' description = matrixsynapse.description diagnostics_module_name = 'matrixsynapse' def dispatch(self, request, *args, **kwargs): """Redirect to setup page if setup is not done yet.""" if not matrixsynapse.is_setup(): return redirect('matrixsynapse:setup') return super().dispatch(request, *args, **kwargs) def get_context_data(self, *args, **kwargs): """Add additional context data for template.""" context = super().get_context_data(**kwargs) context['domain_name'] = matrixsynapse.get_configured_domain_name() return context
agpl-3.0
ZihengJiang/mxnet
python/mxnet/profiler.py
31
2161
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. # coding: utf-8 # pylint: disable=fixme, invalid-name, too-many-arguments, too-many-locals, too-many-lines # pylint: disable=too-many-branches, too-many-statements """Profiler setting methods.""" from __future__ import absolute_import import ctypes from .base import _LIB, check_call, c_str def profiler_set_config(mode='symbolic', filename='profile.json'): """Set up the configure of profiler. Parameters ---------- mode : string, optional Indicates whether to enable the profiler, can be 'symbolic', or 'all'. Defaults to `symbolic`. filename : string, optional The name of output trace file. Defaults to 'profile.json'. """ mode2int = {'symbolic': 0, 'all': 1} check_call(_LIB.MXSetProfilerConfig( ctypes.c_int(mode2int[mode]), c_str(filename))) def profiler_set_state(state='stop'): """Set up the profiler state to record operator. Parameters ---------- state : string, optional Indicates whether to run the profiler, can be 'stop' or 'run'. Default is `stop`. """ state2int = {'stop': 0, 'run': 1} check_call(_LIB.MXSetProfilerState(ctypes.c_int(state2int[state]))) def dump_profile(): """Dump profile and stop profiler. Use this to save profile in advance in case your program cannot exit normally.""" check_call(_LIB.MXDumpProfile())
apache-2.0
critiqjo/key-mon
src/keymon/options.py
15
14492
#!/usr/bin/python # # Copyright 2010 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Options Class for save, restoring and getting parameters from the command line. This provides a class which handles both saving options to disk and gathering options from the command line. It behaves a little like optparse in that you can get or set the attributes by name. It uses ConfigParser to save the variables to disk in ini format. """ __author__ = 'Scott Kirkwood (scott+keymon@forusers.com)' import ConfigParser import gettext import logging import optparse import os import sys LOG = logging.getLogger('options') gettext.install('key-mon', 'locale') class OptionException(Exception): pass class OptionItem(object): """Handles on option. It know both about optparse options and ConfigParser options. By setting opt_short, opt_long to None you won't create an optparse option. By setting ini_group, ini_name to None you won't create a ConfigParser option. """ def __init__(self, dest, _type, default, name, help, opt_group=None, opt_short=None, opt_long=None, ini_group=None, ini_name=None): """Create an option Args: dest: a unique name for this variable, used internally. _type: The data type. default: The default value if none given. name: the translated name. _help: Help text to show. opt_group: Optional option group opt_short: the short name of the option opt_long: the long name for the option ini_group: optional name of the group in the ini file. ini_name: optional name of the name in the ini file """ self._dirty = False self._value = None self._temp_value = None self._dest = dest self._type = _type self._default = default self._name = name self._help = help self._opt_group = opt_group self._opt_short = opt_short if self._opt_short and not self._opt_short.startswith('-'): raise OptionException('Invalid short option %s' % self._opt_short) self._opt_long = opt_long if self._opt_long and not self._opt_long.startswith('--'): raise OptionException('Invalid long option %r' % self._opt_long) self._ini_group = ini_group self._ini_name = ini_name if self._type not in ('int', 'float', 'bool', 'str'): raise OptionException('Unsupported type: %s' % self._type) self._set_value(default) def add_to_parser(self, parser): if not self._opt_short and not self._opt_long: return if self._type == 'bool': self._add_bool_to_parser(parser) return args = [] if self._opt_short: args.append(self._opt_short) if self._opt_long: args.append(self._opt_long) parser.add_option(dest=self._dest, type=self._type, default=self._default, help=self._help, *args) def _add_bool_to_parser(self, parser): """Booleans need special handling.""" args = [] if self._opt_short: args.append(self._opt_short) if self._opt_long: args.append(self._opt_long) parser.add_option(action='store_true', default=self._default, dest=self._dest, help=self._help, *args) if self._ini_group: # Only need the --no version if it could be saved to ini file. parser.add_option('--no' + self._opt_long.lstrip('-'), action='store_false', dest=self._dest, help=_('Opposite of %s') % self._opt_long) def set_from_optparse(self, opts, args): """Try and set an option from optparse. Args: opts: options as returned from parse_args() args: arguments as returned bys sys.args. """ if not self._opt_short and not self._opt_long: return # Was this option actually passed on the command line? found = False if args: for arg in args: if self._type == 'bool' and arg.startswith('--no'): arg = '--' + arg[4:] # Remove the --x=123, if any arg = arg.split('=')[0] if arg == self._opt_short or arg == self._opt_long: found = True break if hasattr(opts, self._dest): opt_val = getattr(opts, self._dest) if not self._ini_name: # For commands like --version which don't have stored values self._set_value(opt_val) if found: self._set_temp_value(opt_val) def reset_to_default(self): """Reset to the default value.""" self._set_value(self._default) self._set_temp_value(None) def get_value(self): """Return the value.""" if self._temp_value is not None: return self._temp_value return self._value def _set_attr_value(self, attr, val): """Set the value via attribute name. Args: attr: attribute name ('_value', or '_temp_value') val: value to set """ old_val = getattr(self, attr) if val is None: setattr(self, attr, val) elif self._type == 'int': setattr(self, attr, int(val)) elif self._type == 'float': setattr(self, attr, float(val)) elif self._type == 'bool': if isinstance(val, basestring): if val.lower() in ('false', 'off', 'no', '0'): setattr(self, attr, False) elif val.lower() in ('true', 'on', 'yes', '1'): setattr(self, attr, True) else: raise OptionException('Unable to convert %s to bool' % val) else: setattr(self, attr, bool(val)) else: setattr(self, attr, val) self._dirty = old_val != getattr(self, attr) def _set_value(self, val): self._set_attr_value('_value', val) self._set_attr_value('_temp_value', None) def _set_temp_value(self, val): self._set_attr_value('_temp_value', val) value = property(get_value, _set_value, doc="Value") @property def dest(self): """Destination variable name.""" return self._dest @property def name(self): """Localized name of the option.""" return self._name @property def help(self): """Long description of the option.""" return self._help @property def type(self): """String name of the type.""" return self._type @property def opt_group(self): """Option group, if any.""" return self._opt_group @property def opt_short(self): """Short option property (ex. '-v').""" return self._opt_short @property def opt_long(self): """Long option property (ex. '--verbose').""" return self._opt_long @property def ini_group(self): """Name of the ini group or None.""" return self._ini_group @property def ini_name(self): """Name in the ini, or None.""" return self._ini_name @property def ini_value(self): """Value to store in the ini, always a string.""" if self._value is None: return None if self._type == 'bool': if self._value is True: return '1' else: return '0' else: return str(self._value) class Options(object): """Store the options in memory, also saves to dist and creates opt_parser.""" def __init__(self): self._options = {} self._ini_filename = None self._opt_group = None self._opt_group_desc = {} self._options_order = [] def __getattr__(self, name): if name not in self.__dict__['_options']: raise AttributeError('Invalid attribute name: %r' % name) return self._options[name].value def __setattr__(self, name, value): if name == '_options' or name not in self.__dict__['_options']: object.__setattr__(self, name, value) else: LOG.info('Setting %r = %r', name, value) self.__dict__['_options'][name].value = value def add_option_group(self, group, desc): self._opt_group = group self._opt_group_desc[group] = desc def add_option(self, dest, type='str', default=None, name=None, help=None, opt_short=None, opt_long=None, ini_group=None, ini_name=None): """Create an option Args: dest: a unique name for this variable, used internally. type: The data type. default: The default value if none given. name: the translated name. help: Help text to show. opt_group: the name of the option group or None opt_short: the short name of the option opt_long: the long name for the option ini_group: the name of the group in the ini file. ini_name: the name of the name in the ini file """ if dest in self._options: raise OptionException('Options %s already added' % dest) self._options_order.append(dest) self._options[dest] = OptionItem(dest, type, default, name, help, opt_group=self._opt_group, opt_short=opt_short, opt_long=opt_long, ini_group=ini_group, ini_name=ini_name) def parse_args(self, desc, args=None): """Add the options to the optparse instance and parse command line Args: desc: Description to use for the program. args: Args for testing or sys.args[1:] otherwise """ parser = optparse.OptionParser(desc) for dest in self._options_order: opt = self._options[dest] opt.add_to_parser(parser) self._opt_ret, self._other_args = parser.parse_args(args) for opt in self._options.values(): opt.set_from_optparse(self._opt_ret, args) def parse_ini(self, fp): """Parser an ini file from fp, which is file-like class.""" config = ConfigParser.SafeConfigParser() config.readfp(fp) checker = {} for opt in self._options.values(): if opt.ini_group: checker[opt.ini_group + '-' + opt.ini_name] = True if (config.has_section(opt.ini_group) and config.has_option(opt.ini_group, opt.ini_name)): opt.value = config.get(opt.ini_group, opt.ini_name) LOG.info('From ini getting %s.%s = %s', opt.ini_group, opt.ini_name, opt.value) for section in config.sections(): for name, value in config.items(section): combined_name = section + '-' + name if not combined_name in checker: LOG.info('Unknown option %r in section [%s]', name, section) # we no longer throw an error to be backward compatible def write_ini(self, fp): """Parser an ini file from fp, which is file-like class.""" config = ConfigParser.SafeConfigParser() for opt in self._options.values(): if not opt.ini_group: continue if not config.has_section(opt.ini_group): config.add_section(opt.ini_group) if opt.ini_value is not None: config.set(opt.ini_group, opt.ini_name, opt.ini_value) config.write(fp) def read_ini_file(self, fname): self._ini_filename = os.path.expanduser(fname) LOG.info('Reading from %r', self._ini_filename) if os.path.exists(self._ini_filename) and os.path.isfile(self._ini_filename): fi = open(self._ini_filename) self.parse_ini(fi) fi.close() else: LOG.info('%r does not exist', self._ini_filename) def save(self): self._write_ini_file(self._ini_filename) def _make_dirs(self, fname): if not os.path.exists(fname): dirname = os.path.dirname(fname) if not os.path.exists(dirname): LOG.info('Creating directory %r', dirname) os.makedirs(dirname) def _write_ini_file(self, fname): self._make_dirs(fname) LOG.info('Writing config file %r', fname) fo = open(fname, 'w') self.write_ini(fo) fo.close() def reset_to_defaults(self): """Reset ini file to defaults.""" for opt in self._options.values(): if not opt.ini_group: continue opt.reset_to_default() if __name__ == '__main__': o = Options() o.add_option(opt_short='-l', opt_long='--larger', dest='larger', default=False, type='bool', help='Make the dialog 25% larger than normal.') o.add_option(opt_short='-m', opt_long='--meta', dest='meta', type='bool', ini_group='buttons', ini_name='meta', default=False, help='Show the meta (windows) key.') o.add_option(opt_long='--scale', dest='scale', type='float', default=1.0, ini_group='ui', ini_name='scale', help='Scale the dialog. ex. 2.0 is 2 times larger, 0.5 is ' 'half the size. Defaults to %default') o.add_option(opt_long='--kbdfile', dest='kbd_file', ini_group='devices', ini_name='map', default='us.kbd', help='Use this kbd filename instead running xmodmap.') o.add_option(opt_short='-v', opt_long='--version', dest='version', type='bool', help='Show version information and exit.') o.add_option(opt_short='-t', opt_long='--theme', dest='theme', type='str', ini_group='ui', ini_name='theme', help='The theme to use when drawing status images (ex. "-t apple").') o.add_option(opt_long='--list-themes', dest='list_themes', type='bool', help='List available themes') o.add_option(opt_long='--old-keys', dest='old_keys', type='int', ini_group='buttons', ini_name='old-keys', help='How many historical keypresses to show (defaults to %default)', default=0) o.add_option(opt_short=None, opt_long=None, type='int', dest='x_pos', default=-1, help='Last X Position', ini_group='position', ini_name='x') o.add_option_group('Developer Options', 'Don\'t use') o.add_option(opt_short='-d', opt_long='--debug', dest='debug', type='bool', help='Output debugging information.') lines = [] lines.append('[ui]') lines.append('scale = 1.0') lines.append('theme = classic') lines.append('[buttons]') lines.append('meta = 0') lines.append('old-keys = 0') lines.append('[devices]') lines.append('map = us.kbd') lines.append('[position]') lines.append('x = -1') import StringIO io = StringIO.StringIO('\n'.join(lines)) o.parse_ini(io) o.parse_args('%prog [options]', sys.argv) io = StringIO.StringIO() o.write_ini(io) print io.getvalue()
apache-2.0
maciekcc/tensorflow
tensorflow/python/ops/linalg_grad.py
23
9040
# Copyright 2015 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Gradients for operators defined in linalg_ops.py. Useful reference for derivative formulas is An extended collection of matrix derivative results for forward and reverse mode algorithmic differentiation by Mike Giles: http://eprints.maths.ox.ac.uk/1079/1/NA-08-01.pdf A detailed derivation of formulas for backpropagating through spectral layers (SVD and Eig) by Ionescu, Vantzos & Sminchisescu: https://arxiv.org/pdf/1509.07838v4.pdf """ from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow.python.framework import ops from tensorflow.python.ops import array_ops from tensorflow.python.ops import control_flow_ops from tensorflow.python.ops import linalg_ops from tensorflow.python.ops import math_ops @ops.RegisterGradient("MatrixInverse") def _MatrixInverseGrad(op, grad): """Gradient for MatrixInverse.""" ainv = op.outputs[0] return -math_ops.matmul( ainv, math_ops.matmul( grad, ainv, adjoint_b=True), adjoint_a=True) @ops.RegisterGradient("MatrixDeterminant") def _MatrixDeterminantGrad(op, grad): """Gradient for MatrixDeterminant.""" a = op.inputs[0] c = op.outputs[0] a_adj_inv = linalg_ops.matrix_inverse(a, adjoint=True) multipliers = array_ops.reshape( grad * c, array_ops.concat([array_ops.shape(c), [1, 1]], 0)) return multipliers * a_adj_inv @ops.RegisterGradient("Cholesky") def _CholeskyGrad(op, grad): """Gradient for Cholesky.""" # Gradient is l^{-H} @ ((l^{H} @ grad) * (tril(ones)-1/2*eye)) @ l^{-1} l = op.outputs[0] num_rows = array_ops.shape(l)[-1] batch_shape = array_ops.shape(l)[:-2] l_inverse = linalg_ops.matrix_triangular_solve( l, linalg_ops.eye(num_rows, batch_shape=batch_shape, dtype=l.dtype)) middle = math_ops.matmul(l, grad, adjoint_a=True) middle = array_ops.matrix_set_diag(middle, 0.5 * array_ops.matrix_diag_part(middle)) middle = array_ops.matrix_band_part(middle, -1, 0) grad_a = math_ops.matmul( math_ops.matmul(l_inverse, middle, adjoint_a=True), l_inverse) grad_a += math_ops.conj(array_ops.matrix_transpose(grad_a)) return grad_a * 0.5 @ops.RegisterGradient("MatrixSolve") def _MatrixSolveGrad(op, grad): """Gradient for MatrixSolve.""" a = op.inputs[0] adjoint_a = op.get_attr("adjoint") c = op.outputs[0] grad_b = linalg_ops.matrix_solve(a, grad, adjoint=not adjoint_a) if adjoint_a: grad_a = -math_ops.matmul(c, grad_b, adjoint_b=True) else: grad_a = -math_ops.matmul(grad_b, c, adjoint_b=True) return (grad_a, grad_b) @ops.RegisterGradient("MatrixSolveLs") def _MatrixSolveLsGrad(op, grad): """Gradients for MatrixSolveLs.""" # TODO(rmlarsen): The implementation could be more efficient: # a) Output the Cholesky factorization from forward op instead of # recomputing it here. # b) Implement a symmetric rank-k update op instead of computing # x*z + transpose(x*z). This pattern occurs other places in TensorFlow. def _overdetermined(op, grad): """Gradients for the overdetermined case of MatrixSolveLs. This is the backprop for the solution to the normal equations of the first kind: X = F(A, B) = (A^T * A + lambda * I)^{-1} * A^T * B which solve the least squares problem min ||A * X - B||_F^2 + lambda ||X||_F^2. """ a = op.inputs[0] b = op.inputs[1] l2_regularizer = math_ops.cast(op.inputs[2], a.dtype.base_dtype) x = op.outputs[0] a_shape = array_ops.shape(a) batch_shape = a_shape[:-2] n = a_shape[-1] identity = linalg_ops.eye(n, batch_shape=batch_shape, dtype=a.dtype) gramian = math_ops.matmul(a, a, adjoint_a=True) + l2_regularizer * identity chol = linalg_ops.cholesky(gramian) # Temporary z = (A^T * A + lambda * I)^{-1} * grad. z = linalg_ops.cholesky_solve(chol, grad) xzt = math_ops.matmul(x, z, adjoint_b=True) zx_sym = xzt + array_ops.matrix_transpose(xzt) grad_a = -math_ops.matmul(a, zx_sym) + math_ops.matmul(b, z, adjoint_b=True) grad_b = math_ops.matmul(a, z) return (grad_a, grad_b, None) def _underdetermined(op, grad): """Gradients for the underdetermined case of MatrixSolveLs. This is the backprop for the solution to the normal equations of the second kind: X = F(A, B) = A * (A*A^T + lambda*I)^{-1} * B that (for lambda=0) solve the least squares problem min ||X||_F subject to A*X = B. """ a = op.inputs[0] b = op.inputs[1] l2_regularizer = math_ops.cast(op.inputs[2], a.dtype.base_dtype) a_shape = array_ops.shape(a) batch_shape = a_shape[:-2] m = a_shape[-2] identity = linalg_ops.eye(m, batch_shape=batch_shape, dtype=a.dtype) gramian = math_ops.matmul(a, a, adjoint_b=True) + l2_regularizer * identity chol = linalg_ops.cholesky(gramian) grad_b = linalg_ops.cholesky_solve(chol, math_ops.matmul(a, grad)) # Temporary tmp = (A * A^T + lambda * I)^{-1} * B. tmp = linalg_ops.cholesky_solve(chol, b) a1 = math_ops.matmul(tmp, a, adjoint_a=True) a1 = -math_ops.matmul(grad_b, a1) a2 = grad - math_ops.matmul(a, grad_b, adjoint_a=True) a2 = math_ops.matmul(tmp, a2, adjoint_b=True) grad_a = a1 + a2 return (grad_a, grad_b, None) fast = op.get_attr("fast") if fast is False: raise ValueError("Gradient not defined for fast=False") matrix_shape = op.inputs[0].get_shape()[-2:] if matrix_shape.is_fully_defined(): if matrix_shape[-2] >= matrix_shape[-1]: return _overdetermined(op, grad) else: return _underdetermined(op, grad) else: # We have to defer determining the shape to runtime and use # conditional execution of the appropriate graph. matrix_shape = array_ops.shape(op.inputs[0])[-2:] return control_flow_ops.cond(matrix_shape[-2] >= matrix_shape[-1], lambda: _overdetermined(op, grad), lambda: _underdetermined(op, grad)) @ops.RegisterGradient("MatrixTriangularSolve") def _MatrixTriangularSolveGrad(op, grad): """Gradient for MatrixTriangularSolve.""" a = op.inputs[0] adjoint_a = op.get_attr("adjoint") lower_a = op.get_attr("lower") c = op.outputs[0] grad_b = linalg_ops.matrix_triangular_solve( a, grad, lower=lower_a, adjoint=not adjoint_a) if adjoint_a: grad_a = -math_ops.matmul(c, grad_b, adjoint_b=True) else: grad_a = -math_ops.matmul(grad_b, c, adjoint_b=True) if lower_a: grad_a = array_ops.matrix_band_part(grad_a, -1, 0) else: grad_a = array_ops.matrix_band_part(grad_a, 0, -1) return (grad_a, grad_b) @ops.RegisterGradient("SelfAdjointEigV2") def _SelfAdjointEigV2Grad(op, grad_e, grad_v): """Gradient for SelfAdjointEigV2.""" e = op.outputs[0] v = op.outputs[1] # a = op.inputs[0], which satisfies # a[...,:,:] * v[...,:,i] = e[...,i] * v[...,i] with ops.control_dependencies([grad_e.op, grad_v.op]): if grad_v is not None: # Construct the matrix f(i,j) = (i != j ? 1 / (e_i - e_j) : 0). # Notice that because of the term involving f, the gradient becomes # infinite (or NaN in practice) when eigenvalues are not unique. # Mathematically this should not be surprising, since for (k-fold) # degenerate eigenvalues, the corresponding eigenvectors are only defined # up to arbitrary rotation in a (k-dimensional) subspace. f = array_ops.matrix_set_diag( math_ops.reciprocal( array_ops.expand_dims(e, -2) - array_ops.expand_dims(e, -1)), array_ops.zeros_like(e)) grad_a = math_ops.matmul( v, math_ops.matmul( array_ops.matrix_diag(grad_e) + f * math_ops.matmul( v, grad_v, adjoint_a=True), v, adjoint_b=True)) else: grad_a = math_ops.matmul( v, math_ops.matmul( array_ops.matrix_diag(grad_e), v, adjoint_b=True)) # The forward op only depends on the lower triangular part of a, so here we # symmetrize and take the lower triangle grad_a = array_ops.matrix_band_part( grad_a + math_ops.conj(array_ops.matrix_transpose(grad_a)), -1, 0) grad_a = array_ops.matrix_set_diag(grad_a, 0.5 * array_ops.matrix_diag_part(grad_a)) return grad_a
apache-2.0
kirkbroadhurst/sudokupy
test_moves.py
1
3391
from moves import find_possible_moves, find_impossible_moves, resolve_moves def test_find_possible_moves_1(): # the first item, with value 0, is 'missing'. find_possible_moves will tell us it should be a 9. state = [((i,i), i) for i in range(9)] moves = find_possible_moves(state) assert len(moves) == 1 assert moves[0][0] == (0,0) assert moves[0][1] == 9 def test_find_possible_moves_2(): # the first and last items have value 0 state = [((i,i), i if i < 8 else 0) for i in range(9)] moves = find_possible_moves(state) moves = sorted(moves) assert len(moves) == 4 assert moves[0] == ((0,0), 8) assert moves[1] == ((0,0), 9) assert moves[2] == ((8,8), 8) assert moves[3] == ((8,8), 9) def test_resolve_moves_0(): possible_moves = [[((0,0), 1)], [((0,0), 1)]] moves = resolve_moves(possible_moves) assert len(moves) == 1 assert moves[0] == ((0,0), 1) def test_resolve_moves_1(): possible_moves = [[((0,0), 1), ((0,0), 2)], [((0,0), 1), ((0,0), 3)]] moves = resolve_moves(possible_moves) assert len(moves) == 1 assert moves[0] == ((0,0), 1) def test_resolve_moves_2(): possible_moves = [[((0,0), 1), ((0,0), 2)], [((0,0), 1), ((0,0), 2), ((0,0), 3)]] moves = resolve_moves(possible_moves) assert len(moves) == 0 def test_resolve_moves_confounding_0(): possible_moves = [[((0,0), 1), ((0,0), 2), ((1,1), 1), ((1,1), 2)], [((0,0), 1), ((0,0), 2), ((0,0), 3), ((1,1), 1), ((1,1), 2)]] moves = resolve_moves(possible_moves) assert len(moves) == 0 def test_find_impossible_moves_1(): # the first item, with value 0, is 'missing'. find_impossible_moves will tell us it cannot be [1,2,3,4,5,6,7,8]. state = [((i,i), i) for i in range(9)] moves = find_impossible_moves(state) assert len(moves) == 1 assert moves[0][0] == (0,0) assert moves[0][1] == [1,2,3,4,5,6,7,8] def test_find_impossible_moves_2(): state = [((i,i), 0 if i%2==0 else i) for i in range(9)] moves = find_impossible_moves(state) assert len(moves) == 5 for i, m in enumerate(moves): assert m[0] == (2*i,2*i) assert m[1] == [1,3, 5, 7] def test_find_impossible_moves_3(): state = [((i,i), 0) for i in range(9)] moves = find_impossible_moves(state) assert len(moves) == 9 for i, m in enumerate(moves): assert m[0] == (i, i) # there are no impossible moves in this case assert m[1] == [] def test_resolve_impossible_moves_1(): possible_moves = [[((0,0), 1), ((0,0), 2)], [((0,0), 1), ((0,0), 2), ((0,0), 3)]] impossible_moves = [((0,0), [2, 3])] moves = resolve_moves(possible_moves, impossible_moves) assert len(moves) == 1 def test_resolve_impossible_moves_2(): possible_moves = [[((0,0), 1), ((0,0), 2), ((1,1), 1), ((1,1), 2)], [((0,0), 1), ((0,0), 2), ((0,0), 3), ((1,1), 1), ((1,1), 2)]] impossible_moves = [((0,0), [1, 5, 6, 7, 8])] moves = resolve_moves(possible_moves, impossible_moves) assert len(moves) == 1 def test_resolve_impossible_moves_3(): possible_moves = [[((0,0), 1), ((0,0), 2), ((1,1), 1), ((1,1), 2)], [((0,0), 1), ((0,0), 2), ((0,0), 3), ((1,1), 1), ((1,1), 2)]] impossible_moves = [((0,0), [1, 5, 6, 7, 8]), ((1,1), [1, 3, 5, 6, 7, 8, 9])] moves = resolve_moves(possible_moves, impossible_moves) assert len(moves) == 2
mit
Fireblend/chromium-crosswalk
tools/cygprofile/cyglog_to_orderfile.py
6
8358
#!/usr/bin/python # Copyright 2015 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Symbolizes a log file produced by cyprofile instrumentation. Given a log file and the binary being profiled, creates an orderfile. """ import logging import multiprocessing import optparse import os import tempfile import string import sys import cygprofile_utils import symbol_extractor def _ParseLogLines(log_file_lines): """Parses a merged cyglog produced by mergetraces.py. Args: log_file_lines: array of lines in log file produced by profiled run lib_name: library or executable containing symbols Below is an example of a small log file: 5086e000-52e92000 r-xp 00000000 b3:02 51276 libchromeview.so secs usecs pid:threadid func START 1314897086 795828 3587:1074648168 0x509e105c 1314897086 795874 3587:1074648168 0x509e0eb4 1314897086 796326 3587:1074648168 0x509e0e3c 1314897086 796552 3587:1074648168 0x509e07bc END Returns: An ordered list of callee offsets. """ call_lines = [] vm_start = 0 line = log_file_lines[0] assert 'r-xp' in line end_index = line.find('-') vm_start = int(line[:end_index], 16) for line in log_file_lines[3:]: fields = line.split() if len(fields) == 4: call_lines.append(fields) else: assert fields[0] == 'END' # Convert strings to int in fields. call_info = [] for call_line in call_lines: addr = int(call_line[3], 16) if vm_start < addr: addr -= vm_start call_info.append(addr) return call_info def _GroupLibrarySymbolInfosByOffset(lib_filename): """Returns a dict {offset: [SymbolInfo]} from a library.""" symbol_infos = symbol_extractor.SymbolInfosFromBinary(lib_filename) return symbol_extractor.GroupSymbolInfosByOffset(symbol_infos) class SymbolNotFoundException(Exception): def __init__(self, value): super(SymbolNotFoundException, self).__init__(value) self.value = value def __str__(self): return repr(self.value) def _FindSymbolInfosAtOffset(offset_to_symbol_infos, offset): """Finds all SymbolInfo at a given offset. Args: offset_to_symbol_infos: {offset: [SymbolInfo]} offset: offset to look the symbols at Returns: The list of SymbolInfo at the given offset Raises: SymbolNotFoundException if the offset doesn't match any symbol. """ if offset in offset_to_symbol_infos: return offset_to_symbol_infos[offset] elif offset % 2 and (offset - 1) in offset_to_symbol_infos: # On ARM, odd addresses are used to signal thumb instruction. They are # generated by setting the LSB to 1 (see # http://infocenter.arm.com/help/index.jsp?topic=/com.arm.doc.dui0471e/Babfjhia.html). # TODO(lizeb): Make sure this hack doesn't propagate to other archs. return offset_to_symbol_infos[offset - 1] else: raise SymbolNotFoundException(offset) def _GetObjectFileNames(obj_dir): """Returns the list of object files in a directory.""" obj_files = [] for (dirpath, _, filenames) in os.walk(obj_dir): for file_name in filenames: if file_name.endswith('.o'): obj_files.append(os.path.join(dirpath, file_name)) return obj_files def _AllSymbolInfos(object_filenames): """Returns a list of SymbolInfo from an iterable of filenames.""" pool = multiprocessing.Pool() # Hopefully the object files are in the page cache at this step, so IO should # not be a problem (hence no concurrency limit on the pool). symbol_infos_nested = pool.map( symbol_extractor.SymbolInfosFromBinary, object_filenames) result = [] for symbol_infos in symbol_infos_nested: result += symbol_infos return result def _GetSymbolToSectionMapFromObjectFiles(obj_dir): """ Creates a mapping from symbol to linker section name by scanning all the object files. """ object_files = _GetObjectFileNames(obj_dir) symbol_to_section_map = {} symbol_warnings = cygprofile_utils.WarningCollector(300) symbol_infos = _AllSymbolInfos(object_files) for symbol_info in symbol_infos: symbol = symbol_info.name if symbol.startswith('.LTHUNK'): continue section = symbol_info.section if ((symbol in symbol_to_section_map) and (symbol_to_section_map[symbol] != symbol_info.section)): symbol_warnings.Write('Symbol ' + symbol + ' in conflicting sections ' + section + ' and ' + symbol_to_section_map[symbol]) elif not section.startswith('.text'): symbol_warnings.Write('Symbol ' + symbol + ' in incorrect section ' + section) else: symbol_to_section_map[symbol] = section symbol_warnings.WriteEnd('bad sections') return symbol_to_section_map def _WarnAboutDuplicates(offsets): """Warns about duplicate offsets. Args: offsets: list of offsets to check for duplicates Returns: True if there are no duplicates, False otherwise. """ seen_offsets = set() ok = True for offset in offsets: if offset not in seen_offsets: seen_offsets.add(offset) else: ok = False logging.warning('Duplicate offset: ' + hex(offset)) return ok def _OutputOrderfile(offsets, offset_to_symbol_infos, symbol_to_section_map, output_file): """Outputs the orderfile to output_file. Args: offsets: Iterable of offsets to match to section names offset_to_symbol_infos: {offset: [SymbolInfo]} symbol_to_section_map: {name: section} output_file: file-like object to write the results to """ success = True unknown_symbol_warnings = cygprofile_utils.WarningCollector(300) symbol_not_found_warnings = cygprofile_utils.WarningCollector(300) output_sections = set() for offset in offsets: try: symbol_infos = _FindSymbolInfosAtOffset(offset_to_symbol_infos, offset) for symbol_info in symbol_infos: if symbol_info.name in symbol_to_section_map: section = symbol_to_section_map[symbol_info.name] if not section in output_sections: output_file.write(section + '\n') output_sections.add(section) else: unknown_symbol_warnings.Write( 'No known section for symbol ' + symbol_info.name) except SymbolNotFoundException: symbol_not_found_warnings.Write( 'Did not find function in binary. offset: ' + hex(offset)) success = False unknown_symbol_warnings.WriteEnd('no known section for symbol.') symbol_not_found_warnings.WriteEnd('symbol not found in the binary.') return success def main(): parser = optparse.OptionParser(usage= 'usage: %prog [options] <merged_cyglog> <library> <output_filename>') parser.add_option('--target-arch', action='store', dest='arch', choices=['arm', 'arm64', 'x86', 'x86_64', 'x64', 'mips'], help='The target architecture for libchrome.so') options, argv = parser.parse_args(sys.argv) if not options.arch: options.arch = cygprofile_utils.DetectArchitecture() if len(argv) != 4: parser.print_help() return 1 (log_filename, lib_filename, output_filename) = argv[1:] symbol_extractor.SetArchitecture(options.arch) obj_dir = os.path.abspath(os.path.join( os.path.dirname(lib_filename), '../obj')) log_file_lines = map(string.rstrip, open(log_filename).readlines()) offsets = _ParseLogLines(log_file_lines) _WarnAboutDuplicates(offsets) offset_to_symbol_infos = _GroupLibrarySymbolInfosByOffset(lib_filename) symbol_to_section_map = _GetSymbolToSectionMapFromObjectFiles(obj_dir) success = False temp_filename = None output_file = None try: (fd, temp_filename) = tempfile.mkstemp(dir=os.path.dirname(output_filename)) output_file = os.fdopen(fd, 'w') ok = _OutputOrderfile( offsets, offset_to_symbol_infos, symbol_to_section_map, output_file) output_file.close() os.rename(temp_filename, output_filename) temp_filename = None success = ok finally: if output_file: output_file.close() if temp_filename: os.remove(temp_filename) return 0 if success else 1 if __name__ == '__main__': logging.basicConfig(level=logging.INFO) sys.exit(main())
bsd-3-clause
sqlalchemy/sqlalchemy
test/orm/test_rel_fn.py
3
45031
from sqlalchemy import and_ from sqlalchemy import Boolean from sqlalchemy import Column from sqlalchemy import exc from sqlalchemy import ForeignKey from sqlalchemy import ForeignKeyConstraint from sqlalchemy import func from sqlalchemy import Integer from sqlalchemy import MetaData from sqlalchemy import select from sqlalchemy import String from sqlalchemy import Table from sqlalchemy.orm import foreign from sqlalchemy.orm import relationship from sqlalchemy.orm import relationships from sqlalchemy.orm import remote from sqlalchemy.orm.interfaces import MANYTOONE from sqlalchemy.orm.interfaces import ONETOMANY from sqlalchemy.testing import assert_raises_message from sqlalchemy.testing import AssertsCompiledSQL from sqlalchemy.testing import eq_ from sqlalchemy.testing import fixtures from sqlalchemy.testing import is_ from sqlalchemy.testing import mock class _JoinFixtures(object): @classmethod def setup_test_class(cls): m = MetaData() cls.left = Table( "lft", m, Column("id", Integer, primary_key=True), Column("x", Integer), Column("y", Integer), ) cls.right = Table( "rgt", m, Column("id", Integer, primary_key=True), Column("lid", Integer, ForeignKey("lft.id")), Column("x", Integer), Column("y", Integer), ) cls.right_multi_fk = Table( "rgt_multi_fk", m, Column("id", Integer, primary_key=True), Column("lid1", Integer, ForeignKey("lft.id")), Column("lid2", Integer, ForeignKey("lft.id")), ) cls.selfref = Table( "selfref", m, Column("id", Integer, primary_key=True), Column("sid", Integer, ForeignKey("selfref.id")), ) cls.composite_selfref = Table( "composite_selfref", m, Column("id", Integer, primary_key=True), Column("group_id", Integer, primary_key=True), Column("parent_id", Integer), ForeignKeyConstraint( ["parent_id", "group_id"], ["composite_selfref.id", "composite_selfref.group_id"], ), ) cls.m2mleft = Table( "m2mlft", m, Column("id", Integer, primary_key=True) ) cls.m2mright = Table( "m2mrgt", m, Column("id", Integer, primary_key=True) ) cls.m2msecondary = Table( "m2msecondary", m, Column("lid", Integer, ForeignKey("m2mlft.id"), primary_key=True), Column("rid", Integer, ForeignKey("m2mrgt.id"), primary_key=True), ) cls.m2msecondary_no_fks = Table( "m2msecondary_no_fks", m, Column("lid", Integer, primary_key=True), Column("rid", Integer, primary_key=True), ) cls.m2msecondary_ambig_fks = Table( "m2msecondary_ambig_fks", m, Column("lid1", Integer, ForeignKey("m2mlft.id"), primary_key=True), Column("rid1", Integer, ForeignKey("m2mrgt.id"), primary_key=True), Column("lid2", Integer, ForeignKey("m2mlft.id"), primary_key=True), Column("rid2", Integer, ForeignKey("m2mrgt.id"), primary_key=True), ) cls.base_w_sub_rel = Table( "base_w_sub_rel", m, Column("id", Integer, primary_key=True), Column("sub_id", Integer, ForeignKey("rel_sub.id")), ) cls.rel_sub = Table( "rel_sub", m, Column( "id", Integer, ForeignKey("base_w_sub_rel.id"), primary_key=True, ), ) cls.base = Table( "base", m, Column("id", Integer, primary_key=True), Column("flag", Boolean), ) cls.sub = Table( "sub", m, Column("id", Integer, ForeignKey("base.id"), primary_key=True), ) cls.sub_w_base_rel = Table( "sub_w_base_rel", m, Column("id", Integer, ForeignKey("base.id"), primary_key=True), Column("base_id", Integer, ForeignKey("base.id")), ) cls.sub_w_sub_rel = Table( "sub_w_sub_rel", m, Column("id", Integer, ForeignKey("base.id"), primary_key=True), Column("sub_id", Integer, ForeignKey("sub.id")), ) cls.right_w_base_rel = Table( "right_w_base_rel", m, Column("id", Integer, primary_key=True), Column("base_id", Integer, ForeignKey("base.id")), ) cls.three_tab_a = Table( "three_tab_a", m, Column("id", Integer, primary_key=True) ) cls.three_tab_b = Table( "three_tab_b", m, Column("id", Integer, primary_key=True), Column("aid", Integer, ForeignKey("three_tab_a.id")), ) cls.three_tab_c = Table( "three_tab_c", m, Column("id", Integer, primary_key=True), Column("aid", Integer, ForeignKey("three_tab_a.id")), Column("bid", Integer, ForeignKey("three_tab_b.id")), ) cls.composite_target = Table( "composite_target", m, Column("uid", Integer, primary_key=True), Column("oid", Integer, primary_key=True), ) cls.composite_multi_ref = Table( "composite_multi_ref", m, Column("uid1", Integer), Column("uid2", Integer), Column("oid", Integer), ForeignKeyConstraint( ("uid1", "oid"), ("composite_target.uid", "composite_target.oid"), ), ForeignKeyConstraint( ("uid2", "oid"), ("composite_target.uid", "composite_target.oid"), ), ) cls.purely_single_col = Table( "purely_single_col", m, Column("path", String) ) def _join_fixture_overlapping_three_tables(self, **kw): def _can_sync(*cols): for c in cols: if self.three_tab_c.c.contains_column(c): return False else: return True return relationships.JoinCondition( self.three_tab_a, self.three_tab_b, self.three_tab_a, self.three_tab_b, support_sync=False, can_be_synced_fn=_can_sync, primaryjoin=and_( self.three_tab_a.c.id == self.three_tab_b.c.aid, self.three_tab_c.c.bid == self.three_tab_b.c.id, self.three_tab_c.c.aid == self.three_tab_a.c.id, ), ) def _join_fixture_m2m(self, **kw): return relationships.JoinCondition( self.m2mleft, self.m2mright, self.m2mleft, self.m2mright, secondary=self.m2msecondary, **kw ) def _join_fixture_m2m_backref(self, **kw): """return JoinCondition in the same way RelationshipProperty calls it for a backref on an m2m. """ j1 = self._join_fixture_m2m() return ( j1, relationships.JoinCondition( self.m2mright, self.m2mleft, self.m2mright, self.m2mleft, secondary=self.m2msecondary, primaryjoin=j1.secondaryjoin_minus_local, secondaryjoin=j1.primaryjoin_minus_local, ), ) def _join_fixture_o2m(self, **kw): return relationships.JoinCondition( self.left, self.right, self.left, self.right, **kw ) def _join_fixture_m2o(self, **kw): return relationships.JoinCondition( self.right, self.left, self.right, self.left, **kw ) def _join_fixture_o2m_selfref(self, **kw): return relationships.JoinCondition( self.selfref, self.selfref, self.selfref, self.selfref, **kw ) def _join_fixture_m2o_selfref(self, **kw): return relationships.JoinCondition( self.selfref, self.selfref, self.selfref, self.selfref, remote_side=set([self.selfref.c.id]), **kw ) def _join_fixture_o2m_composite_selfref(self, **kw): return relationships.JoinCondition( self.composite_selfref, self.composite_selfref, self.composite_selfref, self.composite_selfref, **kw ) def _join_fixture_m2o_composite_selfref(self, **kw): return relationships.JoinCondition( self.composite_selfref, self.composite_selfref, self.composite_selfref, self.composite_selfref, remote_side=set( [ self.composite_selfref.c.id, self.composite_selfref.c.group_id, ] ), **kw ) def _join_fixture_o2m_composite_selfref_func(self, **kw): return relationships.JoinCondition( self.composite_selfref, self.composite_selfref, self.composite_selfref, self.composite_selfref, primaryjoin=and_( self.composite_selfref.c.group_id == func.foo(self.composite_selfref.c.group_id), self.composite_selfref.c.parent_id == self.composite_selfref.c.id, ), **kw ) def _join_fixture_o2m_composite_selfref_func_remote_side(self, **kw): return relationships.JoinCondition( self.composite_selfref, self.composite_selfref, self.composite_selfref, self.composite_selfref, primaryjoin=and_( self.composite_selfref.c.group_id == func.foo(self.composite_selfref.c.group_id), self.composite_selfref.c.parent_id == self.composite_selfref.c.id, ), remote_side=set([self.composite_selfref.c.parent_id]), **kw ) def _join_fixture_o2m_composite_selfref_func_annotated(self, **kw): return relationships.JoinCondition( self.composite_selfref, self.composite_selfref, self.composite_selfref, self.composite_selfref, primaryjoin=and_( remote(self.composite_selfref.c.group_id) == func.foo(self.composite_selfref.c.group_id), remote(self.composite_selfref.c.parent_id) == self.composite_selfref.c.id, ), **kw ) def _join_fixture_compound_expression_1(self, **kw): return relationships.JoinCondition( self.left, self.right, self.left, self.right, primaryjoin=(self.left.c.x + self.left.c.y) == relationships.remote( relationships.foreign(self.right.c.x * self.right.c.y) ), **kw ) def _join_fixture_compound_expression_2(self, **kw): return relationships.JoinCondition( self.left, self.right, self.left, self.right, primaryjoin=(self.left.c.x + self.left.c.y) == relationships.foreign(self.right.c.x * self.right.c.y), **kw ) def _join_fixture_compound_expression_1_non_annotated(self, **kw): return relationships.JoinCondition( self.left, self.right, self.left, self.right, primaryjoin=(self.left.c.x + self.left.c.y) == (self.right.c.x * self.right.c.y), **kw ) def _join_fixture_base_to_joined_sub(self, **kw): # see test/orm/inheritance/test_abc_inheritance:TestaTobM2O # and others there right = self.base_w_sub_rel.join( self.rel_sub, self.base_w_sub_rel.c.id == self.rel_sub.c.id ) return relationships.JoinCondition( self.base_w_sub_rel, right, self.base_w_sub_rel, self.rel_sub, primaryjoin=self.base_w_sub_rel.c.sub_id == self.rel_sub.c.id, **kw ) def _join_fixture_o2m_joined_sub_to_base(self, **kw): left = self.base.join( self.sub_w_base_rel, self.base.c.id == self.sub_w_base_rel.c.id ) return relationships.JoinCondition( left, self.base, self.sub_w_base_rel, self.base, primaryjoin=self.sub_w_base_rel.c.base_id == self.base.c.id, ) def _join_fixture_m2o_joined_sub_to_sub_on_base(self, **kw): # this is a late add - a variant of the test case # in #2491 where we join on the base cols instead. only # m2o has a problem at the time of this test. left = self.base.join(self.sub, self.base.c.id == self.sub.c.id) right = self.base.join( self.sub_w_base_rel, self.base.c.id == self.sub_w_base_rel.c.id ) return relationships.JoinCondition( left, right, self.sub, self.sub_w_base_rel, primaryjoin=self.sub_w_base_rel.c.base_id == self.base.c.id, ) def _join_fixture_o2m_joined_sub_to_sub(self, **kw): left = self.base.join(self.sub, self.base.c.id == self.sub.c.id) right = self.base.join( self.sub_w_sub_rel, self.base.c.id == self.sub_w_sub_rel.c.id ) return relationships.JoinCondition( left, right, self.sub, self.sub_w_sub_rel, primaryjoin=self.sub.c.id == self.sub_w_sub_rel.c.sub_id, ) def _join_fixture_m2o_sub_to_joined_sub(self, **kw): # see test.orm.test_mapper:MapperTest.test_add_column_prop_deannotate, right = self.base.join( self.right_w_base_rel, self.base.c.id == self.right_w_base_rel.c.id ) return relationships.JoinCondition( self.right_w_base_rel, right, self.right_w_base_rel, self.right_w_base_rel, ) def _join_fixture_m2o_sub_to_joined_sub_func(self, **kw): # see test.orm.test_mapper:MapperTest.test_add_column_prop_deannotate, right = self.base.join( self.right_w_base_rel, self.base.c.id == self.right_w_base_rel.c.id ) return relationships.JoinCondition( self.right_w_base_rel, right, self.right_w_base_rel, self.right_w_base_rel, primaryjoin=self.right_w_base_rel.c.base_id == func.foo(self.base.c.id), ) def _join_fixture_o2o_joined_sub_to_base(self, **kw): left = self.base.join(self.sub, self.base.c.id == self.sub.c.id) # see test_relationships->AmbiguousJoinInterpretedAsSelfRef return relationships.JoinCondition(left, self.sub, left, self.sub) def _join_fixture_o2m_to_annotated_func(self, **kw): return relationships.JoinCondition( self.left, self.right, self.left, self.right, primaryjoin=self.left.c.id == foreign(func.foo(self.right.c.lid)), **kw ) def _join_fixture_o2m_to_oldstyle_func(self, **kw): return relationships.JoinCondition( self.left, self.right, self.left, self.right, primaryjoin=self.left.c.id == func.foo(self.right.c.lid), consider_as_foreign_keys={self.right.c.lid}, **kw ) def _join_fixture_overlapping_composite_fks(self, **kw): return relationships.JoinCondition( self.composite_target, self.composite_multi_ref, self.composite_target, self.composite_multi_ref, consider_as_foreign_keys={ self.composite_multi_ref.c.uid2, self.composite_multi_ref.c.oid, }, **kw ) def _join_fixture_o2m_o_side_none(self, **kw): return relationships.JoinCondition( self.left, self.right, self.left, self.right, primaryjoin=and_( self.left.c.id == self.right.c.lid, self.left.c.x == 5 ), **kw ) def _join_fixture_purely_single_o2m(self, **kw): return relationships.JoinCondition( self.purely_single_col, self.purely_single_col, self.purely_single_col, self.purely_single_col, support_sync=False, primaryjoin=self.purely_single_col.c.path.like( remote(foreign(self.purely_single_col.c.path.concat("%"))) ), ) def _join_fixture_purely_single_m2o(self, **kw): return relationships.JoinCondition( self.purely_single_col, self.purely_single_col, self.purely_single_col, self.purely_single_col, support_sync=False, primaryjoin=remote(self.purely_single_col.c.path).like( foreign(self.purely_single_col.c.path.concat("%")) ), ) def _join_fixture_remote_local_multiple_ref(self, **kw): def fn(a, b): return (a == b) | (b == a) return relationships.JoinCondition( self.selfref, self.selfref, self.selfref, self.selfref, support_sync=False, primaryjoin=fn( # we're putting a do-nothing annotation on # "a" so that the left/right is preserved; # annotation vs. non seems to affect __eq__ behavior self.selfref.c.sid._annotate({"foo": "bar"}), foreign(remote(self.selfref.c.sid)), ), ) def _join_fixture_inh_selfref_w_entity(self, **kw): fake_logger = mock.Mock(info=lambda *arg, **kw: None) prop = mock.Mock( parent=mock.Mock(), mapper=mock.Mock(), logger=fake_logger ) local_selectable = self.base.join(self.sub) remote_selectable = self.base.join(self.sub_w_sub_rel) # note this test requires that "parentmapper" annotation is # present in the columns ahead of time sub_w_sub_rel__sub_id = self.sub_w_sub_rel.c.sub_id._annotate( {"parentmapper": prop.mapper} ) sub__id = self.sub.c.id._annotate({"parentmapper": prop.parent}) sub_w_sub_rel__flag = self.base.c.flag._annotate( {"parentmapper": prop.mapper} ) return relationships.JoinCondition( local_selectable, remote_selectable, local_selectable, remote_selectable, primaryjoin=and_( sub_w_sub_rel__sub_id == sub__id, sub_w_sub_rel__flag == True, # noqa ), prop=prop, ) def _assert_non_simple_warning(self, fn): assert_raises_message( exc.SAWarning, "Non-simple column elements in " "primary join condition for property " r"None - consider using remote\(\) " "annotations to mark the remote side.", fn, ) def _assert_raises_no_relevant_fks( self, fn, expr, relname, primary, *arg, **kw ): assert_raises_message( exc.ArgumentError, r"Could not locate any relevant foreign key columns " r"for %s join condition '%s' on relationship %s. " r"Ensure that referencing columns are associated with " r"a ForeignKey or ForeignKeyConstraint, or are annotated " r"in the join condition with the foreign\(\) annotation." % (primary, expr, relname), fn, *arg, **kw ) def _assert_raises_no_equality( self, fn, expr, relname, primary, *arg, **kw ): assert_raises_message( exc.ArgumentError, "Could not locate any simple equality expressions " "involving locally mapped foreign key columns for %s join " "condition '%s' on relationship %s. " "Ensure that referencing columns are associated with a " "ForeignKey or ForeignKeyConstraint, or are annotated in " r"the join condition with the foreign\(\) annotation. " "To allow comparison operators other than '==', " "the relationship can be marked as viewonly=True." % (primary, expr, relname), fn, *arg, **kw ) def _assert_raises_ambig_join( self, fn, relname, secondary_arg, *arg, **kw ): if secondary_arg is not None: assert_raises_message( exc.AmbiguousForeignKeysError, "Could not determine join condition between " "parent/child tables on relationship %s - " "there are multiple foreign key paths linking the " "tables via secondary table '%s'. " "Specify the 'foreign_keys' argument, providing a list " "of those columns which should be counted as " "containing a foreign key reference from the " "secondary table to each of the parent and child tables." % (relname, secondary_arg), fn, *arg, **kw ) else: assert_raises_message( exc.AmbiguousForeignKeysError, "Could not determine join condition between " "parent/child tables on relationship %s - " "there are no foreign keys linking these tables. " % (relname,), fn, *arg, **kw ) def _assert_raises_no_join(self, fn, relname, secondary_arg, *arg, **kw): if secondary_arg is not None: assert_raises_message( exc.NoForeignKeysError, "Could not determine join condition between " "parent/child tables on relationship %s - " "there are no foreign keys linking these tables " "via secondary table '%s'. " "Ensure that referencing columns are associated " "with a ForeignKey " "or ForeignKeyConstraint, or specify 'primaryjoin' and " "'secondaryjoin' expressions" % (relname, secondary_arg), fn, *arg, **kw ) else: assert_raises_message( exc.NoForeignKeysError, "Could not determine join condition between " "parent/child tables on relationship %s - " "there are no foreign keys linking these tables. " "Ensure that referencing columns are associated " "with a ForeignKey " "or ForeignKeyConstraint, or specify a 'primaryjoin' " "expression." % (relname,), fn, *arg, **kw ) class ColumnCollectionsTest( _JoinFixtures, fixtures.TestBase, AssertsCompiledSQL ): def test_determine_local_remote_pairs_o2o_joined_sub_to_base(self): joincond = self._join_fixture_o2o_joined_sub_to_base() eq_(joincond.local_remote_pairs, [(self.base.c.id, self.sub.c.id)]) def test_determine_synchronize_pairs_o2m_to_annotated_func(self): joincond = self._join_fixture_o2m_to_annotated_func() eq_(joincond.synchronize_pairs, [(self.left.c.id, self.right.c.lid)]) def test_determine_synchronize_pairs_o2m_to_oldstyle_func(self): joincond = self._join_fixture_o2m_to_oldstyle_func() eq_(joincond.synchronize_pairs, [(self.left.c.id, self.right.c.lid)]) def test_determinelocal_remote_m2o_joined_sub_to_sub_on_base(self): joincond = self._join_fixture_m2o_joined_sub_to_sub_on_base() eq_( joincond.local_remote_pairs, [(self.base.c.id, self.sub_w_base_rel.c.base_id)], ) def test_determine_local_remote_base_to_joined_sub(self): joincond = self._join_fixture_base_to_joined_sub() eq_( joincond.local_remote_pairs, [(self.base_w_sub_rel.c.sub_id, self.rel_sub.c.id)], ) def test_determine_local_remote_o2m_joined_sub_to_base(self): joincond = self._join_fixture_o2m_joined_sub_to_base() eq_( joincond.local_remote_pairs, [(self.sub_w_base_rel.c.base_id, self.base.c.id)], ) def test_determine_local_remote_m2o_sub_to_joined_sub(self): joincond = self._join_fixture_m2o_sub_to_joined_sub() eq_( joincond.local_remote_pairs, [(self.right_w_base_rel.c.base_id, self.base.c.id)], ) def test_determine_remote_columns_o2m_joined_sub_to_sub(self): joincond = self._join_fixture_o2m_joined_sub_to_sub() eq_( joincond.local_remote_pairs, [(self.sub.c.id, self.sub_w_sub_rel.c.sub_id)], ) def test_determine_remote_columns_compound_1(self): joincond = self._join_fixture_compound_expression_1(support_sync=False) eq_(joincond.remote_columns, set([self.right.c.x, self.right.c.y])) def test_determine_local_remote_compound_1(self): joincond = self._join_fixture_compound_expression_1(support_sync=False) eq_( joincond.local_remote_pairs, [ (self.left.c.x, self.right.c.x), (self.left.c.x, self.right.c.y), (self.left.c.y, self.right.c.x), (self.left.c.y, self.right.c.y), ], ) def test_determine_local_remote_compound_2(self): joincond = self._join_fixture_compound_expression_2(support_sync=False) eq_( joincond.local_remote_pairs, [ (self.left.c.x, self.right.c.x), (self.left.c.x, self.right.c.y), (self.left.c.y, self.right.c.x), (self.left.c.y, self.right.c.y), ], ) def test_determine_local_remote_compound_3(self): joincond = self._join_fixture_compound_expression_1() eq_( joincond.local_remote_pairs, [ (self.left.c.x, self.right.c.x), (self.left.c.x, self.right.c.y), (self.left.c.y, self.right.c.x), (self.left.c.y, self.right.c.y), ], ) def test_err_local_remote_compound_1(self): self._assert_raises_no_relevant_fks( self._join_fixture_compound_expression_1_non_annotated, r"lft.x \+ lft.y = rgt.x \* rgt.y", "None", "primary", ) def test_determine_remote_columns_compound_2(self): joincond = self._join_fixture_compound_expression_2(support_sync=False) eq_(joincond.remote_columns, set([self.right.c.x, self.right.c.y])) def test_determine_remote_columns_o2m(self): joincond = self._join_fixture_o2m() eq_(joincond.remote_columns, set([self.right.c.lid])) def test_determine_remote_columns_o2m_selfref(self): joincond = self._join_fixture_o2m_selfref() eq_(joincond.remote_columns, set([self.selfref.c.sid])) def test_determine_local_remote_pairs_o2m_composite_selfref(self): joincond = self._join_fixture_o2m_composite_selfref() eq_( joincond.local_remote_pairs, [ ( self.composite_selfref.c.group_id, self.composite_selfref.c.group_id, ), ( self.composite_selfref.c.id, self.composite_selfref.c.parent_id, ), ], ) def test_determine_local_remote_pairs_o2m_composite_selfref_func_warning( self, ): self._assert_non_simple_warning( self._join_fixture_o2m_composite_selfref_func ) def test_determine_local_remote_pairs_o2m_composite_selfref_func_rs(self): # no warning self._join_fixture_o2m_composite_selfref_func_remote_side() def test_determine_local_remote_pairs_o2m_overlap_func_warning(self): self._assert_non_simple_warning( self._join_fixture_m2o_sub_to_joined_sub_func ) def test_determine_local_remote_pairs_o2m_composite_selfref_func_annotated( self, ): joincond = self._join_fixture_o2m_composite_selfref_func_annotated() eq_( joincond.local_remote_pairs, [ ( self.composite_selfref.c.group_id, self.composite_selfref.c.group_id, ), ( self.composite_selfref.c.id, self.composite_selfref.c.parent_id, ), ], ) def test_determine_remote_columns_m2o_composite_selfref(self): joincond = self._join_fixture_m2o_composite_selfref() eq_( joincond.remote_columns, set( [ self.composite_selfref.c.id, self.composite_selfref.c.group_id, ] ), ) def test_determine_remote_columns_m2o(self): joincond = self._join_fixture_m2o() eq_(joincond.remote_columns, set([self.left.c.id])) def test_determine_local_remote_pairs_o2m(self): joincond = self._join_fixture_o2m() eq_(joincond.local_remote_pairs, [(self.left.c.id, self.right.c.lid)]) def test_determine_synchronize_pairs_m2m(self): joincond = self._join_fixture_m2m() eq_( joincond.synchronize_pairs, [(self.m2mleft.c.id, self.m2msecondary.c.lid)], ) eq_( joincond.secondary_synchronize_pairs, [(self.m2mright.c.id, self.m2msecondary.c.rid)], ) def test_determine_local_remote_pairs_o2m_backref(self): joincond = self._join_fixture_o2m() joincond2 = self._join_fixture_m2o( primaryjoin=joincond.primaryjoin_reverse_remote ) eq_(joincond2.local_remote_pairs, [(self.right.c.lid, self.left.c.id)]) def test_determine_local_remote_pairs_m2m(self): joincond = self._join_fixture_m2m() eq_( joincond.local_remote_pairs, [ (self.m2mleft.c.id, self.m2msecondary.c.lid), (self.m2mright.c.id, self.m2msecondary.c.rid), ], ) def test_determine_local_remote_pairs_m2m_backref(self): j1, j2 = self._join_fixture_m2m_backref() eq_( j1.local_remote_pairs, [ (self.m2mleft.c.id, self.m2msecondary.c.lid), (self.m2mright.c.id, self.m2msecondary.c.rid), ], ) eq_( j2.local_remote_pairs, [ (self.m2mright.c.id, self.m2msecondary.c.rid), (self.m2mleft.c.id, self.m2msecondary.c.lid), ], ) def test_determine_local_columns_m2m_backref(self): j1, j2 = self._join_fixture_m2m_backref() eq_(j1.local_columns, set([self.m2mleft.c.id])) eq_(j2.local_columns, set([self.m2mright.c.id])) def test_determine_remote_columns_m2m_backref(self): j1, j2 = self._join_fixture_m2m_backref() eq_( j1.remote_columns, set([self.m2msecondary.c.lid, self.m2msecondary.c.rid]), ) eq_( j2.remote_columns, set([self.m2msecondary.c.lid, self.m2msecondary.c.rid]), ) def test_determine_remote_columns_m2o_selfref(self): joincond = self._join_fixture_m2o_selfref() eq_(joincond.remote_columns, set([self.selfref.c.id])) def test_determine_local_remote_cols_three_tab_viewonly(self): joincond = self._join_fixture_overlapping_three_tables() eq_( joincond.local_remote_pairs, [(self.three_tab_a.c.id, self.three_tab_b.c.aid)], ) eq_( joincond.remote_columns, set([self.three_tab_b.c.id, self.three_tab_b.c.aid]), ) def test_determine_local_remote_overlapping_composite_fks(self): joincond = self._join_fixture_overlapping_composite_fks() eq_( joincond.local_remote_pairs, [ (self.composite_target.c.uid, self.composite_multi_ref.c.uid2), (self.composite_target.c.oid, self.composite_multi_ref.c.oid), ], ) def test_determine_local_remote_pairs_purely_single_col_o2m(self): joincond = self._join_fixture_purely_single_o2m() eq_( joincond.local_remote_pairs, [(self.purely_single_col.c.path, self.purely_single_col.c.path)], ) def test_determine_local_remote_pairs_inh_selfref_w_entities(self): joincond = self._join_fixture_inh_selfref_w_entity() eq_( joincond.local_remote_pairs, [(self.sub.c.id, self.sub_w_sub_rel.c.sub_id)], ) eq_( joincond.remote_columns, set([self.base.c.flag, self.sub_w_sub_rel.c.sub_id]), ) class DirectionTest(_JoinFixtures, fixtures.TestBase, AssertsCompiledSQL): def test_determine_direction_compound_2(self): joincond = self._join_fixture_compound_expression_2(support_sync=False) is_(joincond.direction, ONETOMANY) def test_determine_direction_o2m(self): joincond = self._join_fixture_o2m() is_(joincond.direction, ONETOMANY) def test_determine_direction_o2m_selfref(self): joincond = self._join_fixture_o2m_selfref() is_(joincond.direction, ONETOMANY) def test_determine_direction_m2o_selfref(self): joincond = self._join_fixture_m2o_selfref() is_(joincond.direction, MANYTOONE) def test_determine_direction_o2m_composite_selfref(self): joincond = self._join_fixture_o2m_composite_selfref() is_(joincond.direction, ONETOMANY) def test_determine_direction_m2o_composite_selfref(self): joincond = self._join_fixture_m2o_composite_selfref() is_(joincond.direction, MANYTOONE) def test_determine_direction_m2o(self): joincond = self._join_fixture_m2o() is_(joincond.direction, MANYTOONE) def test_determine_direction_purely_single_o2m(self): joincond = self._join_fixture_purely_single_o2m() is_(joincond.direction, ONETOMANY) def test_determine_direction_purely_single_m2o(self): joincond = self._join_fixture_purely_single_m2o() is_(joincond.direction, MANYTOONE) class DetermineJoinTest(_JoinFixtures, fixtures.TestBase, AssertsCompiledSQL): __dialect__ = "default" def test_determine_join_o2m(self): joincond = self._join_fixture_o2m() self.assert_compile(joincond.primaryjoin, "lft.id = rgt.lid") def test_determine_join_o2m_selfref(self): joincond = self._join_fixture_o2m_selfref() self.assert_compile(joincond.primaryjoin, "selfref.id = selfref.sid") def test_determine_join_m2o_selfref(self): joincond = self._join_fixture_m2o_selfref() self.assert_compile(joincond.primaryjoin, "selfref.id = selfref.sid") def test_determine_join_o2m_composite_selfref(self): joincond = self._join_fixture_o2m_composite_selfref() self.assert_compile( joincond.primaryjoin, "composite_selfref.group_id = composite_selfref.group_id " "AND composite_selfref.id = composite_selfref.parent_id", ) def test_determine_join_m2o_composite_selfref(self): joincond = self._join_fixture_m2o_composite_selfref() self.assert_compile( joincond.primaryjoin, "composite_selfref.group_id = composite_selfref.group_id " "AND composite_selfref.id = composite_selfref.parent_id", ) def test_determine_join_m2o(self): joincond = self._join_fixture_m2o() self.assert_compile(joincond.primaryjoin, "lft.id = rgt.lid") def test_determine_join_ambiguous_fks_o2m(self): assert_raises_message( exc.AmbiguousForeignKeysError, "Could not determine join condition between " "parent/child tables on relationship None - " "there are multiple foreign key paths linking " "the tables. Specify the 'foreign_keys' argument, " "providing a list of those columns which " "should be counted as containing a foreign " "key reference to the parent table.", relationships.JoinCondition, self.left, self.right_multi_fk, self.left, self.right_multi_fk, ) def test_determine_join_no_fks_o2m(self): self._assert_raises_no_join( relationships.JoinCondition, "None", None, self.left, self.selfref, self.left, self.selfref, ) def test_determine_join_ambiguous_fks_m2m(self): self._assert_raises_ambig_join( relationships.JoinCondition, "None", self.m2msecondary_ambig_fks, self.m2mleft, self.m2mright, self.m2mleft, self.m2mright, secondary=self.m2msecondary_ambig_fks, ) def test_determine_join_no_fks_m2m(self): self._assert_raises_no_join( relationships.JoinCondition, "None", self.m2msecondary_no_fks, self.m2mleft, self.m2mright, self.m2mleft, self.m2mright, secondary=self.m2msecondary_no_fks, ) def _join_fixture_fks_ambig_m2m(self): return relationships.JoinCondition( self.m2mleft, self.m2mright, self.m2mleft, self.m2mright, secondary=self.m2msecondary_ambig_fks, consider_as_foreign_keys={ self.m2msecondary_ambig_fks.c.lid1, self.m2msecondary_ambig_fks.c.rid1, }, ) def test_determine_join_w_fks_ambig_m2m(self): joincond = self._join_fixture_fks_ambig_m2m() self.assert_compile( joincond.primaryjoin, "m2mlft.id = m2msecondary_ambig_fks.lid1" ) self.assert_compile( joincond.secondaryjoin, "m2mrgt.id = m2msecondary_ambig_fks.rid1" ) class AdaptedJoinTest(_JoinFixtures, fixtures.TestBase, AssertsCompiledSQL): __dialect__ = "default" def test_join_targets_o2m_selfref(self): joincond = self._join_fixture_o2m_selfref() left = select(joincond.parent_persist_selectable).alias("pj") pj, sj, sec, adapter, ds = joincond.join_targets( left, joincond.child_persist_selectable, True ) self.assert_compile(pj, "pj.id = selfref.sid") self.assert_compile(pj, "pj.id = selfref.sid") right = select(joincond.child_persist_selectable).alias("pj") pj, sj, sec, adapter, ds = joincond.join_targets( joincond.parent_persist_selectable, right, True ) self.assert_compile(pj, "selfref.id = pj.sid") self.assert_compile(pj, "selfref.id = pj.sid") def test_join_targets_o2m_plain(self): joincond = self._join_fixture_o2m() pj, sj, sec, adapter, ds = joincond.join_targets( joincond.parent_persist_selectable, joincond.child_persist_selectable, False, ) self.assert_compile(pj, "lft.id = rgt.lid") self.assert_compile(pj, "lft.id = rgt.lid") def test_join_targets_o2m_left_aliased(self): joincond = self._join_fixture_o2m() left = select(joincond.parent_persist_selectable).alias("pj") pj, sj, sec, adapter, ds = joincond.join_targets( left, joincond.child_persist_selectable, True ) self.assert_compile(pj, "pj.id = rgt.lid") self.assert_compile(pj, "pj.id = rgt.lid") def test_join_targets_o2m_right_aliased(self): joincond = self._join_fixture_o2m() right = select(joincond.child_persist_selectable).alias("pj") pj, sj, sec, adapter, ds = joincond.join_targets( joincond.parent_persist_selectable, right, True ) self.assert_compile(pj, "lft.id = pj.lid") self.assert_compile(pj, "lft.id = pj.lid") def test_join_targets_o2m_composite_selfref(self): joincond = self._join_fixture_o2m_composite_selfref() right = select(joincond.child_persist_selectable).alias("pj") pj, sj, sec, adapter, ds = joincond.join_targets( joincond.parent_persist_selectable, right, True ) self.assert_compile( pj, "pj.group_id = composite_selfref.group_id " "AND composite_selfref.id = pj.parent_id", ) def test_join_targets_m2o_composite_selfref(self): joincond = self._join_fixture_m2o_composite_selfref() right = select(joincond.child_persist_selectable).alias("pj") pj, sj, sec, adapter, ds = joincond.join_targets( joincond.parent_persist_selectable, right, True ) self.assert_compile( pj, "pj.group_id = composite_selfref.group_id " "AND pj.id = composite_selfref.parent_id", ) class LazyClauseTest(_JoinFixtures, fixtures.TestBase, AssertsCompiledSQL): __dialect__ = "default" def test_lazy_clause_o2m(self): joincond = self._join_fixture_o2m() lazywhere, bind_to_col, equated_columns = joincond.create_lazy_clause() self.assert_compile(lazywhere, ":param_1 = rgt.lid") def test_lazy_clause_o2m_reverse(self): joincond = self._join_fixture_o2m() lazywhere, bind_to_col, equated_columns = joincond.create_lazy_clause( reverse_direction=True ) self.assert_compile(lazywhere, "lft.id = :param_1") def test_lazy_clause_o2m_o_side_none(self): # test for #2948. When the join is "o.id == m.oid # AND o.something == something", # we don't want 'o' brought into the lazy load for 'm' joincond = self._join_fixture_o2m_o_side_none() lazywhere, bind_to_col, equated_columns = joincond.create_lazy_clause() self.assert_compile( lazywhere, ":param_1 = rgt.lid AND :param_2 = :x_1", checkparams={"param_1": None, "param_2": None, "x_1": 5}, ) def test_lazy_clause_o2m_o_side_none_reverse(self): # continued test for #2948. joincond = self._join_fixture_o2m_o_side_none() lazywhere, bind_to_col, equated_columns = joincond.create_lazy_clause( reverse_direction=True ) self.assert_compile( lazywhere, "lft.id = :param_1 AND lft.x = :x_1", checkparams={"param_1": None, "x_1": 5}, ) def test_lazy_clause_remote_local_multiple_ref(self): joincond = self._join_fixture_remote_local_multiple_ref() lazywhere, bind_to_col, equated_columns = joincond.create_lazy_clause() self.assert_compile( lazywhere, ":param_1 = selfref.sid OR selfref.sid = :param_1", checkparams={"param_1": None}, ) class DeannotateCorrectlyTest(fixtures.TestBase): def test_pj_deannotates(self): from sqlalchemy.ext.declarative import declarative_base Base = declarative_base() class A(Base): __tablename__ = "a" id = Column(Integer, primary_key=True) class B(Base): __tablename__ = "b" id = Column(Integer, primary_key=True) a_id = Column(ForeignKey(A.id)) a = relationship(A) eq_( B.a.property.primaryjoin.left._annotations, {"parentmapper": A.__mapper__, "remote": True}, ) eq_( B.a.property.primaryjoin.right._annotations, {"foreign": True, "local": True, "parentmapper": B.__mapper__}, )
mit
igor-rangel7l/novoigorrangel.repository
plugin.video.SportsDevil/service/asn1crypto/keys.py
13
35048
# coding: utf-8 """ ASN.1 type classes for public and private keys. Exports the following items: - DSAPrivateKey() - ECPrivateKey() - EncryptedPrivateKeyInfo() - PrivateKeyInfo() - PublicKeyInfo() - RSAPrivateKey() - RSAPublicKey() Other type classes are defined that help compose the types listed above. """ from __future__ import unicode_literals, division, absolute_import, print_function import hashlib import math from ._elliptic_curve import ( SECP192R1_BASE_POINT, SECP224R1_BASE_POINT, SECP256R1_BASE_POINT, SECP384R1_BASE_POINT, SECP521R1_BASE_POINT, PrimeCurve, PrimePoint, ) from ._errors import unwrap from ._types import type_name, str_cls, byte_cls from .algos import _ForceNullParameters, DigestAlgorithm, EncryptionAlgorithm from .core import ( Any, Asn1Value, BitString, Choice, Integer, IntegerOctetString, Null, ObjectIdentifier, OctetBitString, OctetString, ParsableOctetString, ParsableOctetBitString, Sequence, SequenceOf, SetOf, ) from .util import int_from_bytes, int_to_bytes class OtherPrimeInfo(Sequence): """ Source: https://tools.ietf.org/html/rfc3447#page-46 """ _fields = [ ('prime', Integer), ('exponent', Integer), ('coefficient', Integer), ] class OtherPrimeInfos(SequenceOf): """ Source: https://tools.ietf.org/html/rfc3447#page-46 """ _child_spec = OtherPrimeInfo class RSAPrivateKeyVersion(Integer): """ Original Name: Version Source: https://tools.ietf.org/html/rfc3447#page-45 """ _map = { 0: 'two-prime', 1: 'multi', } class RSAPrivateKey(Sequence): """ Source: https://tools.ietf.org/html/rfc3447#page-45 """ _fields = [ ('version', RSAPrivateKeyVersion), ('modulus', Integer), ('public_exponent', Integer), ('private_exponent', Integer), ('prime1', Integer), ('prime2', Integer), ('exponent1', Integer), ('exponent2', Integer), ('coefficient', Integer), ('other_prime_infos', OtherPrimeInfos, {'optional': True}) ] class RSAPublicKey(Sequence): """ Source: https://tools.ietf.org/html/rfc3447#page-44 """ _fields = [ ('modulus', Integer), ('public_exponent', Integer) ] class DSAPrivateKey(Sequence): """ The ASN.1 structure that OpenSSL uses to store a DSA private key that is not part of a PKCS#8 structure. Reversed engineered from english-language description on linked OpenSSL documentation page. Original Name: None Source: https://www.openssl.org/docs/apps/dsa.html """ _fields = [ ('version', Integer), ('p', Integer), ('q', Integer), ('g', Integer), ('public_key', Integer), ('private_key', Integer), ] class _ECPoint(): """ In both PublicKeyInfo and PrivateKeyInfo, the EC public key is a byte string that is encoded as a bit string. This class adds convenience methods for converting to and from the byte string to a pair of integers that are the X and Y coordinates. """ @classmethod def from_coords(cls, x, y): """ Creates an ECPoint object from the X and Y integer coordinates of the point :param x: The X coordinate, as an integer :param y: The Y coordinate, as an integer :return: An ECPoint object """ x_bytes = int(math.ceil(math.log(x, 2) / 8.0)) y_bytes = int(math.ceil(math.log(y, 2) / 8.0)) num_bytes = max(x_bytes, y_bytes) byte_string = b'\x04' byte_string += int_to_bytes(x, width=num_bytes) byte_string += int_to_bytes(y, width=num_bytes) return cls(byte_string) def to_coords(self): """ Returns the X and Y coordinates for this EC point, as native Python integers :return: A 2-element tuple containing integers (X, Y) """ data = self.native first_byte = data[0:1] # Uncompressed if first_byte == b'\x04': remaining = data[1:] field_len = len(remaining) // 2 x = int_from_bytes(remaining[0:field_len]) y = int_from_bytes(remaining[field_len:]) return (x, y) if first_byte not in set([b'\x02', b'\x03']): raise ValueError(unwrap( ''' Invalid EC public key - first byte is incorrect ''' )) raise ValueError(unwrap( ''' Compressed representations of EC public keys are not supported due to patent US6252960 ''' )) class ECPoint(OctetString, _ECPoint): pass class ECPointBitString(OctetBitString, _ECPoint): pass class SpecifiedECDomainVersion(Integer): """ Source: http://www.secg.org/sec1-v2.pdf page 104 """ _map = { 1: 'ecdpVer1', 2: 'ecdpVer2', 3: 'ecdpVer3', } class FieldType(ObjectIdentifier): """ Original Name: None Source: http://www.secg.org/sec1-v2.pdf page 101 """ _map = { '1.2.840.10045.1.1': 'prime_field', '1.2.840.10045.1.2': 'characteristic_two_field', } class CharacteristicTwoBasis(ObjectIdentifier): """ Original Name: None Source: http://www.secg.org/sec1-v2.pdf page 102 """ _map = { '1.2.840.10045.1.2.1.1': 'gn_basis', '1.2.840.10045.1.2.1.2': 'tp_basis', '1.2.840.10045.1.2.1.3': 'pp_basis', } class Pentanomial(Sequence): """ Source: http://www.secg.org/sec1-v2.pdf page 102 """ _fields = [ ('k1', Integer), ('k2', Integer), ('k3', Integer), ] class CharacteristicTwo(Sequence): """ Original Name: Characteristic-two Source: http://www.secg.org/sec1-v2.pdf page 101 """ _fields = [ ('m', Integer), ('basis', CharacteristicTwoBasis), ('parameters', Any), ] _oid_pair = ('basis', 'parameters') _oid_specs = { 'gn_basis': Null, 'tp_basis': Integer, 'pp_basis': Pentanomial, } class FieldID(Sequence): """ Source: http://www.secg.org/sec1-v2.pdf page 100 """ _fields = [ ('field_type', FieldType), ('parameters', Any), ] _oid_pair = ('field_type', 'parameters') _oid_specs = { 'prime_field': Integer, 'characteristic_two_field': CharacteristicTwo, } class Curve(Sequence): """ Source: http://www.secg.org/sec1-v2.pdf page 104 """ _fields = [ ('a', OctetString), ('b', OctetString), ('seed', OctetBitString, {'optional': True}), ] class SpecifiedECDomain(Sequence): """ Source: http://www.secg.org/sec1-v2.pdf page 103 """ _fields = [ ('version', SpecifiedECDomainVersion), ('field_id', FieldID), ('curve', Curve), ('base', ECPoint), ('order', Integer), ('cofactor', Integer, {'optional': True}), ('hash', DigestAlgorithm, {'optional': True}), ] class NamedCurve(ObjectIdentifier): """ Various named curves Original Name: None Source: https://tools.ietf.org/html/rfc3279#page-23, https://tools.ietf.org/html/rfc5480#page-5 """ _map = { # https://tools.ietf.org/html/rfc3279#page-23 '1.2.840.10045.3.0.1': 'c2pnb163v1', '1.2.840.10045.3.0.2': 'c2pnb163v2', '1.2.840.10045.3.0.3': 'c2pnb163v3', '1.2.840.10045.3.0.4': 'c2pnb176w1', '1.2.840.10045.3.0.5': 'c2tnb191v1', '1.2.840.10045.3.0.6': 'c2tnb191v2', '1.2.840.10045.3.0.7': 'c2tnb191v3', '1.2.840.10045.3.0.8': 'c2onb191v4', '1.2.840.10045.3.0.9': 'c2onb191v5', '1.2.840.10045.3.0.10': 'c2pnb208w1', '1.2.840.10045.3.0.11': 'c2tnb239v1', '1.2.840.10045.3.0.12': 'c2tnb239v2', '1.2.840.10045.3.0.13': 'c2tnb239v3', '1.2.840.10045.3.0.14': 'c2onb239v4', '1.2.840.10045.3.0.15': 'c2onb239v5', '1.2.840.10045.3.0.16': 'c2pnb272w1', '1.2.840.10045.3.0.17': 'c2pnb304w1', '1.2.840.10045.3.0.18': 'c2tnb359v1', '1.2.840.10045.3.0.19': 'c2pnb368w1', '1.2.840.10045.3.0.20': 'c2tnb431r1', '1.2.840.10045.3.1.2': 'prime192v2', '1.2.840.10045.3.1.3': 'prime192v3', '1.2.840.10045.3.1.4': 'prime239v1', '1.2.840.10045.3.1.5': 'prime239v2', '1.2.840.10045.3.1.6': 'prime239v3', # https://tools.ietf.org/html/rfc5480#page-5 '1.3.132.0.1': 'sect163k1', '1.3.132.0.15': 'sect163r2', '1.2.840.10045.3.1.1': 'secp192r1', '1.3.132.0.33': 'secp224r1', '1.3.132.0.26': 'sect233k1', '1.2.840.10045.3.1.7': 'secp256r1', '1.3.132.0.27': 'sect233r1', '1.3.132.0.16': 'sect283k1', '1.3.132.0.17': 'sect283r1', '1.3.132.0.34': 'secp384r1', '1.3.132.0.36': 'sect409k1', '1.3.132.0.37': 'sect409r1', '1.3.132.0.35': 'secp521r1', '1.3.132.0.38': 'sect571k1', '1.3.132.0.39': 'sect571r1', } class ECDomainParameters(Choice): """ Source: http://www.secg.org/sec1-v2.pdf page 102 """ _alternatives = [ ('specified', SpecifiedECDomain), ('named', NamedCurve), ('implicit_ca', Null), ] class ECPrivateKeyVersion(Integer): """ Original Name: None Source: http://www.secg.org/sec1-v2.pdf page 108 """ _map = { 1: 'ecPrivkeyVer1', } class ECPrivateKey(Sequence): """ Source: http://www.secg.org/sec1-v2.pdf page 108 """ _fields = [ ('version', ECPrivateKeyVersion), ('private_key', IntegerOctetString), ('parameters', ECDomainParameters, {'tag_type': 'explicit', 'tag': 0, 'optional': True}), ('public_key', ECPointBitString, {'tag_type': 'explicit', 'tag': 1, 'optional': True}), ] class DSAParams(Sequence): """ Parameters for a DSA public or private key Original Name: Dss-Parms Source: https://tools.ietf.org/html/rfc3279#page-9 """ _fields = [ ('p', Integer), ('q', Integer), ('g', Integer), ] class Attribute(Sequence): """ Source: https://www.itu.int/rec/dologin_pub.asp?lang=e&id=T-REC-X.501-198811-S!!PDF-E&type=items page 8 """ _fields = [ ('type', ObjectIdentifier), ('values', SetOf, {'spec': Any}), ] class Attributes(SetOf): """ Source: https://tools.ietf.org/html/rfc5208#page-3 """ _child_spec = Attribute class PrivateKeyAlgorithmId(ObjectIdentifier): """ These OIDs for various public keys are reused when storing private keys inside of a PKCS#8 structure Original Name: None Source: https://tools.ietf.org/html/rfc3279 """ _map = { # https://tools.ietf.org/html/rfc3279#page-19 '1.2.840.113549.1.1.1': 'rsa', # https://tools.ietf.org/html/rfc3279#page-18 '1.2.840.10040.4.1': 'dsa', # https://tools.ietf.org/html/rfc3279#page-13 '1.2.840.10045.2.1': 'ec', } class PrivateKeyAlgorithm(_ForceNullParameters, Sequence): """ Original Name: PrivateKeyAlgorithmIdentifier Source: https://tools.ietf.org/html/rfc5208#page-3 """ _fields = [ ('algorithm', PrivateKeyAlgorithmId), ('parameters', Any, {'optional': True}), ] _oid_pair = ('algorithm', 'parameters') _oid_specs = { 'dsa': DSAParams, 'ec': ECDomainParameters, } class PrivateKeyInfo(Sequence): """ Source: https://tools.ietf.org/html/rfc5208#page-3 """ _fields = [ ('version', Integer), ('private_key_algorithm', PrivateKeyAlgorithm), ('private_key', ParsableOctetString), ('attributes', Attributes, {'tag_type': 'implicit', 'tag': 0, 'optional': True}), ] def _private_key_spec(self): algorithm = self['private_key_algorithm']['algorithm'].native return { 'rsa': RSAPrivateKey, 'dsa': Integer, 'ec': ECPrivateKey, }[algorithm] _spec_callbacks = { 'private_key': _private_key_spec } _algorithm = None _bit_size = None _public_key = None _fingerprint = None @classmethod def wrap(cls, private_key, algorithm): """ Wraps a private key in a PrivateKeyInfo structure :param private_key: A byte string or Asn1Value object of the private key :param algorithm: A unicode string of "rsa", "dsa" or "ec" :return: A PrivateKeyInfo object """ if not isinstance(private_key, byte_cls) and not isinstance(private_key, Asn1Value): raise TypeError(unwrap( ''' private_key must be a byte string or Asn1Value, not %s ''', type_name(private_key) )) if algorithm == 'rsa': if not isinstance(private_key, RSAPrivateKey): private_key = RSAPrivateKey.load(private_key) params = Null() elif algorithm == 'dsa': if not isinstance(private_key, DSAPrivateKey): private_key = DSAPrivateKey.load(private_key) params = DSAParams() params['p'] = private_key['p'] params['q'] = private_key['q'] params['g'] = private_key['g'] public_key = private_key['public_key'] private_key = private_key['private_key'] elif algorithm == 'ec': if not isinstance(private_key, ECPrivateKey): private_key = ECPrivateKey.load(private_key) else: private_key = private_key.copy() params = private_key['parameters'] del private_key['parameters'] else: raise ValueError(unwrap( ''' algorithm must be one of "rsa", "dsa", "ec", not %s ''', repr(algorithm) )) private_key_algo = PrivateKeyAlgorithm() private_key_algo['algorithm'] = PrivateKeyAlgorithmId(algorithm) private_key_algo['parameters'] = params container = cls() container._algorithm = algorithm container['version'] = Integer(0) container['private_key_algorithm'] = private_key_algo container['private_key'] = private_key # Here we save the DSA public key if possible since it is not contained # within the PKCS#8 structure for a DSA key if algorithm == 'dsa': container._public_key = public_key return container def _compute_public_key(self): """ Computes the public key corresponding to the current private key. :return: For RSA keys, an RSAPublicKey object. For DSA keys, an Integer object. For EC keys, an ECPointBitString. """ if self.algorithm == 'dsa': params = self['private_key_algorithm']['parameters'] return Integer(pow( params['g'].native, self['private_key'].parsed.native, params['p'].native )) if self.algorithm == 'rsa': key = self['private_key'].parsed return RSAPublicKey({ 'modulus': key['modulus'], 'public_exponent': key['public_exponent'], }) if self.algorithm == 'ec': curve_type, details = self.curve if curve_type == 'implicit_ca': raise ValueError(unwrap( ''' Unable to compute public key for EC key using Implicit CA parameters ''' )) if curve_type == 'specified': if details['field_id']['field_type'] == 'characteristic_two_field': raise ValueError(unwrap( ''' Unable to compute public key for EC key over a characteristic two field ''' )) curve = PrimeCurve( details['field_id']['parameters'], int_from_bytes(details['curve']['a']), int_from_bytes(details['curve']['b']) ) base_x, base_y = self['private_key_algorithm']['parameters'].chosen['base'].to_coords() base_point = PrimePoint(curve, base_x, base_y) elif curve_type == 'named': if details not in ('secp192r1', 'secp224r1', 'secp256r1', 'secp384r1', 'secp521r1'): raise ValueError(unwrap( ''' Unable to compute public key for EC named curve %s, parameters not currently included ''', details )) base_point = { 'secp192r1': SECP192R1_BASE_POINT, 'secp224r1': SECP224R1_BASE_POINT, 'secp256r1': SECP256R1_BASE_POINT, 'secp384r1': SECP384R1_BASE_POINT, 'secp521r1': SECP521R1_BASE_POINT, }[details] public_point = base_point * self['private_key'].parsed['private_key'].native return ECPointBitString.from_coords(public_point.x, public_point.y) def unwrap(self): """ Unwraps the private key into an RSAPrivateKey, DSAPrivateKey or ECPrivateKey object :return: An RSAPrivateKey, DSAPrivateKey or ECPrivateKey object """ if self.algorithm == 'rsa': return self['private_key'].parsed if self.algorithm == 'dsa': params = self['private_key_algorithm']['parameters'] return DSAPrivateKey({ 'version': 0, 'p': params['p'], 'q': params['q'], 'g': params['g'], 'public_key': self.public_key, 'private_key': self['private_key'].parsed, }) if self.algorithm == 'ec': output = self['private_key'].parsed output['parameters'] = self['private_key_algorithm']['parameters'] output['public_key'] = self.public_key return output @property def curve(self): """ Returns information about the curve used for an EC key :raises: ValueError - when the key is not an EC key :return: A two-element tuple, with the first element being a unicode string of "implicit_ca", "specified" or "named". If the first element is "implicit_ca", the second is None. If "specified", the second is an OrderedDict that is the native version of SpecifiedECDomain. If "named", the second is a unicode string of the curve name. """ if self.algorithm != 'ec': raise ValueError(unwrap( ''' Only EC keys have a curve, this key is %s ''', self.algorithm.upper() )) params = self['private_key_algorithm']['parameters'] chosen = params.chosen if params.name == 'implicit_ca': value = None else: value = chosen.native return (params.name, value) @property def hash_algo(self): """ Returns the name of the family of hash algorithms used to generate a DSA key :raises: ValueError - when the key is not a DSA key :return: A unicode string of "sha1" or "sha2" """ if self.algorithm != 'dsa': raise ValueError(unwrap( ''' Only DSA keys are generated using a hash algorithm, this key is %s ''', self.algorithm.upper() )) byte_len = math.log(self['private_key_algorithm']['parameters']['q'].native, 2) / 8 return 'sha1' if byte_len <= 20 else 'sha2' @property def algorithm(self): """ :return: A unicode string of "rsa", "dsa" or "ec" """ if self._algorithm is None: self._algorithm = self['private_key_algorithm']['algorithm'].native return self._algorithm @property def bit_size(self): """ :return: The bit size of the private key, as an integer """ if self._bit_size is None: if self.algorithm == 'rsa': prime = self['private_key'].parsed['modulus'].native elif self.algorithm == 'dsa': prime = self['private_key_algorithm']['parameters']['p'].native elif self.algorithm == 'ec': prime = self['private_key'].parsed['private_key'].native self._bit_size = int(math.ceil(math.log(prime, 2))) modulus = self._bit_size % 8 if modulus != 0: self._bit_size += 8 - modulus return self._bit_size @property def byte_size(self): """ :return: The byte size of the private key, as an integer """ return int(math.ceil(self.bit_size / 8)) @property def public_key(self): """ :return: If an RSA key, an RSAPublicKey object. If a DSA key, an Integer object. If an EC key, an ECPointBitString object. """ if self._public_key is None: if self.algorithm == 'ec': key = self['private_key'].parsed if key['public_key']: self._public_key = key['public_key'].untag() else: self._public_key = self._compute_public_key() else: self._public_key = self._compute_public_key() return self._public_key @property def public_key_info(self): """ :return: A PublicKeyInfo object derived from this private key. """ return PublicKeyInfo({ 'algorithm': { 'algorithm': self.algorithm, 'parameters': self['private_key_algorithm']['parameters'] }, 'public_key': self.public_key }) @property def fingerprint(self): """ Creates a fingerprint that can be compared with a public key to see if the two form a pair. This fingerprint is not compatiable with fingerprints generated by any other software. :return: A byte string that is a sha256 hash of selected components (based on the key type) """ if self._fingerprint is None: params = self['private_key_algorithm']['parameters'] key = self['private_key'].parsed if self.algorithm == 'rsa': to_hash = '%d:%d' % ( key['modulus'].native, key['public_exponent'].native, ) elif self.algorithm == 'dsa': public_key = self.public_key to_hash = '%d:%d:%d:%d' % ( params['p'].native, params['q'].native, params['g'].native, public_key.native, ) elif self.algorithm == 'ec': public_key = key['public_key'].native if public_key is None: public_key = self.public_key.native if params.name == 'named': to_hash = '%s:' % params.chosen.native to_hash = to_hash.encode('utf-8') to_hash += public_key elif params.name == 'implicit_ca': to_hash = public_key elif params.name == 'specified': to_hash = '%s:' % params.chosen['field_id']['parameters'].native to_hash = to_hash.encode('utf-8') to_hash += b':' + params.chosen['curve']['a'].native to_hash += b':' + params.chosen['curve']['b'].native to_hash += public_key if isinstance(to_hash, str_cls): to_hash = to_hash.encode('utf-8') self._fingerprint = hashlib.sha256(to_hash).digest() return self._fingerprint class EncryptedPrivateKeyInfo(Sequence): """ Source: https://tools.ietf.org/html/rfc5208#page-4 """ _fields = [ ('encryption_algorithm', EncryptionAlgorithm), ('encrypted_data', OctetString), ] # These structures are from https://tools.ietf.org/html/rfc3279 class ValidationParms(Sequence): """ Source: https://tools.ietf.org/html/rfc3279#page-10 """ _fields = [ ('seed', BitString), ('pgen_counter', Integer), ] class DomainParameters(Sequence): """ Source: https://tools.ietf.org/html/rfc3279#page-10 """ _fields = [ ('p', Integer), ('g', Integer), ('q', Integer), ('j', Integer, {'optional': True}), ('validation_params', ValidationParms, {'optional': True}), ] class PublicKeyAlgorithmId(ObjectIdentifier): """ Original Name: None Source: https://tools.ietf.org/html/rfc3279 """ _map = { # https://tools.ietf.org/html/rfc3279#page-19 '1.2.840.113549.1.1.1': 'rsa', # https://tools.ietf.org/html/rfc3279#page-18 '1.2.840.10040.4.1': 'dsa', # https://tools.ietf.org/html/rfc3279#page-13 '1.2.840.10045.2.1': 'ec', # https://tools.ietf.org/html/rfc3279#page-10 '1.2.840.10046.2.1': 'dh', } class PublicKeyAlgorithm(_ForceNullParameters, Sequence): """ Original Name: AlgorithmIdentifier Source: https://tools.ietf.org/html/rfc5280#page-18 """ _fields = [ ('algorithm', PublicKeyAlgorithmId), ('parameters', Any, {'optional': True}), ] _oid_pair = ('algorithm', 'parameters') _oid_specs = { 'dsa': DSAParams, 'ec': ECDomainParameters, 'dh': DomainParameters, } class PublicKeyInfo(Sequence): """ Original Name: SubjectPublicKeyInfo Source: https://tools.ietf.org/html/rfc5280#page-17 """ _fields = [ ('algorithm', PublicKeyAlgorithm), ('public_key', ParsableOctetBitString), ] def _public_key_spec(self): algorithm = self['algorithm']['algorithm'].native return { 'rsa': RSAPublicKey, 'dsa': Integer, # We override the field spec with ECPoint so that users can easily # decompose the byte string into the constituent X and Y coords 'ec': (ECPointBitString, None), 'dh': Integer, }[algorithm] _spec_callbacks = { 'public_key': _public_key_spec } _algorithm = None _bit_size = None _fingerprint = None _sha1 = None _sha256 = None @classmethod def wrap(cls, public_key, algorithm): """ Wraps a public key in a PublicKeyInfo structure :param public_key: A byte string or Asn1Value object of the public key :param algorithm: A unicode string of "rsa" :return: A PublicKeyInfo object """ if not isinstance(public_key, byte_cls) and not isinstance(public_key, Asn1Value): raise TypeError(unwrap( ''' public_key must be a byte string or Asn1Value, not %s ''', type_name(public_key) )) if algorithm != 'rsa': raise ValueError(unwrap( ''' algorithm must "rsa", not %s ''', repr(algorithm) )) algo = PublicKeyAlgorithm() algo['algorithm'] = PublicKeyAlgorithmId(algorithm) algo['parameters'] = Null() container = cls() container['algorithm'] = algo if isinstance(public_key, Asn1Value): public_key = public_key.untag().dump() container['public_key'] = ParsableOctetBitString(public_key) return container def unwrap(self): """ Unwraps an RSA public key into an RSAPublicKey object. Does not support DSA or EC public keys since they do not have an unwrapped form. :return: An RSAPublicKey object """ if self.algorithm == 'rsa': return self['public_key'].parsed key_type = self.algorithm.upper() a_an = 'an' if key_type == 'EC' else 'a' raise ValueError(unwrap( ''' Only RSA public keys may be unwrapped - this key is %s %s public key ''', a_an, key_type )) @property def curve(self): """ Returns information about the curve used for an EC key :raises: ValueError - when the key is not an EC key :return: A two-element tuple, with the first element being a unicode string of "implicit_ca", "specified" or "named". If the first element is "implicit_ca", the second is None. If "specified", the second is an OrderedDict that is the native version of SpecifiedECDomain. If "named", the second is a unicode string of the curve name. """ if self.algorithm != 'ec': raise ValueError(unwrap( ''' Only EC keys have a curve, this key is %s ''', self.algorithm.upper() )) params = self['algorithm']['parameters'] chosen = params.chosen if params.name == 'implicit_ca': value = None else: value = chosen.native return (params.name, value) @property def hash_algo(self): """ Returns the name of the family of hash algorithms used to generate a DSA key :raises: ValueError - when the key is not a DSA key :return: A unicode string of "sha1" or "sha2" or None if no parameters are present """ if self.algorithm != 'dsa': raise ValueError(unwrap( ''' Only DSA keys are generated using a hash algorithm, this key is %s ''', self.algorithm.upper() )) parameters = self['algorithm']['parameters'] if parameters.native is None: return None byte_len = math.log(parameters['q'].native, 2) / 8 return 'sha1' if byte_len <= 20 else 'sha2' @property def algorithm(self): """ :return: A unicode string of "rsa", "dsa" or "ec" """ if self._algorithm is None: self._algorithm = self['algorithm']['algorithm'].native return self._algorithm @property def bit_size(self): """ :return: The bit size of the public key, as an integer """ if self._bit_size is None: if self.algorithm == 'ec': self._bit_size = ((len(self['public_key'].native) - 1) / 2) * 8 else: if self.algorithm == 'rsa': prime = self['public_key'].parsed['modulus'].native elif self.algorithm == 'dsa': prime = self['algorithm']['parameters']['p'].native self._bit_size = int(math.ceil(math.log(prime, 2))) modulus = self._bit_size % 8 if modulus != 0: self._bit_size += 8 - modulus return self._bit_size @property def byte_size(self): """ :return: The byte size of the public key, as an integer """ return int(math.ceil(self.bit_size / 8)) @property def sha1(self): """ :return: The SHA1 hash of the DER-encoded bytes of this public key info """ if self._sha1 is None: self._sha1 = hashlib.sha1(byte_cls(self['public_key'])).digest() return self._sha1 @property def sha256(self): """ :return: The SHA-256 hash of the DER-encoded bytes of this public key info """ if self._sha256 is None: self._sha256 = hashlib.sha256(byte_cls(self['public_key'])).digest() return self._sha256 @property def fingerprint(self): """ Creates a fingerprint that can be compared with a private key to see if the two form a pair. This fingerprint is not compatiable with fingerprints generated by any other software. :return: A byte string that is a sha256 hash of selected components (based on the key type) """ if self._fingerprint is None: key_type = self['algorithm']['algorithm'].native params = self['algorithm']['parameters'] if key_type == 'rsa': key = self['public_key'].parsed to_hash = '%d:%d' % ( key['modulus'].native, key['public_exponent'].native, ) elif key_type == 'dsa': key = self['public_key'].parsed to_hash = '%d:%d:%d:%d' % ( params['p'].native, params['q'].native, params['g'].native, key.native, ) elif key_type == 'ec': key = self['public_key'] if params.name == 'named': to_hash = '%s:' % params.chosen.native to_hash = to_hash.encode('utf-8') to_hash += key.native elif params.name == 'implicit_ca': to_hash = key.native elif params.name == 'specified': to_hash = '%s:' % params.chosen['field_id']['parameters'].native to_hash = to_hash.encode('utf-8') to_hash += b':' + params.chosen['curve']['a'].native to_hash += b':' + params.chosen['curve']['b'].native to_hash += key.native if isinstance(to_hash, str_cls): to_hash = to_hash.encode('utf-8') self._fingerprint = hashlib.sha256(to_hash).digest() return self._fingerprint
gpl-2.0
bhansa/fireball
pyvenv/Lib/site-packages/pip/vcs/bazaar.py
514
3803
from __future__ import absolute_import import logging import os import tempfile # TODO: Get this into six.moves.urllib.parse try: from urllib import parse as urllib_parse except ImportError: import urlparse as urllib_parse from pip.utils import rmtree, display_path from pip.vcs import vcs, VersionControl from pip.download import path_to_url logger = logging.getLogger(__name__) class Bazaar(VersionControl): name = 'bzr' dirname = '.bzr' repo_name = 'branch' schemes = ( 'bzr', 'bzr+http', 'bzr+https', 'bzr+ssh', 'bzr+sftp', 'bzr+ftp', 'bzr+lp', ) def __init__(self, url=None, *args, **kwargs): super(Bazaar, self).__init__(url, *args, **kwargs) # Python >= 2.7.4, 3.3 doesn't have uses_fragment or non_hierarchical # Register lp but do not expose as a scheme to support bzr+lp. if getattr(urllib_parse, 'uses_fragment', None): urllib_parse.uses_fragment.extend(['lp']) urllib_parse.non_hierarchical.extend(['lp']) def export(self, location): """ Export the Bazaar repository at the url to the destination location """ temp_dir = tempfile.mkdtemp('-export', 'pip-') self.unpack(temp_dir) if os.path.exists(location): # Remove the location to make sure Bazaar can export it correctly rmtree(location) try: self.run_command(['export', location], cwd=temp_dir, show_stdout=False) finally: rmtree(temp_dir) def switch(self, dest, url, rev_options): self.run_command(['switch', url], cwd=dest) def update(self, dest, rev_options): self.run_command(['pull', '-q'] + rev_options, cwd=dest) def obtain(self, dest): url, rev = self.get_url_rev() if rev: rev_options = ['-r', rev] rev_display = ' (to revision %s)' % rev else: rev_options = [] rev_display = '' if self.check_destination(dest, url, rev_options, rev_display): logger.info( 'Checking out %s%s to %s', url, rev_display, display_path(dest), ) self.run_command(['branch', '-q'] + rev_options + [url, dest]) def get_url_rev(self): # hotfix the URL scheme after removing bzr+ from bzr+ssh:// readd it url, rev = super(Bazaar, self).get_url_rev() if url.startswith('ssh://'): url = 'bzr+' + url return url, rev def get_url(self, location): urls = self.run_command(['info'], show_stdout=False, cwd=location) for line in urls.splitlines(): line = line.strip() for x in ('checkout of branch: ', 'parent branch: '): if line.startswith(x): repo = line.split(x)[1] if self._is_local_repository(repo): return path_to_url(repo) return repo return None def get_revision(self, location): revision = self.run_command( ['revno'], show_stdout=False, cwd=location) return revision.splitlines()[-1] def get_src_requirement(self, dist, location): repo = self.get_url(location) if not repo: return None if not repo.lower().startswith('bzr:'): repo = 'bzr+' + repo egg_project_name = dist.egg_name().split('-', 1)[0] current_rev = self.get_revision(location) return '%s@%s#egg=%s' % (repo, current_rev, egg_project_name) def check_version(self, dest, rev_options): """Always assume the versions don't match""" return False vcs.register(Bazaar)
gpl-3.0
codingjoe/viewflow
tests/integration/tests/test_flow_func.py
2
2022
from django.test import TestCase from viewflow import flow from viewflow.base import Flow, this from .. import integration_test def create_test_flow(activation): activation.prepare() activation.done() return activation @flow.flow_func(task_loader=lambda flow_task, process: process.get_task(FunctionFlow.func_task)) def function_task(activation, process): activation.prepare() activation.done() def handler(activation): pass class FunctionFlow(Flow): start = flow.StartFunction(create_test_flow).Next(this.func_task) default_start = flow.StartFunction().Next(this.func_task) inline_start = flow.StartFunction().Next(this.func_task) func_task = flow.Function(function_task).Next(this.handler_task) handler_task = flow.Handler(handler).Next(this.end) end = flow.End() def inline_start_func(self, activation): activation.prepare() activation.done() self.inline_start_func_called = True return activation @integration_test class TestFunctionFlow(TestCase): def test_function_flow(self): act = FunctionFlow.start.run() FunctionFlow.func_task.run(act.process) tasks = act.process.task_set.all() self.assertEqual(4, tasks.count()) self.assertTrue(all(task.finished is not None for task in tasks)) def test_function_default_start(self): act = FunctionFlow.default_start.run() FunctionFlow.func_task.run(act.process) tasks = act.process.task_set.all() self.assertEqual(4, tasks.count()) self.assertTrue(all(task.finished is not None for task in tasks)) def test_function_inline_start(self): act = FunctionFlow.inline_start.run() self.assertTrue(getattr(FunctionFlow.instance, 'inline_start_func_called', False)) FunctionFlow.func_task.run(act.process) tasks = act.process.task_set.all() self.assertEqual(4, tasks.count()) self.assertTrue(all(task.finished is not None for task in tasks))
agpl-3.0
noroutine/ansible
lib/ansible/playbook/role_include.py
23
5507
# # This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <http://www.gnu.org/licenses/>. # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) __metaclass__ = type from os.path import basename from ansible.errors import AnsibleParserError from ansible.playbook.attribute import FieldAttribute from ansible.playbook.task_include import TaskInclude from ansible.playbook.role import Role from ansible.playbook.role.include import RoleInclude try: from __main__ import display except ImportError: from ansible.utils.display import Display display = Display() __all__ = ['IncludeRole'] class IncludeRole(TaskInclude): """ A Role include is derived from a regular role to handle the special circumstances related to the `- include_role: ...` """ BASE = ('name', 'role') # directly assigned FROM_ARGS = ('tasks_from', 'vars_from', 'defaults_from') # used to populate from dict in role OTHER_ARGS = ('private', 'allow_duplicates') # assigned to matching property VALID_ARGS = tuple(frozenset(BASE + FROM_ARGS + OTHER_ARGS)) # all valid args _inheritable = False # ================================================================================= # ATTRIBUTES # private as this is a 'module options' vs a task property _allow_duplicates = FieldAttribute(isa='bool', default=True, private=True) _private = FieldAttribute(isa='bool', default=None, private=True) def __init__(self, block=None, role=None, task_include=None): super(IncludeRole, self).__init__(block=block, role=role, task_include=task_include) self._from_files = {} self._parent_role = role self._role_name = None self._role_path = None def get_block_list(self, play=None, variable_manager=None, loader=None): # only need play passed in when dynamic if play is None: myplay = self._parent._play else: myplay = play ri = RoleInclude.load(self._role_name, play=myplay, variable_manager=variable_manager, loader=loader) ri.vars.update(self.vars) # build role actual_role = Role.load(ri, myplay, parent_role=self._parent_role, from_files=self._from_files) actual_role._metadata.allow_duplicates = self.allow_duplicates # save this for later use self._role_path = actual_role._role_path # compile role with parent roles as dependencies to ensure they inherit # variables if not self._parent_role: dep_chain = [] else: dep_chain = list(self._parent_role._parents) dep_chain.append(self._parent_role) blocks = actual_role.compile(play=myplay, dep_chain=dep_chain) for b in blocks: b._parent = self # updated available handlers in play handlers = actual_role.get_handler_blocks(play=myplay) myplay.handlers = myplay.handlers + handlers return blocks, handlers @staticmethod def load(data, block=None, role=None, task_include=None, variable_manager=None, loader=None): ir = IncludeRole(block, role, task_include=task_include).load_data(data, variable_manager=variable_manager, loader=loader) # Validate options my_arg_names = frozenset(ir.args.keys()) # name is needed, or use role as alias ir._role_name = ir.args.get('name', ir.args.get('role')) if ir._role_name is None: raise AnsibleParserError("'name' is a required field for include_role.") # validate bad args, otherwise we silently ignore bad_opts = my_arg_names.difference(IncludeRole.VALID_ARGS) if bad_opts: raise AnsibleParserError('Invalid options for include_role: %s' % ','.join(list(bad_opts))) # build options for role includes for key in my_arg_names.intersection(IncludeRole.FROM_ARGS): from_key = key.replace('_from', '') ir._from_files[from_key] = basename(ir.args.get(key)) # manual list as otherwise the options would set other task parameters we don't want. for option in my_arg_names.intersection(IncludeRole.OTHER_ARGS): setattr(ir, option, ir.args.get(option)) return ir def copy(self, exclude_parent=False, exclude_tasks=False): new_me = super(IncludeRole, self).copy(exclude_parent=exclude_parent, exclude_tasks=exclude_tasks) new_me.statically_loaded = self.statically_loaded new_me._from_files = self._from_files.copy() new_me._parent_role = self._parent_role new_me._role_name = self._role_name new_me._role_path = self._role_path return new_me def get_include_params(self): v = super(IncludeRole, self).get_include_params() if self._parent_role: v.update(self._parent_role.get_role_params()) return v
gpl-3.0
hnousiainen/pghoard
pghoard/pgutil.py
2
3543
# Copied from https://github.com/ohmu/ohmu_common_py ohmu_common_py/pgutil.py version 0.0.1-0-unknown-fa54b44 """ pghoard - postgresql utility functions Copyright (c) 2015 Ohmu Ltd See LICENSE for details """ try: from urllib.parse import urlparse, parse_qs # pylint: disable=no-name-in-module, import-error except ImportError: from urlparse import urlparse, parse_qs # pylint: disable=no-name-in-module, import-error def create_connection_string(connection_info): return " ".join("{}='{}'".format(k, str(v).replace("'", "\\'")) for k, v in sorted(connection_info.items())) def mask_connection_info(info): masked_info = get_connection_info(info) password = masked_info.pop("password", None) return "{0}; {1} password".format( create_connection_string(masked_info), "no" if password is None else "hidden") def get_connection_info_from_config_line(line): _, value = line.split("=", 1) value = value.strip()[1:-1].replace("''", "'") return get_connection_info(value) def get_connection_info(info): """turn a connection info object into a dict or return it if it was a dict already. supports both the traditional libpq format and the new url format""" if isinstance(info, dict): return info.copy() elif info.startswith("postgres://") or info.startswith("postgresql://"): return parse_connection_string_url(info) else: return parse_connection_string_libpq(info) def parse_connection_string_url(url): # drop scheme from the url as some versions of urlparse don't handle # query and path properly for urls with a non-http scheme schemeless_url = url.split(":", 1)[1] p = urlparse(schemeless_url) fields = {} if p.hostname: fields["host"] = p.hostname if p.port: fields["port"] = str(p.port) if p.username: fields["user"] = p.username if p.password is not None: fields["password"] = p.password if p.path and p.path != "/": fields["dbname"] = p.path[1:] for k, v in parse_qs(p.query).items(): fields[k] = v[-1] return fields def parse_connection_string_libpq(connection_string): """parse a postgresql connection string as defined in http://www.postgresql.org/docs/current/static/libpq-connect.html#LIBPQ-CONNSTRING""" fields = {} while True: connection_string = connection_string.strip() if not connection_string: break if "=" not in connection_string: raise ValueError("expecting key=value format in connection_string fragment {!r}".format(connection_string)) key, rem = connection_string.split("=", 1) if rem.startswith("'"): asis, value = False, "" for i in range(1, len(rem)): if asis: value += rem[i] asis = False elif rem[i] == "'": break # end of entry elif rem[i] == "\\": asis = True else: value += rem[i] else: raise ValueError("invalid connection_string fragment {!r}".format(rem)) connection_string = rem[i + 1:] # pylint: disable=undefined-loop-variable else: res = rem.split(None, 1) if len(res) > 1: value, connection_string = res else: value, connection_string = rem, "" fields[key] = value return fields
apache-2.0
vgkholla/ambry
ambry-tools/scripts/layout-analyzer.py
7
7025
#!/usr/bin/python2.7 import argparse import json import sys from collections import defaultdict class Node(object): def __init__(self, node, datacenter): self.node = node self.datacenter = datacenter self.partitions = set() @property def hostname(self): return self.node["hostname"] @property def port(self): return self.node["port"] @property def rack_id(self): if "rackId" in self.node: return self.node["rackId"] return -1 @property def datacenter_name(self): return self.datacenter["name"] def add_partition(self, partition): self.partitions.add(partition) def __repr__(self): return "[hostname: {}, port: {}, dc: {}]".format( self.hostname, self.port, self.datacenter_name) __str__ = __repr__ class Partition(object): def __init__(self, partition): self.partition = partition self.nodes_by_datacenter = defaultdict(set) @property def id(self): return self.partition["id"] def add_node(self, node): self.nodes_by_datacenter[node.datacenter_name].add(node) node.add_partition(self) def racks_used(self, datacenter_name): return {node.rack_id for node in self.nodes_by_datacenter[datacenter_name]} def __repr__(self): return "[id: {}]".format(self.id) __str__ = __repr__ class Layout(object): BALANCE_THRESHOLD = 4.0 def __init__(self, hardware_layout_filename, partition_layout_filename): with open(hardware_layout_filename) as f: self.hardware_layout = json.load(f) with open(partition_layout_filename) as f: self.partition_layout = json.load(f) self.setup() def setup(self): self.node_map = {} self.partition_map = {} self.dc_node_combo_map = defaultdict(lambda: defaultdict(set)) for datacenter_struct in self.hardware_layout["datacenters"]: for node_struct in datacenter_struct["dataNodes"]: k = (node_struct["hostname"], node_struct["port"]) self.node_map[k] = Node(node_struct, datacenter_struct) for partition_struct in self.partition_layout["partitions"]: partition = Partition(partition_struct) if len(partition_struct["replicas"]) == 0: raise Exception("No replicas assigned to partition {}".format(partition.id)) for replica_struct in partition_struct["replicas"]: k = (replica_struct["hostname"], replica_struct["port"]) node = self.node_map[k] partition.add_node(node) for dc, nodes in partition.nodes_by_datacenter.items(): self.dc_node_combo_map[dc][frozenset(nodes)].add(partition) self.partition_map[partition_struct["id"]] = partition def rack_id(self, node_host, node_port): k = (node_host, node_port) if k in self.node_map: return self.node_map[k].rack_id raise Exception("Node {}:{} not found".format(node_host, node_port)) def racks_used(self, partition_id, datacenter_name): return self.partition_map[partition_id].racks_used(datacenter_name) def shared_partitions(self, *nodes): return set.intersection( *(self.node_map[node].partitions for node in nodes) ) def print_report(self): for dc, node_combo_map in self.dc_node_combo_map.items(): print("In datacenter: {}".format(dc)) max_combo = max(node_combo_map, key=lambda k: len(node_combo_map[k])) avg_per_combo = sum(len(partitions) for partitions in node_combo_map.values()) / float(len(node_combo_map)) max_per_combo = len(node_combo_map[max_combo]) print("Num node combos used: {}".format(len(node_combo_map))) print("Average partitions sharing a node combo: {}".format(avg_per_combo)) print("Max partitions sharing a node combo: {} on the following nodes:".format(max_per_combo)) for node in max_combo: print(node) if (float(max_per_combo) / avg_per_combo) > self.BALANCE_THRESHOLD: print("The ratio of max to average number of partitions sharing a node combo " + "exceeds the threshold: {} on this datacenter".format(self.BALANCE_THRESHOLD)) sum_racks, n_partitions, min_racks = 0, 0, sys.maxsize for partition in self.partition_map.values(): num_racks = len(partition.racks_used(dc)) n_partitions += 1 sum_racks += num_racks if num_racks < min_racks: min_racks = num_racks print("Min racks used: {}".format(min_racks)) print("Average racks used: {}".format( float(sum_racks) / n_partitions)) partitions_per_node = [len(node.partitions) for node in self.node_map.values() if node.datacenter_name == dc] print("") def interactive(self): while True: cmd = raw_input(">> ").split() try: if len(cmd) == 0: continue elif cmd[0] == "report": self.print_report() elif cmd[0] == "rack_id": print("Node {}:{} is on rack {}".format( cmd[1], cmd[2], self.rack_id(cmd[1], int(cmd[2])))) elif cmd[0] == "racks_used": print("Partition {} in datacenter {} uses the following racks: {}".format( cmd[1], cmd[2], self.racks_used(int(cmd[1]), cmd[2]))) elif cmd[0] == "shared_partitions": args = [(cmd[i + 1], int(cmd[i + 2])) for i in range(0, len(cmd) - 1, 2)] print("The following nodes:") for hostname, port in args: print(" {}:{}".format(hostname, port)) print("share the following partitions:") print(self.shared_partitions(*args)) else: print("Command not recognized") except Exception: print("Invalid input") print("") def main(): parser = argparse.ArgumentParser( description='Analyze node distribution in a partition layout') parser.add_argument("--interactive", "-i", action="store_true") parser.add_argument('hardware_layout', help='the path to the hardware layout file') parser.add_argument('partition_layout', help='the path to the partition layout file') args = parser.parse_args() layout = Layout(args.hardware_layout, args.partition_layout) if args.interactive: layout.interactive() else: layout.print_report() if __name__ == "__main__": main()
apache-2.0
mlperf/training_results_v0.7
Fujitsu/benchmarks/resnet/implementations/implementation_open/mxnet/3rdparty/tvm/topi/python/topi/testing/roi_align_python.py
2
3126
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. # pylint: disable=invalid-name, too-many-nested-blocks "Roi align in python" import math import numpy as np def roi_align_nchw_python(a_np, rois_np, pooled_size, spatial_scale, sample_ratio): """Roi align in python""" _, channel, height, width = a_np.shape num_roi = rois_np.shape[0] b_np = np.zeros((num_roi, channel, pooled_size, pooled_size), dtype=a_np.dtype) if isinstance(pooled_size, int): pooled_size_h = pooled_size_w = pooled_size else: pooled_size_h, pooled_size_w = pooled_size def _bilinear(b, c, y, x): if y < -1 or y > height or x < -1 or x > width: return 0 y = max(y, 0.0) x = max(x, 0.0) y_low = int(y) x_low = int(x) y_high = min(y_low + 1, height - 1) x_high = min(x_low + 1, width - 1) ly = y - y_low lx = x - x_low return (1 - ly) * (1 - lx) * a_np[b, c, y_low, x_low] + \ (1 - ly) * lx * a_np[b, c, y_low, x_high] + \ ly * (1 - lx) * a_np[b, c, y_high, x_low] + \ ly * lx * a_np[b, c, y_high, x_high] for i in range(num_roi): roi = rois_np[i] batch_index = int(roi[0]) roi_start_w, roi_start_h, roi_end_w, roi_end_h = roi[1:] * spatial_scale roi_h = max(roi_end_h - roi_start_h, 1.0) roi_w = max(roi_end_w - roi_start_w, 1.0) bin_h = roi_h / pooled_size_h bin_w = roi_w / pooled_size_w if sample_ratio > 0: roi_bin_grid_h = roi_bin_grid_w = int(sample_ratio) else: roi_bin_grid_h = int(math.ceil(roi_h / pooled_size)) roi_bin_grid_w = int(math.ceil(roi_w / pooled_size)) count = roi_bin_grid_h * roi_bin_grid_w for c in range(channel): for ph in range(pooled_size_h): for pw in range(pooled_size_w): total = 0. for iy in range(roi_bin_grid_h): for ix in range(roi_bin_grid_w): y = roi_start_h + ph * bin_h + (iy + 0.5) * bin_h / roi_bin_grid_h x = roi_start_w + pw * bin_w + (ix + 0.5) * bin_w / roi_bin_grid_w total += _bilinear(batch_index, c, y, x) b_np[i, c, ph, pw] = total / count return b_np
apache-2.0
zhouxiao-coder/Paddle
demo/introduction/trainer_config.py
3
1344
# Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from paddle.trainer_config_helpers import * # 1. read data. Suppose you saved above python code as dataprovider.py data_file = 'empty.list' with open(data_file, 'w') as f: f.writelines(' ') define_py_data_sources2( train_list=data_file, test_list=None, module='dataprovider', obj='process', args={}) # 2. learning algorithm settings(batch_size=12, learning_rate=1e-3, learning_method=MomentumOptimizer()) # 3. Network configuration x = data_layer(name='x', size=1) y = data_layer(name='y', size=1) y_predict = fc_layer( input=x, param_attr=ParamAttr(name='w'), size=1, act=LinearActivation(), bias_attr=ParamAttr(name='b')) cost = regression_cost(input=y_predict, label=y) outputs(cost)
apache-2.0
caesar2164/edx-platform
common/djangoapps/track/transformers.py
51
16677
""" EventTransformers are data structures that represents events, and modify those events to match the format desired for the tracking logs. They are registered by name (or name prefix) in the EventTransformerRegistry, which is used to apply them to the appropriate events. """ import json import logging from opaque_keys import InvalidKeyError from opaque_keys.edx.keys import UsageKey log = logging.getLogger(__name__) class DottedPathMapping(object): """ Dictionary-like object for creating keys of dotted paths. If a key is created that ends with a dot, it will be treated as a path prefix. Any value whose prefix matches the dotted path can be used as a key for that value, but only the most specific match will be used. """ # TODO: The current implementation of the prefix registry requires # O(number of prefix event transformers) to access an event. If we get a # large number of EventTransformers, it may be worth writing a tree-based # map structure where each node is a segment of the match key, which would # reduce access time to O(len(match.key.split('.'))), or essentially constant # time. def __init__(self, registry=None): self._match_registry = {} self._prefix_registry = {} self.update(registry or {}) def __contains__(self, key): try: _ = self[key] return True except KeyError: return False def __getitem__(self, key): if key in self._match_registry: return self._match_registry[key] if isinstance(key, basestring): # Reverse-sort the keys to find the longest matching prefix. for prefix in sorted(self._prefix_registry, reverse=True): if key.startswith(prefix): return self._prefix_registry[prefix] raise KeyError('Key {} not found in {}'.format(key, type(self))) def __setitem__(self, key, value): if key.endswith('.'): self._prefix_registry[key] = value else: self._match_registry[key] = value def __delitem__(self, key): if key.endswith('.'): del self._prefix_registry[key] else: del self._match_registry[key] def get(self, key, default=None): """ Return `self[key]` if it exists, otherwise, return `None` or `default` if it is specified. """ try: self[key] except KeyError: return default def update(self, dict_): """ Update the mapping with the values in the supplied `dict`. """ for key, value in dict_: self[key] = value def keys(self): """ Return the keys of the mapping, including both exact matches and prefix matches. """ return self._match_registry.keys() + self._prefix_registry.keys() class EventTransformerRegistry(object): """ Registry to track which EventTransformers handle which events. The EventTransformer must define a `match_key` attribute which contains the name or prefix of the event names it tracks. Any `match_key` that ends with a `.` will match all events that share its prefix. A transformer name without a trailing dot only processes exact matches. """ mapping = DottedPathMapping() @classmethod def register(cls, transformer): """ Decorator to register an EventTransformer. It must have a `match_key` class attribute defined. """ cls.mapping[transformer.match_key] = transformer return transformer @classmethod def create_transformer(cls, event): """ Create an EventTransformer of the given event. If no transformer is registered to handle the event, this raises a KeyError. """ name = event.get(u'name') return cls.mapping[name](event) class EventTransformer(dict): """ Creates a transformer to modify analytics events based on event type. To use the transformer, instantiate it using the `EventTransformer.create_transformer()` classmethod with the event dictionary as the sole argument, and then call `transformer.transform()` on the created object to modify the event to the format required for output. Custom transformers will want to define some or all of the following values Attributes: match_key: This is the name of the event you want to transform. If the name ends with a `'.'`, it will be treated as a *prefix transformer*. All other names denote *exact transformers*. A *prefix transformer* will handle any event whose name begins with the name of the prefix transformer. Only the most specific match will be used, so if a transformer exists with a name of `'edx.ui.lms.'` and another transformer has the name `'edx.ui.lms.sequence.'` then an event called `'edx.ui.lms.sequence.tab_selected'` will be handled by the `'edx.ui.lms.sequence.'` transformer. An *exact transformer* will only handle events whose name matches name of the transformer exactly. Exact transformers always take precedence over prefix transformers. Transformers without a name will not be added to the registry, and cannot be accessed via the `EventTransformer.create_transformer()` classmethod. is_legacy_event: If an event is a legacy event, it needs to set event_type to the legacy name for the event, and may need to set certain event fields to maintain backward compatiblity. If an event needs to provide legacy support in some contexts, `is_legacy_event` can be defined as a property to add dynamic behavior. Default: False legacy_event_type: If the event is or can be a legacy event, it should define the legacy value for the event_type field here. Processing methods. Override these to provide the behavior needed for your particular EventTransformer: self.process_legacy_fields(): This method should modify the event payload in any way necessary to support legacy event types. It will only be run if `is_legacy_event` returns a True value. self.process_event() This method modifies the event payload unconditionally. It will always be run. """ def __init__(self, *args, **kwargs): super(EventTransformer, self).__init__(*args, **kwargs) self.load_payload() # Properties to be overridden is_legacy_event = False @property def legacy_event_type(self): """ Override this as an attribute or property to provide the value for the event's `event_type`, if it does not match the event's `name`. """ raise NotImplementedError # Convenience properties @property def name(self): """ Returns the event's name. """ return self[u'name'] @property def context(self): """ Returns the event's context dict. """ return self.get(u'context', {}) # Transform methods def load_payload(self): """ Create a data version of self[u'event'] at self.event """ if u'event' in self: if isinstance(self[u'event'], basestring): self.event = json.loads(self[u'event']) else: self.event = self[u'event'] def dump_payload(self): """ Write self.event back to self[u'event']. Keep the same format we were originally given. """ if isinstance(self.get(u'event'), basestring): self[u'event'] = json.dumps(self.event) else: self[u'event'] = self.event def transform(self): """ Transform the event with legacy fields and other necessary modifications. """ if self.is_legacy_event: self._set_legacy_event_type() self.process_legacy_fields() self.process_event() self.dump_payload() def _set_legacy_event_type(self): """ Update the event's `event_type` to the value specified by `self.legacy_event_type`. """ self['event_type'] = self.legacy_event_type def process_legacy_fields(self): """ Override this method to specify how to update event fields to maintain compatibility with legacy events. """ pass def process_event(self): """ Override this method to make unconditional modifications to event fields. """ pass @EventTransformerRegistry.register class SequenceTabSelectedEventTransformer(EventTransformer): """ Transformer to maintain backward compatiblity with seq_goto events. """ match_key = u'edx.ui.lms.sequence.tab_selected' is_legacy_event = True legacy_event_type = u'seq_goto' def process_legacy_fields(self): self.event[u'old'] = self.event[u'current_tab'] self.event[u'new'] = self.event[u'target_tab'] class _BaseLinearSequenceEventTransformer(EventTransformer): """ Common functionality for transforming `edx.ui.lms.sequence.{next,previous}_selected` events. """ offset = None @property def is_legacy_event(self): """ Linear sequence events are legacy events if the origin and target lie within the same sequence. """ return not self.crosses_boundary() def process_legacy_fields(self): """ Set legacy payload fields: old: equal to the new current_tab field new: the tab to which the user is navigating """ self.event[u'old'] = self.event[u'current_tab'] self.event[u'new'] = self.event[u'current_tab'] + self.offset def crosses_boundary(self): """ Returns true if the navigation takes the focus out of the current sequence. """ raise NotImplementedError @EventTransformerRegistry.register class NextSelectedEventTransformer(_BaseLinearSequenceEventTransformer): """ Transformer to maintain backward compatiblity with seq_next events. """ match_key = u'edx.ui.lms.sequence.next_selected' offset = 1 legacy_event_type = u'seq_next' def crosses_boundary(self): """ Returns true if the navigation moves the focus to the next sequence. """ return self.event[u'current_tab'] == self.event[u'tab_count'] @EventTransformerRegistry.register class PreviousSelectedEventTransformer(_BaseLinearSequenceEventTransformer): """ Transformer to maintain backward compatiblity with seq_prev events. """ match_key = u'edx.ui.lms.sequence.previous_selected' offset = -1 legacy_event_type = u'seq_prev' def crosses_boundary(self): """ Returns true if the navigation moves the focus to the previous sequence. """ return self.event[u'current_tab'] == 1 @EventTransformerRegistry.register class VideoEventTransformer(EventTransformer): """ Converts new format video events into the legacy video event format. Mobile devices cannot actually emit events that exactly match their counterparts emitted by the LMS javascript video player. Instead of attempting to get them to do that, we instead insert a transformer here that converts the events they *can* easily emit and converts them into the legacy format. """ match_key = u'edx.video.' name_to_event_type_map = { u'edx.video.played': u'play_video', u'edx.video.paused': u'pause_video', u'edx.video.stopped': u'stop_video', u'edx.video.loaded': u'load_video', u'edx.video.position.changed': u'seek_video', u'edx.video.seeked': u'seek_video', u'edx.video.transcript.shown': u'show_transcript', u'edx.video.transcript.hidden': u'hide_transcript', u'edx.video.language_menu.shown': u'video_show_cc_menu', u'edx.video.language_menu.hidden': u'video_hide_cc_menu', } is_legacy_event = True @property def legacy_event_type(self): """ Return the legacy event_type of the current event """ return self.name_to_event_type_map[self.name] def transform(self): """ Transform the event with necessary modifications if it is one of the expected types of events. """ if self.name in self.name_to_event_type_map: super(VideoEventTransformer, self).transform() def process_event(self): """ Modify event fields. """ # Convert edx.video.seeked to edx.video.position.changed because edx.video.seeked was not intended to actually # ever be emitted. if self.name == "edx.video.seeked": self['name'] = "edx.video.position.changed" if not self.event: return self.set_id_to_usage_key() self.capcase_current_time() self.convert_seek_type() self.disambiguate_skip_and_seek() self.set_page_to_browser_url() self.handle_ios_seek_bug() def set_id_to_usage_key(self): """ Validate that the module_id is a valid usage key, and set the id field accordingly. """ if 'module_id' in self.event: module_id = self.event['module_id'] try: usage_key = UsageKey.from_string(module_id) except InvalidKeyError: log.warning('Unable to parse module_id "%s"', module_id, exc_info=True) else: self.event['id'] = usage_key.html_id() del self.event['module_id'] def capcase_current_time(self): """ Convert the current_time field to currentTime. """ if 'current_time' in self.event: self.event['currentTime'] = self.event.pop('current_time') def convert_seek_type(self): """ Converts seek_type to seek and converts skip|slide to onSlideSeek|onSkipSeek. """ if 'seek_type' in self.event: seek_type = self.event['seek_type'] if seek_type == 'slide': self.event['type'] = "onSlideSeek" elif seek_type == 'skip': self.event['type'] = "onSkipSeek" del self.event['seek_type'] def disambiguate_skip_and_seek(self): """ For the Android build that isn't distinguishing between skip/seek. """ if 'requested_skip_interval' in self.event: if abs(self.event['requested_skip_interval']) != 30: if 'type' in self.event: self.event['type'] = 'onSlideSeek' def set_page_to_browser_url(self): """ If `open_in_browser_url` is specified, set the page to the base of the specified url. """ if 'open_in_browser_url' in self.context: self['page'] = self.context.pop('open_in_browser_url').rpartition('/')[0] def handle_ios_seek_bug(self): """ Handle seek bug in iOS. iOS build 1.0.02 has a bug where it returns a +30 second skip when it should be returning -30. """ if self._build_requests_plus_30_for_minus_30(): if self._user_requested_plus_30_skip(): self.event[u'requested_skip_interval'] = -30 def _build_requests_plus_30_for_minus_30(self): """ Returns True if this build contains the seek bug """ if u'application' in self.context: if all(key in self.context[u'application'] for key in (u'version', u'name')): app_version = self.context[u'application'][u'version'] app_name = self.context[u'application'][u'name'] return app_version == u'1.0.02' and app_name == u'edx.mobileapp.iOS' return False def _user_requested_plus_30_skip(self): """ If the user requested a +30 second skip, return True. """ if u'requested_skip_interval' in self.event and u'type' in self.event: interval = self.event[u'requested_skip_interval'] action = self.event[u'type'] return interval == 30 and action == u'onSkipSeek' else: return False
agpl-3.0
luci/luci-py
appengine/swarming/server/task_request_test.py
2
86434
#!/usr/bin/env vpython # Copyright 2014 The LUCI Authors. All rights reserved. # Use of this source code is governed under the Apache License, Version 2.0 # that can be found in the LICENSE file. import datetime import logging import random import string import sys import unittest import test_env test_env.setup_test_env() from google.protobuf import duration_pb2 from google.protobuf import timestamp_pb2 from google.appengine.api import datastore_errors from google.appengine.ext import ndb from components import auth_testing from components import utils from test_support import test_case from proto.api import swarming_pb2 from server import bq_state from server import config from server import pools_config from server import task_pack from server import task_request # pylint: disable=W0212 def _gen_cipd_input(**kwargs): """Creates a CipdInput.""" args = { u'client_package': task_request.CipdPackage( package_name=u'infra/tools/cipd/${platform}', version=u'git_revision:deadbeef'), u'packages': [ task_request.CipdPackage( package_name=u'rm', path=u'bin', version=u'git_revision:deadbeef'), ], u'server': u'https://chrome-infra-packages.appspot.com' } args.update(kwargs) return task_request.CipdInput(**args) def _gen_properties(**kwargs): """Creates a TaskProperties.""" args = { u'cipd_input': _gen_cipd_input(), u'command': [u'command1', u'arg1'], u'containment': { u'lower_priority': False, u'containment_type': None, u'limit_processes': None, u'limit_total_committed_memory': None, }, u'dimensions': { u'OS': [u'Windows-3.1.1'], u'hostname': [u'localhost'], u'pool': [u'default'], }, u'env': { u'foo': u'bar', u'joe': u'2' }, u'env_prefixes': { u'PATH': [u'local/path'] }, u'execution_timeout_secs': 30, u'grace_period_secs': 30, u'idempotent': False, u'inputs_ref': task_request.FilesRef( isolatedserver=u'https://isolateserver.appspot.com', namespace=u'default-gzip'), u'io_timeout_secs': None, u'has_secret_bytes': False, } args.update(kwargs) args[u'dimensions_data'] = args.pop(u'dimensions') return task_request.TaskProperties(**args) def _gen_request_slices(**kwargs): """Creates a TaskRequest.""" template_apply = kwargs.pop('_template_apply', task_request.TEMPLATE_AUTO) now = utils.utcnow() args = { u'created_ts': now, u'manual_tags': [u'tag:1'], u'name': u'Request name', u'priority': 50, u'task_slices': [ task_request.TaskSlice( expiration_secs=30, properties=_gen_properties()), ], u'user': u'Jesus', u'bot_ping_tolerance_secs': 120, } args.update(kwargs) # Note that ndb model constructor accepts dicts for structured properties. req = task_request.TaskRequest(**args) task_request.init_new_request(req, True, template_apply) return req def _gen_request(properties=None, **kwargs): """Creates a TaskRequest with a single TaskSlice.""" return _gen_request_slices( task_slices=[ task_request.TaskSlice( expiration_secs=30, properties=properties or _gen_properties()), ], **kwargs) def _gen_secret(req, secret_bytes): assert req.key sb = task_request.SecretBytes(secret_bytes=secret_bytes) sb.key = req.secret_bytes_key return sb def _gen_task_template(cache=None, cipd_package=None, env=None): """Builds an unverified pools_config.TaskTemplate for use with _set_pool_config_with_templates. Args: cache (None|dict{name: path}) - cache entries to set. cipd_package (None|dict{(path, pkg): version}) - cipd packages to set. env (None|dict{var: value|(value, prefix)|(value, prefix, soft)}) - envvars to set. The key is always the envvar to set, and the value may be: * the envvar value as a string (prefix=() and soft=False) * A (value, prefix) tuple (soft=False) * A (value, prefix, soft) tuple Returns constructed pools_config.TaskTemplate. """ def env_value(var, combo_value): prefix, soft = (), False if isinstance(combo_value, tuple): assert len(combo_value) in (2, 3), ('unexpected tuple length: %r' % combo_value) if len(combo_value) == 2: value, prefix = combo_value else: value, prefix, soft = combo_value else: value = unicode(combo_value) return pools_config.Env(var, value, tuple(map(unicode, prefix)), soft) return pools_config.TaskTemplate( cache=sorted( pools_config.CacheEntry(unicode(name), unicode(path)) for name, path in (cache or {}).items()), cipd_package=sorted( pools_config.CipdPackage( unicode(path), unicode(pkg), unicode(version)) for (path, pkg), version in (cipd_package or {}).items()), env=sorted( env_value(unicode(var), value) for var, value in (env or {}).items()), inclusions=(), ) class Prop(object): _name = 'foo' class TestCase(test_case.TestCase): def setUp(self): super(TestCase, self).setUp() auth_testing.mock_get_current_identity(self) class TaskRequestPrivateTest(TestCase): def test_validate_task_run_id(self): self.assertEqual( '1d69b9f088008811', task_request._validate_task_run_id(Prop(), '1d69b9f088008811')) self.assertEqual(None, task_request._validate_task_run_id(Prop(), '')) with self.assertRaises(ValueError): task_request._validate_task_run_id(Prop(), '1') def test_validate_cas_instance(self): valid_cas_instance = 'projects/chromium-swarm/instances/default_instance' self.assertEqual( valid_cas_instance, task_request._validate_cas_instance(Prop(), valid_cas_instance)) self.assertEqual(None, task_request._validate_cas_instance(Prop(), '')) with self.assertRaises(datastore_errors.BadValueError): task_request._validate_cas_instance(Prop(), 'invalid') def test_apply_template_simple(self): tt = _gen_task_template( cache={'cache': 'c'}, cipd_package={('cipd', 'some/pkg'): 'latest'}, env={'ENV': ('1', ['a'])}, ) p = task_request.TaskProperties() task_request._apply_task_template(tt, p) self.assertEqual( p, task_request.TaskProperties( env={u'ENV': u'1'}, env_prefixes={u'ENV': [u'a']}, caches=[task_request.CacheEntry(name=u'cache', path=u'c')], cipd_input=task_request.CipdInput(packages=[ task_request.CipdPackage( package_name=u'some/pkg', path=u'cipd', version=u'latest') ]))) def test_apply_template_env_set_error(self): tt = _gen_task_template(env={'ENV': ('1', ['a'])}) p = task_request.TaskProperties(env={u'ENV': u'10'}) with self.assertRaises(ValueError) as ex: task_request._apply_task_template(tt, p) self.assertEqual(ex.exception.message, "request.env[u'ENV'] conflicts with pool's template") def test_apply_template_env_prefix_set_error(self): tt = _gen_task_template(env={'ENV': ('1', ['a'])}) p = task_request.TaskProperties(env_prefixes={u'ENV': [u'b']}) with self.assertRaises(ValueError) as ex: task_request._apply_task_template(tt, p) self.assertEqual( ex.exception.message, "request.env_prefixes[u'ENV'] conflicts with pool's template") def test_apply_template_env_override_soft(self): tt = _gen_task_template(env={'ENV': ('1', ['a'], True)}) p = task_request.TaskProperties(env={u'ENV': u'2'}) task_request._apply_task_template(tt, p) self.assertEqual( p, task_request.TaskProperties( env={u'ENV': u'2'}, env_prefixes={u'ENV': [u'a']}, )) def test_apply_template_env_prefixes_append_soft(self): tt = _gen_task_template(env={'ENV': ('1', ['a'], True)}) p = task_request.TaskProperties(env_prefixes={u'ENV': [u'b']}) task_request._apply_task_template(tt, p) self.assertEqual( p, task_request.TaskProperties( env={u'ENV': u'1'}, env_prefixes={u'ENV': [u'a', u'b']}, )) def test_apply_template_conflicting_cache(self): tt = _gen_task_template(cache={'c': 'C'}) p = task_request.TaskProperties( caches=[task_request.CacheEntry(name='c', path='B')]) with self.assertRaises(ValueError) as ex: task_request._apply_task_template(tt, p) self.assertEqual(ex.exception.message, "request.cache['c'] conflicts with pool's template") def test_apply_template_conflicting_cache_path(self): tt = _gen_task_template(cache={'c': 'C'}) p = task_request.TaskProperties( caches=[task_request.CacheEntry(name='other', path='C')]) with self.assertRaises(ValueError) as ex: task_request._apply_task_template(tt, p) self.assertEqual( ex.exception.message, "u'C': directory has conflicting owners: task cache 'other' " "and task template cache u'c'") def test_apply_template_conflicting_cache_cipd_path(self): tt = _gen_task_template(cache={'c': 'C'}) p = task_request.TaskProperties( cipd_input=task_request.CipdInput(packages=[ task_request.CipdPackage( path='C', package_name='pkg', version='latest') ])) with self.assertRaises(ValueError) as ex: task_request._apply_task_template(tt, p) self.assertEqual( ex.exception.message, "u'C': directory has conflicting owners: task cipd['pkg:latest'] " "and task template cache u'c'") def test_apply_template_conflicting_cipd_package(self): tt = _gen_task_template(cipd_package={('C', 'pkg'): 'latest'}) p = task_request.TaskProperties( cipd_input=task_request.CipdInput(packages=[ task_request.CipdPackage( path='C', package_name='other', version='latest') ])) with self.assertRaises(ValueError) as ex: task_request._apply_task_template(tt, p) self.assertEqual( ex.exception.message, "u'C': directory has conflicting owners: task cipd['other:latest'] " "and task template cipd[u'pkg:latest']") def test_apply_template_conflicting_cipd_cache_path(self): tt = _gen_task_template(cipd_package={('C', 'pkg'): 'latest'}) p = task_request.TaskProperties( caches=[task_request.CacheEntry(name='other', path='C')]) with self.assertRaises(ValueError) as ex: task_request._apply_task_template(tt, p) self.assertEqual( ex.exception.message, "u'C': directory has conflicting owners: task cache 'other' " "and task template cipd[u'pkg:latest']") class TaskRequestApiTest(TestCase): def setUp(self): super(TaskRequestApiTest, self).setUp() # pool_configs is a mapping of pool name -> pools_config.PoolConfig. Tests # can modify this to have pools_config.get_pool_config return the # appropriate data. self._pool_configs = {} self.mock(pools_config, 'get_pool_config', self._pool_configs.get) self._enqueue_calls = [] self._enqueue_orig = self.mock(utils, 'enqueue_task', self._enqueue) def tearDown(self): try: self.assertFalse(self._enqueue_calls) finally: super(TaskRequestApiTest, self).tearDown() def _enqueue(self, *args, **kwargs): self._enqueue_calls.append((args, kwargs)) return self._enqueue_orig(*args, use_dedicated_module=False, **kwargs) def test_all_apis_are_tested(self): # Ensures there's a test for each public API. module = task_request expected = frozenset( i for i in dir(module) if i[0] != '_' and hasattr(getattr(module, i), 'func_name')) missing = expected - frozenset( i[5:] for i in dir(self) if i.startswith('test_')) self.assertFalse(missing) def test_get_automatic_tags(self): req = _gen_request() expected = set((u'hostname:localhost', u'OS:Windows-3.1.1', u'pool:default', u'priority:50', u'service_account:none', u'user:Jesus', u'use_isolate_1143123:1', u'use_cas_1143123:0')) self.assertEqual(expected, task_request.get_automatic_tags(req, 0)) with self.assertRaises(IndexError): task_request.get_automatic_tags(req, 1) def test_get_automatic_tags_slices(self): # Repeated TaskSlice. slices = [ task_request.TaskSlice( expiration_secs=60, properties=_gen_properties(dimensions={ u'gpu': [u'1234:5678'], u'pool': [u'GPU'] })), task_request.TaskSlice( expiration_secs=60, properties=_gen_properties(dimensions={ u'gpu': [u'none'], u'pool': [u'GPU'] })), ] req = _gen_request_slices(task_slices=slices) expected = set( (u'gpu:1234:5678', u'pool:GPU', u'priority:50', u'service_account:none', u'user:Jesus', u'use_isolate_1143123:1', u'use_cas_1143123:0')) self.assertEqual(expected, task_request.get_automatic_tags(req, 0)) expected = set( (u'gpu:none', u'pool:GPU', u'priority:50', u'service_account:none', u'user:Jesus', u'use_isolate_1143123:1', u'use_cas_1143123:0')) self.assertEqual(expected, task_request.get_automatic_tags(req, 1)) with self.assertRaises(IndexError): task_request.get_automatic_tags(req, 2) def test_get_automatic_tags_or_dim(self): slices = [ task_request.TaskSlice( expiration_secs=60, properties=_gen_properties(dimensions={ u'gpu': [u'nv|amd'], u'pool': [u'foo'] })), task_request.TaskSlice( expiration_secs=60, properties=_gen_properties(dimensions={ u'os': [u'linux|mac|win'], u'pool': [u'bar'] })), ] req = _gen_request_slices(task_slices=slices) expected = set((u'gpu:nv', u'gpu:amd', u'pool:foo', u'priority:50', u'service_account:none', u'user:Jesus', u'use_isolate_1143123:1', u'use_cas_1143123:0')) self.assertEqual(expected, task_request.get_automatic_tags(req, 0)) expected = set((u'os:linux', u'os:mac', u'os:win', u'pool:bar', u'priority:50', u'service_account:none', u'user:Jesus', u'use_isolate_1143123:1', u'use_cas_1143123:0')) self.assertEqual(expected, task_request.get_automatic_tags(req, 1)) def test_create_termination_task(self): request = task_request.create_termination_task( u'some-bot', wait_for_capacity=True) self.assertTrue(request.task_slice(0).properties.is_terminate) def test_new_request_key(self): for _ in range(3): delta = utils.utcnow() - task_request._BEGINING_OF_THE_WORLD now = int(round(delta.total_seconds() * 1000.)) key = task_request.new_request_key() # Remove the XOR. key_id = key.integer_id() ^ task_pack.TASK_REQUEST_KEY_ID_MASK timestamp = key_id >> 20 randomness = (key_id >> 4) & 0xFFFF version = key_id & 0xF self.assertLess(abs(timestamp - now), 1000) self.assertEqual(1, version) if randomness: break else: self.fail('Failed to find randomness') def test_new_request_key_zero(self): def getrandbits(i): self.assertEqual(i, 16) return 0x7766 self.mock(random, 'getrandbits', getrandbits) self.mock_now(task_request._BEGINING_OF_THE_WORLD) key = task_request.new_request_key() # Remove the XOR. key_id = key.integer_id() ^ task_pack.TASK_REQUEST_KEY_ID_MASK # 00000000000 7766 1 # ^ ^ ^ # | | | # since 2010 | schema version # | # rand self.assertEqual('0x0000000000077661', '0x%016x' % key_id) def test_new_request_key_end(self): def getrandbits(i): self.assertEqual(i, 16) return 0x7766 self.mock(random, 'getrandbits', getrandbits) days_until_end_of_the_world = 2**43 / 24. / 60. / 60. / 1000. num_days = int(days_until_end_of_the_world) # Remove 1ms to not overflow. num_seconds = ((days_until_end_of_the_world - num_days) * 24. * 60. * 60. - 0.001) self.assertEqual(101806, num_days) self.assertEqual(278, int(num_days / 365.3)) now = ( task_request._BEGINING_OF_THE_WORLD + datetime.timedelta(days=num_days, seconds=num_seconds)) self.mock_now(now) key = task_request.new_request_key() # Remove the XOR. key_id = key.integer_id() ^ task_pack.TASK_REQUEST_KEY_ID_MASK # 7ffffffffff 7766 1 # ^ ^ ^ # | | | # since 2010 | schema version # | # rand self.assertEqual('0x7ffffffffff77661', '0x%016x' % key_id) def test_validate_request_key(self): task_request.validate_request_key(task_pack.unpack_request_key('11')) with self.assertRaises(ValueError): task_request.validate_request_key(ndb.Key('TaskRequest', 1)) def test_init_new_request(self): parent = _gen_request() # Parent entity must have a valid key id and be stored. parent.key = task_request.new_request_key() parent.put() # The reference is to the TaskRunResult. parent_id = task_pack.pack_request_key(parent.key) + u'1' req = _gen_request( properties=_gen_properties( idempotent=True, relative_cwd=u'deeep', has_secret_bytes=True), parent_task_id=parent_id) # TaskRequest with secret must have a valid key. req.key = task_request.new_request_key() # Needed for the get() call below. req.put() sb = _gen_secret(req, 'I am a banana') # Needed for properties_hash() call. sb.put() expected_properties = { 'caches': [], 'cipd_input': { 'client_package': { 'package_name': u'infra/tools/cipd/${platform}', 'path': None, 'version': u'git_revision:deadbeef', }, 'packages': [{ 'package_name': u'rm', 'path': u'bin', 'version': u'git_revision:deadbeef', }], 'server': u'https://chrome-infra-packages.appspot.com' }, 'command': [u'command1', u'arg1'], 'containment': { u'lower_priority': False, u'containment_type': None, u'limit_processes': None, u'limit_total_committed_memory': None, }, 'relative_cwd': u'deeep', 'dimensions': { u'OS': [u'Windows-3.1.1'], u'hostname': [u'localhost'], u'pool': [u'default'], }, 'env': { u'foo': u'bar', u'joe': u'2' }, 'env_prefixes': { u'PATH': [u'local/path'] }, 'extra_args': [], 'execution_timeout_secs': 30, 'grace_period_secs': 30, 'has_secret_bytes': True, 'idempotent': True, 'inputs_ref': { 'isolated': None, 'isolatedserver': u'https://isolateserver.appspot.com', 'namespace': u'default-gzip', }, 'cas_input_root': None, 'io_timeout_secs': None, 'outputs': [], } expected_request = { 'authenticated': auth_testing.DEFAULT_MOCKED_IDENTITY, 'name': u'Request name', 'parent_task_id': unicode(parent_id), 'priority': 50, 'pubsub_topic': None, 'pubsub_userdata': None, 'service_account': u'none', 'tags': [ u'OS:Windows-3.1.1', u'authenticated:user:mocked@example.com', u'hostname:localhost', u'parent_task_id:%s' % parent_id, u'pool:default', u'priority:50', u'realm:none', u'service_account:none', u'swarming.pool.template:no_config', u'tag:1', u'use_cas_1143123:0', u'use_isolate_1143123:1', u'user:Jesus', ], 'task_slices': [{ 'expiration_secs': 30, 'properties': expected_properties, 'wait_for_capacity': False, },], 'user': u'Jesus', 'realm': None, 'realms_enabled': False, 'bot_ping_tolerance_secs': 120, 'resultdb': None, } actual = req.to_dict() actual.pop('created_ts') actual.pop('expiration_ts') self.assertEqual(expected_request, actual) self.assertEqual(30, req.expiration_secs) # Intentionally hard code the hash value since it has to be deterministic. # Other unit tests should use the calculated value. self.assertEqual( 'c262bae20e9b1a265fa5937d67aa36f690612b0e28c8af7e38b347dd6746da65', req.task_slice(0).properties_hash(req).encode('hex')) # TODO(crbug.com/1115778): remove after RBE-CAS migration. def test_init_new_request_isolated(self): parent = _gen_request( properties=_gen_properties( command=[u'command1', u'arg1'], inputs_ref={ 'isolated': '0123456789012345678901234567890123456789', 'isolatedserver': 'http://localhost:1', 'namespace': 'default-gzip', })) # Parent entity must have a valid key id and be stored. parent.key = task_request.new_request_key() parent.put() # The reference is to the TaskRunResult. parent_id = task_pack.pack_request_key(parent.key) + u'1' req = _gen_request( properties=_gen_properties(idempotent=True, has_secret_bytes=True), parent_task_id=parent_id) # TaskRequest with secret must have a valid key. req.key = task_request.new_request_key() # Needed for the get() call below. req.put() sb = _gen_secret(req, 'I am not a banana') # Needed for properties_hash() call. sb.put() expected_properties = { 'caches': [], 'cipd_input': { 'client_package': { 'package_name': u'infra/tools/cipd/${platform}', 'path': None, 'version': u'git_revision:deadbeef', }, 'packages': [{ 'package_name': u'rm', 'path': u'bin', 'version': u'git_revision:deadbeef', }], 'server': u'https://chrome-infra-packages.appspot.com' }, 'command': [u'command1', u'arg1'], 'containment': { u'lower_priority': False, u'containment_type': None, u'limit_processes': None, u'limit_total_committed_memory': None, }, 'relative_cwd': None, 'dimensions': { u'OS': [u'Windows-3.1.1'], u'hostname': [u'localhost'], u'pool': [u'default'], }, 'env': { u'foo': u'bar', u'joe': u'2' }, 'env_prefixes': { u'PATH': [u'local/path'] }, 'extra_args': [], 'execution_timeout_secs': 30, 'grace_period_secs': 30, 'idempotent': True, 'inputs_ref': { 'isolated': None, 'isolatedserver': u'https://isolateserver.appspot.com', 'namespace': u'default-gzip', }, 'cas_input_root': None, 'io_timeout_secs': None, 'outputs': [], 'has_secret_bytes': True, } expected_request = { 'authenticated': auth_testing.DEFAULT_MOCKED_IDENTITY, 'name': u'Request name', 'parent_task_id': unicode(parent_id), 'priority': 50, 'pubsub_topic': None, 'pubsub_userdata': None, 'service_account': u'none', 'tags': [ u'OS:Windows-3.1.1', u'authenticated:user:mocked@example.com', u'hostname:localhost', u'parent_task_id:%s' % parent_id, u'pool:default', u'priority:50', u'realm:none', u'service_account:none', u'swarming.pool.template:no_config', u'tag:1', u'use_cas_1143123:0', u'use_isolate_1143123:1', u'user:Jesus', ], 'task_slices': [{ 'expiration_secs': 30, 'properties': expected_properties, 'wait_for_capacity': False, },], 'user': u'Jesus', 'realm': None, 'realms_enabled': False, 'bot_ping_tolerance_secs': 120, 'resultdb': None, } actual = req.to_dict() # expiration_ts - created_ts == scheduling_expiration_secs. actual.pop('created_ts') actual.pop('expiration_ts') self.assertEqual(expected_request, actual) self.assertEqual(30, req.expiration_secs) # Intentionally hard code the hash value since it has to be deterministic. # Other unit tests should use the calculated value. self.assertEqual( 'f9254eae480e442121919c503c685319ab3a903c2d7b76eac79a947afd09d425', req.task_slice(0).properties_hash(req).encode('hex')) def test_init_new_request_cas_input(self): parent = _gen_request() # Parent entity must have a valid key id and be stored. parent.key = task_request.new_request_key() parent.put() # The reference is to the TaskRunResult. parent_id = task_pack.pack_request_key(parent.key) + u'1' cas_input_root = { 'cas_instance': u'projects/test/instances/default', 'digest': { 'hash': u'12345', 'size_bytes': 1, } } req = _gen_request( parent_task_id=parent_id, properties=_gen_properties( idempotent=True, has_secret_bytes=True, inputs_ref=None, cas_input_root=cas_input_root, )) # TaskRequest with secret must have a valid key. req.key = task_request.new_request_key() # Needed for the get() call below. req.put() sb = _gen_secret(req, 'I am not a banana') # Needed for properties_hash() call. sb.put() expected_properties = { 'caches': [], 'cipd_input': { 'client_package': { 'package_name': u'infra/tools/cipd/${platform}', 'path': None, 'version': u'git_revision:deadbeef', }, 'packages': [{ 'package_name': u'rm', 'path': u'bin', 'version': u'git_revision:deadbeef', }], 'server': u'https://chrome-infra-packages.appspot.com' }, 'command': [u'command1', u'arg1'], 'containment': { u'lower_priority': False, u'containment_type': None, u'limit_processes': None, u'limit_total_committed_memory': None, }, 'relative_cwd': None, 'dimensions': { u'OS': [u'Windows-3.1.1'], u'hostname': [u'localhost'], u'pool': [u'default'], }, 'env': { u'foo': u'bar', u'joe': u'2' }, 'env_prefixes': { u'PATH': [u'local/path'] }, 'extra_args': [], 'execution_timeout_secs': 30, 'grace_period_secs': 30, 'idempotent': True, 'inputs_ref': None, 'cas_input_root': cas_input_root, 'io_timeout_secs': None, 'outputs': [], 'has_secret_bytes': True, } expected_request = { 'authenticated': auth_testing.DEFAULT_MOCKED_IDENTITY, 'name': u'Request name', 'parent_task_id': parent_id, 'priority': 50, 'pubsub_topic': None, 'pubsub_userdata': None, 'service_account': u'none', 'tags': [ u'OS:Windows-3.1.1', u'authenticated:user:mocked@example.com', u'hostname:localhost', u'parent_task_id:%s' % parent_id, u'pool:default', u'priority:50', u'realm:none', u'service_account:none', u'swarming.pool.template:no_config', u'tag:1', u'use_cas_1143123:1', u'use_isolate_1143123:0', u'user:Jesus', ], 'task_slices': [{ 'expiration_secs': 30, 'properties': expected_properties, 'wait_for_capacity': False, },], 'user': u'Jesus', 'realm': None, 'realms_enabled': False, 'bot_ping_tolerance_secs': 120, 'resultdb': None, } actual = req.to_dict() # expiration_ts - created_ts == scheduling_expiration_secs. actual.pop('created_ts') actual.pop('expiration_ts') self.assertEqual(expected_request, actual) self.assertEqual(30, req.expiration_secs) # Intentionally hard code the hash value since it has to be deterministic. # Other unit tests should use the calculated value. self.assertEqual( '9e1b99c20a5c523ea1ade51276230781f9ddfd3ae396e66c810612a1c5c8062a', req.task_slice(0).properties_hash(req).encode('hex')) def test_init_new_request_parent(self): parent = _gen_request() # Parent entity must have a valid key id and be stored. parent.key = task_request.new_request_key() parent.put() # The reference is to the TaskRunResult. parent_id = task_pack.pack_request_key(parent.key) + '1' child = _gen_request(parent_task_id=parent_id) self.assertEqual(parent_id, child.parent_task_id) def test_init_new_request_invalid_parent_id(self): # Must ends with '1' or '2', not '0' with self.assertRaises(ValueError): _gen_request(parent_task_id='1d69b9f088008810') def test_init_new_request_missing_name(self): req = _gen_request(name=None) with self.assertRaisesRegexp(datastore_errors.BadValueError, '^name is missing$'): req.put() def test_init_new_request_idempotent(self): request = _gen_request(properties=_gen_properties(idempotent=True)) as_dict = request.to_dict() self.assertEqual(True, as_dict['task_slices'][0]['properties']['idempotent']) # Intentionally hard code the hash value since it has to be deterministic. # Other unit tests should use the calculated value. # Ensure the algorithm is deterministic. self.assertEqual( 'b1230281cc4bcc8d9458dab0810c86fcfaf8e4124351f4d39517833eb9541465', request.task_slice(0).properties_hash(request).encode('hex')) def test_init_new_request_bot_service_account(self): request = _gen_request(service_account='bot') request.put() as_dict = request.to_dict() self.assertEqual('bot', as_dict['service_account']) self.assertIn(u'service_account:bot', as_dict['tags']) def test_init_new_request_RBE_CAS(self): request = _gen_request( properties=_gen_properties( inputs_ref=task_request.FilesRef( isolated='dead' * (64 / 4), isolatedserver='astuce-service', namespace='sha256-GCP'))) request.put() as_dict = request.to_dict() expected = { 'isolated': u'deaddeaddeaddeaddeaddeaddeaddeaddeaddeaddeaddeaddeaddeaddeaddead', 'isolatedserver': u'astuce-service', 'namespace': u'sha256-GCP', } self.assertEqual(expected, as_dict['task_slices'][0]['properties']['inputs_ref']) def _set_pool_config_with_templates(self, prod=None, canary=None, canary_chance=None, pool_name=u'default'): """Builds a new pools_config.PoolConfig populated with the given pools_config.TaskTemplate objects and assigns it into the mocked `pools_confi.get_pool_config()` method. If prod is None, this omits the TaskTemplateDeployment entirely. canary_chance may be supplied as >9999 (normally illegal) in order to force the selection of canary.""" deployment = None if prod is not None: deployment = pools_config.TaskTemplateDeployment( prod=prod, canary=canary, canary_chance=canary_chance) self._pool_configs[pool_name] = pools_config.init_pool_config( name=pool_name, rev=u'testVersion1', task_template_deployment=deployment, ) def test_init_new_request_skip_template(self): self._set_pool_config_with_templates(_gen_task_template(env={'hi': 'prod'})) request = _gen_request(_template_apply=task_request.TEMPLATE_SKIP) as_dict = request.to_dict() self.assertIn(u'swarming.pool.version:testVersion1', as_dict['tags']) self.assertIn(u'swarming.pool.template:skip', as_dict['tags']) def test_init_new_request_missing_template(self): self._set_pool_config_with_templates() request = _gen_request() as_dict = request.to_dict() self.assertIn(u'swarming.pool.version:testVersion1', as_dict['tags']) self.assertIn(u'swarming.pool.template:none', as_dict['tags']) def test_init_new_request_prod_template(self): self._set_pool_config_with_templates( _gen_task_template(env={'hi': 'prod'}), canary=None, canary_chance=0, # always prefer prod serverside ) request = _gen_request() as_dict = request.to_dict() self.assertIn(u'swarming.pool.version:testVersion1', as_dict['tags']) self.assertIn(u'swarming.pool.template:prod', as_dict['tags']) self.assertEqual(as_dict['task_slices'][0]['properties']['env']['hi'], 'prod') def test_init_new_request_canary_template(self): self._set_pool_config_with_templates( _gen_task_template(env={'hi': 'prod'}), _gen_task_template(env={'hi': 'canary'}), canary_chance=10000, # always prefer canary serverside ) request = _gen_request() as_dict = request.to_dict() self.assertIn(u'swarming.pool.version:testVersion1', as_dict['tags']) self.assertIn(u'swarming.pool.template:canary', as_dict['tags']) self.assertEqual(as_dict['task_slices'][0]['properties']['env']['hi'], 'canary') def test_init_new_request_canary_never_template(self): self._set_pool_config_with_templates( _gen_task_template(env={'hi': 'prod'}), _gen_task_template(env={'hi': 'canary'}), canary_chance=10000, # always prefer canary serverside ) request = _gen_request(_template_apply=task_request.TEMPLATE_CANARY_NEVER) as_dict = request.to_dict() self.assertIn(u'swarming.pool.version:testVersion1', as_dict['tags']) self.assertIn(u'swarming.pool.template:prod', as_dict['tags']) self.assertEqual(as_dict['task_slices'][0]['properties']['env']['hi'], 'prod') def test_init_new_request_canary_prefer_template(self): self._set_pool_config_with_templates( _gen_task_template(env={'hi': 'prod'}), _gen_task_template(env={'hi': 'canary'}), canary_chance=0, # always prefer prod serverside ) request = _gen_request(_template_apply=task_request.TEMPLATE_CANARY_PREFER) as_dict = request.to_dict() self.assertIn(u'swarming.pool.version:testVersion1', as_dict['tags']) self.assertIn(u'swarming.pool.template:canary', as_dict['tags']) self.assertEqual(as_dict['task_slices'][0]['properties']['env']['hi'], 'canary') def test_init_new_request_canary_prefer_prod_template(self): self._set_pool_config_with_templates( _gen_task_template(env={'hi': 'prod'}), # No canary defined, even though caller would prefer it, if available. ) request = _gen_request(_template_apply=task_request.TEMPLATE_CANARY_PREFER) as_dict = request.to_dict() self.assertIn(u'swarming.pool.version:testVersion1', as_dict['tags']) self.assertIn(u'swarming.pool.template:prod', as_dict['tags']) self.assertEqual(as_dict['task_slices'][0]['properties']['env']['hi'], 'prod') def test_duped(self): # Two TestRequest with the same properties. request_1 = _gen_request(properties=_gen_properties(idempotent=True)) now = utils.utcnow() request_2 = _gen_request_slices( name='Other', user='Other', priority=201, created_ts=now, manual_tags=['tag:2'], task_slices=[ task_request.TaskSlice( expiration_secs=129, properties=_gen_properties(idempotent=True)), ]) self.assertEqual( request_1.task_slice(0).properties_hash(request_1), request_2.task_slice(0).properties_hash(request_2)) self.assertTrue(request_1.task_slice(0).properties_hash(request_1)) def test_different(self): # Two TestRequest with different properties. request_1 = _gen_request( properties=_gen_properties(execution_timeout_secs=30, idempotent=True)) request_2 = _gen_request( properties=_gen_properties(execution_timeout_secs=129, idempotent=True)) self.assertNotEqual( request_1.task_slice(0).properties_hash(request_1), request_2.task_slice(0).properties_hash(request_2)) # TODO(crbug.com/1115778): remove after RBE-CAS migration. def test_TaskRequest_to_proto_isolated(self): # Try to set as much things as possible to exercise most code paths. def getrandbits(i): self.assertEqual(i, 16) return 0x7766 self.mock(random, 'getrandbits', getrandbits) self.mock_now(task_request._BEGINING_OF_THE_WORLD) # Parent entity must have a valid key id and be stored. # This task uses user:Jesus, which will be inherited automatically. parent = _gen_request() parent.key = task_request.new_request_key() parent.put() # The reference is to the TaskRunResult. parent_id = task_pack.pack_request_key(parent.key) + u'0' parent_run_id = task_pack.pack_request_key(parent.key) + u'1' self.mock_now(task_request._BEGINING_OF_THE_WORLD, 2) request_props = _gen_properties( inputs_ref={ 'isolated': '0123456789012345678901234567890123456789', 'isolatedserver': 'http://localhost:1', 'namespace': 'default-gzip', }, relative_cwd=u'subdir', caches=[ task_request.CacheEntry(name=u'git_chromium', path=u'git_cache'), ], cipd_input=_gen_cipd_input( packages=[ task_request.CipdPackage( package_name=u'foo', path=u'tool', version=u'git:12345'), ],), idempotent=True, outputs=[u'foo'], has_secret_bytes=True, containment=task_request.Containment( lower_priority=True, containment_type=task_request.ContainmentType.JOB_OBJECT, limit_processes=1000, limit_total_committed_memory=1024**3, ), ) request = _gen_request_slices( task_slices=[ task_request.TaskSlice( expiration_secs=30, properties=request_props, wait_for_capacity=True, ), ], # The user is ignored; the value is overridden by the parent task's # user. user=u'Joe', parent_task_id=parent_run_id, service_account=u'foo@gserviceaccount.com', pubsub_topic=u'projects/a/topics/abc', pubsub_auth_token=u'sekret', pubsub_userdata=u'obscure_reference', ) # Necessary to have a valid task_id: request.key = task_request.new_request_key() # Necessary to attach a secret to the request: request.put() _gen_secret(request, 'I am a banana').put() expected_props = swarming_pb2.TaskProperties( cas_inputs=swarming_pb2.CASTree( digest=u'0123456789012345678901234567890123456789', server=u'http://localhost:1', namespace=u'default-gzip', ), cipd_inputs=[ swarming_pb2.CIPDPackage( package_name=u'foo', version=u'git:12345', dest_path=u'tool'), ], named_caches=[ swarming_pb2.NamedCacheEntry( name=u'git_chromium', dest_path=u'git_cache'), ], containment=swarming_pb2.Containment( lower_priority=True, containment_type=swarming_pb2.Containment.JOB_OBJECT, limit_processes=1000, limit_total_committed_memory=1024**3, ), command=[u'command1', u'arg1'], relative_cwd=u'subdir', # extra_args cannot be specified with command. # secret_bytes cannot be retrieved, but is included in properties_hash. has_secret_bytes=True, dimensions=[ swarming_pb2.StringListPair(key=u'OS', values=[u'Windows-3.1.1']), swarming_pb2.StringListPair(key=u'hostname', values=[u'localhost']), swarming_pb2.StringListPair(key=u'pool', values=[u'default']), ], env=[ swarming_pb2.StringPair(key=u'foo', value=u'bar'), swarming_pb2.StringPair(key=u'joe', value=u'2'), ], env_paths=[ swarming_pb2.StringListPair(key=u'PATH', values=[u'local/path']), ], execution_timeout=duration_pb2.Duration(seconds=30), grace_period=duration_pb2.Duration(seconds=30), idempotent=True, outputs=[u'foo'], ) # To be updated every time the schema changes. props_h = 'e8718f59959d2c17d9ab1084b6fc9b3ee63e998a704de579543dd84bc1ef603a' expected = swarming_pb2.TaskRequest( # Scheduling. task_slices=[ swarming_pb2.TaskSlice( properties=expected_props, expiration=duration_pb2.Duration(seconds=30), wait_for_capacity=True, properties_hash=props_h, ), ], priority=50, service_account=u'foo@gserviceaccount.com', # Information. create_time=timestamp_pb2.Timestamp(seconds=1262304002), name=u'Request name', authenticated='user:mocked@example.com', tags=[ u'OS:Windows-3.1.1', u"authenticated:user:mocked@example.com", u'hostname:localhost', u'parent_task_id:%s' % parent_run_id, u'pool:default', u'priority:50', u'realm:none', u'service_account:foo@gserviceaccount.com', u'swarming.pool.template:no_config', u'tag:1', u"use_cas_1143123:0", u"use_isolate_1143123:1", u'user:Jesus', ], user=u'Jesus', # Hierarchy. task_id=u'7d0776610', parent_task_id=parent_id, parent_run_id=parent_run_id, # Notification. auth_token cannot be retrieved. pubsub_notification=swarming_pb2.PubSub( topic=u'projects/a/topics/abc', userdata=u'obscure_reference'), ) actual = swarming_pb2.TaskRequest() request.to_proto(actual) self.assertEqual(unicode(expected), unicode(actual)) # with append_root_ids=True. actual = swarming_pb2.TaskRequest() request.to_proto(actual, append_root_ids=True) expected.root_task_id = parent_id expected.root_run_id = parent_run_id self.assertEqual(unicode(expected), unicode(actual)) def test_TaskRequest_to_proto(self): # Try to set as much things as possible to exercise most code paths. def getrandbits(i): self.assertEqual(i, 16) return 0x7766 self.mock(random, 'getrandbits', getrandbits) self.mock_now(task_request._BEGINING_OF_THE_WORLD) # Grand parent entity must have a valid key id and be stored. # This task uses user:Jesus, which will be inherited automatically. grand_parent = _gen_request() grand_parent.key = task_request.new_request_key() grand_parent.put() # Parent entity must have a valid key id and be stored. self.mock_now(task_request._BEGINING_OF_THE_WORLD, 1) parent = _gen_request(parent_task_id=grand_parent.task_id[:-1] + u'1') parent.key = task_request.new_request_key() parent.put() parent_run_id = parent.task_id[:-1] + u'1' self.mock_now(task_request._BEGINING_OF_THE_WORLD, 2) request_props = _gen_properties( inputs_ref=None, cas_input_root={ 'cas_instance': u'projects/test/instances/default', 'digest': { 'hash': u'12345', 'size_bytes': 1, } }, relative_cwd=u'subdir', caches=[ task_request.CacheEntry(name=u'git_chromium', path=u'git_cache'), ], cipd_input=_gen_cipd_input( packages=[ task_request.CipdPackage( package_name=u'foo', path=u'tool', version=u'git:12345'), ],), idempotent=True, outputs=[u'foo'], has_secret_bytes=True, containment=task_request.Containment( lower_priority=True, containment_type=task_request.ContainmentType.JOB_OBJECT, limit_processes=1000, limit_total_committed_memory=1024**3, ), ) request = _gen_request_slices( task_slices=[ task_request.TaskSlice( expiration_secs=30, properties=request_props, wait_for_capacity=True, ), ], # The user is ignored; the value is overridden by the parent task's # user. user=u'Joe', parent_task_id=parent.task_id[:-1] + u'1', service_account=u'foo@gserviceaccount.com', pubsub_topic=u'projects/a/topics/abc', pubsub_auth_token=u'sekret', pubsub_userdata=u'obscure_reference', ) # Necessary to have a valid task_id: request.key = task_request.new_request_key() # Necessary to attach a secret to the request: request.put() _gen_secret(request, 'I am a banana').put() expected_props = swarming_pb2.TaskProperties( cas_input_root=swarming_pb2.CASReference( cas_instance='projects/test/instances/default', digest=swarming_pb2.Digest( hash='12345', size_bytes=1), ), cipd_inputs=[ swarming_pb2.CIPDPackage( package_name=u'foo', version=u'git:12345', dest_path=u'tool'), ], named_caches=[ swarming_pb2.NamedCacheEntry( name=u'git_chromium', dest_path=u'git_cache'), ], containment=swarming_pb2.Containment( lower_priority=True, containment_type=swarming_pb2.Containment.JOB_OBJECT, limit_processes=1000, limit_total_committed_memory=1024**3, ), command=[u'command1', u'arg1'], relative_cwd=u'subdir', # extra_args cannot be specified with command. # secret_bytes cannot be retrieved, but is included in properties_hash. has_secret_bytes=True, dimensions=[ swarming_pb2.StringListPair(key=u'OS', values=[u'Windows-3.1.1']), swarming_pb2.StringListPair(key=u'hostname', values=[u'localhost']), swarming_pb2.StringListPair(key=u'pool', values=[u'default']), ], env=[ swarming_pb2.StringPair(key=u'foo', value=u'bar'), swarming_pb2.StringPair(key=u'joe', value=u'2'), ], env_paths=[ swarming_pb2.StringListPair(key=u'PATH', values=[u'local/path']), ], execution_timeout=duration_pb2.Duration(seconds=30), grace_period=duration_pb2.Duration(seconds=30), idempotent=True, outputs=[u'foo'], ) # To be updated every time the schema changes. props_h = '516b5f86592b0e5e3bdd9fbf715305ee6f7ddad36320775d5a945e60df67c360' expected = swarming_pb2.TaskRequest( # Scheduling. task_slices=[ swarming_pb2.TaskSlice( properties=expected_props, expiration=duration_pb2.Duration(seconds=30), wait_for_capacity=True, properties_hash=props_h, ), ], priority=50, service_account=u'foo@gserviceaccount.com', # Information. create_time=timestamp_pb2.Timestamp(seconds=1262304002), name=u'Request name', authenticated='user:mocked@example.com', tags=[ u'OS:Windows-3.1.1', u"authenticated:user:mocked@example.com", u'hostname:localhost', u'parent_task_id:%s' % parent_run_id, u'pool:default', u'priority:50', u'realm:none', u'service_account:foo@gserviceaccount.com', u'swarming.pool.template:no_config', u'tag:1', u"use_cas_1143123:1", u"use_isolate_1143123:0", u'user:Jesus', ], user=u'Jesus', # Hierarchy. task_id=u'7d0776610', parent_task_id=parent.task_id, parent_run_id=parent_run_id, # Notification. auth_token cannot be retrieved. pubsub_notification=swarming_pb2.PubSub( topic=u'projects/a/topics/abc', userdata=u'obscure_reference'), ) actual = swarming_pb2.TaskRequest() request.to_proto(actual) self.assertEqual(unicode(expected), unicode(actual)) actual = swarming_pb2.TaskRequest() expected.root_task_id = grand_parent.task_id expected.root_run_id = grand_parent.task_id[:-1] + u'1' request.to_proto(actual, append_root_ids=True) self.assertEqual(unicode(expected), unicode(actual)) # With append_root_ids=True. actual = swarming_pb2.TaskRequest() request.to_proto(actual, append_root_ids=True) expected.root_task_id = grand_parent.task_id expected.root_run_id = grand_parent.task_id[:-1] + u'1' self.assertEqual(unicode(expected), unicode(actual)) def test_TaskRequest_to_proto_empty(self): # Assert that it doesn't throw on empty entity. actual = swarming_pb2.TaskRequest() task_request.TaskRequest().to_proto(actual) self.assertEqual(swarming_pb2.TaskRequest(), actual) def test_TaskSlice_to_proto_empty(self): # Assert that it doesn't throw on empty entity. request = task_request.TaskRequest() actual = swarming_pb2.TaskSlice() task_request.TaskSlice().to_proto(actual, request) self.assertEqual(swarming_pb2.TaskSlice(), actual) def test_TaskProperties_to_proto_empty(self): # Assert that it doesn't throw on empty entity. actual = swarming_pb2.TaskProperties() task_request.TaskProperties().to_proto(actual) expected = swarming_pb2.TaskProperties() expected.grace_period.seconds = 30 self.assertEqual(expected, actual) def test_request_bad_values(self): with self.assertRaises(AttributeError): _gen_request(properties=_gen_properties(foo='bar')) def test_request_bad_values_stale_style(self): # Old TaskRequest.properties style. # Hack a bit the call to force the incorrect call. req = _gen_request_slices( task_slices=[], expiration_ts=utils.utcnow() + datetime.timedelta(hours=1), properties_old=_gen_properties()) with self.assertRaises(datastore_errors.BadValueError): req.put() def test_request_bad_values_task_slices(self): with self.assertRaises(ValueError): # No TaskSlice _gen_request_slices(task_slices=[]) def _gen_slice(**props): return task_request.TaskSlice( expiration_secs=60, properties=_gen_properties(**props)) slices = [_gen_slice(dimensions={u'pool': [u'GPU']})] _gen_request_slices(task_slices=slices).put() # Limit on the maximum number of TaskSlice in a TaskRequest. slices = [ _gen_slice(dimensions={ u'pool': [u'GPU'], u'v': [unicode(i)] }) for i in range(8) ] _gen_request_slices(task_slices=slices).put() slices = [ _gen_slice(dimensions={ u'pool': [u'GPU'], u'v': [unicode(i)] }) for i in range(9) ] req = _gen_request_slices(task_slices=slices) with self.assertRaises(datastore_errors.BadValueError): req.put() # Different pools. slices = [ task_request.TaskSlice( expiration_secs=60, properties=_gen_properties(dimensions={u'pool': [u'GPU']})), task_request.TaskSlice( expiration_secs=60, properties=_gen_properties(dimensions={u'pool': [u'other']})), ] req = _gen_request_slices(task_slices=slices) with self.assertRaises(datastore_errors.BadValueError): req.put() def test_request_bad_command(self): req = _gen_request(properties=_gen_properties(command=[], inputs_ref=None)) with self.assertRaises(datastore_errors.BadValueError): req.put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(properties=_gen_properties(command={'a': 'b'})) with self.assertRaises(datastore_errors.BadValueError): _gen_request(properties=_gen_properties(command='python')) _gen_request(properties=_gen_properties(command=['python'])).put() _gen_request(properties=_gen_properties(command=[u'python'])).put() _gen_request(properties=_gen_properties(command=[u'python'] * 128)).put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(properties=_gen_properties(command=[u'python'] * 129)).put() def test_request_extra_args(self): with self.assertRaises(datastore_errors.BadValueError): _gen_request( properties=_gen_properties( command=[], extra_args=[u'python'], inputs_ref=task_request.FilesRef( isolated='deadbeefdeadbeefdeadbeefdeadbeefdeadbeef', isolatedserver='http://localhost:1', namespace='default-gzip'))).put() with self.assertRaises(datastore_errors.BadValueError): _gen_request( properties=_gen_properties( command=[u'python'], extra_args=[u'python'], inputs_ref=task_request.FilesRef( isolated='deadbeefdeadbeefdeadbeefdeadbeefdeadbeef', isolatedserver='http://localhost:1', namespace='default-gzip'))).put() def test_request_bad_cipd_input(self): def mkcipdreq(idempotent=False, **cipd_input): return _gen_request( properties=_gen_properties( idempotent=idempotent, cipd_input=_gen_cipd_input(**cipd_input))) req = mkcipdreq(packages=[{}]) with self.assertRaises(datastore_errors.BadValueError): req.put() with self.assertRaises(datastore_errors.BadValueError): mkcipdreq(packages=[ task_request.CipdPackage( package_name='infra|rm', path='.', version='latest'), ]) req = mkcipdreq( packages=[task_request.CipdPackage(package_name='rm', path='.')]) with self.assertRaises(datastore_errors.BadValueError): req.put() req = mkcipdreq(packages=[ task_request.CipdPackage(package_name='rm', version='latest'), ]) with self.assertRaises(datastore_errors.BadValueError): req.put() with self.assertRaises(datastore_errors.BadValueError): mkcipdreq(packages=[ task_request.CipdPackage( package_name='rm', path='/', version='latest'), ]) with self.assertRaises(datastore_errors.BadValueError): mkcipdreq(packages=[ task_request.CipdPackage( package_name='rm', path='/a', version='latest'), ]) with self.assertRaises(datastore_errors.BadValueError): mkcipdreq(packages=[ task_request.CipdPackage( package_name='rm', path='a/..', version='latest'), ]) with self.assertRaises(datastore_errors.BadValueError): mkcipdreq(packages=[ task_request.CipdPackage( package_name='rm', path='a/./b', version='latest'), ]) req = mkcipdreq(packages=[ task_request.CipdPackage(package_name='rm', path='.', version='latest'), task_request.CipdPackage(package_name='rm', path='.', version='canary'), ]) with self.assertRaises(datastore_errors.BadValueError): req.put() req = mkcipdreq( idempotent=True, packages=[ task_request.CipdPackage( package_name='rm', path='.', version='latest'), ]) with self.assertRaises(datastore_errors.BadValueError): req.put() with self.assertRaises(datastore_errors.BadValueError): mkcipdreq(server='abc') with self.assertRaises(datastore_errors.BadValueError): mkcipdreq( client_package=task_request.CipdPackage( package_name='--bad package--')) mkcipdreq().put() mkcipdreq(packages=[ task_request.CipdPackage(package_name='rm', path='.', version='latest'), ]).put() mkcipdreq( client_package=task_request.CipdPackage( package_name='infra/tools/cipd/${platform}', version='git_revision:daedbeef'), packages=[ task_request.CipdPackage( package_name='rm', path='.', version='latest'), ], server='https://chrome-infra-packages.appspot.com').put() def test_request_bad_named_cache(self): mkcachereq = lambda *c: _gen_request(properties=_gen_properties(caches=c) ).put() with self.assertRaises(datastore_errors.BadValueError): mkcachereq(task_request.CacheEntry(name='', path='git_cache')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq(task_request.CacheEntry(name='git_chromium', path='')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq( task_request.CacheEntry(name='git_chromium', path='git_cache'), task_request.CacheEntry(name='git_v8', path='git_cache')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq( task_request.CacheEntry(name='git_chromium', path='git_cache'), task_request.CacheEntry(name='git_chromium', path='git_cache2')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq( task_request.CacheEntry(name='git_chromium', path='/git_cache')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq( task_request.CacheEntry(name='git_chromium', path='../git_cache')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq( task_request.CacheEntry( name='git_chromium', path='git_cache/../../a')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq( task_request.CacheEntry(name='git_chromium', path='../git_cache')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq( task_request.CacheEntry(name='git_chromium', path='git_cache//a')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq( task_request.CacheEntry(name='git_chromium', path='a/./git_cache')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq(task_request.CacheEntry(name='has space', path='git_cache')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq(task_request.CacheEntry(name='CAPITAL', path='git_cache')) mkcachereq() mkcachereq(task_request.CacheEntry(name='git_chromium', path='git_cache')) mkcachereq( task_request.CacheEntry(name='git_chromium', path='git_cache'), task_request.CacheEntry(name='build_chromium', path='out')) mkcachereq(task_request.CacheEntry(name=u'g' * 128, path=u'git_cache')) with self.assertRaises(datastore_errors.BadValueError): mkcachereq(task_request.CacheEntry(name=u'g' * 129, path=u'git_cache')) mkcachereq(task_request.CacheEntry(name=u'g', path=u'p' * 256)) with self.assertRaises(datastore_errors.BadValueError): mkcachereq(task_request.CacheEntry(name=u'g', path=u'p' * 257)) # Too many. c = [ task_request.CacheEntry(name=unicode(i), path=unicode(i)) for i in range(32) ] _gen_request(properties=_gen_properties(caches=c)).put() with self.assertRaises(datastore_errors.BadValueError): c = [ task_request.CacheEntry(name=unicode(i), path=unicode(i)) for i in range(33) ] _gen_request(properties=_gen_properties(caches=c)).put() def test_request_bad_named_cache_and_cipd_input(self): # A CIPD package and named caches cannot be mapped to the same path. req = _gen_request( properties=_gen_properties( caches=[ task_request.CacheEntry(name='git_chromium', path='git_cache'), ], cipd_input=_gen_cipd_input(packages=[ task_request.CipdPackage( package_name='foo', path='git_cache', version='latest'), ]))) with self.assertRaises(datastore_errors.BadValueError): req.put() req = _gen_request( properties=_gen_properties( caches=[ task_request.CacheEntry(name='git_chromium', path='git_cache1'), ], cipd_input=_gen_cipd_input(packages=[ task_request.CipdPackage( package_name='foo', path='git_cache2', version='latest'), ]))).put() def test_request_bad_dimensions(self): # Type error. with self.assertRaises(TypeError): _gen_request(properties=_gen_properties(dimensions=[])) with self.assertRaises(datastore_errors.BadValueError): _gen_request(properties=_gen_properties(dimensions={})) with self.assertRaises(datastore_errors.BadValueError): _gen_request( properties=_gen_properties(dimensions={ u'id': u'b', u'a:': u'b' })) with self.assertRaises(datastore_errors.BadValueError): _gen_request( properties=_gen_properties(dimensions={ u'id': u'b', u'a.': u'b' })) with self.assertRaises(datastore_errors.BadValueError): _gen_request( properties=_gen_properties(dimensions={ u'id': u'b', u'a': [u'b'] })) # >1 value for id. with self.assertRaises(datastore_errors.BadValueError): _gen_request(properties=_gen_properties(dimensions={u'id': [u'a', u'b']})) # >1 value for pool. with self.assertRaises(datastore_errors.BadValueError): _gen_request( properties=_gen_properties(dimensions={u'pool': [u'b', u'b']})) _gen_request( properties=_gen_properties(dimensions={ u'id': [u'b'], u'pool': [u'b'] })).put() _gen_request( properties=_gen_properties(dimensions={ u'id': [u'b'], u'pool': [u'b'], u'a.': [u'c'] })).put() _gen_request( properties=_gen_properties(dimensions={ u'pool': [u'b'], u'a.': [u'b', u'c'] })).put() def test_request_bad_dimensions_key(self): # Max # keys. d = {u'a%s' % string.ascii_letters[i]: [unicode(i)] for i in range(31)} d[u'pool'] = [u'a'] _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaises(datastore_errors.BadValueError): d = {u'a%s' % string.ascii_letters[i]: [unicode(i)] for i in range(32)} d[u'pool'] = [u'a'] _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaises(datastore_errors.BadValueError): # Key regexp. d = {u'pool': [u'default'], u'1': [u'value']} _gen_request(properties=_gen_properties(dimensions=d)).put() # Key length. d = { u'pool': [u'default'], u'v' * config.DIMENSION_KEY_LENGTH: [u'v'], } _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaises(datastore_errors.BadValueError): d = { u'pool': [u'default'], u'v' * (config.DIMENSION_KEY_LENGTH + 1): [u'value'], } _gen_request(properties=_gen_properties(dimensions=d)).put() def test_request_bad_dimensions_value(self): # Max # values. d = {u'pool': [u'b'], u'a.': [unicode(i) for i in range(16)]} _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaises(datastore_errors.BadValueError): d = {u'pool': [u'b'], u'a.': [unicode(i) for i in range(17)]} _gen_request(properties=_gen_properties(dimensions=d)).put() # Value length. d = { u'pool': [u'default'], u'v': [u'v' * config.DIMENSION_VALUE_LENGTH], } _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaises(datastore_errors.BadValueError): d = { u'pool': [u'default'], u'v': [u'v' * (config.DIMENSION_VALUE_LENGTH + 1)], } _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaises(datastore_errors.BadValueError): # Value with space. d = {u'pool': [u'default'], u'v': [u'v ']} _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaises(datastore_errors.BadValueError): # Duplicate value. d = {u'pool': [u'default'], u'v': [u'v', u'v']} _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaisesRegexp( datastore_errors.BadValueError, "^dimension key u'v' has invalid value u'v||c'$"): # Empty 'or' dimension value. d = {u'pool': [u'default'], u'v': [u'v||c']} _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaisesRegexp( datastore_errors.BadValueError, "^'pool' cannot be specified more than once in dimensions " "\[u'default|non-default'\]$"): # Use 'or' dimension in pool. d = {u'pool': [u'default|non-default'], u'v': [u'v']} _gen_request(properties=_gen_properties(dimensions=d)).put() with self.assertRaisesRegexp( datastore_errors.BadValueError, 'possible dimension subset for \'or\' dimensions ' 'should not be more than 8, but 9'): # Too many combinations for 'or' d = { u'pool': [u'default'], u'x': [u'1|2|3'], u'y': [u'1|2|3'], } _gen_request(properties=_gen_properties(dimensions=d)).put() d = { u'pool': [u'default'], u'x': [u'1|2'], u'y': [u'1|2'], u'z': [u'1|2'], } _gen_request(properties=_gen_properties(dimensions=d)).put() def test_request_bad_env(self): # Type error. with self.assertRaises(TypeError): _gen_request(properties=_gen_properties(env=[])) with self.assertRaises(TypeError): _gen_request(properties=_gen_properties(env={u'a': 1})) _gen_request(properties=_gen_properties(env={})).put() e = {u'k': u'v'} _gen_request(properties=_gen_properties(env=e)).put() # Key length. e = {u'k' * 64: u'v'} _gen_request(properties=_gen_properties(env=e)).put() with self.assertRaises(datastore_errors.BadValueError): e = {u'k' * 65: u'v'} _gen_request(properties=_gen_properties(env=e)).put() # # keys. e = {u'k%s' % i: u'v' for i in range(64)} _gen_request(properties=_gen_properties(env=e)).put() with self.assertRaises(datastore_errors.BadValueError): e = {u'k%s' % i: u'v' for i in range(65)} _gen_request(properties=_gen_properties(env=e)).put() # Value length. e = {u'k': u'v' * 1024} _gen_request(properties=_gen_properties(env=e)).put() with self.assertRaises(datastore_errors.BadValueError): e = {u'k': u'v' * 1025} _gen_request(properties=_gen_properties(env=e)).put() def test_request_bad_env_prefixes(self): # Type error. with self.assertRaises(TypeError): _gen_request(properties=_gen_properties(env_prefixes=[])) with self.assertRaises(TypeError): _gen_request(properties=_gen_properties(env_prefixes={u'a': 1})) _gen_request(properties=_gen_properties(env_prefixes={})).put() e = {u'k': [u'v']} _gen_request(properties=_gen_properties(env_prefixes=e)).put() # Key length. e = {u'k' * 64: [u'v']} _gen_request(properties=_gen_properties(env_prefixes=e)).put() with self.assertRaises(datastore_errors.BadValueError): e = {u'k' * 65: [u'v']} _gen_request(properties=_gen_properties(env_prefixes=e)).put() # # keys. e = {u'k%s' % i: [u'v'] for i in range(64)} _gen_request(properties=_gen_properties(env_prefixes=e)).put() with self.assertRaises(datastore_errors.BadValueError): e = {u'k%s' % i: [u'v'] for i in range(65)} _gen_request(properties=_gen_properties(env_prefixes=e)).put() # Value length. e = {u'k': [u'v' * 1024]} _gen_request(properties=_gen_properties(env_prefixes=e)).put() with self.assertRaises(datastore_errors.BadValueError): e = {u'k': [u'v' * 1025]} _gen_request(properties=_gen_properties(env_prefixes=e)).put() def test_request_bad_priority(self): with self.assertRaises(datastore_errors.BadValueError): _gen_request(priority=task_request.MAXIMUM_PRIORITY + 1) _gen_request(priority=task_request.MAXIMUM_PRIORITY).put() def test_request_bad_bot_ping_tolerance(self): with self.assertRaises(datastore_errors.BadValueError): _gen_request( bot_ping_tolerance_secs=task_request._MAX_BOT_PING_TOLERANCE_SECS + 1) _gen_request( bot_ping_tolerance_secs=task_request._MIN_BOT_PING_TOLERANCE_SECS - 1) def test_request_bad_execution_timeout(self): # When used locally, it is set to 1, which means it's impossible to test # below _MIN_TIMEOUT_SECS but above 0. self.mock(task_request, '_MIN_TIMEOUT_SECS', 30) p = _gen_request(properties=_gen_properties(execution_timeout_secs=0)) with self.assertRaises(datastore_errors.BadValueError): # Only termination task may have 0. p.put() with self.assertRaises(datastore_errors.BadValueError): _gen_request( properties=_gen_properties( execution_timeout_secs=task_request._MIN_TIMEOUT_SECS - 1)) _gen_request( properties=_gen_properties( execution_timeout_secs=task_request._MIN_TIMEOUT_SECS)) with self.assertRaises(datastore_errors.BadValueError): _gen_request( properties=_gen_properties( execution_timeout_secs=task_request.MAX_TIMEOUT_SECS + 1)) _gen_request( properties=_gen_properties( execution_timeout_secs=task_request.MAX_TIMEOUT_SECS)).put() def test_request_bad_expiration(self): now = utils.utcnow() with self.assertRaises(ValueError): _gen_request_slices( created_ts=now, task_slices=[ task_request.TaskSlice( expiration_secs=None, properties=_gen_properties()), ]) with self.assertRaises(datastore_errors.BadValueError): _gen_request_slices( created_ts=now, task_slices=[ task_request.TaskSlice( expiration_secs=task_request._MIN_TIMEOUT_SECS - 1, properties=_gen_properties()), ]) with self.assertRaises(datastore_errors.BadValueError): _gen_request_slices( created_ts=now, task_slices=[ task_request.TaskSlice( expiration_secs=task_request.MAX_EXPIRATION_SECS + 1, properties=_gen_properties()), ]) _gen_request_slices( created_ts=now, task_slices=[ task_request.TaskSlice( expiration_secs=task_request._MIN_TIMEOUT_SECS, properties=_gen_properties()), ]).put() _gen_request_slices( created_ts=now, task_slices=[ task_request.TaskSlice( expiration_secs=task_request.MAX_EXPIRATION_SECS, properties=_gen_properties()), ]).put() def test_request_bad_inputs_ref(self): # Both command and inputs_ref.isolated. _gen_request( properties=_gen_properties( command=['python'], inputs_ref=task_request.FilesRef( isolated='deadbeefdeadbeefdeadbeefdeadbeefdeadbeef', isolatedserver='http://localhost:1', namespace='default-gzip'))).put() # Bad digest. req = _gen_request( properties=_gen_properties( command=['see', 'spot', 'run'], inputs_ref=task_request.FilesRef( isolated='deadbeef', isolatedserver='http://localhost:1', namespace='default-gzip'))) with self.assertRaises(datastore_errors.BadValueError): req.put() # inputs_ref without server/namespace. req = _gen_request( properties=_gen_properties(inputs_ref=task_request.FilesRef())) with self.assertRaises(datastore_errors.BadValueError): req.put() # Without digest nor command. req = _gen_request( properties=_gen_properties( command=[], inputs_ref=task_request.FilesRef( isolatedserver='https://isolateserver.appspot.com', namespace='default-gzip^^^'))) with self.assertRaises(datastore_errors.BadValueError): req.put() # For 'sha256-GCP', the length must be 64. req = _gen_request( properties=_gen_properties( command=[], inputs_ref=task_request.FilesRef( isolated='deadbeefdeadbeefdeadbeefdeadbeefdeadbeef', isolatedserver='foo-bar', namespace='sha256-GCP'))) with self.assertRaises(datastore_errors.BadValueError): req.put() # For 'sha256-GCP', the isolatedserver value must not contain '://'. req = _gen_request( properties=_gen_properties( command=[], inputs_ref=task_request.FilesRef( isolated='dead' * (64 / 4), isolatedserver='foo://bar', namespace='sha256-GCP'))) with self.assertRaises(datastore_errors.BadValueError): req.put() def test_request_bad_cas_input_root(self): def _gen_request_with_cas_input_root(cas_instance, digest): return _gen_request( properties=_gen_properties( inputs_ref=None, # inputs_ref can't be set with cas_input_root. cas_input_root=task_request.CASReference( cas_instance=cas_instance, digest=digest))) valid_cas_instance = 'projects/test/instances/default' valid_digest = task_request.Digest(hash='12345', size_bytes=1) # TaskRequest with a valid cas_input_root. _gen_request_with_cas_input_root( cas_instance=valid_cas_instance, digest=valid_digest).put() # Missing cas_instance. with self.assertRaises(datastore_errors.BadValueError): _gen_request_with_cas_input_root( cas_instance=None, digest=valid_digest).put() # Invalid cas_instance. with self.assertRaises(datastore_errors.BadValueError): _gen_request_with_cas_input_root( cas_instance='invalid_instance_name', digest=valid_digest).put() # Missing digest. with self.assertRaises(datastore_errors.BadValueError): _gen_request_with_cas_input_root( cas_instance=valid_cas_instance, digest=None).put() # Missing digest.hash. with self.assertRaises(datastore_errors.BadValueError): _gen_request_with_cas_input_root( cas_instance=valid_cas_instance, digest=task_request.Digest(hash=None, size_bytes=1)).put() # Missing digest.size_bytes. with self.assertRaises(datastore_errors.BadValueError): _gen_request_with_cas_input_root( cas_instance=valid_cas_instance, digest=task_request.Digest(hash='12345', size_bytes=None)).put() def test_request_conflict_inputs(self): req = _gen_request( properties=_gen_properties( inputs_ref=task_request.FilesRef( isolated='0123456789012345678901234567890123456789', isolatedserver=u'https://isolateserver.appspot.com', namespace=u'default-gzip'), cas_input_root=task_request.CASReference( cas_instance='projects/test/instances/default', digest=task_request.Digest(hash='12345', size_bytes=1)), )) with self.assertRaises(datastore_errors.BadValueError) as e: req.put() self.assertEqual(e.exception.message, "can't set both inputs_ref and cas_input_root") def test_request_bad_pubsub(self): _gen_request(pubsub_topic=u'projects/a/topics/abc').put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(pubsub_topic=u'a') with self.assertRaises(datastore_errors.BadValueError): _gen_request(pubsub_topic=u'projects/a/topics/ab').put() _gen_request(pubsub_topic=u'projects/' + u'a' * 1004 + u'/topics/abc').put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(pubsub_topic=u'projects/' + u'a' * 1005 + u'/topics/abc').put() def test_request_bad_service_account(self): _gen_request(service_account=u'none').put() _gen_request(service_account=u'bot').put() _gen_request(service_account=u'joe@localhost').put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(service_account=u'joe').put() _gen_request(service_account=u'joe@' + u'l' * 124).put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(service_account=u'joe@' + u'l' * 125).put() def test_request_bad_tags(self): with self.assertRaises(datastore_errors.BadValueError): _gen_request(manual_tags=['a']).put() def test_request_bad_tags_too_many(self): _gen_request(manual_tags=['a:b'] * 256).put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(manual_tags=['a:b'] * 257).put() def test_request_bad_tags_too_long(self): # Minus 2 for the 'a:' prefix. l = task_request._TAG_LENGTH - 2 _gen_request(manual_tags=['a:' + 'b' * l]).put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(manual_tags=['a:' + 'a' * (l + 1)]).put() def test_request_bad_realm(self): _gen_request(realm=None).put() _gen_request(realm='test:realm').put() with self.assertRaises(datastore_errors.BadValueError): _gen_request(realm='invalid_realm').put() def test_resultdb_enable(self): request = _gen_request(resultdb=task_request.ResultDBCfg(enable=True)) actual = swarming_pb2.TaskRequest() request.to_proto(actual) self.assertTrue(actual.resultdb.enable) def test_execution_deadline(self): self.mock_now(datetime.datetime(2020, 1, 2, 3, 4, 5)) request = _gen_request() self.assertEqual(request.execution_deadline, datetime.datetime(2020, 1, 2, 3, 5, 35)) def test_validate_priority(self): with self.assertRaises(TypeError): task_request.validate_priority(None) with self.assertRaises(TypeError): task_request.validate_priority('1') with self.assertRaises(datastore_errors.BadValueError): task_request.validate_priority(-1) with self.assertRaises(datastore_errors.BadValueError): task_request.validate_priority(task_request.MAXIMUM_PRIORITY + 1) task_request.validate_priority(0) task_request.validate_priority(1) task_request.validate_priority(task_request.MAXIMUM_PRIORITY) def test_datetime_to_request_base_id(self): now = datetime.datetime(2012, 1, 2, 3, 4, 5, 123456) self.assertEqual(0xeb5313d0300000, task_request.datetime_to_request_base_id(now)) def test_convert_to_request_key(self): """Indirectly tested by API.""" now = datetime.datetime(2012, 1, 2, 3, 4, 5, 123456) key = task_request.convert_to_request_key(now) self.assertEqual(9157134072765480958, key.id()) def test_request_key_to_datetime(self): key = ndb.Key(task_request.TaskRequest, 0x7f14acec2fcfffff) # Resolution is only kept at millisecond level compared to # datetime_to_request_base_id() by design. self.assertEqual( datetime.datetime(2012, 1, 2, 3, 4, 5, 123000), task_request.request_key_to_datetime(key)) def test_request_id_to_key(self): # Simple XOR. self.assertEqual( ndb.Key(task_request.TaskRequest, 0x7f14acec2fcfffff), task_request.request_id_to_key(0xeb5313d0300000)) def test_secret_bytes(self): task_request.SecretBytes(secret_bytes='a' * (20 * 1024)).put() with self.assertRaises(datastore_errors.BadValueError): task_request.SecretBytes(secret_bytes='a' * (20 * 1024 + 1)).put() def test_cron_delete_old_task_requests(self): # Creating 1000 tasks would make this test significantly slower. self.mock(task_request, '_TASKS_DELETE_CHUNK_SIZE', 5) now = utils.utcnow() task_ids = [] for i in range(14): self.mock_now(now, i) request = _gen_request_slices() request.key = task_request.new_request_key() request.put() task_ids.append(task_pack.pack_request_key(request.key)) # Use 11 seconds offset, so that entities 12, 13 are not deleted. Yet create # 3 GAE tasks to delete the chunks limited at 5 items. self.mock_now(now + task_request._OLD_TASK_REQUEST_CUT_OFF, 11) self.assertEqual(12, task_request.cron_delete_old_task_requests()) expected = [ ( ('/internal/taskqueue/cleanup/tasks/delete', 'delete-tasks'), { 'payload': utils.encode_to_json({u'task_ids': task_ids[0:5]}) }, ), ( ('/internal/taskqueue/cleanup/tasks/delete', 'delete-tasks'), { 'payload': utils.encode_to_json({u'task_ids': task_ids[5:10]}) }, ), ( ('/internal/taskqueue/cleanup/tasks/delete', 'delete-tasks'), { 'payload': utils.encode_to_json({u'task_ids': task_ids[10:12]}) }, ), ] # task_ids[12:14] are not touched. self.assertEqual(expected, self._enqueue_calls) self._enqueue_calls = [] def test_task_delete_tasks(self): # The data here should be the same as what is passed to the task queue in # test_cron_delete_old_task_requests. class Foo(ndb.Model): pass task_ids = [] for _ in range(5): request = _gen_request_slices() request.key = task_request.new_request_key() request.put() # Create a dummy child entity to ensure it's deleted too. Foo(parent=request.key, id=1).put() task_ids.append(task_pack.pack_request_key(request.key)) self.assertEqual(5, task_request.task_delete_tasks(task_ids)) self.assertEqual(0, task_request.TaskRequest.query().count()) self.assertEqual(0, Foo.query().count()) def test_task_bq_empty(self): # Empty, nothing is done. start = utils.utcnow() end = start + datetime.timedelta(seconds=60) self.assertEqual(0, task_request.task_bq(start, end)) def test_task_bq(self): def getrandbits(i): self.assertEqual(i, 16) return 0x7766 self.mock(random, 'getrandbits', getrandbits) payloads = [] def send_to_bq(table_name, rows): self.assertEqual('task_requests', table_name) payloads.append(rows) self.mock(bq_state, 'send_to_bq', send_to_bq) # Generate two tasks requests. now = datetime.datetime(2014, 1, 2, 3, 4, 5, 6) start = self.mock_now(now, 10) request_1 = _gen_request() request_1.key = task_request.new_request_key() run_1_id = request_1.task_id[:-1] + '1' request_1.put() self.mock_now(now, 20) request_2 = _gen_request(parent_task_id=run_1_id) request_2.key = task_request.new_request_key() request_2.put() end = self.mock_now(now, 30) self.assertEqual(2, task_request.task_bq(start, end)) self.assertEqual(1, len(payloads), payloads) actual_rows = payloads[0] self.assertEqual(2, len(actual_rows)) expected_ids = [ # No root IDs on task 1. (request_1.task_id, '', ''), # Task 1 is the root of Task 2. (request_2.task_id, request_1.task_id, run_1_id), ] self.assertEqual( expected_ids, [(t.task_id, t.root_task_id, t.root_run_id) for _, t in actual_rows]) def test_yield_request_keys_by_parent_task_id(self): parent_request = _gen_request() parent_request.key = task_request.new_request_key() parent_request.put() parent_summary_key = task_pack.request_key_to_result_summary_key( parent_request.key) parent_summary_id = task_pack.pack_result_summary_key(parent_summary_key) parent_run_key = task_pack.result_summary_key_to_run_result_key( parent_summary_key, 1) parent_run_id = task_pack.pack_run_result_key(parent_run_key) child_request_1_key = _gen_request(parent_task_id=parent_run_id).put() child_request_2_key = _gen_request(parent_task_id=parent_run_id).put() it = task_request.yield_request_keys_by_parent_task_id(parent_summary_id) expected = [child_request_1_key, child_request_2_key] self.assertEqual(sorted(expected), sorted([k for k in it])) def test_normalize_or_dimensions(self): dim1 = _gen_request( properties=_gen_properties(dimensions={u'foo': [u'a|c|b', u'xyz']})).task_slice(0).properties.dimensions dim2 = _gen_request( properties=_gen_properties(dimensions={u'foo': [u'xyz', u'c|b|a']})).task_slice(0).properties.dimensions expected = {u'foo': [u'a|b|c', u'xyz']} self.assertEqual(dim1, expected) self.assertEqual(dim1, dim2) if __name__ == '__main__': if '-v' in sys.argv: unittest.TestCase.maxDiff = None logging.basicConfig( level=logging.DEBUG if '-v' in sys.argv else logging.ERROR) unittest.main()
apache-2.0
Clemson-DPA/dpa-pipe
dpa/app/session.py
1
9786
from abc import ABCMeta, abstractmethod, abstractproperty import importlib import os import shlex import subprocess import socket import time import rpyc from dpa.app.entity import EntityRegistry from dpa.env.vars import DpaVars from dpa.ptask.area import PTaskArea from dpa.ptask import PTaskError, PTask from dpa.singleton import Singleton # ----------------------------------------------------------------------------- class SessionRegistry(Singleton): # ------------------------------------------------------------------------- def init(self): self._registry = {} # ------------------------------------------------------------------------- def current(self): for registered_cls in self._registry.values(): if registered_cls.current(): return registered_cls() return None # ------------------------------------------------------------------------- def register(self, cls): self._registry[cls.app_name] = cls # ----------------------------------------------------------------------------- class Session(object): __metaclass__ = ABCMeta app_name = None # ------------------------------------------------------------------------- @classmethod def current(cls): return None # ------------------------------------------------------------------------- def __init__(self): pass # ------------------------------------------------------------------------- @abstractmethod def close(self): """Close the current file.""" # ------------------------------------------------------------------------- def list_entities(self, categories=None): """List entities in the session.""" entities = [] entity_classes = EntityRegistry().get_entity_classes( self.__class__.app_name) for entity_class in entity_classes: entities.extend(entity_class.list(self)) if categories: filtered = [e for e in entities if e.category in categories] else: filtered = entities return filtered # ------------------------------------------------------------------------- @classmethod def open_file(self, filepath): """Open a new session with the supplied file.""" # ------------------------------------------------------------------------- @abstractmethod def save(self, filepath=None): """Save the current session. Save to the file path if provided.""" # ------------------------------------------------------------------------- @abstractproperty def in_session(self): """Returns True if inside a current app session.""" # ------------------------------------------------------------------------- def init_module(self, module_path): _module = None if self.in_session: try: _module = importlib.import_module(module_path) except ImportError: pass # will raise below if not _module: raise SessionError( "Failed to initialize session. " + \ "'{mod}' module could not be imported.".format(mod=module_path) ) return _module # ------------------------------------------------------------------------- def require_executable(self, executable): """Returns the full path for the supplied executable name.""" (path, file_name) = os.path.split(executable) # path already included if path: if not os.path.isfile(executable): raise SessionError("Unable to locate executable: " + executable) elif not os.access(executable, os.X_OK): raise SessionError("File is not executable: " + executable) else: return executable else: bin_paths = DpaVars.path() bin_paths.get() for path in bin_paths.list: executable_path = os.path.join(path, executable) if (os.path.isfile(executable_path) and os.access(executable_path, os.X_OK)): return executable_path raise SessionError("Unable to locate executable: " + executable) # ------------------------------------------------------------------------- @property def app_name(self): return self.__class__.app_name # ------------------------------------------------------------------------- @property def ptask_area(self): """Return the current ptask area for this session.""" if not hasattr(self, '_ptask_area'): self._ptask_area = PTaskArea.current() return self._ptask_area # ------------------------------------------------------------------------- @property def ptask(self): if not hasattr(self, '_ptask'): ptask_area = self.ptask_area if not ptask_area.spec: self._ptask = None else: try: self._ptask = PTask.get(ptask_area.spec) except PTaskError as e: raise SessionError("Unable to determine ptask.") return self._ptask # ------------------------------------------------------------------------- @property def ptask_version(self): """Return the current ptask version for this session.""" if not hasattr(self, '_ptask_version'): ptask = self.ptask if not ptask: self._ptask_version = None else: self._ptask_version = ptask.latest_version return self._ptask_version # ----------------------------------------------------------------------------- class RemoteMixin(object): __metaclass__ = ABCMeta # ------------------------------------------------------------------------- def __init__(self, remote=False): self._remote = remote # ------------------------------------------------------------------------- def __del__(self): self.shutdown() # ------------------------------------------------------------------------- def __enter__(self): return self # ------------------------------------------------------------------------- def __exit__(self, exc_type, exc_value, traceback): self.shutdown() # ------------------------------------------------------------------------- @staticmethod def _get_port(): sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM) sock.bind(("",0)) port = sock.getsockname()[1] sock.close() return port # ------------------------------------------------------------------------- @property def remote(self): """Returns True if in a session, False otherwise.""" return self._remote # ------------------------------------------------------------------------- @property def remote_connection(self): if not hasattr(self, '_remote_connection'): self._remote_connection = self._connect_remote() return self._remote_connection # ------------------------------------------------------------------------- @abstractproperty def server_executable(self): """The executable for starting the remote app server.""" # ------------------------------------------------------------------------- def shutdown(self): if hasattr(self, '_remote_connection'): try: self._remote_connection.root.shutdown() except EOFError: # this is the expected error on shutdown pass else: self._remote_connection = None # ------------------------------------------------------------------------- def init_module(self, module_path): _module = None if self.remote: # need to give time for standalone app to import properly tries = 0 while not _module or tries < 30: try: self.remote_connection.execute("import " + module_path) _module = getattr( self.remote_connection.modules, module_path) break except ImportError: tries += 1 time.sleep(1) if not _module: self.shutdown() elif self.in_session: try: _module = importlib.import_module(module_path) except ImportError: pass # will raise below if not _module: raise SessionError( "Failed to initialize session. " + \ "'{mod}' module could not be imported.".format(mod=module_path) ) return _module # ------------------------------------------------------------------------- def _connect_remote(self): port = self._get_port() cmd = "{cmd} {port}".format(cmd=self.server_executable, port=port) args = shlex.split(cmd) subprocess.Popen(args) connection = None tries = 0 while not connection or tries < 30: try: connection = rpyc.classic.connect("localhost", port) break except socket.error: tries += 1 time.sleep(1) if not connection: raise SessionError("Unable connect to remote session.") return connection # ----------------------------------------------------------------------------- class SessionError(Exception): pass
mit
h2non/paco
paco/wait.py
1
3118
# -*- coding: utf-8 -*- import asyncio from .assertions import isiter from .concurrent import ConcurrentExecutor @asyncio.coroutine def wait(*coros_or_futures, limit=0, timeout=None, loop=None, return_exceptions=False, return_when='ALL_COMPLETED'): """ Wait for the Futures and coroutine objects given by the sequence futures to complete, with optional concurrency limit. Coroutines will be wrapped in Tasks. ``timeout`` can be used to control the maximum number of seconds to wait before returning. timeout can be an int or float. If timeout is not specified or None, there is no limit to the wait time. If ``return_exceptions`` is True, exceptions in the tasks are treated the same as successful results, and gathered in the result list; otherwise, the first raised exception will be immediately propagated to the returned future. ``return_when`` indicates when this function should return. It must be one of the following constants of the concurrent.futures module. All futures must share the same event loop. This functions is mostly compatible with Python standard ``asyncio.wait()``. Arguments: *coros_or_futures (iter|list): an iterable collection yielding coroutines functions. limit (int): optional concurrency execution limit. Use ``0`` for no limit. timeout (int/float): maximum number of seconds to wait before returning. return_exceptions (bool): exceptions in the tasks are treated the same as successful results, instead of raising them. return_when (str): indicates when this function should return. loop (asyncio.BaseEventLoop): optional event loop to use. *args (mixed): optional variadic argument to pass to the coroutines function. Returns: tuple: Returns two sets of Future: (done, pending). Raises: TypeError: in case of invalid coroutine object. ValueError: in case of empty set of coroutines or futures. TimeoutError: if execution takes more than expected. Usage:: async def sum(x, y): return x + y done, pending = await paco.wait( sum(1, 2), sum(3, 4)) [task.result() for task in done] # => [3, 7] """ # Support iterable as first argument for better interoperability if len(coros_or_futures) == 1 and isiter(coros_or_futures[0]): coros_or_futures = coros_or_futures[0] # If no coroutines to schedule, return empty list # Mimics asyncio behaviour. if len(coros_or_futures) == 0: raise ValueError('paco: set of coroutines/futures is empty') # Create concurrent executor pool = ConcurrentExecutor(limit=limit, loop=loop, coros=coros_or_futures) # Wait until all the tasks finishes return (yield from pool.run(timeout=timeout, return_when=return_when, return_exceptions=return_exceptions))
mit
BehavioralInsightsTeam/edx-platform
lms/djangoapps/grades/apps.py
16
1502
""" Grades Application Configuration Signal handlers are connected here. """ from django.apps import AppConfig from django.conf import settings from edx_proctoring.runtime import set_runtime_service from openedx.core.djangoapps.plugins.constants import ProjectType, SettingsType, PluginURLs, PluginSettings class GradesConfig(AppConfig): """ Application Configuration for Grades. """ name = u'lms.djangoapps.grades' plugin_app = { PluginURLs.CONFIG: { ProjectType.LMS: { PluginURLs.NAMESPACE: u'grades_api', PluginURLs.REGEX: u'api/grades/', PluginURLs.RELATIVE_PATH: u'api.urls', } }, PluginSettings.CONFIG: { ProjectType.LMS: { SettingsType.AWS: {PluginSettings.RELATIVE_PATH: u'settings.aws'}, SettingsType.COMMON: {PluginSettings.RELATIVE_PATH: u'settings.common'}, SettingsType.TEST: {PluginSettings.RELATIVE_PATH: u'settings.test'}, } } } def ready(self): """ Connect handlers to recalculate grades. """ # Can't import models at module level in AppConfigs, and models get # included from the signal handlers from .signals import handlers # pylint: disable=unused-variable if settings.FEATURES.get('ENABLE_SPECIAL_EXAMS'): from .services import GradesService set_runtime_service('grades', GradesService())
agpl-3.0
akopich/spark
python/pyspark/java_gateway.py
56
5851
# # Licensed to the Apache Software Foundation (ASF) under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import atexit import os import sys import select import signal import shlex import socket import platform from subprocess import Popen, PIPE if sys.version >= '3': xrange = range from py4j.java_gateway import java_import, JavaGateway, GatewayClient from pyspark.find_spark_home import _find_spark_home from pyspark.serializers import read_int def launch_gateway(conf=None): """ launch jvm gateway :param conf: spark configuration passed to spark-submit :return: """ if "PYSPARK_GATEWAY_PORT" in os.environ: gateway_port = int(os.environ["PYSPARK_GATEWAY_PORT"]) else: SPARK_HOME = _find_spark_home() # Launch the Py4j gateway using Spark's run command so that we pick up the # proper classpath and settings from spark-env.sh on_windows = platform.system() == "Windows" script = "./bin/spark-submit.cmd" if on_windows else "./bin/spark-submit" command = [os.path.join(SPARK_HOME, script)] if conf: for k, v in conf.getAll(): command += ['--conf', '%s=%s' % (k, v)] submit_args = os.environ.get("PYSPARK_SUBMIT_ARGS", "pyspark-shell") if os.environ.get("SPARK_TESTING"): submit_args = ' '.join([ "--conf spark.ui.enabled=false", submit_args ]) command = command + shlex.split(submit_args) # Start a socket that will be used by PythonGatewayServer to communicate its port to us callback_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) callback_socket.bind(('127.0.0.1', 0)) callback_socket.listen(1) callback_host, callback_port = callback_socket.getsockname() env = dict(os.environ) env['_PYSPARK_DRIVER_CALLBACK_HOST'] = callback_host env['_PYSPARK_DRIVER_CALLBACK_PORT'] = str(callback_port) # Launch the Java gateway. # We open a pipe to stdin so that the Java gateway can die when the pipe is broken if not on_windows: # Don't send ctrl-c / SIGINT to the Java gateway: def preexec_func(): signal.signal(signal.SIGINT, signal.SIG_IGN) proc = Popen(command, stdin=PIPE, preexec_fn=preexec_func, env=env) else: # preexec_fn not supported on Windows proc = Popen(command, stdin=PIPE, env=env) gateway_port = None # We use select() here in order to avoid blocking indefinitely if the subprocess dies # before connecting while gateway_port is None and proc.poll() is None: timeout = 1 # (seconds) readable, _, _ = select.select([callback_socket], [], [], timeout) if callback_socket in readable: gateway_connection = callback_socket.accept()[0] # Determine which ephemeral port the server started on: gateway_port = read_int(gateway_connection.makefile(mode="rb")) gateway_connection.close() callback_socket.close() if gateway_port is None: raise Exception("Java gateway process exited before sending the driver its port number") # In Windows, ensure the Java child processes do not linger after Python has exited. # In UNIX-based systems, the child process can kill itself on broken pipe (i.e. when # the parent process' stdin sends an EOF). In Windows, however, this is not possible # because java.lang.Process reads directly from the parent process' stdin, contending # with any opportunity to read an EOF from the parent. Note that this is only best # effort and will not take effect if the python process is violently terminated. if on_windows: # In Windows, the child process here is "spark-submit.cmd", not the JVM itself # (because the UNIX "exec" command is not available). This means we cannot simply # call proc.kill(), which kills only the "spark-submit.cmd" process but not the # JVMs. Instead, we use "taskkill" with the tree-kill option "/t" to terminate all # child processes in the tree (http://technet.microsoft.com/en-us/library/bb491009.aspx) def killChild(): Popen(["cmd", "/c", "taskkill", "/f", "/t", "/pid", str(proc.pid)]) atexit.register(killChild) # Connect to the gateway gateway = JavaGateway(GatewayClient(port=gateway_port), auto_convert=True) # Import the classes used by PySpark java_import(gateway.jvm, "org.apache.spark.SparkConf") java_import(gateway.jvm, "org.apache.spark.api.java.*") java_import(gateway.jvm, "org.apache.spark.api.python.*") java_import(gateway.jvm, "org.apache.spark.ml.python.*") java_import(gateway.jvm, "org.apache.spark.mllib.api.python.*") # TODO(davies): move into sql java_import(gateway.jvm, "org.apache.spark.sql.*") java_import(gateway.jvm, "org.apache.spark.sql.hive.*") java_import(gateway.jvm, "scala.Tuple2") return gateway
apache-2.0
MatriX-Coder/wsploit
modules/wjoomlacomponents.py
1
8322
#!/usr/bin/python # Wsploit Project ''' this is simple joomla components scanner ''' try: import urllib2, Queue except: print 'You need urllib2 and Queue librarys installed.' try: from threading import Thread except: print 'You need threading library installed.' try: from time import sleep except: print 'You need time library installed.' paths = [ '/components/com_tag', '/components/com_virtuemart', '/components/com_jvehicles', '/components/com_s5clanroster', '/components/com_fireboard', '/components/com_fabrik', '/components/com_jinc', '/components/com_xcloner-backupandrestore', '/components/com_dshop', '/components/com_ponygallery', '/components/com_bearleague', '/components/com_obsuggest', '/components/com_alameda', '/components/com_estateagent', '/components/com_collector', '/components/com_qcontacts', '/components/com_niceajaxpoll', '/components/com_xmap', '/components/com_team', '/components/com_joomnik', '/components/com_question', '/components/com_jmsfileseller', '/components/com_rsfiles', '/components/com_versioning', '/components/com_hello', '/components/com_calcbuilder', '/components/com_jmsfileseller', '/components/com_xmovie', '/components/com_people', '/components/com_idoblog', '/components/com_adsmanager', '/components/com_xgallery', '/components/com_alfurqan15x', '/components/com_alfurqan', '/components/com_billyportfolio', '/components/com_jimtawl', '/components/com_content', '/components/com_jfuploader', '/components/com_kunena', '/components/com_jooproperty', '/components/com_jsupport', '/components/com_markt', '/components/com_img', '/components/com_clanlist', '/components/com_clan', '/components/com_ckforms', '/components/com_dcnews', '/components/com_connect', '/components/com_rsappt_pro2', '/components/com_techfolio', '/components/com_zcalendar', '/components/com_tpjobs', '/components/com_simpleshop', '/components/com_sef', '/components/com_searchlog', '/components/com_contact', '/components/com_enmasse', '/components/com_elite_experts', '/components/com_ezautos', '/components/com_jgen', '/components/com_jphone', '/components/com_mosets', '/components/com_jefaqpro', '/components/com_picsell', '/components/com_ongallery', '/components/com_equipment', '/components/com_zoomportfolio', '/components/com_amblog', '/components/com_joltcard', '/components/com_jp_jobs', '/components/com_bfquiztrial', '/components/com_qpersonel', '/components/com_pandafminigames', '/components/com_golfcourseguid', '/components/com_jejob', '/components/com_jeajaxeventcalendar', '/components/com_jradio', '/components/com_spidercatalog', '/components/com_/components/commedia', '/components/com_fss', '/components/com_icagenda', '/components/com_spidercalendar', '/components/com_joomgalaxy', '/components/com_ornekek', '/components/com_weblinks', '/components/com_rokmodule', '/components/com_discussions', '/components/com_hm/components/community', '/components/com_eslamiat', '/components/com_listing', '/components/com_jeemasms', '/components/com_yjcontactus', '/components/com_timereturns', '/components/com_jce', '/components/com_joomtouch', '/components/com_jdirectory', '/components/com_jesubmit', '/components/com_sobi2', '/components/com_acooldebate', '/components/com_booklibrary', '/components/com_acymailing', '/components/com_doqment', '/components/com_allcinevid', '/components/com_jotloader', '/components/com_jeauto', '/components/com_ccboard', '/components/com_ccinvoices', '/components/com_flipwall', '/components/com_sponsorwall', '/components/com_cbe', '/components/com_jscalendar', '/components/com_restaurantguide', '/components/com_nkc', '/components/com_aardvertiser', '/components/com_clantools', '/components/com_remository', '/components/com_dateconverter', '/components/com_wmtpic', '/components/com_donateprocess', '/components/com_gamesbox', '/components/com_jcafe', '/components/com_awd_song', '/components/com_picasa2gallery', '/components/com_ybggal', '/components/com_joomdocs', '/components/com_answers', '/components/com_galleryxml', '/components/com_oziogallery2', '/components/com_listbingo', '/components/com_easygb', '/components/com_jtickets', '/components/com_jesectionfinder', '/components/com_realtyna', '/components/com_/components/community', '/components/com_jomestate', '/components/com_jtickets', '/components/com_cinema', '/components/com_jstore', '/components/com_annonces', '/components/com_lead', '/components/com_sar_news', '/components/com_chronocontact', '/components/com_chronoconnectivity', '/components/com_djartgallery', '/components/com_quran', '/components/com_g2bridge', '/components/com_reservations', '/components/com_jepoll', '/components/com_mycar', '/components/com_mediqna', '/components/com_zelig', '/components/com_bookmarks', '/components/com_hotproperty', '/components/com_jombib', '/components/com_store', '/components/com_mosforms', '/components/com_/components/comprofiler', '/components/com_crowdsource', '/components/com_camp', '/components/com_ms/components/comment', '/components/com_extcalendar', '/components/com_imoti', '/components/com_product', '/components/com_event', '/components/com_simpledownload', '/components/com_news', '/components/com_article', '/components/com_jequoteform', '/components/com_konsultasi', '/components/com_sebercart', '/components/com_php', '/components/com_jinc', '/components/com_mytube', '/components/com_jbudgetsmagic', '/components/com_surveymanager', '/components/com_jreservation', '/components/com_foobla_suggestions', '/components/com_djcatalog', '/components/com_turtushout', '/components/com_alphauserpoints', '/components/com_lucygames', '/components/com_bfsurvey_profree', '/components/com_tpdugg', '/components/com_joomloc', '/components/com_joomlub', '/components/com_artportal', '/components/com_agora', '/components/com_gameserver', '/components/com_digifolio', '/components/com_bca-rss-syndicator', '/components/com_expose', '/components/com_equotes', '/components/com_media', '/components/com_misterestate', '/components/com_wrapper', '/components/com_mailto', '/components/com_autartimonial', '/components/com_artforms', '/components/com_redshop', '/components/com_staticxt', '/components/com_spa', '/components/com_jomtube', '/components/com_golfcourseguide', '/components/com_huruhelpdesk', '/components/com_joomdle', '/components/com_youtube', '/components/com_joomla-visites', '/components/com_ttvideo', '/components/com_appointinator', '/components/com_photomapgallery', '/components/com_spielothek', '/components/com_pbbooking', '/components/com_beamospetition', '/components/com_neorecruit', '/components/com_cgtestimonial', '/components/com_jgrid', '/components/com_zina', '/components/com_pro_desk', '/components/com_user', '/components/com_k2', '/components/com_rsbook_15', '/components/com_gk3_photoslide', '/components/com_jvideodirect', '/components/com_jcalpro', '/components/com_banners', '/components/com_datsogallery', '/components/com_joomradio', '/components/com_jfbconnect', '/components/com_myblog', '/components/com_phocamaps', '/components/com_contact_enhanced', '/components/com_aicontactsafe', '/components/com_poll'] def one(): def test(target,path): if 'http://' not in target: target = 'http://'+target bb = target+path try: a = urllib2.urlopen(bb) c = a.getcode() if c == 200: print 'Found ---> '+path[12:] except urllib2.URLError: pass thrdlst = [] target = raw_input('\nEnter site : ') for path in paths: t = Thread(target=test , args=(target,path)) t.start() thrdlst.append(t) sleep(0.009) for b in thrdlst: b.join() def lista(): path = raw_input('\nEnter List Path : ') sites = open(path,'r') sites = sites.readlines() print '\n' for site in sites: if 'http://' not in site: site = 'http://'+site site = site.strip() print '\n[*] Target : %s\n' % site for path in paths: bb = site+path try: a = urllib2.urlopen(bb) c = a.getcode() if c == 200: print 'Found ---> '+path[12:] except urllib2.URLError: pass def init(): print '\n[1]-Single URL' print '[2]-List Of URLs\n' line_1 = "Enter Option : " choose = raw_input(line_1) if choose.isdigit(): choose = int(choose) pass else : print "Choose From List Bro" exit() if choose == 1: one() if choose == 2: lista()
artistic-2.0
anentropic/django-oscar
tests/integration/offer/availability_tests.py
40
1143
from django.test import TestCase from oscar.apps.offer import models from oscar.test.factories import ( create_order, OrderDiscountFactory, UserFactory) class TestAPerUserConditionalOffer(TestCase): def setUp(self): self.offer = models.ConditionalOffer(max_user_applications=1) self.user = UserFactory() def test_is_available_with_no_applications(self): self.assertTrue(self.offer.is_available()) def test_max_applications_is_correct_when_no_applications(self): self.assertEqual(1, self.offer.get_max_applications(self.user)) def test_max_applications_is_correct_when_equal_applications(self): order = create_order(user=self.user) OrderDiscountFactory( order=order, offer_id=self.offer.id, frequency=1) self.assertEqual(0, self.offer.get_max_applications(self.user)) def test_max_applications_is_correct_when_more_applications(self): order = create_order(user=self.user) OrderDiscountFactory( order=order, offer_id=self.offer.id, frequency=5) self.assertEqual(0, self.offer.get_max_applications(self.user))
bsd-3-clause
uskudnik/ggrc-core
src/ggrc/migrations/versions/20130910232450_53ef72c8a867_remove_old_audit_tab.py
2
13296
"""Remove old audit tables Revision ID: 53ef72c8a867 Revises: 526117e15ce4 Create Date: 2013-09-10 23:24:50.751098 """ # revision identifiers, used by Alembic. revision = '53ef72c8a867' down_revision = '526117e15ce4' from alembic import op import sqlalchemy as sa from sqlalchemy.dialects import mysql NOT_NULL_COLS = [ ('control_assessments', 'pbc_list_id'), ('control_assessments', 'control_id'), ('system_controls', 'system_id'), ('system_controls', 'control_id'), ('responses', 'request_id'), ('responses', 'system_id'), ] EXPLICIT_INDEXES = [ ('control_assessments', 'control_id', 'controls', 'control_assessments_ibfk_1'), ('control_assessments', 'pbc_list_id', 'pbc_lists', 'control_assessments_ibfk_2'), ('system_controls', 'system_id', 'systems', 'system_controls_ibfk_3'), ('system_controls', 'control_id', 'controls', 'system_controls_ibfk_1'), ('responses', 'request_id', 'requests', 'responses_ibfk_1'), ('responses', 'system_id', 'systems', 'responses_ibfk_2'), ] UNIQUE_CONSTRAINTS = [('control_assessments', ['pbc_list_id', 'control_id']), ('system_controls', ['system_id', 'control_id']), ('responses', ['request_id', 'system_id']), ] def create_explicit_index(table, column, referred_table, constraint_name): " Explicit indexes need to be created to work around http://bugs.mysql.com/bug.php?id=21395 " op.drop_constraint(constraint_name, table, type_='foreignkey') op.create_index('ix_' + column, table, [column]) op.create_foreign_key(constraint_name, table, referred_table, [column], ['id']) def drop_explicit_index(table, column, referred_table, constraint_name): op.drop_constraint(constraint_name, table, type_='foreignkey') op.drop_index('ix_' + column, table) op.create_foreign_key(constraint_name, table, referred_table, [column], ['id']) def upgrade(): op.drop_table(u'system_controls') op.drop_table(u'meetings') op.drop_table(u'population_samples') op.drop_table(u'responses') op.drop_table(u'requests') op.drop_table(u'control_assessments') op.drop_table(u'pbc_lists') op.drop_table(u'cycles') op.drop_table(u'transactions') def downgrade(): op.create_table(u'transactions', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'description', sa.TEXT(), nullable=True), sa.Column(u'title', sa.VARCHAR(length=250), nullable=True), sa.Column(u'system_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_transactions_contexts'), sa.ForeignKeyConstraint(['system_id'], [u'systems.id'], name=u'transactions_ibfk_1'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) op.create_table(u'cycles', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'description', sa.TEXT(), nullable=True), sa.Column(u'start_at', sa.DATE(), nullable=True), sa.Column(u'complete', sa.BOOLEAN, nullable=False), sa.Column(u'title', sa.VARCHAR(length=250), nullable=True), sa.Column(u'audit_firm', sa.VARCHAR(length=250), nullable=True), sa.Column(u'audit_lead', sa.VARCHAR(length=250), nullable=True), sa.Column(u'status', sa.VARCHAR(length=250), nullable=True), sa.Column(u'notes', sa.TEXT(), nullable=True), sa.Column(u'end_at', sa.DATE(), nullable=True), sa.Column(u'program_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'report_due_at', sa.DATE(), nullable=True), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_cycles_contexts'), sa.ForeignKeyConstraint(['program_id'], [u'programs.id'], name=u'cycles_ibfk_1'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) op.create_table(u'pbc_lists', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'audit_cycle_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['audit_cycle_id'], [u'cycles.id'], name=u'pbc_lists_ibfk_1'), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_pbc_lists_contexts'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) op.create_table(u'control_assessments', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'pbc_list_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'control_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'control_version', sa.VARCHAR(length=250), nullable=True), sa.Column(u'internal_tod', sa.BOOLEAN, nullable=True), sa.Column(u'internal_toe', sa.BOOLEAN, nullable=True), sa.Column(u'external_tod', sa.BOOLEAN, nullable=True), sa.Column(u'external_toe', sa.BOOLEAN, nullable=True), sa.Column(u'notes', sa.TEXT(), nullable=True), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_control_assessments_contexts'), sa.ForeignKeyConstraint(['control_id'], [u'controls.id'], name=u'control_assessments_ibfk_1'), sa.ForeignKeyConstraint(['pbc_list_id'], [u'pbc_lists.id'], name=u'control_assessments_ibfk_2'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) op.create_table(u'requests', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'pbc_list_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'type_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'pbc_control_code', sa.VARCHAR(length=250), nullable=True), sa.Column(u'pbc_control_desc', sa.TEXT(), nullable=True), sa.Column(u'request', sa.TEXT(), nullable=True), sa.Column(u'test', sa.TEXT(), nullable=True), sa.Column(u'notes', sa.TEXT(), nullable=True), sa.Column(u'company_responsible', sa.VARCHAR(length=250), nullable=True), sa.Column(u'auditor_responsible', sa.VARCHAR(length=250), nullable=True), sa.Column(u'date_requested', sa.DATETIME(), nullable=True), sa.Column(u'status', sa.VARCHAR(length=250), nullable=True), sa.Column(u'control_assessment_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'response_due_at', sa.DATE(), nullable=True), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_requests_contexts'), sa.ForeignKeyConstraint(['control_assessment_id'], [u'control_assessments.id'], name=u'requests_ibfk_1'), sa.ForeignKeyConstraint(['pbc_list_id'], [u'pbc_lists.id'], name=u'requests_ibfk_2'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) op.create_table(u'responses', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'request_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'system_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'status', sa.VARCHAR(length=250), nullable=True), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_responses_contexts'), sa.ForeignKeyConstraint(['request_id'], [u'requests.id'], name=u'responses_ibfk_1'), sa.ForeignKeyConstraint(['system_id'], [u'systems.id'], name=u'responses_ibfk_2'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) op.create_table(u'population_samples', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'response_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'population_document_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'population', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'sample_worksheet_document_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'samples', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'sample_evidence_document_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_population_samples_contexts'), sa.ForeignKeyConstraint(['population_document_id'], [u'documents.id'], name=u'population_samples_ibfk_1'), sa.ForeignKeyConstraint(['response_id'], [u'responses.id'], name=u'population_samples_ibfk_2'), sa.ForeignKeyConstraint(['sample_evidence_document_id'], [u'documents.id'], name=u'population_samples_ibfk_3'), sa.ForeignKeyConstraint(['sample_worksheet_document_id'], [u'documents.id'], name=u'population_samples_ibfk_4'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) op.create_table(u'meetings', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'response_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'start_at', sa.DATETIME(), nullable=True), sa.Column(u'calendar_url', sa.VARCHAR(length=250), nullable=True), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_meetings_contexts'), sa.ForeignKeyConstraint(['response_id'], [u'responses.id'], name=u'meetings_ibfk_1'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) op.create_table(u'system_controls', sa.Column(u'id', sa.INTEGER(), nullable=False), sa.Column(u'modified_by_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'created_at', sa.DATETIME(), nullable=True), sa.Column(u'updated_at', sa.DATETIME(), nullable=True), sa.Column(u'system_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'control_id', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'state', sa.INTEGER(), autoincrement=False, nullable=False), sa.Column(u'cycle_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.Column(u'context_id', sa.INTEGER(), autoincrement=False, nullable=True), sa.ForeignKeyConstraint(['context_id'], [u'contexts.id'], name=u'fk_system_controls_contexts'), sa.ForeignKeyConstraint(['control_id'], [u'controls.id'], name=u'system_controls_ibfk_1'), sa.ForeignKeyConstraint(['cycle_id'], [u'cycles.id'], name=u'system_controls_ibfk_2'), sa.ForeignKeyConstraint(['system_id'], [u'systems.id'], name=u'system_controls_ibfk_3'), sa.PrimaryKeyConstraint(u'id'), mysql_default_charset=u'utf8', mysql_engine=u'InnoDB' ) for table, column in NOT_NULL_COLS: op.alter_column(table, column, nullable=False, existing_type = sa.INTEGER) for table, column, referred_table, constraint_name in EXPLICIT_INDEXES: create_explicit_index(table, column, referred_table, constraint_name) for table, columns in UNIQUE_CONSTRAINTS: op.create_unique_constraint('uq_' + table, table, columns)
apache-2.0
ryokochang/Slab-GCS
ExtLibs/Mavlink/pymavlink/generator/lib/genxmlif/xmlifApi.py
12
54395
# # genxmlif, Release 0.9.0 # file: xmlifapi.py # # API (interface) classes for generic interface package # # history: # 2007-06-29 rl created, classes extracted from xmlifbase.py # # Copyright (c) 2005-2008 by Roland Leuthe. All rights reserved. # # -------------------------------------------------------------------- # The generic XML interface is # # Copyright (c) 2005-2008 by Roland Leuthe # # By obtaining, using, and/or copying this software and/or its # associated documentation, you agree that you have read, understood, # and will comply with the following terms and conditions: # # Permission to use, copy, modify, and distribute this software and # its associated documentation for any purpose and without fee is # hereby granted, provided that the above copyright notice appears in # all copies, and that both that copyright notice and this permission # notice appear in supporting documentation, and that the name of # the author not be used in advertising or publicity # pertaining to distribution of the software without specific, written # prior permission. # # THE AUTHOR DISCLAIMS ALL WARRANTIES WITH REGARD # TO THIS SOFTWARE, INCLUDING ALL IMPLIED WARRANTIES OF MERCHANT- # ABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR # BE LIABLE FOR ANY SPECIAL, INDIRECT OR CONSEQUENTIAL DAMAGES OR ANY # DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, # WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS # ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE # OF THIS SOFTWARE. # -------------------------------------------------------------------- __author__ = "Roland Leuthe <roland@leuthe-net.de>" __date__ = "08. August 2008" __version__ = "0.9.0" import string import os import re import copy from types import TupleType, StringTypes from xml.dom import EMPTY_PREFIX, EMPTY_NAMESPACE from xmlifUtils import processWhitespaceAction, NsNameTupleFactory, splitQName, nsNameToQName, escapeCdata, escapeAttribute ######################################## # XML interface base class # All not implemented methods have to be overloaded by the derived class!! # class XmlInterfaceBase: """XML interface base class. All not implemented methods have to be overloaded by the derived class!! """ def __init__(self, verbose, useCaching, processXInclude): """Constructor of class XmlInterfaceBase. Input parameter: 'verbose': 0 or 1: controls verbose print output for module genxmlif 'useCaching': 0 or 1: controls usage of caching for module genxmlif 'processXInclude': 0 or 1: controls XInclude processing during parsing """ self.verbose = verbose self.useCaching = useCaching self.processXInclude = processXInclude # set default wrapper classes self.setTreeWrapperClass (XmlTreeWrapper) self.setElementWrapperClass (XmlElementWrapper) def createXmlTree (self, namespace, xmlRootTagName, attributeDict={}, publicId=None, systemId=None): """Create a new XML TreeWrapper object (wrapper for DOM document or elementtree). Input parameter: 'namespace': not yet handled (for future use) 'xmlRootTagName': specifies the tag name of the root element 'attributeDict': contains the attributes of the root node (optional) 'publicId': forwarded to contained DOM tree (unused for elementtree) 'systemId': forwarded to contained DOM tree (unused for elementtree) Returns the created XML tree wrapper object. Method has to be implemented by derived classes! """ raise NotImplementedError def parse (self, filePath, baseUrl="", ownerDoc=None): """Call the XML parser for 'file'. Input parameter: 'filePath': a file path or an URI 'baseUrl': if specified, it is used e.g. as base path for schema files referenced inside the XML file. 'ownerDoc': only used in case of 4DOM (forwarded to 4DOM parser). Returns the respective XML tree wrapper object for the parsed XML file. Method has to be implemented by derived classes! """ raise NotImplementedError def parseString (self, text, baseUrl="", ownerDoc=None): """Call the XML parser for 'text'. Input parameter: 'text': contains the XML string to be parsed 'baseUrl': if specified, it is used e.g. as base path for schema files referenced inside the XML string. 'ownerDoc': only used in case of 4DOM (forwarded to 4DOM parser). Returns the respective XML tree wrapper object for the parsed XML 'text' string. Method has to be implemented by derived classes! """ raise NotImplementedError def setTreeWrapperClass (self, treeWrapperClass): """Set the tree wrapper class which shall be used by this interface. Input parameter: treeWrapperClass: tree wrapper class """ self.treeWrapperClass = treeWrapperClass def setElementWrapperClass (self, elementWrapperClass): """Set the element wrapper classes which shall be used by this interface. Input parameter: elementWrapperClass: element wrapper class """ self.elementWrapperClass = elementWrapperClass def getXmlIfType (self): """Retrieve the type of the XML interface.""" return self.xmlIfType ######################################## # Tree wrapper API (interface class) # class XmlTreeWrapper: """XML tree wrapper API. Contains a DOM tree or an elementtree (depending on used XML parser) """ def __init__(self, xmlIf, tree, useCaching): """Constructor of wrapper class XmlTreeWrapper. Input parameter: 'xmlIf': used XML interface class 'tree': DOM tree or elementtree which is wrapped by this object 'useCaching': 1 if caching shall be used inside genxmlif, otherwise 0 """ self.xmlIf = xmlIf self.__tree = tree self.__useCaching = useCaching def createElement (self, tupleOrLocalName, attributeDict=None, curNs=[]): """Create an ElementWrapper object. Input parameter: tupleOrLocalName: tag name of element node to be created (tuple of namespace and localName or only localName if no namespace is used) attributeDict: attributes for this elements curNs: namespaces for scope of this element Returns an ElementWrapper object containing the created element node. """ nsName = NsNameTupleFactory(tupleOrLocalName) elementNode = self.__tree.xmlIfExtCreateElement(nsName, attributeDict, curNs) return self.xmlIf.elementWrapperClass(elementNode, self, curNs) def cloneTree (self): """Creates a copy of a whole XML DOM tree.""" rootElementWrapperCopy = self.getRootNode().cloneNode(deep=1) treeWrapperCopy = self.__class__(self.xmlIf, self.__tree.xmlIfExtCloneTree(rootElementWrapperCopy.element), self.__useCaching) for elementWrapper in rootElementWrapperCopy.getIterator(): elementWrapper.treeWrapper = treeWrapperCopy return treeWrapperCopy def getRootNode (self): """Retrieve the wrapper object of the root element of the contained XML tree. Returns the ElementWrapper object of the root element. """ return self.__tree.xmlIfExtGetRootNode().xmlIfExtElementWrapper def getTree (self): """Retrieve the contained XML tree. Returns the contained XML tree object (internal DOM tree wrapper or elementtree). """ return self.__tree def printTree (self, prettyPrint=0, printElementValue=1, encoding=None): """Return the string representation of the contained XML tree. Input parameter: 'prettyPrint': aligns the columns of the attributes of childNodes 'printElementValue': controls if the lement values are printed or not. Returns a string with the string representation of the whole XML tree. """ if not encoding: encoding = "utf-8" if encoding != "utf-8" and encoding != "us-ascii": text = "<?xml version='1.0' encoding='%s'?>\n" % encoding else: text = "" return text + self.getRootNode().printNode(deep=1, prettyPrint=prettyPrint, printElementValue=printElementValue, encoding=encoding) def useCaching (self): """Return 1 if caching should be used for the contained XML tree.""" return self.__useCaching def setExternalCacheUsage (self, used): """Set external cache usage for the whole tree unlink commands are ignored if used by an external cache Input parameter: used: 0 or 1 (used by external cache) """ self.getRootNode().setExternalCacheUsage (used, deep=1) def unlink (self): """Break circular references of the complete XML tree. To be called if the XML tree is not longer used => garbage collection! """ self.getRootNode().unlink() def __str__ (self): """Return the string representation of the contained XML tree.""" return self.printTree() ######################################## # Element wrapper API (interface class) # class XmlElementWrapper: """XML element wrapper API. Contains a XML element node All not implemented methods have to be overloaded by the derived class!! """ def __init__(self, element, treeWrapper, curNs=[], initAttrSeq=1): """Constructor of wrapper class XmlElementWrapper. Input parameter: element: XML element node which is wrapped by this object treeWrapper: XML tree wrapper class the current element belongs to curNs: namespaces for scope of this element """ self.element = element self.element.xmlIfExtElementWrapper = self self.treeWrapper = treeWrapper self.nodeUsedByExternalCache = 0 if self.__useCaching(): self.__childrenCache = {} self.__firstChildCache = {} self.__qNameAttrCache = {} self.baseUrl = None self.absUrl = None self.filePath = None self.startLineNumber = None self.endLineNumber = None self.curNs = curNs[:] self.attributeSequence = [] if initAttrSeq: self.attributeSequence = self.getAttributeDict().keys() def unlink (self): """Break circular references of this element and its children.""" for childWrapper in self.getChildren(): childWrapper.unlink() if not self.isUsedByExternalCache(): self.element.xmlIfExtUnlink() def cloneNode (self, deep, cloneCallback=None): """Create a copy of the current element wrapper. The reference to the parent node is set to None!""" elementCopy = self.element.xmlIfExtCloneNode() elementWrapperCopy = self.__class__(elementCopy, self.treeWrapper, initAttrSeq=0) elementWrapperCopy.treeWrapper = None elementWrapperCopy.baseUrl = self.baseUrl elementWrapperCopy.absUrl = self.absUrl elementWrapperCopy.filePath = self.filePath elementWrapperCopy.startLineNumber = self.startLineNumber elementWrapperCopy.endLineNumber = self.endLineNumber elementWrapperCopy.curNs = self.curNs[:] elementWrapperCopy.attributeSequence = self.attributeSequence[:] if cloneCallback: cloneCallback(elementWrapperCopy) if deep: for childElement in self.element.xmlIfExtGetChildren(): childWrapperElementCopy = childElement.xmlIfExtElementWrapper.cloneNode(deep, cloneCallback) childWrapperElementCopy.element.xmlIfExtSetParentNode(elementWrapperCopy.element) elementWrapperCopy.element.xmlIfExtAppendChild(childWrapperElementCopy.element) return elementWrapperCopy def clearNodeCache (self): """Clear all caches used by this element wrapper which contains element wrapper references.""" self.__clearChildrenCache() def isUsedByExternalCache (self): """Check if this node is used by an external cache. unlink commands are ignored if used by an external cache""" return self.nodeUsedByExternalCache def setExternalCacheUsage (self, used, deep=1): """Set external cache usage for this node and its children unlink commands are ignored if used by an external cache Input parameter: used: 0 or 1 (used by external cache) deep: 0 or 1: controls if the child elements are also marked as used by external cache """ self.nodeUsedByExternalCache = used if deep: for childWrapper in self.getChildren(): childWrapper.setExternalCacheUsage (used, deep) ########################################################## # attributes of the current node can be accessed via key operator def __getitem__(self, tupleOrAttrName): """Attributes of the contained element node can be accessed via key operator. Input parameter: tupleOrAttrName: name of the attribute (tuple of namespace and attributeName or only attributeName) Returns the attribute value. """ attrValue = self.getAttribute (tupleOrAttrName) if attrValue != None: return attrValue else: raise AttributeError, "Attribute %s not found!" %(repr(tupleOrAttrName)) def __setitem__(self, tupleOrAttrName, attributeValue): """Attributes of the contained element node can be accessed via key operator. Input parameter: tupleOrAttrName: name of the attribute (tuple of namespace and attributeName or only attributeName) attributeValue: attribute value to be set """ self.setAttribute (tupleOrAttrName, attributeValue) #++++++++++++ methods concerning the tag name ++++++++++++++++++++++++ def getTagName (self): """Retrieve the (complete) tag name of the contained element node Returns the (complete) tag name of the contained element node """ return self.element.xmlIfExtGetTagName() def getLocalName (self): """Retrieve the local name (without namespace) of the contained element node Returns the local name (without namespace) of the contained element node """ try: return self.__localNameCache except: prefix, localName = splitQName (self.getTagName()) if self.__useCaching(): self.__localNameCache = localName return localName def getNamespaceURI (self): """Retrieve the namespace URI of the contained element node Returns the namespace URI of the contained element node (None if no namespace is used). """ try: return self.__nsUriCache except: prefix = self.element.xmlIfExtGetNamespaceURI() if self.__useCaching(): self.__nsUriCache = prefix return prefix def getNsName (self): """Retrieve a tuple (namespace, localName) of the contained element node Returns a tuple (namespace, localName) of the contained element node (namespace is None if no namespace is used). """ try: return self.__nsNameCache except: nsName = NsNameTupleFactory( (self.getNamespaceURI(), self.getLocalName()) ) if self.__useCaching(): self.__nsNameCache = nsName return nsName def getQName (self): """Retrieve a string prefix and localName of the contained element node Returns a string "prefix:localName" of the contained element node """ return self.nsName2QName(self.getNsName()) def getPrefix (self): """Retrieve the namespace prefix of the contained element node Returns the namespace prefix of the contained element node (None if no namespace is used). """ return self.getNsPrefix(self.getNsName()) #++++++++++++ methods concerning print support ++++++++++++++++++++++++ def __str__ (self): """Retrieve the textual representation of the contained element node.""" return self.printNode() def printNode (self, indent="", deep=0, prettyPrint=0, attrMaxLengthDict={}, printElementValue=1, encoding=None): """Retrieve the textual representation of the contained element node. Input parameter: indent: indentation to be used for string representation deep: 0 or 1: controls if the child element nodes are also printed prettyPrint: aligns the columns of the attributes of childNodes attrMaxLengthDict: dictionary containing the length of the attribute values (used for prettyprint) printElementValue: 0 or 1: controls if the element value is printed Returns the string representation """ patternXmlTagShort = '''\ %(indent)s<%(qName)s%(attributeString)s/>%(tailText)s%(lf)s''' patternXmlTagLong = '''\ %(indent)s<%(qName)s%(attributeString)s>%(elementValueString)s\ %(lf)s%(subTreeString)s\ %(indent)s</%(qName)s>%(tailText)s%(lf)s''' subTreeStringList = [] tailText = "" addIndent = "" lf = "" if deep: childAttrMaxLengthDict = {} if prettyPrint: for childNode in self.getChildren(): childNode.__updateAttrMaxLengthDict(childAttrMaxLengthDict) lf = "\n" addIndent = " " for childNode in self.getChildren(): subTreeStringList.append (childNode.printNode(indent + addIndent, deep, prettyPrint, childAttrMaxLengthDict, printElementValue)) tailText = escapeCdata(self.element.xmlIfExtGetElementTailText(), encoding) attributeStringList = [] for attrName in self.getAttributeList(): attrValue = escapeAttribute(self.getAttribute(attrName), encoding) if prettyPrint: try: align = attrMaxLengthDict[attrName] except: align = len(attrValue) else: align = len(attrValue) qName = self.nsName2QName(attrName) attributeStringList.append (' %s="%s"%*s' %(qName, attrValue, align - len(attrValue), "")) attributeString = string.join (attributeStringList, "") qName = self.getQName() if printElementValue: if deep: elementValueString = escapeCdata(self.element.xmlIfExtGetElementText(), encoding) else: elementValueString = escapeCdata(self.getElementValue(ignoreEmtpyStringFragments=1), encoding) else: elementValueString = "" if subTreeStringList == [] and elementValueString == "": printPattern = patternXmlTagShort else: if subTreeStringList != []: subTreeString = string.join (subTreeStringList, "") else: subTreeString = "" printPattern = patternXmlTagLong return printPattern % vars() #++++++++++++ methods concerning the parent of the current node ++++++++++++++++++++++++ def getParentNode (self): """Retrieve the ElementWrapper object of the parent element node. Returns the ElementWrapper object of the parent element node. """ parent = self.element.xmlIfExtGetParentNode() if parent != None: return parent.xmlIfExtElementWrapper else: return None #++++++++++++ methods concerning the children of the current node ++++++++++++++++++++++++ def getChildren (self, tagFilter=None): """Retrieve the ElementWrapper objects of the children element nodes. Input parameter: tagFilter: retrieve only the children with this tag name ('*' or None returns all children) Returns all children of this element node which match 'tagFilter' (list) """ if tagFilter in (None, '*', (None, '*')): children = self.element.xmlIfExtGetChildren() elif tagFilter[1] == '*': # handle (namespace, '*') children = filter(lambda child:child.xmlIfExtElementWrapper.getNamespaceURI() == tagFilter[0], self.element.xmlIfExtGetChildren()) else: nsNameFilter = NsNameTupleFactory(tagFilter) try: children = self.__childrenCache[nsNameFilter] except: children = self.element.xmlIfExtGetChildren(nsNameFilter) if self.__useCaching(): self.__childrenCache[nsNameFilter] = children return map(lambda child: child.xmlIfExtElementWrapper, children) def getChildrenNS (self, namespaceURI, tagFilter=None): """Retrieve the ElementWrapper objects of the children element nodes using a namespace. Input parameter: namespaceURI: the namespace URI of the children or None tagFilter: retrieve only the children with this localName ('*' or None returns all children) Returns all children of this element node which match 'namespaceURI' and 'tagFilter' (list) """ return self.getChildren((namespaceURI, tagFilter)) def getChildrenWithKey (self, tagFilter=None, keyAttr=None, keyValue=None): """Retrieve the ElementWrapper objects of the children element nodes. Input parameter: tagFilter: retrieve only the children with this tag name ('*' or None returns all children) keyAttr: name of the key attribute keyValue: value of the key Returns all children of this element node which match 'tagFilter' (list) """ children = self.getChildren(tagFilter) return filter(lambda child:child[keyAttr]==keyValue, children) def getFirstChild (self, tagFilter=None): """Retrieve the ElementWrapper objects of the first child element node. Input parameter: tagFilter: retrieve only the first child with this tag name ('*' or None: no filter) Returns the first child of this element node which match 'tagFilter' or None if no suitable child element was found """ if tagFilter in (None, '*', (None, '*')): element = self.element.xmlIfExtGetFirstChild() elif tagFilter[1] == '*': # handle (namespace, '*') children = filter(lambda child:child.xmlIfExtElementWrapper.getNamespaceURI() == tagFilter[0], self.element.xmlIfExtGetChildren()) try: element = children[0] except: element = None else: nsNameFilter = NsNameTupleFactory(tagFilter) try: element = self.__firstChildCache[nsNameFilter] except: element = self.element.xmlIfExtGetFirstChild(nsNameFilter) if self.__useCaching(): self.__firstChildCache[nsNameFilter] = element if element != None: return element.xmlIfExtElementWrapper else: return None def getFirstChildNS (self, namespaceURI, tagFilter=None): """Retrieve the ElementWrapper objects of the first child element node using a namespace. Input parameter: namespaceURI: the namespace URI of the children or None tagFilter: retrieve only the first child with this localName ('*' or None: no filter) Returns the first child of this element node which match 'namespaceURI' and 'tagFilter' or None if no suitable child element was found """ return self.getFirstChild ((namespaceURI, tagFilter)) def getFirstChildWithKey (self, tagFilter=None, keyAttr=None, keyValue=None): """Retrieve the ElementWrapper objects of the children element nodes. Input parameter: tagFilter: retrieve only the children with this tag name ('*' or None returns all children) keyAttr: name of the key attribute keyValue: value of the key Returns all children of this element node which match 'tagFilter' (list) """ children = self.getChildren(tagFilter) childrenWithKey = filter(lambda child:child[keyAttr]==keyValue, children) if childrenWithKey != []: return childrenWithKey[0] else: return None def getElementsByTagName (self, tagFilter=None): """Retrieve all descendant ElementWrapper object of current node whose tag name match 'tagFilter'. Input parameter: tagFilter: retrieve only the children with this tag name ('*' or None returns all descendants) Returns all descendants of this element node which match 'tagFilter' (list) """ if tagFilter in (None, '*', (None, '*'), (None, None)): descendants = self.element.xmlIfExtGetElementsByTagName() elif tagFilter[1] == '*': # handle (namespace, '*') descendants = filter(lambda desc:desc.xmlIfExtElementWrapper.getNamespaceURI() == tagFilter[0], self.element.xmlIfExtGetElementsByTagName()) else: nsNameFilter = NsNameTupleFactory(tagFilter) descendants = self.element.xmlIfExtGetElementsByTagName(nsNameFilter) return map(lambda descendant: descendant.xmlIfExtElementWrapper, descendants) def getElementsByTagNameNS (self, namespaceURI, tagFilter=None): """Retrieve all descendant ElementWrapper object of current node whose tag name match 'namespaceURI' and 'tagFilter'. Input parameter: namespaceURI: the namespace URI of the descendants or None tagFilter: retrieve only the descendants with this localName ('*' or None returns all descendants) Returns all descendants of this element node which match 'namespaceURI' and 'tagFilter' (list) """ return self.getElementsByTagName((namespaceURI, tagFilter)) def getIterator (self, tagFilter=None): """Creates a tree iterator. The iterator loops over this element and all subelements, in document order, and returns all elements whose tag name match 'tagFilter'. Input parameter: tagFilter: retrieve only the children with this tag name ('*' or None returns all descendants) Returns all element nodes which match 'tagFilter' (list) """ if tagFilter in (None, '*', (None, '*'), (None, None)): matchingElements = self.element.xmlIfExtGetIterator() elif tagFilter[1] == '*': # handle (namespace, '*') matchingElements = filter(lambda desc:desc.xmlIfExtElementWrapper.getNamespaceURI() == tagFilter[0], self.element.xmlIfExtGetIterator()) else: nsNameFilter = NsNameTupleFactory(tagFilter) matchingElements = self.element.xmlIfExtGetIterator(nsNameFilter) return map(lambda e: e.xmlIfExtElementWrapper, matchingElements) def appendChild (self, tupleOrLocalNameOrElement, attributeDict={}): """Append an element node to the children of the current node. Input parameter: tupleOrLocalNameOrElement: (namespace, localName) or tagName or ElementWrapper object of the new child attributeDict: attribute dictionary containing the attributes of the new child (optional) If not an ElementWrapper object is given, a new ElementWrapper object is created with tupleOrLocalName Returns the ElementWrapper object of the new child. """ if not isinstance(tupleOrLocalNameOrElement, self.__class__): childElementWrapper = self.__createElement (tupleOrLocalNameOrElement, attributeDict) else: childElementWrapper = tupleOrLocalNameOrElement self.element.xmlIfExtAppendChild (childElementWrapper.element) self.__clearChildrenCache(childElementWrapper.getNsName()) return childElementWrapper def insertBefore (self, tupleOrLocalNameOrElement, refChild, attributeDict={}): """Insert an child element node before the given reference child of the current node. Input parameter: tupleOrLocalNameOrElement: (namespace, localName) or tagName or ElementWrapper object of the new child refChild: reference child ElementWrapper object attributeDict: attribute dictionary containing the attributes of the new child (optional) If not an ElementWrapper object is given, a new ElementWrapper object is created with tupleOrLocalName Returns the ElementWrapper object of the new child. """ if not isinstance(tupleOrLocalNameOrElement, self.__class__): childElementWrapper = self.__createElement (tupleOrLocalNameOrElement, attributeDict) else: childElementWrapper = tupleOrLocalNameOrElement if refChild == None: self.appendChild (childElementWrapper) else: self.element.xmlIfExtInsertBefore(childElementWrapper.element, refChild.element) self.__clearChildrenCache(childElementWrapper.getNsName()) return childElementWrapper def removeChild (self, childElementWrapper): """Remove the given child element node from the children of the current node. Input parameter: childElementWrapper: ElementWrapper object to be removed """ self.element.xmlIfExtRemoveChild(childElementWrapper.element) self.__clearChildrenCache(childElementWrapper.getNsName()) def insertSubtree (self, refChildWrapper, subTreeWrapper, insertSubTreeRootNode=1): """Insert the given subtree before 'refChildWrapper' ('refChildWrapper' is not removed!) Input parameter: refChildWrapper: reference child ElementWrapper object subTreeWrapper: subtree wrapper object which contains the subtree to be inserted insertSubTreeRootNode: if 1, root node of subtree is inserted into parent tree, otherwise not """ if refChildWrapper != None: self.element.xmlIfExtInsertSubtree (refChildWrapper.element, subTreeWrapper.getTree(), insertSubTreeRootNode) else: self.element.xmlIfExtInsertSubtree (None, subTreeWrapper.getTree(), insertSubTreeRootNode) self.__clearChildrenCache() def replaceChildBySubtree (self, childElementWrapper, subTreeWrapper, insertSubTreeRootNode=1): """Replace child element node by XML subtree (e.g. expanding included XML files) Input parameter: childElementWrapper: ElementWrapper object to be replaced subTreeWrapper: XML subtree wrapper object to be inserted insertSubTreeRootNode: if 1, root node of subtree is inserted into parent tree, otherwise not """ self.insertSubtree (childElementWrapper, subTreeWrapper, insertSubTreeRootNode) self.removeChild(childElementWrapper) #++++++++++++ methods concerning the attributes of the current node ++++++++++++++++++++++++ def getAttributeDict (self): """Retrieve a dictionary containing all attributes of the current element node. Returns a dictionary (copy) containing all attributes of the current element node. """ return self.element.xmlIfExtGetAttributeDict() def getAttributeList (self): """Retrieve a list containing all attributes of the current element node in the sequence specified in the input XML file. Returns a list (copy) containing all attributes of the current element node in the sequence specified in the input XML file (TODO: does currently not work for 4DOM/pyXML interface). """ attrList = map(lambda a: NsNameTupleFactory(a), self.attributeSequence) return attrList def getAttribute (self, tupleOrAttrName): """Retrieve an attribute value of the current element node. Input parameter: tupleOrAttrName: tuple '(namespace, attributeName)' or 'attributeName' if no namespace is used Returns the value of the specified attribute. """ nsName = NsNameTupleFactory(tupleOrAttrName) return self.element.xmlIfExtGetAttribute(nsName) def getAttributeOrDefault (self, tupleOrAttrName, defaultValue): """Retrieve an attribute value of the current element node or the given default value if the attribute doesn't exist. Input parameter: tupleOrAttrName: tuple '(namespace, attributeName)' or 'attributeName' if no namespace is used Returns the value of the specified attribute or the given default value if the attribute doesn't exist. """ attributeValue = self.getAttribute (tupleOrAttrName) if attributeValue == None: attributeValue = defaultValue return attributeValue def getQNameAttribute (self, tupleOrAttrName): """Retrieve a QName attribute value of the current element node. Input parameter: tupleOrAttrName: tuple '(namespace, attributeName)' or 'attributeName' if no namespace is used Returns the value of the specified QName attribute as tuple (namespace, localName), i.e. the prefix is converted into the corresponding namespace value. """ nsNameAttrName = NsNameTupleFactory(tupleOrAttrName) try: return self.__qNameAttrCache[nsNameAttrName] except: qNameValue = self.getAttribute (nsNameAttrName) nsNameValue = self.qName2NsName(qNameValue, useDefaultNs=1) if self.__useCaching(): self.__qNameAttrCache[nsNameAttrName] = nsNameValue return nsNameValue def hasAttribute (self, tupleOrAttrName): """Checks if the requested attribute exist for the current element node. Returns 1 if the attribute exists, otherwise 0. """ nsName = NsNameTupleFactory(tupleOrAttrName) attrValue = self.element.xmlIfExtGetAttribute(nsName) if attrValue != None: return 1 else: return 0 def setAttribute (self, tupleOrAttrName, attributeValue): """Sets an attribute value of the current element node. If the attribute does not yet exist, it will be created. Input parameter: tupleOrAttrName: tuple '(namespace, attributeName)' or 'attributeName' if no namespace is used attributeValue: attribute value to be set """ if not isinstance(attributeValue, StringTypes): raise TypeError, "%s (attribute %s) must be a string!" %(repr(attributeValue), repr(tupleOrAttrName)) nsNameAttrName = NsNameTupleFactory(tupleOrAttrName) if nsNameAttrName not in self.attributeSequence: self.attributeSequence.append(nsNameAttrName) if self.__useCaching(): if self.__qNameAttrCache.has_key(nsNameAttrName): del self.__qNameAttrCache[nsNameAttrName] self.element.xmlIfExtSetAttribute(nsNameAttrName, attributeValue, self.getCurrentNamespaces()) def setAttributeDefault (self, tupleOrAttrName, defaultValue): """Create attribute and set value to default if it does not yet exist for the current element node. If the attribute is already existing nothing is done. Input parameter: tupleOrAttrName: tuple '(namespace, attributeName)' or 'attributeName' if no namespace is used defaultValue: default attribute value to be set """ if not self.hasAttribute(tupleOrAttrName): self.setAttribute(tupleOrAttrName, defaultValue) def removeAttribute (self, tupleOrAttrName): """Removes an attribute from the current element node. No exception is raised if there is no matching attribute. Input parameter: tupleOrAttrName: tuple '(namespace, attributeName)' or 'attributeName' if no namespace is used """ nsNameAttrName = NsNameTupleFactory(tupleOrAttrName) if self.__useCaching(): if self.__qNameAttrCache.has_key(nsNameAttrName): del self.__qNameAttrCache[nsNameAttrName] self.element.xmlIfExtRemoveAttribute(nsNameAttrName) def processWsAttribute (self, tupleOrAttrName, wsAction): """Process white space action for the specified attribute according to requested 'wsAction'. Input parameter: tupleOrAttrName: tuple '(namespace, attributeName)' or 'attributeName' if no namespace is used wsAction: 'collapse': substitute multiple whitespace characters by a single ' ' 'replace': substitute each whitespace characters by a single ' ' """ attributeValue = self.getAttribute(tupleOrAttrName) newValue = processWhitespaceAction (attributeValue, wsAction) if newValue != attributeValue: self.setAttribute(tupleOrAttrName, newValue) return newValue #++++++++++++ methods concerning the content of the current node ++++++++++++++++++++++++ def getElementValue (self, ignoreEmtpyStringFragments=0): """Retrieve the content of the current element node. Returns the content of the current element node as string. The content of multiple text nodes / CDATA nodes are concatenated to one string. Input parameter: ignoreEmtpyStringFragments: if 1, text nodes containing only whitespaces are ignored """ return "".join (self.getElementValueFragments(ignoreEmtpyStringFragments)) def getElementValueFragments (self, ignoreEmtpyStringFragments=0): """Retrieve the content of the current element node as value fragment list. Returns the content of the current element node as list of string fragments. Each list element represents one text nodes / CDATA node. Input parameter: ignoreEmtpyStringFragments: if 1, text nodes containing only whitespaces are ignored Method has to be implemented by derived classes! """ return self.element.xmlIfExtGetElementValueFragments (ignoreEmtpyStringFragments) def setElementValue (self, elementValue): """Set the content of the current element node. Input parameter: elementValue: string containing the new element value If multiple text nodes / CDATA nodes are existing, 'elementValue' is set for the first text node / CDATA node. All other text nodes /CDATA nodes are set to ''. """ self.element.xmlIfExtSetElementValue(elementValue) def processWsElementValue (self, wsAction): """Process white space action for the content of the current element node according to requested 'wsAction'. Input parameter: wsAction: 'collapse': substitute multiple whitespace characters by a single ' ' 'replace': substitute each whitespace characters by a single ' ' """ self.element.xmlIfExtProcessWsElementValue(wsAction) return self.getElementValue() #++++++++++++ methods concerning the info about the current node in the XML file ++++++++++++++++++++ def getStartLineNumber (self): """Retrieve the start line number of the current element node. Returns the start line number of the current element node in the XML file """ return self.startLineNumber def getEndLineNumber (self): """Retrieve the end line number of the current element node. Returns the end line number of the current element node in the XML file """ return self.endLineNumber def getAbsUrl (self): """Retrieve the absolute URL of the XML file the current element node belongs to. Returns the absolute URL of the XML file the current element node belongs to. """ return self.absUrl def getBaseUrl (self): """Retrieve the base URL of the XML file the current element node belongs to. Returns the base URL of the XML file the current element node belongs to. """ return self.baseUrl def getFilePath (self): """Retrieve the file path of the XML file the current element node belongs to. Returns the file path of the XML file the current element node belongs to. """ return self.filePath def getLocation (self, end=0, fullpath=0): """Retrieve a string containing file name and line number of the current element node. Input parameter: end: 1 if end line number shall be shown, 0 for start line number fullpath: 1 if the full path of the XML file shall be shown, 0 for only the file name Returns a string containing file name and line number of the current element node. (e.g. to be used for traces or error messages) """ lineMethod = (self.getStartLineNumber, self.getEndLineNumber) pathFunc = (os.path.basename, os.path.abspath) return "%s, %d" % (pathFunc[fullpath](self.getFilePath()), lineMethod[end]()) #++++++++++++ miscellaneous methods concerning namespaces ++++++++++++++++++++ def getCurrentNamespaces (self): """Retrieve the namespace prefixes visible for the current element node Returns a list of the namespace prefixes visible for the current node. """ return self.curNs def qName2NsName (self, qName, useDefaultNs): """Convert a qName 'prefix:localName' to a tuple '(namespace, localName)'. Input parameter: qName: qName to be converted useDefaultNs: 1 if default namespace shall be used Returns the corresponding tuple '(namespace, localName)' for 'qName'. """ if qName != None: qNamePrefix, qNameLocalName = splitQName (qName) for prefix, namespaceURI in self.getCurrentNamespaces(): if qNamePrefix == prefix: if prefix != EMPTY_PREFIX or useDefaultNs: nsName = (namespaceURI, qNameLocalName) break else: if qNamePrefix == None: nsName = (EMPTY_NAMESPACE, qNameLocalName) else: raise ValueError, "Namespace prefix '%s' not bound to a namespace!" % (qNamePrefix) else: nsName = (None, None) return NsNameTupleFactory(nsName) def nsName2QName (self, nsLocalName): """Convert a tuple '(namespace, localName)' to a string 'prefix:localName' Input parameter: nsLocalName: tuple '(namespace, localName)' to be converted Returns the corresponding string 'prefix:localName' for 'nsLocalName'. """ qName = nsNameToQName (nsLocalName, self.getCurrentNamespaces()) if qName == "xmlns:None": qName = "xmlns" return qName def getNamespace (self, qName): """Retrieve namespace for a qName 'prefix:localName'. Input parameter: qName: qName 'prefix:localName' Returns the corresponding namespace for the prefix of 'qName'. """ if qName != None: qNamePrefix, qNameLocalName = splitQName (qName) for prefix, namespaceURI in self.getCurrentNamespaces(): if qNamePrefix == prefix: namespace = namespaceURI break else: if qNamePrefix == None: namespace = EMPTY_NAMESPACE else: raise LookupError, "Namespace for QName '%s' not found!" % (qName) else: namespace = EMPTY_NAMESPACE return namespace def getNsPrefix (self, nsLocalName): """Retrieve prefix for a tuple '(namespace, localName)'. Input parameter: nsLocalName: tuple '(namespace, localName)' Returns the corresponding prefix for the namespace of 'nsLocalName'. """ ns = nsLocalName[0] for prefix, namespace in self.getCurrentNamespaces(): if ns == namespace: return prefix else: if ns == None: return None else: raise LookupError, "Prefix for namespaceURI '%s' not found!" % (ns) #++++++++++++ limited XPath support ++++++++++++++++++++ def getXPath (self, xPath, namespaceRef=None, useDefaultNs=1, attrIgnoreList=[]): """Retrieve node list or attribute list for specified XPath Input parameter: xPath: string containing xPath specification namespaceRef: scope for namespaces (default is own element node) useDefaultNs: 1, if default namespace shall be used if no prefix is available attrIgnoreList: list of attributes to be ignored if wildcard is specified for attributes Returns all nodes which match xPath specification or list of attribute values if xPath specifies an attribute """ return self.getXPathList(xPath, namespaceRef, useDefaultNs, attrIgnoreList)[0] def getXPathList (self, xPath, namespaceRef=None, useDefaultNs=1, attrIgnoreList=[]): """Retrieve node list or attribute list for specified XPath Input parameter: xPath: string containing xPath specification namespaceRef: scope for namespaces (default is own element node) useDefaultNs: 1, if default namespace shall be used if no prefix is available attrIgnoreList: list of attributes to be ignored if wildcard is specified for attributes Returns tuple (completeChildList, attrNodeList, attrNsNameFirst). completeChildList: contains all child node which match xPath specification or list of attribute values if xPath specifies an attribute attrNodeList: contains all child nodes where the specified attribute was found attrNsNameFirst: contains the name of the first attribute which was found TODO: Re-design namespace and attribute handling of this method """ reChild = re.compile('child *::') reAttribute = re.compile('attribute *::') if namespaceRef == None: namespaceRef = self xPath = reChild.sub('./', xPath) xPath = reAttribute.sub('@', xPath) xPathList = string.split (xPath, "|") completeChildDict = {} completeChildList = [] attrNodeList = [] attrNsNameFirst = None for xRelPath in xPathList: xRelPath = string.strip(xRelPath) descendantOrSelf = 0 if xRelPath[:3] == ".//": descendantOrSelf = 1 xRelPath = xRelPath[3:] xPathLocalStepList = string.split (xRelPath, "/") childList = [self, ] for localStep in xPathLocalStepList: localStep = string.strip(localStep) stepChildList = [] if localStep == "": raise IOError ("Invalid xPath '%s'!" %(xRelPath)) elif localStep == ".": continue elif localStep[0] == '@': if len(localStep) == 1: raise ValueError ("Attribute name is missing in xPath!") if descendantOrSelf: childList = self.getElementsByTagName() attrName = localStep[1:] for childNode in childList: if attrName == '*': attrNodeList.append (childNode) attrDict = childNode.getAttributeDict() for attrIgnore in attrIgnoreList: if attrDict.has_key(attrIgnore): del attrDict[attrIgnore] stepChildList.extend(attrDict.values()) try: attrNsNameFirst = attrDict.keys()[0] except: pass else: attrNsName = namespaceRef.qName2NsName (attrName, useDefaultNs=0) if attrNsName[1] == '*': for attr in childNode.getAttributeDict().keys(): if attr[0] == attrNsName[0]: if attrNodeList == []: attrNsNameFirst = attrNsName attrNodeList.append (childNode) stepChildList.append (childNode.getAttribute(attr)) elif childNode.hasAttribute(attrNsName): if attrNodeList == []: attrNsNameFirst = attrNsName attrNodeList.append (childNode) stepChildList.append (childNode.getAttribute(attrNsName)) childList = stepChildList else: nsLocalName = namespaceRef.qName2NsName (localStep, useDefaultNs=useDefaultNs) if descendantOrSelf: descendantOrSelf = 0 if localStep == "*": stepChildList = self.getElementsByTagName() else: stepChildList = self.getElementsByTagName(nsLocalName) else: for childNode in childList: if localStep == "*": stepChildList.extend (childNode.getChildren()) else: stepChildList.extend (childNode.getChildrenNS(nsLocalName[0], nsLocalName[1])) childList = stepChildList # filter duplicated childs for child in childList: try: childKey = child.element except: childKey = child if not completeChildDict.has_key(childKey): completeChildList.append(child) completeChildDict[childKey] = 1 return completeChildList, attrNodeList, attrNsNameFirst ############################################################### # PRIVATE methods ############################################################### def __createElement (self, tupleOrLocalName, attributeDict): """Create a new ElementWrapper object. Input parameter: tupleOrLocalName: tuple '(namespace, localName)' or 'localName' if no namespace is used attributeDict: dictionary which contains the attributes and their values of the element node to be created Returns the created ElementWrapper object """ childElementWrapper = self.treeWrapper.createElement (tupleOrLocalName, attributeDict, self.curNs[:]) # TODO: when to be adapted???) childElementWrapper.element.xmlIfExtSetParentNode(self.element) return childElementWrapper def __updateAttrMaxLengthDict (self, attrMaxLengthDict): """Update dictionary which contains the maximum length of node attributes. Used for pretty print to align the attributes of child nodes. attrMaxLengthDict is in/out parameter. """ for attrName, attrValue in self.getAttributeDict().items(): attrLength = len(attrValue) if not attrMaxLengthDict.has_key(attrName): attrMaxLengthDict[attrName] = attrLength else: attrMaxLengthDict[attrName] = max(attrMaxLengthDict[attrName], attrLength) def __clearChildrenCache (self, childNsName=None): """Clear children cache. """ if self.__useCaching(): if childNsName != None: if self.__childrenCache.has_key(childNsName): del self.__childrenCache[childNsName] if self.__firstChildCache.has_key(childNsName): del self.__firstChildCache[childNsName] else: self.__childrenCache.clear() self.__firstChildCache.clear() def __useCaching(self): return self.treeWrapper.useCaching()
gpl-3.0
shaistaansari/django
tests/admin_widgets/tests.py
55
57921
# -*- coding: utf-8 -*- from __future__ import unicode_literals import gettext import os from datetime import datetime, timedelta from importlib import import_module from unittest import skipIf from django import forms from django.conf import settings from django.contrib import admin from django.contrib.admin import widgets from django.contrib.admin.tests import AdminSeleniumWebDriverTestCase from django.contrib.auth.models import User from django.core.files.storage import default_storage from django.core.files.uploadedfile import SimpleUploadedFile from django.core.urlresolvers import reverse from django.db.models import CharField, DateField from django.test import SimpleTestCase, TestCase, override_settings from django.utils import six, translation from . import models from .widgetadmin import site as widget_admin_site try: import pytz except ImportError: pytz = None class TestDataMixin(object): @classmethod def setUpTestData(cls): cls.u1 = User.objects.create( pk=100, username='super', first_name='Super', last_name='User', email='super@example.com', password='sha1$995a3$6011485ea3834267d719b4c801409b8b1ddd0158', is_active=True, is_superuser=True, is_staff=True, last_login=datetime(2007, 5, 30, 13, 20, 10), date_joined=datetime(2007, 5, 30, 13, 20, 10) ) cls.u2 = User.objects.create( pk=101, username='testser', first_name='Add', last_name='User', email='auser@example.com', password='sha1$995a3$6011485ea3834267d719b4c801409b8b1ddd0158', is_active=True, is_superuser=False, is_staff=True, last_login=datetime(2007, 5, 30, 13, 20, 10), date_joined=datetime(2007, 5, 30, 13, 20, 10) ) models.Car.objects.create(id=1, owner=cls.u1, make='Volkswagon', model='Passat') models.Car.objects.create(id=2, owner=cls.u2, make='BMW', model='M3') class SeleniumDataMixin(object): def setUp(self): self.u1 = User.objects.create( pk=100, username='super', first_name='Super', last_name='User', email='super@example.com', password='sha1$995a3$6011485ea3834267d719b4c801409b8b1ddd0158', is_active=True, is_superuser=True, is_staff=True, last_login=datetime(2007, 5, 30, 13, 20, 10), date_joined=datetime(2007, 5, 30, 13, 20, 10) ) class AdminFormfieldForDBFieldTests(SimpleTestCase): """ Tests for correct behavior of ModelAdmin.formfield_for_dbfield """ def assertFormfield(self, model, fieldname, widgetclass, **admin_overrides): """ Helper to call formfield_for_dbfield for a given model and field name and verify that the returned formfield is appropriate. """ # Override any settings on the model admin class MyModelAdmin(admin.ModelAdmin): pass for k in admin_overrides: setattr(MyModelAdmin, k, admin_overrides[k]) # Construct the admin, and ask it for a formfield ma = MyModelAdmin(model, admin.site) ff = ma.formfield_for_dbfield(model._meta.get_field(fieldname), request=None) # "unwrap" the widget wrapper, if needed if isinstance(ff.widget, widgets.RelatedFieldWidgetWrapper): widget = ff.widget.widget else: widget = ff.widget # Check that we got a field of the right type self.assertTrue( isinstance(widget, widgetclass), "Wrong widget for %s.%s: expected %s, got %s" % ( model.__class__.__name__, fieldname, widgetclass, type(widget), ) ) # Return the formfield so that other tests can continue return ff def test_DateField(self): self.assertFormfield(models.Event, 'start_date', widgets.AdminDateWidget) def test_DateTimeField(self): self.assertFormfield(models.Member, 'birthdate', widgets.AdminSplitDateTime) def test_TimeField(self): self.assertFormfield(models.Event, 'start_time', widgets.AdminTimeWidget) def test_TextField(self): self.assertFormfield(models.Event, 'description', widgets.AdminTextareaWidget) def test_URLField(self): self.assertFormfield(models.Event, 'link', widgets.AdminURLFieldWidget) def test_IntegerField(self): self.assertFormfield(models.Event, 'min_age', widgets.AdminIntegerFieldWidget) def test_CharField(self): self.assertFormfield(models.Member, 'name', widgets.AdminTextInputWidget) def test_EmailField(self): self.assertFormfield(models.Member, 'email', widgets.AdminEmailInputWidget) def test_FileField(self): self.assertFormfield(models.Album, 'cover_art', widgets.AdminFileWidget) def test_ForeignKey(self): self.assertFormfield(models.Event, 'main_band', forms.Select) def test_raw_id_ForeignKey(self): self.assertFormfield(models.Event, 'main_band', widgets.ForeignKeyRawIdWidget, raw_id_fields=['main_band']) def test_radio_fields_ForeignKey(self): ff = self.assertFormfield(models.Event, 'main_band', widgets.AdminRadioSelect, radio_fields={'main_band': admin.VERTICAL}) self.assertEqual(ff.empty_label, None) def test_many_to_many(self): self.assertFormfield(models.Band, 'members', forms.SelectMultiple) def test_raw_id_many_to_many(self): self.assertFormfield(models.Band, 'members', widgets.ManyToManyRawIdWidget, raw_id_fields=['members']) def test_filtered_many_to_many(self): self.assertFormfield(models.Band, 'members', widgets.FilteredSelectMultiple, filter_vertical=['members']) def test_formfield_overrides(self): self.assertFormfield(models.Event, 'start_date', forms.TextInput, formfield_overrides={DateField: {'widget': forms.TextInput}}) def test_formfield_overrides_widget_instances(self): """ Test that widget instances in formfield_overrides are not shared between different fields. (#19423) """ class BandAdmin(admin.ModelAdmin): formfield_overrides = { CharField: {'widget': forms.TextInput(attrs={'size': '10'})} } ma = BandAdmin(models.Band, admin.site) f1 = ma.formfield_for_dbfield(models.Band._meta.get_field('name'), request=None) f2 = ma.formfield_for_dbfield(models.Band._meta.get_field('style'), request=None) self.assertNotEqual(f1.widget, f2.widget) self.assertEqual(f1.widget.attrs['maxlength'], '100') self.assertEqual(f2.widget.attrs['maxlength'], '20') self.assertEqual(f2.widget.attrs['size'], '10') def test_field_with_choices(self): self.assertFormfield(models.Member, 'gender', forms.Select) def test_choices_with_radio_fields(self): self.assertFormfield(models.Member, 'gender', widgets.AdminRadioSelect, radio_fields={'gender': admin.VERTICAL}) def test_inheritance(self): self.assertFormfield(models.Album, 'backside_art', widgets.AdminFileWidget) def test_m2m_widgets(self): """m2m fields help text as it applies to admin app (#9321).""" class AdvisorAdmin(admin.ModelAdmin): filter_vertical = ['companies'] self.assertFormfield(models.Advisor, 'companies', widgets.FilteredSelectMultiple, filter_vertical=['companies']) ma = AdvisorAdmin(models.Advisor, admin.site) f = ma.formfield_for_dbfield(models.Advisor._meta.get_field('companies'), request=None) self.assertEqual(six.text_type(f.help_text), 'Hold down "Control", or "Command" on a Mac, to select more than one.') @override_settings(PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls') class AdminFormfieldForDBFieldWithRequestTests(TestDataMixin, TestCase): def test_filter_choices_by_request_user(self): """ Ensure the user can only see their own cars in the foreign key dropdown. """ self.client.login(username="super", password="secret") response = self.client.get(reverse('admin:admin_widgets_cartire_add')) self.assertNotContains(response, "BMW M3") self.assertContains(response, "Volkswagon Passat") @override_settings(PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls') class AdminForeignKeyWidgetChangeList(TestDataMixin, TestCase): def setUp(self): self.client.login(username="super", password="secret") def test_changelist_ForeignKey(self): response = self.client.get(reverse('admin:admin_widgets_car_changelist')) self.assertContains(response, '/auth/user/add/') @override_settings(PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls') class AdminForeignKeyRawIdWidget(TestDataMixin, TestCase): def setUp(self): self.client.login(username="super", password="secret") def test_nonexistent_target_id(self): band = models.Band.objects.create(name='Bogey Blues') pk = band.pk band.delete() post_data = { "main_band": '%s' % pk, } # Try posting with a non-existent pk in a raw id field: this # should result in an error message, not a server exception. response = self.client.post(reverse('admin:admin_widgets_event_add'), post_data) self.assertContains(response, 'Select a valid choice. That choice is not one of the available choices.') def test_invalid_target_id(self): for test_str in ('Iñtërnâtiônàlizætiøn', "1234'", -1234): # This should result in an error message, not a server exception. response = self.client.post(reverse('admin:admin_widgets_event_add'), {"main_band": test_str}) self.assertContains(response, 'Select a valid choice. That choice is not one of the available choices.') def test_url_params_from_lookup_dict_any_iterable(self): lookup1 = widgets.url_params_from_lookup_dict({'color__in': ('red', 'blue')}) lookup2 = widgets.url_params_from_lookup_dict({'color__in': ['red', 'blue']}) self.assertEqual(lookup1, {'color__in': 'red,blue'}) self.assertEqual(lookup1, lookup2) def test_url_params_from_lookup_dict_callable(self): def my_callable(): return 'works' lookup1 = widgets.url_params_from_lookup_dict({'myfield': my_callable}) lookup2 = widgets.url_params_from_lookup_dict({'myfield': my_callable()}) self.assertEqual(lookup1, lookup2) class FilteredSelectMultipleWidgetTest(SimpleTestCase): def test_render(self): # Backslash in verbose_name to ensure it is JavaScript escaped. w = widgets.FilteredSelectMultiple('test\\', False) self.assertHTMLEqual( w.render('test', 'test'), '<select multiple="multiple" name="test" class="selectfilter">\n</select>' '<script type="text/javascript">addEvent(window, "load", function(e) ' '{SelectFilter.init("id_test", "test\\u005C", 0); });</script>\n' ) def test_stacked_render(self): # Backslash in verbose_name to ensure it is JavaScript escaped. w = widgets.FilteredSelectMultiple('test\\', True) self.assertHTMLEqual( w.render('test', 'test'), '<select multiple="multiple" name="test" class="selectfilterstacked">\n</select>' '<script type="text/javascript">addEvent(window, "load", function(e) ' '{SelectFilter.init("id_test", "test\\u005C", 1); });</script>\n' ) class AdminDateWidgetTest(SimpleTestCase): def test_attrs(self): """ Ensure that user-supplied attrs are used. Refs #12073. """ w = widgets.AdminDateWidget() self.assertHTMLEqual( w.render('test', datetime(2007, 12, 1, 9, 30)), '<input value="2007-12-01" type="text" class="vDateField" name="test" size="10" />', ) # pass attrs to widget w = widgets.AdminDateWidget(attrs={'size': 20, 'class': 'myDateField'}) self.assertHTMLEqual( w.render('test', datetime(2007, 12, 1, 9, 30)), '<input value="2007-12-01" type="text" class="myDateField" name="test" size="20" />', ) class AdminTimeWidgetTest(SimpleTestCase): def test_attrs(self): """ Ensure that user-supplied attrs are used. Refs #12073. """ w = widgets.AdminTimeWidget() self.assertHTMLEqual( w.render('test', datetime(2007, 12, 1, 9, 30)), '<input value="09:30:00" type="text" class="vTimeField" name="test" size="8" />', ) # pass attrs to widget w = widgets.AdminTimeWidget(attrs={'size': 20, 'class': 'myTimeField'}) self.assertHTMLEqual( w.render('test', datetime(2007, 12, 1, 9, 30)), '<input value="09:30:00" type="text" class="myTimeField" name="test" size="20" />', ) class AdminSplitDateTimeWidgetTest(SimpleTestCase): def test_render(self): w = widgets.AdminSplitDateTime() self.assertHTMLEqual( w.render('test', datetime(2007, 12, 1, 9, 30)), '<p class="datetime">Date: <input value="2007-12-01" type="text" class="vDateField" name="test_0" size="10" /><br />Time: <input value="09:30:00" type="text" class="vTimeField" name="test_1" size="8" /></p>', ) def test_localization(self): w = widgets.AdminSplitDateTime() with self.settings(USE_L10N=True), translation.override('de-at'): w.is_localized = True self.assertHTMLEqual( w.render('test', datetime(2007, 12, 1, 9, 30)), '<p class="datetime">Datum: <input value="01.12.2007" type="text" class="vDateField" name="test_0" size="10" /><br />Zeit: <input value="09:30:00" type="text" class="vTimeField" name="test_1" size="8" /></p>', ) class AdminURLWidgetTest(SimpleTestCase): def test_render(self): w = widgets.AdminURLFieldWidget() self.assertHTMLEqual( w.render('test', ''), '<input class="vURLField" name="test" type="url" />' ) self.assertHTMLEqual( w.render('test', 'http://example.com'), '<p class="url">Currently:<a href="http://example.com">http://example.com</a><br />Change:<input class="vURLField" name="test" type="url" value="http://example.com" /></p>' ) def test_render_idn(self): w = widgets.AdminURLFieldWidget() self.assertHTMLEqual( w.render('test', 'http://example-äüö.com'), '<p class="url">Currently: <a href="http://xn--example--7za4pnc.com">http://example-äüö.com</a><br />Change:<input class="vURLField" name="test" type="url" value="http://example-äüö.com" /></p>' ) def test_render_quoting(self): # WARNING: Don't use assertHTMLEqual in that testcase! # assertHTMLEqual will get rid of some escapes which are tested here! w = widgets.AdminURLFieldWidget() self.assertEqual( w.render('test', 'http://example.com/<sometag>some text</sometag>'), '<p class="url">Currently: <a href="http://example.com/%3Csometag%3Esome%20text%3C/sometag%3E">http://example.com/&lt;sometag&gt;some text&lt;/sometag&gt;</a><br />Change: <input class="vURLField" name="test" type="url" value="http://example.com/&lt;sometag&gt;some text&lt;/sometag&gt;" /></p>' ) self.assertEqual( w.render('test', 'http://example-äüö.com/<sometag>some text</sometag>'), '<p class="url">Currently: <a href="http://xn--example--7za4pnc.com/%3Csometag%3Esome%20text%3C/sometag%3E">http://example-äüö.com/&lt;sometag&gt;some text&lt;/sometag&gt;</a><br />Change: <input class="vURLField" name="test" type="url" value="http://example-äüö.com/&lt;sometag&gt;some text&lt;/sometag&gt;" /></p>' ) self.assertEqual( w.render('test', 'http://www.example.com/%C3%A4"><script>alert("XSS!")</script>"'), '<p class="url">Currently: <a href="http://www.example.com/%C3%A4%22%3E%3Cscript%3Ealert(%22XSS!%22)%3C/script%3E%22">http://www.example.com/%C3%A4&quot;&gt;&lt;script&gt;alert(&quot;XSS!&quot;)&lt;/script&gt;&quot;</a><br />Change: <input class="vURLField" name="test" type="url" value="http://www.example.com/%C3%A4&quot;&gt;&lt;script&gt;alert(&quot;XSS!&quot;)&lt;/script&gt;&quot;" /></p>' ) @override_settings( PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls', ) class AdminFileWidgetTests(TestDataMixin, TestCase): @classmethod def setUpTestData(cls): super(AdminFileWidgetTests, cls).setUpTestData() band = models.Band.objects.create(name='Linkin Park') cls.album = band.album_set.create( name='Hybrid Theory', cover_art=r'albums\hybrid_theory.jpg' ) def test_render(self): w = widgets.AdminFileWidget() self.assertHTMLEqual( w.render('test', self.album.cover_art), '<p class="file-upload">Currently: <a href="%(STORAGE_URL)salbums/' 'hybrid_theory.jpg">albums\hybrid_theory.jpg</a> ' '<span class="clearable-file-input">' '<input type="checkbox" name="test-clear" id="test-clear_id" /> ' '<label for="test-clear_id">Clear</label></span><br />' 'Change: <input type="file" name="test" /></p>' % { 'STORAGE_URL': default_storage.url(''), }, ) self.assertHTMLEqual( w.render('test', SimpleUploadedFile('test', b'content')), '<input type="file" name="test" />', ) def test_readonly_fields(self): """ File widgets should render as a link when they're marked "read only." """ self.client.login(username="super", password="secret") response = self.client.get(reverse('admin:admin_widgets_album_change', args=(self.album.id,))) self.assertContains( response, '<p><a href="%(STORAGE_URL)salbums/hybrid_theory.jpg">' 'albums\hybrid_theory.jpg</a></p>' % {'STORAGE_URL': default_storage.url('')}, html=True, ) self.assertNotContains( response, '<input type="file" name="cover_art" id="id_cover_art" />', html=True, ) response = self.client.get(reverse('admin:admin_widgets_album_add')) self.assertContains( response, '<p></p>', html=True, ) @override_settings(ROOT_URLCONF='admin_widgets.urls') class ForeignKeyRawIdWidgetTest(TestCase): def test_render(self): band = models.Band.objects.create(name='Linkin Park') band.album_set.create( name='Hybrid Theory', cover_art=r'albums\hybrid_theory.jpg' ) rel = models.Album._meta.get_field('band').remote_field w = widgets.ForeignKeyRawIdWidget(rel, widget_admin_site) self.assertHTMLEqual( w.render('test', band.pk, attrs={}), ( '<input type="text" name="test" value="%(bandpk)s" class="vForeignKeyRawIdAdminField" />' '<a href="/admin_widgets/band/?_to_field=id" class="related-lookup" id="lookup_id_test" title="Lookup"></a>' '&nbsp;<strong>Linkin Park</strong>' ) % {'bandpk': band.pk} ) def test_relations_to_non_primary_key(self): # Check that ForeignKeyRawIdWidget works with fields which aren't # related to the model's primary key. apple = models.Inventory.objects.create(barcode=86, name='Apple') models.Inventory.objects.create(barcode=22, name='Pear') core = models.Inventory.objects.create( barcode=87, name='Core', parent=apple ) rel = models.Inventory._meta.get_field('parent').remote_field w = widgets.ForeignKeyRawIdWidget(rel, widget_admin_site) self.assertHTMLEqual( w.render('test', core.parent_id, attrs={}), ( '<input type="text" name="test" value="86" class="vForeignKeyRawIdAdminField" />' '<a href="/admin_widgets/inventory/?_to_field=barcode" class="related-lookup" id="lookup_id_test" title="Lookup">' '</a>&nbsp;<strong>Apple</strong>' ) ) def test_fk_related_model_not_in_admin(self): # FK to a model not registered with admin site. Raw ID widget should # have no magnifying glass link. See #16542 big_honeycomb = models.Honeycomb.objects.create(location='Old tree') big_honeycomb.bee_set.create() rel = models.Bee._meta.get_field('honeycomb').remote_field w = widgets.ForeignKeyRawIdWidget(rel, widget_admin_site) self.assertHTMLEqual( w.render('honeycomb_widget', big_honeycomb.pk, attrs={}), '<input type="text" name="honeycomb_widget" value="%(hcombpk)s" />&nbsp;<strong>Honeycomb object</strong>' % {'hcombpk': big_honeycomb.pk} ) def test_fk_to_self_model_not_in_admin(self): # FK to self, not registered with admin site. Raw ID widget should have # no magnifying glass link. See #16542 subject1 = models.Individual.objects.create(name='Subject #1') models.Individual.objects.create(name='Child', parent=subject1) rel = models.Individual._meta.get_field('parent').remote_field w = widgets.ForeignKeyRawIdWidget(rel, widget_admin_site) self.assertHTMLEqual( w.render('individual_widget', subject1.pk, attrs={}), '<input type="text" name="individual_widget" value="%(subj1pk)s" />&nbsp;<strong>Individual object</strong>' % {'subj1pk': subject1.pk} ) def test_proper_manager_for_label_lookup(self): # see #9258 rel = models.Inventory._meta.get_field('parent').remote_field w = widgets.ForeignKeyRawIdWidget(rel, widget_admin_site) hidden = models.Inventory.objects.create( barcode=93, name='Hidden', hidden=True ) child_of_hidden = models.Inventory.objects.create( barcode=94, name='Child of hidden', parent=hidden ) self.assertHTMLEqual( w.render('test', child_of_hidden.parent_id, attrs={}), ( '<input type="text" name="test" value="93" class="vForeignKeyRawIdAdminField" />' '<a href="/admin_widgets/inventory/?_to_field=barcode" class="related-lookup" id="lookup_id_test" title="Lookup">' '</a>&nbsp;<strong>Hidden</strong>' ) ) @override_settings(ROOT_URLCONF='admin_widgets.urls') class ManyToManyRawIdWidgetTest(TestCase): def test_render(self): band = models.Band.objects.create(name='Linkin Park') m1 = models.Member.objects.create(name='Chester') m2 = models.Member.objects.create(name='Mike') band.members.add(m1, m2) rel = models.Band._meta.get_field('members').remote_field w = widgets.ManyToManyRawIdWidget(rel, widget_admin_site) self.assertHTMLEqual( w.render('test', [m1.pk, m2.pk], attrs={}), ( '<input type="text" name="test" value="%(m1pk)s,%(m2pk)s" class="vManyToManyRawIdAdminField" />' '<a href="/admin_widgets/member/" class="related-lookup" id="lookup_id_test" title="Lookup"></a>' ) % dict(m1pk=m1.pk, m2pk=m2.pk) ) self.assertHTMLEqual( w.render('test', [m1.pk]), ( '<input type="text" name="test" value="%(m1pk)s" class="vManyToManyRawIdAdminField">' '<a href="/admin_widgets/member/" class="related-lookup" id="lookup_id_test" title="Lookup"></a>' ) % dict(m1pk=m1.pk) ) def test_m2m_related_model_not_in_admin(self): # M2M relationship with model not registered with admin site. Raw ID # widget should have no magnifying glass link. See #16542 consultor1 = models.Advisor.objects.create(name='Rockstar Techie') c1 = models.Company.objects.create(name='Doodle') c2 = models.Company.objects.create(name='Pear') consultor1.companies.add(c1, c2) rel = models.Advisor._meta.get_field('companies').remote_field w = widgets.ManyToManyRawIdWidget(rel, widget_admin_site) self.assertHTMLEqual( w.render('company_widget1', [c1.pk, c2.pk], attrs={}), '<input type="text" name="company_widget1" value="%(c1pk)s,%(c2pk)s" />' % {'c1pk': c1.pk, 'c2pk': c2.pk} ) self.assertHTMLEqual( w.render('company_widget2', [c1.pk]), '<input type="text" name="company_widget2" value="%(c1pk)s" />' % {'c1pk': c1.pk} ) class RelatedFieldWidgetWrapperTests(SimpleTestCase): def test_no_can_add_related(self): rel = models.Individual._meta.get_field('parent').remote_field w = widgets.AdminRadioSelect() # Used to fail with a name error. w = widgets.RelatedFieldWidgetWrapper(w, rel, widget_admin_site) self.assertFalse(w.can_add_related) def test_select_multiple_widget_cant_change_delete_related(self): rel = models.Individual._meta.get_field('parent').remote_field widget = forms.SelectMultiple() wrapper = widgets.RelatedFieldWidgetWrapper( widget, rel, widget_admin_site, can_add_related=True, can_change_related=True, can_delete_related=True, ) self.assertTrue(wrapper.can_add_related) self.assertFalse(wrapper.can_change_related) self.assertFalse(wrapper.can_delete_related) def test_on_delete_cascade_rel_cant_delete_related(self): rel = models.Individual._meta.get_field('soulmate').remote_field widget = forms.Select() wrapper = widgets.RelatedFieldWidgetWrapper( widget, rel, widget_admin_site, can_add_related=True, can_change_related=True, can_delete_related=True, ) self.assertTrue(wrapper.can_add_related) self.assertTrue(wrapper.can_change_related) self.assertFalse(wrapper.can_delete_related) @override_settings(PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls') class DateTimePickerSeleniumFirefoxTests(SeleniumDataMixin, AdminSeleniumWebDriverTestCase): available_apps = ['admin_widgets'] + AdminSeleniumWebDriverTestCase.available_apps webdriver_class = 'selenium.webdriver.firefox.webdriver.WebDriver' def test_show_hide_date_time_picker_widgets(self): """ Ensure that pressing the ESC key closes the date and time picker widgets. Refs #17064. """ from selenium.webdriver.common.keys import Keys self.admin_login(username='super', password='secret', login_url='/') # Open a page that has a date and time picker widgets self.selenium.get('%s%s' % (self.live_server_url, reverse('admin:admin_widgets_member_add'))) # First, with the date picker widget --------------------------------- # Check that the date picker is hidden self.assertEqual( self.get_css_value('#calendarbox0', 'display'), 'none') # Click the calendar icon self.selenium.find_element_by_id('calendarlink0').click() # Check that the date picker is visible self.assertEqual( self.get_css_value('#calendarbox0', 'display'), 'block') # Press the ESC key self.selenium.find_element_by_tag_name('body').send_keys([Keys.ESCAPE]) # Check that the date picker is hidden again self.assertEqual( self.get_css_value('#calendarbox0', 'display'), 'none') # Then, with the time picker widget ---------------------------------- # Check that the time picker is hidden self.assertEqual( self.get_css_value('#clockbox0', 'display'), 'none') # Click the time icon self.selenium.find_element_by_id('clocklink0').click() # Check that the time picker is visible self.assertEqual( self.get_css_value('#clockbox0', 'display'), 'block') self.assertEqual( [ x.text for x in self.selenium.find_elements_by_xpath("//ul[@class='timelist']/li/a") ], ['Now', 'Midnight', '6 a.m.', 'Noon', '6 p.m.'] ) # Press the ESC key self.selenium.find_element_by_tag_name('body').send_keys([Keys.ESCAPE]) # Check that the time picker is hidden again self.assertEqual( self.get_css_value('#clockbox0', 'display'), 'none') def test_calendar_nonday_class(self): """ Ensure cells that are not days of the month have the `nonday` CSS class. Refs #4574. """ self.admin_login(username='super', password='secret', login_url='/') # Open a page that has a date and time picker widgets self.selenium.get('%s%s' % (self.live_server_url, reverse('admin:admin_widgets_member_add'))) # fill in the birth date. self.selenium.find_element_by_id('id_birthdate_0').send_keys('2013-06-01') # Click the calendar icon self.selenium.find_element_by_id('calendarlink0').click() # get all the tds within the calendar calendar0 = self.selenium.find_element_by_id('calendarin0') tds = calendar0.find_elements_by_tag_name('td') # make sure the first and last 6 cells have class nonday for td in tds[:6] + tds[-6:]: self.assertEqual(td.get_attribute('class'), 'nonday') def test_calendar_selected_class(self): """ Ensure cell for the day in the input has the `selected` CSS class. Refs #4574. """ self.admin_login(username='super', password='secret', login_url='/') # Open a page that has a date and time picker widgets self.selenium.get('%s%s' % (self.live_server_url, reverse('admin:admin_widgets_member_add'))) # fill in the birth date. self.selenium.find_element_by_id('id_birthdate_0').send_keys('2013-06-01') # Click the calendar icon self.selenium.find_element_by_id('calendarlink0').click() # get all the tds within the calendar calendar0 = self.selenium.find_element_by_id('calendarin0') tds = calendar0.find_elements_by_tag_name('td') # verify the selected cell selected = tds[6] self.assertEqual(selected.get_attribute('class'), 'selected') self.assertEqual(selected.text, '1') def test_calendar_no_selected_class(self): """ Ensure no cells are given the selected class when the field is empty. Refs #4574. """ self.admin_login(username='super', password='secret', login_url='/') # Open a page that has a date and time picker widgets self.selenium.get('%s%s' % (self.live_server_url, reverse('admin:admin_widgets_member_add'))) # Click the calendar icon self.selenium.find_element_by_id('calendarlink0').click() # get all the tds within the calendar calendar0 = self.selenium.find_element_by_id('calendarin0') tds = calendar0.find_elements_by_tag_name('td') # verify there are no cells with the selected class selected = [td for td in tds if td.get_attribute('class') == 'selected'] self.assertEqual(len(selected), 0) def test_calendar_show_date_from_input(self): """ Ensure that the calendar show the date from the input field for every locale supported by django. """ self.admin_login(username='super', password='secret', login_url='/') # Enter test data member = models.Member.objects.create(name='Bob', birthdate=datetime(1984, 5, 15), gender='M') # Get month names translations for every locales month_string = 'January February March April May June July August September October November December' path = os.path.join(os.path.dirname(import_module('django.contrib.admin').__file__), 'locale') for language_code, language_name in settings.LANGUAGES: try: catalog = gettext.translation('djangojs', path, [language_code]) except IOError: continue if month_string in catalog._catalog: month_names = catalog._catalog[month_string] else: month_names = month_string # Get the expected caption may_translation = month_names.split(' ')[4] expected_caption = '{0:s} {1:d}'.format(may_translation, 1984) # Test with every locale with override_settings(LANGUAGE_CODE=language_code, USE_L10N=True): # Open a page that has a date picker widget self.selenium.get('{}{}'.format(self.live_server_url, reverse('admin:admin_widgets_member_change', args=(member.pk,)))) # Click on the calendar icon self.selenium.find_element_by_id('calendarlink0').click() # Get the calendar caption calendar0 = self.selenium.find_element_by_id('calendarin0') caption = calendar0.find_element_by_tag_name('caption') # Make sure that the right month and year are displayed self.assertEqual(caption.text, expected_caption) class DateTimePickerSeleniumChromeTests(DateTimePickerSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.chrome.webdriver.WebDriver' class DateTimePickerSeleniumIETests(DateTimePickerSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.ie.webdriver.WebDriver' @skipIf(pytz is None, "this test requires pytz") @override_settings(TIME_ZONE='Asia/Singapore') @override_settings(PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls') class DateTimePickerShortcutsSeleniumFirefoxTests(SeleniumDataMixin, AdminSeleniumWebDriverTestCase): available_apps = ['admin_widgets'] + AdminSeleniumWebDriverTestCase.available_apps webdriver_class = 'selenium.webdriver.firefox.webdriver.WebDriver' def test_date_time_picker_shortcuts(self): """ Ensure that date/time/datetime picker shortcuts work in the current time zone. Refs #20663. This test case is fairly tricky, it relies on selenium still running the browser in the default time zone "America/Chicago" despite `override_settings` changing the time zone to "Asia/Singapore". """ self.admin_login(username='super', password='secret', login_url='/') error_margin = timedelta(seconds=10) # If we are neighbouring a DST, we add an hour of error margin. tz = pytz.timezone('America/Chicago') utc_now = datetime.now(pytz.utc) tz_yesterday = (utc_now - timedelta(days=1)).astimezone(tz).tzname() tz_tomorrow = (utc_now + timedelta(days=1)).astimezone(tz).tzname() if tz_yesterday != tz_tomorrow: error_margin += timedelta(hours=1) now = datetime.now() self.selenium.get('%s%s' % (self.live_server_url, reverse('admin:admin_widgets_member_add'))) self.selenium.find_element_by_id('id_name').send_keys('test') # Click on the "today" and "now" shortcuts. shortcuts = self.selenium.find_elements_by_css_selector( '.field-birthdate .datetimeshortcuts') for shortcut in shortcuts: shortcut.find_element_by_tag_name('a').click() # Check that there is a time zone mismatch warning. # Warning: This would effectively fail if the TIME_ZONE defined in the # settings has the same UTC offset as "Asia/Singapore" because the # mismatch warning would be rightfully missing from the page. self.selenium.find_elements_by_css_selector( '.field-birthdate .timezonewarning') # Submit the form. self.selenium.find_element_by_tag_name('form').submit() self.wait_page_loaded() # Make sure that "now" in javascript is within 10 seconds # from "now" on the server side. member = models.Member.objects.get(name='test') self.assertGreater(member.birthdate, now - error_margin) self.assertLess(member.birthdate, now + error_margin) class DateTimePickerShortcutsSeleniumChromeTests(DateTimePickerShortcutsSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.chrome.webdriver.WebDriver' class DateTimePickerShortcutsSeleniumIETests(DateTimePickerShortcutsSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.ie.webdriver.WebDriver' @override_settings(PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls') class HorizontalVerticalFilterSeleniumFirefoxTests(SeleniumDataMixin, AdminSeleniumWebDriverTestCase): available_apps = ['admin_widgets'] + AdminSeleniumWebDriverTestCase.available_apps webdriver_class = 'selenium.webdriver.firefox.webdriver.WebDriver' def setUp(self): super(HorizontalVerticalFilterSeleniumFirefoxTests, self).setUp() self.lisa = models.Student.objects.create(name='Lisa') self.john = models.Student.objects.create(name='John') self.bob = models.Student.objects.create(name='Bob') self.peter = models.Student.objects.create(name='Peter') self.jenny = models.Student.objects.create(name='Jenny') self.jason = models.Student.objects.create(name='Jason') self.cliff = models.Student.objects.create(name='Cliff') self.arthur = models.Student.objects.create(name='Arthur') self.school = models.School.objects.create(name='School of Awesome') def assertActiveButtons(self, mode, field_name, choose, remove, choose_all=None, remove_all=None): choose_link = '#id_%s_add_link' % field_name choose_all_link = '#id_%s_add_all_link' % field_name remove_link = '#id_%s_remove_link' % field_name remove_all_link = '#id_%s_remove_all_link' % field_name self.assertEqual(self.has_css_class(choose_link, 'active'), choose) self.assertEqual(self.has_css_class(remove_link, 'active'), remove) if mode == 'horizontal': self.assertEqual(self.has_css_class(choose_all_link, 'active'), choose_all) self.assertEqual(self.has_css_class(remove_all_link, 'active'), remove_all) def execute_basic_operations(self, mode, field_name): from_box = '#id_%s_from' % field_name to_box = '#id_%s_to' % field_name choose_link = 'id_%s_add_link' % field_name choose_all_link = 'id_%s_add_all_link' % field_name remove_link = 'id_%s_remove_link' % field_name remove_all_link = 'id_%s_remove_all_link' % field_name # Initial positions --------------------------------------------------- self.assertSelectOptions(from_box, [str(self.arthur.id), str(self.bob.id), str(self.cliff.id), str(self.jason.id), str(self.jenny.id), str(self.john.id)]) self.assertSelectOptions(to_box, [str(self.lisa.id), str(self.peter.id)]) self.assertActiveButtons(mode, field_name, False, False, True, True) # Click 'Choose all' -------------------------------------------------- if mode == 'horizontal': self.selenium.find_element_by_id(choose_all_link).click() elif mode == 'vertical': # There 's no 'Choose all' button in vertical mode, so individually # select all options and click 'Choose'. for option in self.selenium.find_elements_by_css_selector(from_box + ' > option'): option.click() self.selenium.find_element_by_id(choose_link).click() self.assertSelectOptions(from_box, []) self.assertSelectOptions(to_box, [str(self.lisa.id), str(self.peter.id), str(self.arthur.id), str(self.bob.id), str(self.cliff.id), str(self.jason.id), str(self.jenny.id), str(self.john.id)]) self.assertActiveButtons(mode, field_name, False, False, False, True) # Click 'Remove all' -------------------------------------------------- if mode == 'horizontal': self.selenium.find_element_by_id(remove_all_link).click() elif mode == 'vertical': # There 's no 'Remove all' button in vertical mode, so individually # select all options and click 'Remove'. for option in self.selenium.find_elements_by_css_selector(to_box + ' > option'): option.click() self.selenium.find_element_by_id(remove_link).click() self.assertSelectOptions(from_box, [str(self.lisa.id), str(self.peter.id), str(self.arthur.id), str(self.bob.id), str(self.cliff.id), str(self.jason.id), str(self.jenny.id), str(self.john.id)]) self.assertSelectOptions(to_box, []) self.assertActiveButtons(mode, field_name, False, False, True, False) # Choose some options ------------------------------------------------ from_lisa_select_option = self.get_select_option(from_box, str(self.lisa.id)) # Check the title attribute is there for tool tips: ticket #20821 self.assertEqual(from_lisa_select_option.get_attribute('title'), from_lisa_select_option.get_attribute('text')) from_lisa_select_option.click() self.get_select_option(from_box, str(self.jason.id)).click() self.get_select_option(from_box, str(self.bob.id)).click() self.get_select_option(from_box, str(self.john.id)).click() self.assertActiveButtons(mode, field_name, True, False, True, False) self.selenium.find_element_by_id(choose_link).click() self.assertActiveButtons(mode, field_name, False, False, True, True) self.assertSelectOptions(from_box, [str(self.peter.id), str(self.arthur.id), str(self.cliff.id), str(self.jenny.id)]) self.assertSelectOptions(to_box, [str(self.lisa.id), str(self.bob.id), str(self.jason.id), str(self.john.id)]) # Check the tooltip is still there after moving: ticket #20821 to_lisa_select_option = self.get_select_option(to_box, str(self.lisa.id)) self.assertEqual(to_lisa_select_option.get_attribute('title'), to_lisa_select_option.get_attribute('text')) # Remove some options ------------------------------------------------- self.get_select_option(to_box, str(self.lisa.id)).click() self.get_select_option(to_box, str(self.bob.id)).click() self.assertActiveButtons(mode, field_name, False, True, True, True) self.selenium.find_element_by_id(remove_link).click() self.assertActiveButtons(mode, field_name, False, False, True, True) self.assertSelectOptions(from_box, [str(self.peter.id), str(self.arthur.id), str(self.cliff.id), str(self.jenny.id), str(self.lisa.id), str(self.bob.id)]) self.assertSelectOptions(to_box, [str(self.jason.id), str(self.john.id)]) # Choose some more options -------------------------------------------- self.get_select_option(from_box, str(self.arthur.id)).click() self.get_select_option(from_box, str(self.cliff.id)).click() self.selenium.find_element_by_id(choose_link).click() self.assertSelectOptions(from_box, [str(self.peter.id), str(self.jenny.id), str(self.lisa.id), str(self.bob.id)]) self.assertSelectOptions(to_box, [str(self.jason.id), str(self.john.id), str(self.arthur.id), str(self.cliff.id)]) def test_basic(self): self.school.students = [self.lisa, self.peter] self.school.alumni = [self.lisa, self.peter] self.school.save() self.admin_login(username='super', password='secret', login_url='/') self.selenium.get('%s%s' % ( self.live_server_url, reverse('admin:admin_widgets_school_change', args=(self.school.id,)))) self.wait_page_loaded() self.execute_basic_operations('vertical', 'students') self.execute_basic_operations('horizontal', 'alumni') # Save and check that everything is properly stored in the database --- self.selenium.find_element_by_xpath('//input[@value="Save"]').click() self.wait_page_loaded() self.school = models.School.objects.get(id=self.school.id) # Reload from database self.assertEqual(list(self.school.students.all()), [self.arthur, self.cliff, self.jason, self.john]) self.assertEqual(list(self.school.alumni.all()), [self.arthur, self.cliff, self.jason, self.john]) def test_filter(self): """ Ensure that typing in the search box filters out options displayed in the 'from' box. """ from selenium.webdriver.common.keys import Keys self.school.students = [self.lisa, self.peter] self.school.alumni = [self.lisa, self.peter] self.school.save() self.admin_login(username='super', password='secret', login_url='/') self.selenium.get( '%s%s' % (self.live_server_url, reverse('admin:admin_widgets_school_change', args=(self.school.id,)))) for field_name in ['students', 'alumni']: from_box = '#id_%s_from' % field_name to_box = '#id_%s_to' % field_name choose_link = '#id_%s_add_link' % field_name remove_link = '#id_%s_remove_link' % field_name input = self.selenium.find_element_by_css_selector('#id_%s_input' % field_name) # Initial values self.assertSelectOptions(from_box, [str(self.arthur.id), str(self.bob.id), str(self.cliff.id), str(self.jason.id), str(self.jenny.id), str(self.john.id)]) # Typing in some characters filters out non-matching options input.send_keys('a') self.assertSelectOptions(from_box, [str(self.arthur.id), str(self.jason.id)]) input.send_keys('R') self.assertSelectOptions(from_box, [str(self.arthur.id)]) # Clearing the text box makes the other options reappear input.send_keys([Keys.BACK_SPACE]) self.assertSelectOptions(from_box, [str(self.arthur.id), str(self.jason.id)]) input.send_keys([Keys.BACK_SPACE]) self.assertSelectOptions(from_box, [str(self.arthur.id), str(self.bob.id), str(self.cliff.id), str(self.jason.id), str(self.jenny.id), str(self.john.id)]) # ----------------------------------------------------------------- # Check that choosing a filtered option sends it properly to the # 'to' box. input.send_keys('a') self.assertSelectOptions(from_box, [str(self.arthur.id), str(self.jason.id)]) self.get_select_option(from_box, str(self.jason.id)).click() self.selenium.find_element_by_css_selector(choose_link).click() self.assertSelectOptions(from_box, [str(self.arthur.id)]) self.assertSelectOptions(to_box, [str(self.lisa.id), str(self.peter.id), str(self.jason.id)]) self.get_select_option(to_box, str(self.lisa.id)).click() self.selenium.find_element_by_css_selector(remove_link).click() self.assertSelectOptions(from_box, [str(self.arthur.id), str(self.lisa.id)]) self.assertSelectOptions(to_box, [str(self.peter.id), str(self.jason.id)]) input.send_keys([Keys.BACK_SPACE]) # Clear text box self.assertSelectOptions(from_box, [str(self.arthur.id), str(self.bob.id), str(self.cliff.id), str(self.jenny.id), str(self.john.id), str(self.lisa.id)]) self.assertSelectOptions(to_box, [str(self.peter.id), str(self.jason.id)]) # ----------------------------------------------------------------- # Check that pressing enter on a filtered option sends it properly # to the 'to' box. self.get_select_option(to_box, str(self.jason.id)).click() self.selenium.find_element_by_css_selector(remove_link).click() input.send_keys('ja') self.assertSelectOptions(from_box, [str(self.jason.id)]) input.send_keys([Keys.ENTER]) self.assertSelectOptions(to_box, [str(self.peter.id), str(self.jason.id)]) input.send_keys([Keys.BACK_SPACE, Keys.BACK_SPACE]) # Save and check that everything is properly stored in the database --- self.selenium.find_element_by_xpath('//input[@value="Save"]').click() self.wait_page_loaded() self.school = models.School.objects.get(id=self.school.id) # Reload from database self.assertEqual(list(self.school.students.all()), [self.jason, self.peter]) self.assertEqual(list(self.school.alumni.all()), [self.jason, self.peter]) class HorizontalVerticalFilterSeleniumChromeTests(HorizontalVerticalFilterSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.chrome.webdriver.WebDriver' class HorizontalVerticalFilterSeleniumIETests(HorizontalVerticalFilterSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.ie.webdriver.WebDriver' @override_settings(PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls') class AdminRawIdWidgetSeleniumFirefoxTests(SeleniumDataMixin, AdminSeleniumWebDriverTestCase): available_apps = ['admin_widgets'] + AdminSeleniumWebDriverTestCase.available_apps webdriver_class = 'selenium.webdriver.firefox.webdriver.WebDriver' def setUp(self): super(AdminRawIdWidgetSeleniumFirefoxTests, self).setUp() models.Band.objects.create(id=42, name='Bogey Blues') models.Band.objects.create(id=98, name='Green Potatoes') def test_ForeignKey(self): self.admin_login(username='super', password='secret', login_url='/') self.selenium.get( '%s%s' % (self.live_server_url, reverse('admin:admin_widgets_event_add'))) main_window = self.selenium.current_window_handle # No value has been selected yet self.assertEqual( self.selenium.find_element_by_id('id_main_band').get_attribute('value'), '') # Open the popup window and click on a band self.selenium.find_element_by_id('lookup_id_main_band').click() self.selenium.switch_to.window('id_main_band') self.wait_page_loaded() link = self.selenium.find_element_by_link_text('Bogey Blues') self.assertIn('/band/42/', link.get_attribute('href')) link.click() # The field now contains the selected band's id self.selenium.switch_to.window(main_window) self.wait_for_value('#id_main_band', '42') # Reopen the popup window and click on another band self.selenium.find_element_by_id('lookup_id_main_band').click() self.selenium.switch_to.window('id_main_band') self.wait_page_loaded() link = self.selenium.find_element_by_link_text('Green Potatoes') self.assertIn('/band/98/', link.get_attribute('href')) link.click() # The field now contains the other selected band's id self.selenium.switch_to.window(main_window) self.wait_for_value('#id_main_band', '98') def test_many_to_many(self): self.admin_login(username='super', password='secret', login_url='/') self.selenium.get( '%s%s' % (self.live_server_url, reverse('admin:admin_widgets_event_add'))) main_window = self.selenium.current_window_handle # No value has been selected yet self.assertEqual( self.selenium.find_element_by_id('id_supporting_bands').get_attribute('value'), '') # Open the popup window and click on a band self.selenium.find_element_by_id('lookup_id_supporting_bands').click() self.selenium.switch_to.window('id_supporting_bands') self.wait_page_loaded() link = self.selenium.find_element_by_link_text('Bogey Blues') self.assertIn('/band/42/', link.get_attribute('href')) link.click() # The field now contains the selected band's id self.selenium.switch_to.window(main_window) self.wait_for_value('#id_supporting_bands', '42') # Reopen the popup window and click on another band self.selenium.find_element_by_id('lookup_id_supporting_bands').click() self.selenium.switch_to.window('id_supporting_bands') self.wait_page_loaded() link = self.selenium.find_element_by_link_text('Green Potatoes') self.assertIn('/band/98/', link.get_attribute('href')) link.click() # The field now contains the two selected bands' ids self.selenium.switch_to.window(main_window) self.wait_for_value('#id_supporting_bands', '42,98') class AdminRawIdWidgetSeleniumChromeTests(AdminRawIdWidgetSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.chrome.webdriver.WebDriver' class AdminRawIdWidgetSeleniumIETests(AdminRawIdWidgetSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.ie.webdriver.WebDriver' @override_settings(PASSWORD_HASHERS=['django.contrib.auth.hashers.SHA1PasswordHasher'], ROOT_URLCONF='admin_widgets.urls') class RelatedFieldWidgetSeleniumFirefoxTests(SeleniumDataMixin, AdminSeleniumWebDriverTestCase): available_apps = ['admin_widgets'] + AdminSeleniumWebDriverTestCase.available_apps webdriver_class = 'selenium.webdriver.firefox.webdriver.WebDriver' def test_ForeignKey_using_to_field(self): self.admin_login(username='super', password='secret', login_url='/') self.selenium.get('%s%s' % ( self.live_server_url, reverse('admin:admin_widgets_profile_add'))) main_window = self.selenium.current_window_handle # Click the Add User button to add new self.selenium.find_element_by_id('add_id_user').click() self.selenium.switch_to.window('id_user') self.wait_for('#id_password') password_field = self.selenium.find_element_by_id('id_password') password_field.send_keys('password') username_field = self.selenium.find_element_by_id('id_username') username_value = 'newuser' username_field.send_keys(username_value) save_button_css_selector = '.submit-row > input[type=submit]' self.selenium.find_element_by_css_selector(save_button_css_selector).click() self.selenium.switch_to.window(main_window) # The field now contains the new user self.wait_for('#id_user option[value="newuser"]') # Click the Change User button to change it self.selenium.find_element_by_id('change_id_user').click() self.selenium.switch_to_window('id_user') self.wait_page_loaded() username_field = self.selenium.find_element_by_id('id_username') username_value = 'changednewuser' username_field.clear() username_field.send_keys(username_value) save_button_css_selector = '.submit-row > input[type=submit]' self.selenium.find_element_by_css_selector(save_button_css_selector).click() self.selenium.switch_to_window(main_window) # Wait up to 2 seconds for the new option to show up after clicking save in the popup. self.selenium.implicitly_wait(2) self.selenium.find_element_by_css_selector('#id_user option[value=changednewuser]') self.selenium.implicitly_wait(0) # Go ahead and submit the form to make sure it works self.selenium.find_element_by_css_selector(save_button_css_selector).click() self.wait_for_text('li.success', 'The profile "changednewuser" was added successfully.') profiles = models.Profile.objects.all() self.assertEqual(len(profiles), 1) self.assertEqual(profiles[0].user.username, username_value) class RelatedFieldWidgetSeleniumChromeTests(RelatedFieldWidgetSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.chrome.webdriver.WebDriver' class RelatedFieldWidgetSeleniumIETests(RelatedFieldWidgetSeleniumFirefoxTests): webdriver_class = 'selenium.webdriver.ie.webdriver.WebDriver'
bsd-3-clause
stefelisabeth/dplace
dplace_app/api_views.py
1
21716
import json import re import datetime from itertools import groupby import logging from django.db import connection from django.db.models import Prefetch, Q, Count from django.shortcuts import get_object_or_404 from django.http import Http404 from rest_framework import viewsets from rest_framework.pagination import PageNumberPagination from rest_framework.decorators import api_view, permission_classes, renderer_classes from rest_framework.permissions import AllowAny from rest_framework.views import Response from rest_framework.renderers import JSONRenderer from dplace_app.filters import GeographicRegionFilter from dplace_app.renderers import DPLACECSVRenderer from dplace_app import serializers from dplace_app import models from dplace_app.tree import update_newick log = logging.getLogger('profile') class VariableViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.VariableSerializer filter_fields = ('label', 'name', 'index_categories', 'niche_categories', 'source') queryset = models.Variable.objects\ .prefetch_related('index_categories', 'niche_categories') # Override retrieve to use the detail serializer, which includes categories def retrieve(self, request, *args, **kwargs): self.object = self.get_object() serializer = serializers.VariableDetailSerializer(self.object) return Response(serializer.data) class CategoryViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.CategorySerializer filter_fields = ('name', 'type', 'index_variables', 'niche_variables') queryset = models.Category.objects.all() # Override retrieve to use the detail serializer, which includes variables def retrieve(self, request, *args, **kwargs): self.object = self.get_object() serializer = serializers.CategoryDetailSerializer(self.object) return Response(serializer.data) class CodeDescriptionViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.CodeDescriptionSerializer filter_fields = ('variable',) queryset = models.CodeDescription.objects.all() class ValueViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.ValueSerializer filter_fields = ('variable', 'coded_value', 'code', 'society',) # Avoid additional database trips by select_related for foreign keys queryset = models.Value.objects.filter(variable__type='cultural')\ .select_related('variable', 'code', 'source').all() class SocietyViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.SocietySerializer queryset = models.Society.objects.all().select_related( 'source', 'language__iso_code', 'language__family') lookup_field = 'ext_id' def detail(self, request, society_id): # block spider attacks if len(request.GET) > 0 and request.path.startswith('/society'): raise Http404 society = get_object_or_404(models.Society, ext_id=society_id) # gets the society's location for inset map location = {} if society.location: location = { 'lat': society.location['coordinates'][1], 'lng': society.location['coordinates'][0] } # gets other societies in database with the same xd_id xd_id = models.Society.objects.filter( xd_id=society.xd_id).exclude(ext_id=society_id) if society.hraf_link and '(' in society.hraf_link: hraf_link = society.hraf_link.split('(')[len(society.hraf_link.split('('))-1] else: hraf_link = '' environmentals = society.get_environmental_data() cultural_traits = society.get_cultural_trait_data() references = society.get_data_references() language_classification = None if society.language: # just glottolog at the moment language_classification = models.LanguageFamily.objects\ .filter(name=society.language.family.name) return Response( { 'society': society, 'hraf_link': hraf_link[0:len(hraf_link)-1], 'xd_id': xd_id, 'location': location, 'language_classification': language_classification, 'environmentals': dict(environmentals), 'cultural_traits': dict(cultural_traits), 'references': references }, template_name='society.html' ) class ISOCodeViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.ISOCodeSerializer filter_fields = ('iso_code',) queryset = models.ISOCode.objects.all() class LargeResultsSetPagination(PageNumberPagination): page_size = 1000 page_size_query_param = 'page_size' max_page_size = 1000 class LanguageViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.LanguageSerializerWithSocieties filter_fields = ('name', 'iso_code', 'societies', 'family',) queryset = models.Language.objects.all()\ .select_related('family', 'iso_code')\ .prefetch_related(Prefetch( 'societies', queryset=models.Society.objects.exclude(value__isnull=True) )) pagination_class = LargeResultsSetPagination class LanguageFamilyViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.LanguageFamilySerializer filter_fields = ('name',) queryset = models.LanguageFamily.objects.all()\ .annotate(language_count=Count('language__societies'))\ .order_by('name') pagination_class = LargeResultsSetPagination class TreeResultsSetPagination(PageNumberPagination): """ Since trees may have *many* languages, which are serialized as well, we limit the page size to just 1. """ page_size = 1 page_size_query_param = 'page_size' max_page_size = 10 class LanguageTreeViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.LanguageTreeSerializer filter_fields = ('name',) queryset = models.LanguageTree.objects.all() pagination_class = TreeResultsSetPagination class LanguageTreeLabelsViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.LanguageTreeLabelsSerializer filter_fields = ('label',) queryset = models.LanguageTreeLabels.objects.all() pagination_class = LargeResultsSetPagination class SourceViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.SourceSerializer filter_fields = ('author', 'name') queryset = models.Source.objects.all() def get_query_from_json(request): query_string = request.query_params.get('query') if query_string is None: raise Http404('missing query parameter') try: query_dict = json.loads(query_string) except ValueError: raise Http404('malformed query parameter') if not isinstance(query_dict, dict): raise Http404('malformed query parameter') return query_dict def result_set_from_query_dict(query_dict): from time import time _s = time() log.info('enter result_set_from_query_dict') result_set = serializers.SocietyResultSet() sql_joins, sql_where = [], [] def id_array(l): return '(%s)' % ','.join('%s' % int(i) for i in l) if 'l' in query_dict: sql_joins.append(('language', 'l', 'l.id = s.language_id')) sql_where.append('l.id IN ' + id_array(query_dict['l'])) for lang in models.Language.objects.filter(id__in=query_dict['l']): result_set.languages.add(lang) if 'c' in query_dict: variables = { v.id: v for v in models.Variable.objects .filter(id__in=[int(x.split('-')[0]) for x in query_dict['c']]) .prefetch_related(Prefetch( 'codes', queryset=models.CodeDescription.objects .filter(id__in=[int(x.split('-')[1]) for x in query_dict['c'] if len(x.split('-')) == 2]))) } for variable, codes in groupby( sorted(query_dict['c'], key=lambda c: int(c.split('-')[0])), key=lambda x: int(str(x).split('-')[0]) ): variable = variables[variable] codes = [{ 'id': None if (len(c.split('-')) > 2 or len(c.split('-')) == 1) else int(c.split('-')[1]), 'min': None if len(c.split('-')) < 3 else float(c.split('-')[1]), 'max': None if len(c.split('-')) < 3 else float(c.split('-')[2]) } for c in list(codes)] alias = 'cv%s' % variable.id sql_joins.append(( "value", alias, "{0}.society_id = s.id AND {0}.variable_id = {1}".format(alias, variable.id) )) if variable.data_type and variable.data_type == 'Continuous': include_NA = not all((c['min'] is not None) for c in codes) ors = [ "({0}.coded_value_float >= %(min)f AND {0}.coded_value_float <= %(max)f)".format(alias) % c for c in codes if ('min' in c and c['min'] is not None)] if include_NA: ors.append("%s.coded_value = 'NA'" % alias) sql_where.append("(%s)" % ' OR '.join(ors)) if not include_NA: sql_where.append("{0}.coded_value != 'NA'".format(alias)) else: assert all('id' in c for c in codes) sql_where.append("{0}.code_id IN %s".format(alias) % id_array([x['id'] for x in codes])) result_set.variable_descriptions.add(serializers.VariableCode(variable.codes, variable)) if 'e' in query_dict: # There can be multiple filters, so we must aggregate the results. for varid, criteria in groupby( sorted(query_dict['e'], key=lambda c: c[0]), key=lambda x: x[0] ): alias = 'ev%s' % varid sql_joins.append(( "value", alias, "{0}.society_id = s.id AND {0}.variable_id = {1}".format(alias, int(varid)))) for varid, operator, params in criteria: params = map(float, params) if operator == 'inrange': sql_where.append("{0}.coded_value_float >= {1:f} AND {0}.coded_value_float <= {2:f}".format(alias, params[0], params[1])) elif operator == 'outrange': sql_where.append("{0}.coded_value_float >= {1:f} AND {0}.coded_value_float <= {2:f}".format(alias, params[1], params[0])) elif operator == 'gt': sql_where.append("{0}.coded_value_float >= {1:f}".format(alias, params[0])) elif operator == 'lt': sql_where.append("{0}.coded_value_float <= {1:f}".format(alias, params[0])) for variable in models.Variable.objects.filter(id__in=[x[0] for x in query_dict['e']]): result_set.environmental_variables.add(variable) if 'p' in query_dict: sql_joins.append(('geographicregion', 'r', 'r.id = s.region_id')) sql_where.append('r.id IN %s' % id_array(query_dict['p'])) for region in models.GeographicRegion.objects.filter(id__in=query_dict['p']): result_set.geographic_regions.add(region) if sql_where: cursor = connection.cursor() sql = "select distinct s.id from dplace_app_society as s %s where %s" % ( ' '.join('join dplace_app_%s as %s on %s' % t for t in sql_joins), ' AND '.join(sql_where)) cursor.execute(sql) soc_ids = [r[0] for r in cursor.fetchall()] else: soc_ids = [] soc_query = models.Society.objects.filter(id__in=soc_ids)\ .select_related('source', 'language__family', 'language__iso_code', 'region') if result_set.geographic_regions: soc_query = soc_query.select_related('region') if result_set.variable_descriptions: soc_query = soc_query.prefetch_related(Prefetch( 'value_set', to_attr='selected_cvalues', queryset=models.Value.objects # FIXME: this selects possibly too many values, in case there are multiple # values for the same variable, not all of them matching the criteria. .filter(variable_id__in=[v.variable.id for v in result_set.variable_descriptions]) .select_related('code') .prefetch_related('references'))) if result_set.environmental_variables: soc_query = soc_query.prefetch_related(Prefetch( 'value_set', to_attr='selected_evalues', queryset=models.Value.objects .filter(variable_id__in=[v.id for v in result_set.environmental_variables]) .prefetch_related('references'))) for i, soc in enumerate(soc_query): soc_result = serializers.SocietyResult(soc) if result_set.variable_descriptions: for cval in soc.selected_cvalues: soc_result.variable_coded_values.add(cval) if result_set.environmental_variables: for eval in soc.selected_evalues: soc_result.environmental_values.add(eval) result_set.societies.add(soc_result) log.info('mid 1: %s' % (time() - _s,)) # Filter the results to those that matched all criteria #result_set.finalize(criteria) log.info('mid 2: %s' % (time() - _s,)) return result_set @api_view(['GET']) @permission_classes((AllowAny,)) def trees_from_societies(request): language_trees = [] for k, v in request.query_params.lists(): soc_ids = v labels = models.LanguageTreeLabels.objects.filter(societies__id__in=soc_ids).all() global_tree = None global_newick = [] global_isolates = [] for t in models.LanguageTree.objects\ .filter(taxa__societies__id__in=soc_ids)\ .prefetch_related( 'taxa__languagetreelabelssequence_set__labels', 'taxa__languagetreelabelssequence_set__society', )\ .distinct(): if 'global' in t.name: global_tree = t # TODO ask @Bibiko once the isolates are in the db under global.tree as string: isol1,isol2,isol3,... # global_isolates.extend(t.newick_string.split(',')) global_isolates.extend(['alse1251','amas1236','bana1292','calu1239','chim1301','chit1248','chon1248','coah1252','coos1249','furr1244','gaga1251','guai1237','guat1253','hadz1240','high1242','kara1289','karo1304','klam1254','kute1249','lara1258','mull1237','natc1249','nort2938','paez1247','pume1238','pura1257','pure1242','sali1253','sand1273','seri1257','shom1245','sius1254','sout1439','take1257','ticu1245','timu1245','tiwi1244','toll1241','trum1247','uruu1244','wara1303','wash1253','yama1264','yuch1247','zuni1245']) else: if update_newick(t, labels): language_trees.append(t) if 'glotto' in t.name: #remove last ; in order to be able to join the trees global_newick.append(t.newick_string[:-1]) if global_tree: langs_in_tree = [str(l.label) for l in labels] #add isolates if present in current selection [global_newick.append('(' + isolate + ':1)') for isolate in global_isolates if isolate in langs_in_tree] #join all pruned glottolog trees into the global one global_tree.newick_string = '(' + ','.join(global_newick) + ');' language_trees.append(global_tree) return Response(serializers.LanguageTreeSerializer(language_trees, many=True).data) @api_view(['GET']) @permission_classes((AllowAny,)) def find_societies(request): """ View to find the societies that match an input request. Currently expects { language_filters: [{language_ids: [1,2,3]}], variable_codes: [4,5,6...], environmental_filters: [{id: 1, operator: 'gt', params: [0.0]}, {id:3, operator 'inrange', params: [10.0,20.0] }] } Returns serialized collection of SocietyResult objects """ from time import time from django.db import connection s = time() log.info('%s find_societies 1: %s queries' % (time() - s, len(connection.queries))) query = {} if 'name' in request.query_params: result_set = serializers.SocietyResultSet() q = request.query_params['name'] if q: soc = models.Society.objects.filter( Q(name__icontains=q) | Q(alternate_names__unaccent__icontains=q)) for s in soc: if s.value_set.count(): result_set.societies.add(serializers.SocietyResult(s)) return Response(serializers.SocietyResultSetSerializer(result_set).data) for k, v in request.query_params.lists(): if str(k) == 'c': query[k] = v else: query[k] = [json.loads(vv) for vv in v] result_set = result_set_from_query_dict(query) log.info('%s find_societies 2: %s queries' % (time() - s, len(connection.queries))) d = serializers.SocietyResultSetSerializer(result_set).data log.info('%s find_societies 3: %s queries' % (time() - s, len(connection.queries))) for i, q in enumerate( sorted(connection.queries, key=lambda q: q['time'], reverse=True)): if 10 < i < 20: # pragma: no cover log.info('%s for %s' % (q['time'], q['sql'][:500])) return Response(d) @api_view(['GET']) @permission_classes((AllowAny,)) def get_categories(request): """ Filters categories for sources, as some categories are empty for some sources """ query_dict = get_query_from_json(request) categories = models.Category.objects.filter(type='cultural') source_categories = [] if 'source' in query_dict: source = models.Source.objects.filter(id=query_dict['source']) variables = models.Variable.objects.filter(source=source) for c in categories: if variables.filter(index_categories=c.id): source_categories.append(c) return Response( serializers.CategorySerializer(source_categories, many=True).data) return Response(serializers.CategorySerializer(categories, many=True).data) @api_view(['GET']) @permission_classes((AllowAny,)) def get_dataset_sources(request): return Response( serializers.SourceSerializer( models.Source.objects.all().exclude(name=""), many=True).data) class GeographicRegionViewSet(viewsets.ReadOnlyModelViewSet): serializer_class = serializers.GeographicRegionSerializer model = models.GeographicRegion filter_class = GeographicRegionFilter queryset = models.GeographicRegion.objects.all() @api_view(['GET']) @permission_classes((AllowAny,)) @renderer_classes((JSONRenderer,)) def get_min_and_max(request): res = {} varid = get_query_from_json(request).get('environmental_id') if varid: values = [ v.coded_value_float for v in models.Value.objects.filter(variable__id=varid) if v.coded_value_float is not None] vmin = min(values) if values else 0.0 vmax = max(values) if values else 0.0 res = {'min': format(vmin, '.4f'), 'max': format(vmax, '.4f')} return Response(res) @api_view(['GET']) @permission_classes((AllowAny,)) @renderer_classes((JSONRenderer,)) def bin_cont_data(request): # MAKE THIS GENERIC bf_id = get_query_from_json(request).get('bf_id') bins = [] if bf_id: values = models.Value.objects.filter(variable__id=bf_id) min_value = None max_value = 0.0 missing_data_option = False for v in values: if re.search('[a-zA-Z]', v.coded_value): if not missing_data_option: bins.append({ 'code': v.coded_value, 'description': v.code.description, 'variable': bf_id, }) missing_data_option = True continue else: v.coded_value = v.coded_value.replace(',', '') if min_value is None: min_value = float(v.coded_value) elif float(v.coded_value) < min_value: min_value = float(v.coded_value) elif float(v.coded_value) > max_value: max_value = float(v.coded_value) min_value = min_value or 0.0 # This is the case when there are no values! data_range = max_value - min_value bin_size = data_range / 5 min_bin = min_value for x in range(0, 5): min = min_bin max = min_bin + bin_size bins.append({ 'code': x, 'description': str(min) + ' - ' + str(max), 'min': min_bin, 'max': min_bin + bin_size, 'variable': bf_id, }) min_bin = min_bin + bin_size + 1 return Response(bins) @api_view(['GET']) @permission_classes((AllowAny,)) @renderer_classes((DPLACECSVRenderer,)) def csv_download(request): query_dict = get_query_from_json(request) result_set = result_set_from_query_dict(query_dict) response = Response(serializers.SocietyResultSetSerializer(result_set).data) filename = "dplace-societies-%s.csv" % datetime.datetime.now().strftime("%Y-%m-%d") response['Content-Disposition'] = 'attachment; filename="%s"' % filename return response
mit
vjmac15/Lyilis
lib/youtube_dl/extractor/aparat.py
15
2195
# coding: utf-8 from __future__ import unicode_literals from .common import InfoExtractor from ..utils import ( int_or_none, mimetype2ext, ) class AparatIE(InfoExtractor): _VALID_URL = r'https?://(?:www\.)?aparat\.com/(?:v/|video/video/embed/videohash/)(?P<id>[a-zA-Z0-9]+)' _TEST = { 'url': 'http://www.aparat.com/v/wP8On', 'md5': '131aca2e14fe7c4dcb3c4877ba300c89', 'info_dict': { 'id': 'wP8On', 'ext': 'mp4', 'title': 'تیم گلکسی 11 - زومیت', 'age_limit': 0, }, # 'skip': 'Extremely unreliable', } def _real_extract(self, url): video_id = self._match_id(url) # Note: There is an easier-to-parse configuration at # http://www.aparat.com/video/video/config/videohash/%video_id # but the URL in there does not work webpage = self._download_webpage( 'http://www.aparat.com/video/video/embed/vt/frame/showvideo/yes/videohash/' + video_id, video_id) title = self._search_regex(r'\s+title:\s*"([^"]+)"', webpage, 'title') file_list = self._parse_json( self._search_regex( r'fileList\s*=\s*JSON\.parse\(\'([^\']+)\'\)', webpage, 'file list'), video_id) formats = [] for item in file_list[0]: file_url = item.get('file') if not file_url: continue ext = mimetype2ext(item.get('type')) label = item.get('label') formats.append({ 'url': file_url, 'ext': ext, 'format_id': label or ext, 'height': int_or_none(self._search_regex( r'(\d+)[pP]', label or '', 'height', default=None)), }) self._sort_formats(formats) thumbnail = self._search_regex( r'image:\s*"([^"]+)"', webpage, 'thumbnail', fatal=False) return { 'id': video_id, 'title': title, 'thumbnail': thumbnail, 'age_limit': self._family_friendly_search(webpage), 'formats': formats, }
gpl-3.0
denys-duchier/django
tests/file_uploads/tests.py
45
23742
import base64 import hashlib import os import shutil import sys import tempfile as sys_tempfile import unittest from io import BytesIO, StringIO from urllib.parse import quote from django.core.files import temp as tempfile from django.core.files.uploadedfile import SimpleUploadedFile from django.http.multipartparser import MultiPartParser, parse_header from django.test import SimpleTestCase, TestCase, client, override_settings from django.utils.encoding import force_bytes from . import uploadhandler from .models import FileModel UNICODE_FILENAME = 'test-0123456789_中文_Orléans.jpg' MEDIA_ROOT = sys_tempfile.mkdtemp() UPLOAD_TO = os.path.join(MEDIA_ROOT, 'test_upload') @override_settings(MEDIA_ROOT=MEDIA_ROOT, ROOT_URLCONF='file_uploads.urls', MIDDLEWARE=[]) class FileUploadTests(TestCase): @classmethod def setUpClass(cls): super().setUpClass() if not os.path.isdir(MEDIA_ROOT): os.makedirs(MEDIA_ROOT) @classmethod def tearDownClass(cls): shutil.rmtree(MEDIA_ROOT) super().tearDownClass() def test_simple_upload(self): with open(__file__, 'rb') as fp: post_data = { 'name': 'Ringo', 'file_field': fp, } response = self.client.post('/upload/', post_data) self.assertEqual(response.status_code, 200) def test_large_upload(self): file = tempfile.NamedTemporaryFile with file(suffix=".file1") as file1, file(suffix=".file2") as file2: file1.write(b'a' * (2 ** 21)) file1.seek(0) file2.write(b'a' * (10 * 2 ** 20)) file2.seek(0) post_data = { 'name': 'Ringo', 'file_field1': file1, 'file_field2': file2, } for key in list(post_data): try: post_data[key + '_hash'] = hashlib.sha1(post_data[key].read()).hexdigest() post_data[key].seek(0) except AttributeError: post_data[key + '_hash'] = hashlib.sha1(force_bytes(post_data[key])).hexdigest() response = self.client.post('/verify/', post_data) self.assertEqual(response.status_code, 200) def _test_base64_upload(self, content, encode=base64.b64encode): payload = client.FakePayload("\r\n".join([ '--' + client.BOUNDARY, 'Content-Disposition: form-data; name="file"; filename="test.txt"', 'Content-Type: application/octet-stream', 'Content-Transfer-Encoding: base64', ''])) payload.write(b"\r\n" + encode(force_bytes(content)) + b"\r\n") payload.write('--' + client.BOUNDARY + '--\r\n') r = { 'CONTENT_LENGTH': len(payload), 'CONTENT_TYPE': client.MULTIPART_CONTENT, 'PATH_INFO': "/echo_content/", 'REQUEST_METHOD': 'POST', 'wsgi.input': payload, } response = self.client.request(**r) self.assertEqual(response.json()['file'], content) def test_base64_upload(self): self._test_base64_upload("This data will be transmitted base64-encoded.") def test_big_base64_upload(self): self._test_base64_upload("Big data" * 68000) # > 512Kb def test_big_base64_newlines_upload(self): self._test_base64_upload("Big data" * 68000, encode=base64.encodebytes) def test_unicode_file_name(self): with sys_tempfile.TemporaryDirectory() as temp_dir: # This file contains Chinese symbols and an accented char in the name. with open(os.path.join(temp_dir, UNICODE_FILENAME), 'w+b') as file1: file1.write(b'b' * (2 ** 10)) file1.seek(0) response = self.client.post('/unicode_name/', {'file_unicode': file1}) self.assertEqual(response.status_code, 200) def test_unicode_file_name_rfc2231(self): """ Test receiving file upload when filename is encoded with RFC2231 (#22971). """ payload = client.FakePayload() payload.write('\r\n'.join([ '--' + client.BOUNDARY, 'Content-Disposition: form-data; name="file_unicode"; filename*=UTF-8\'\'%s' % quote(UNICODE_FILENAME), 'Content-Type: application/octet-stream', '', 'You got pwnd.\r\n', '\r\n--' + client.BOUNDARY + '--\r\n' ])) r = { 'CONTENT_LENGTH': len(payload), 'CONTENT_TYPE': client.MULTIPART_CONTENT, 'PATH_INFO': "/unicode_name/", 'REQUEST_METHOD': 'POST', 'wsgi.input': payload, } response = self.client.request(**r) self.assertEqual(response.status_code, 200) def test_unicode_name_rfc2231(self): """ Test receiving file upload when filename is encoded with RFC2231 (#22971). """ payload = client.FakePayload() payload.write( '\r\n'.join([ '--' + client.BOUNDARY, 'Content-Disposition: form-data; name*=UTF-8\'\'file_unicode; filename*=UTF-8\'\'%s' % quote( UNICODE_FILENAME ), 'Content-Type: application/octet-stream', '', 'You got pwnd.\r\n', '\r\n--' + client.BOUNDARY + '--\r\n' ]) ) r = { 'CONTENT_LENGTH': len(payload), 'CONTENT_TYPE': client.MULTIPART_CONTENT, 'PATH_INFO': "/unicode_name/", 'REQUEST_METHOD': 'POST', 'wsgi.input': payload, } response = self.client.request(**r) self.assertEqual(response.status_code, 200) def test_blank_filenames(self): """ Receiving file upload when filename is blank (before and after sanitization) should be okay. """ # The second value is normalized to an empty name by # MultiPartParser.IE_sanitize() filenames = ['', 'C:\\Windows\\'] payload = client.FakePayload() for i, name in enumerate(filenames): payload.write('\r\n'.join([ '--' + client.BOUNDARY, 'Content-Disposition: form-data; name="file%s"; filename="%s"' % (i, name), 'Content-Type: application/octet-stream', '', 'You got pwnd.\r\n' ])) payload.write('\r\n--' + client.BOUNDARY + '--\r\n') r = { 'CONTENT_LENGTH': len(payload), 'CONTENT_TYPE': client.MULTIPART_CONTENT, 'PATH_INFO': '/echo/', 'REQUEST_METHOD': 'POST', 'wsgi.input': payload, } response = self.client.request(**r) self.assertEqual(response.status_code, 200) # Empty filenames should be ignored received = response.json() for i, name in enumerate(filenames): self.assertIsNone(received.get('file%s' % i)) def test_dangerous_file_names(self): """Uploaded file names should be sanitized before ever reaching the view.""" # This test simulates possible directory traversal attacks by a # malicious uploader We have to do some monkeybusiness here to construct # a malicious payload with an invalid file name (containing os.sep or # os.pardir). This similar to what an attacker would need to do when # trying such an attack. scary_file_names = [ "/tmp/hax0rd.txt", # Absolute path, *nix-style. "C:\\Windows\\hax0rd.txt", # Absolute path, win-style. "C:/Windows/hax0rd.txt", # Absolute path, broken-style. "\\tmp\\hax0rd.txt", # Absolute path, broken in a different way. "/tmp\\hax0rd.txt", # Absolute path, broken by mixing. "subdir/hax0rd.txt", # Descendant path, *nix-style. "subdir\\hax0rd.txt", # Descendant path, win-style. "sub/dir\\hax0rd.txt", # Descendant path, mixed. "../../hax0rd.txt", # Relative path, *nix-style. "..\\..\\hax0rd.txt", # Relative path, win-style. "../..\\hax0rd.txt" # Relative path, mixed. ] payload = client.FakePayload() for i, name in enumerate(scary_file_names): payload.write('\r\n'.join([ '--' + client.BOUNDARY, 'Content-Disposition: form-data; name="file%s"; filename="%s"' % (i, name), 'Content-Type: application/octet-stream', '', 'You got pwnd.\r\n' ])) payload.write('\r\n--' + client.BOUNDARY + '--\r\n') r = { 'CONTENT_LENGTH': len(payload), 'CONTENT_TYPE': client.MULTIPART_CONTENT, 'PATH_INFO': "/echo/", 'REQUEST_METHOD': 'POST', 'wsgi.input': payload, } response = self.client.request(**r) # The filenames should have been sanitized by the time it got to the view. received = response.json() for i, name in enumerate(scary_file_names): got = received["file%s" % i] self.assertEqual(got, "hax0rd.txt") def test_filename_overflow(self): """File names over 256 characters (dangerous on some platforms) get fixed up.""" long_str = 'f' * 300 cases = [ # field name, filename, expected ('long_filename', '%s.txt' % long_str, '%s.txt' % long_str[:251]), ('long_extension', 'foo.%s' % long_str, '.%s' % long_str[:254]), ('no_extension', long_str, long_str[:255]), ('no_filename', '.%s' % long_str, '.%s' % long_str[:254]), ('long_everything', '%s.%s' % (long_str, long_str), '.%s' % long_str[:254]), ] payload = client.FakePayload() for name, filename, _ in cases: payload.write("\r\n".join([ '--' + client.BOUNDARY, 'Content-Disposition: form-data; name="{}"; filename="{}"', 'Content-Type: application/octet-stream', '', 'Oops.', '' ]).format(name, filename)) payload.write('\r\n--' + client.BOUNDARY + '--\r\n') r = { 'CONTENT_LENGTH': len(payload), 'CONTENT_TYPE': client.MULTIPART_CONTENT, 'PATH_INFO': "/echo/", 'REQUEST_METHOD': 'POST', 'wsgi.input': payload, } response = self.client.request(**r) result = response.json() for name, _, expected in cases: got = result[name] self.assertEqual(expected, got, 'Mismatch for {}'.format(name)) self.assertLess(len(got), 256, "Got a long file name (%s characters)." % len(got)) def test_file_content(self): file = tempfile.NamedTemporaryFile with file(suffix=".ctype_extra") as no_content_type, file(suffix=".ctype_extra") as simple_file: no_content_type.write(b'no content') no_content_type.seek(0) simple_file.write(b'text content') simple_file.seek(0) simple_file.content_type = 'text/plain' string_io = StringIO('string content') bytes_io = BytesIO(b'binary content') response = self.client.post('/echo_content/', { 'no_content_type': no_content_type, 'simple_file': simple_file, 'string': string_io, 'binary': bytes_io, }) received = response.json() self.assertEqual(received['no_content_type'], 'no content') self.assertEqual(received['simple_file'], 'text content') self.assertEqual(received['string'], 'string content') self.assertEqual(received['binary'], 'binary content') def test_content_type_extra(self): """Uploaded files may have content type parameters available.""" file = tempfile.NamedTemporaryFile with file(suffix=".ctype_extra") as no_content_type, file(suffix=".ctype_extra") as simple_file: no_content_type.write(b'something') no_content_type.seek(0) simple_file.write(b'something') simple_file.seek(0) simple_file.content_type = 'text/plain; test-key=test_value' response = self.client.post('/echo_content_type_extra/', { 'no_content_type': no_content_type, 'simple_file': simple_file, }) received = response.json() self.assertEqual(received['no_content_type'], {}) self.assertEqual(received['simple_file'], {'test-key': 'test_value'}) def test_truncated_multipart_handled_gracefully(self): """ If passed an incomplete multipart message, MultiPartParser does not attempt to read beyond the end of the stream, and simply will handle the part that can be parsed gracefully. """ payload_str = "\r\n".join([ '--' + client.BOUNDARY, 'Content-Disposition: form-data; name="file"; filename="foo.txt"', 'Content-Type: application/octet-stream', '', 'file contents' '--' + client.BOUNDARY + '--', '', ]) payload = client.FakePayload(payload_str[:-10]) r = { 'CONTENT_LENGTH': len(payload), 'CONTENT_TYPE': client.MULTIPART_CONTENT, 'PATH_INFO': '/echo/', 'REQUEST_METHOD': 'POST', 'wsgi.input': payload, } self.assertEqual(self.client.request(**r).json(), {}) def test_empty_multipart_handled_gracefully(self): """ If passed an empty multipart message, MultiPartParser will return an empty QueryDict. """ r = { 'CONTENT_LENGTH': 0, 'CONTENT_TYPE': client.MULTIPART_CONTENT, 'PATH_INFO': '/echo/', 'REQUEST_METHOD': 'POST', 'wsgi.input': client.FakePayload(b''), } self.assertEqual(self.client.request(**r).json(), {}) def test_custom_upload_handler(self): file = tempfile.NamedTemporaryFile with file() as smallfile, file() as bigfile: # A small file (under the 5M quota) smallfile.write(b'a' * (2 ** 21)) smallfile.seek(0) # A big file (over the quota) bigfile.write(b'a' * (10 * 2 ** 20)) bigfile.seek(0) # Small file posting should work. self.assertIn('f', self.client.post('/quota/', {'f': smallfile}).json()) # Large files don't go through. self.assertNotIn('f', self.client.post("/quota/", {'f': bigfile}).json()) def test_broken_custom_upload_handler(self): with tempfile.NamedTemporaryFile() as file: file.write(b'a' * (2 ** 21)) file.seek(0) # AttributeError: You cannot alter upload handlers after the upload has been processed. with self.assertRaises(AttributeError): self.client.post('/quota/broken/', {'f': file}) def test_fileupload_getlist(self): file = tempfile.NamedTemporaryFile with file() as file1, file() as file2, file() as file2a: file1.write(b'a' * (2 ** 23)) file1.seek(0) file2.write(b'a' * (2 * 2 ** 18)) file2.seek(0) file2a.write(b'a' * (5 * 2 ** 20)) file2a.seek(0) response = self.client.post('/getlist_count/', { 'file1': file1, 'field1': 'test', 'field2': 'test3', 'field3': 'test5', 'field4': 'test6', 'field5': 'test7', 'file2': (file2, file2a) }) got = response.json() self.assertEqual(got.get('file1'), 1) self.assertEqual(got.get('file2'), 2) def test_fileuploads_closed_at_request_end(self): file = tempfile.NamedTemporaryFile with file() as f1, file() as f2a, file() as f2b: response = self.client.post('/fd_closing/t/', { 'file': f1, 'file2': (f2a, f2b), }) request = response.wsgi_request # The files were parsed. self.assertTrue(hasattr(request, '_files')) file = request._files['file'] self.assertTrue(file.closed) files = request._files.getlist('file2') self.assertTrue(files[0].closed) self.assertTrue(files[1].closed) def test_no_parsing_triggered_by_fd_closing(self): file = tempfile.NamedTemporaryFile with file() as f1, file() as f2a, file() as f2b: response = self.client.post('/fd_closing/f/', { 'file': f1, 'file2': (f2a, f2b), }) request = response.wsgi_request # The fd closing logic doesn't trigger parsing of the stream self.assertFalse(hasattr(request, '_files')) def test_file_error_blocking(self): """ The server should not block when there are upload errors (bug #8622). This can happen if something -- i.e. an exception handler -- tries to access POST while handling an error in parsing POST. This shouldn't cause an infinite loop! """ class POSTAccessingHandler(client.ClientHandler): """A handler that'll access POST during an exception.""" def handle_uncaught_exception(self, request, resolver, exc_info): ret = super().handle_uncaught_exception(request, resolver, exc_info) request.POST # evaluate return ret # Maybe this is a little more complicated that it needs to be; but if # the django.test.client.FakePayload.read() implementation changes then # this test would fail. So we need to know exactly what kind of error # it raises when there is an attempt to read more than the available bytes: try: client.FakePayload(b'a').read(2) except Exception as err: reference_error = err # install the custom handler that tries to access request.POST self.client.handler = POSTAccessingHandler() with open(__file__, 'rb') as fp: post_data = { 'name': 'Ringo', 'file_field': fp, } try: self.client.post('/upload_errors/', post_data) except reference_error.__class__ as err: self.assertFalse( str(err) == str(reference_error), "Caught a repeated exception that'll cause an infinite loop in file uploads." ) except Exception as err: # CustomUploadError is the error that should have been raised self.assertEqual(err.__class__, uploadhandler.CustomUploadError) def test_filename_case_preservation(self): """ The storage backend shouldn't mess with the case of the filenames uploaded. """ # Synthesize the contents of a file upload with a mixed case filename # so we don't have to carry such a file in the Django tests source code # tree. vars = {'boundary': 'oUrBoUnDaRyStRiNg'} post_data = [ '--%(boundary)s', 'Content-Disposition: form-data; name="file_field"; filename="MiXeD_cAsE.txt"', 'Content-Type: application/octet-stream', '', 'file contents\n' '', '--%(boundary)s--\r\n', ] response = self.client.post( '/filename_case/', '\r\n'.join(post_data) % vars, 'multipart/form-data; boundary=%(boundary)s' % vars ) self.assertEqual(response.status_code, 200) id = int(response.content) obj = FileModel.objects.get(pk=id) # The name of the file uploaded and the file stored in the server-side # shouldn't differ. self.assertEqual(os.path.basename(obj.testfile.path), 'MiXeD_cAsE.txt') @override_settings(MEDIA_ROOT=MEDIA_ROOT) class DirectoryCreationTests(SimpleTestCase): """ Tests for error handling during directory creation via _save_FIELD_file (ticket #6450) """ @classmethod def setUpClass(cls): super().setUpClass() if not os.path.isdir(MEDIA_ROOT): os.makedirs(MEDIA_ROOT) @classmethod def tearDownClass(cls): shutil.rmtree(MEDIA_ROOT) super().tearDownClass() def setUp(self): self.obj = FileModel() @unittest.skipIf(sys.platform == 'win32', "Python on Windows doesn't have working os.chmod().") def test_readonly_root(self): """Permission errors are not swallowed""" os.chmod(MEDIA_ROOT, 0o500) self.addCleanup(os.chmod, MEDIA_ROOT, 0o700) with self.assertRaises(PermissionError): self.obj.testfile.save('foo.txt', SimpleUploadedFile('foo.txt', b'x'), save=False) def test_not_a_directory(self): """The correct IOError is raised when the upload directory name exists but isn't a directory""" # Create a file with the upload directory name open(UPLOAD_TO, 'wb').close() self.addCleanup(os.remove, UPLOAD_TO) with self.assertRaises(IOError) as exc_info: with SimpleUploadedFile('foo.txt', b'x') as file: self.obj.testfile.save('foo.txt', file, save=False) # The test needs to be done on a specific string as IOError # is raised even without the patch (just not early enough) self.assertEqual(exc_info.exception.args[0], "%s exists and is not a directory." % UPLOAD_TO) class MultiParserTests(unittest.TestCase): def test_empty_upload_handlers(self): # We're not actually parsing here; just checking if the parser properly # instantiates with empty upload handlers. MultiPartParser({ 'CONTENT_TYPE': 'multipart/form-data; boundary=_foo', 'CONTENT_LENGTH': '1' }, StringIO('x'), [], 'utf-8') def test_rfc2231_parsing(self): test_data = ( (b"Content-Type: application/x-stuff; title*=us-ascii'en-us'This%20is%20%2A%2A%2Afun%2A%2A%2A", "This is ***fun***"), (b"Content-Type: application/x-stuff; title*=UTF-8''foo-%c3%a4.html", "foo-ä.html"), (b"Content-Type: application/x-stuff; title*=iso-8859-1''foo-%E4.html", "foo-ä.html"), ) for raw_line, expected_title in test_data: parsed = parse_header(raw_line) self.assertEqual(parsed[1]['title'], expected_title) def test_rfc2231_wrong_title(self): """ Test wrongly formatted RFC 2231 headers (missing double single quotes). Parsing should not crash (#24209). """ test_data = ( (b"Content-Type: application/x-stuff; title*='This%20is%20%2A%2A%2Afun%2A%2A%2A", b"'This%20is%20%2A%2A%2Afun%2A%2A%2A"), (b"Content-Type: application/x-stuff; title*='foo.html", b"'foo.html"), (b"Content-Type: application/x-stuff; title*=bar.html", b"bar.html"), ) for raw_line, expected_title in test_data: parsed = parse_header(raw_line) self.assertEqual(parsed[1]['title'], expected_title)
bsd-3-clause
onyxfish/agate
agate/table/bar_chart.py
3
1303
#!/usr/bin/env python # pylint: disable=W0212 import leather from agate import utils def bar_chart(self, label=0, value=1, path=None, width=None, height=None): """ Render a bar chart using :class:`leather.Chart`. :param label: The name or index of a column to plot as the labels of the chart. Defaults to the first column in the table. :param value: The name or index of a column to plot as the values of the chart. Defaults to the second column in the table. :param path: If specified, the resulting SVG will be saved to this location. If :code:`None` and running in IPython, then the SVG will be rendered inline. Otherwise, the SVG data will be returned as a string. :param width: The width of the output SVG. :param height: The height of the output SVG. """ if type(label) is int: label_name = self.column_names[label] else: label_name = label if type(value) is int: value_name = self.column_names[value] else: value_name = value chart = leather.Chart() chart.add_x_axis(name=value_name) chart.add_y_axis(name=label_name) chart.add_bars(self, x=value, y=label) return chart.to_svg(path=path, width=width, height=height)
mit
heiscsy/evolutus_ros_src
evolutus_filter/build/catkin_generated/installspace/_setup_util.py
4
12275
#!/usr/bin/python # Software License Agreement (BSD License) # # Copyright (c) 2012, Willow Garage, Inc. # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions # are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above # copyright notice, this list of conditions and the following # disclaimer in the documentation and/or other materials provided # with the distribution. # * Neither the name of Willow Garage, Inc. nor the names of its # contributors may be used to endorse or promote products derived # from this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS # FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE # COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, # INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, # BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER # CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT # LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN # ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE # POSSIBILITY OF SUCH DAMAGE. '''This file generates shell code for the setup.SHELL scripts to set environment variables''' from __future__ import print_function import argparse import copy import errno import os import platform import sys CATKIN_MARKER_FILE = '.catkin' system = platform.system() IS_DARWIN = (system == 'Darwin') IS_WINDOWS = (system == 'Windows') # subfolder of workspace prepended to CMAKE_PREFIX_PATH ENV_VAR_SUBFOLDERS = { 'CMAKE_PREFIX_PATH': '', 'CPATH': 'include', 'LD_LIBRARY_PATH' if not IS_DARWIN else 'DYLD_LIBRARY_PATH': ['lib', os.path.join('lib', 'x86_64-linux-gnu')], 'PATH': 'bin', 'PKG_CONFIG_PATH': [os.path.join('lib', 'pkgconfig'), os.path.join('lib', 'x86_64-linux-gnu', 'pkgconfig')], 'PYTHONPATH': 'lib/python2.7/dist-packages', } def rollback_env_variables(environ, env_var_subfolders): ''' Generate shell code to reset environment variables by unrolling modifications based on all workspaces in CMAKE_PREFIX_PATH. This does not cover modifications performed by environment hooks. ''' lines = [] unmodified_environ = copy.copy(environ) for key in sorted(env_var_subfolders.keys()): subfolders = env_var_subfolders[key] if not isinstance(subfolders, list): subfolders = [subfolders] for subfolder in subfolders: value = _rollback_env_variable(unmodified_environ, key, subfolder) if value is not None: environ[key] = value lines.append(assignment(key, value)) if lines: lines.insert(0, comment('reset environment variables by unrolling modifications based on all workspaces in CMAKE_PREFIX_PATH')) return lines def _rollback_env_variable(environ, name, subfolder): ''' For each catkin workspace in CMAKE_PREFIX_PATH remove the first entry from env[NAME] matching workspace + subfolder. :param subfolder: str '' or subfoldername that may start with '/' :returns: the updated value of the environment variable. ''' value = environ[name] if name in environ else '' env_paths = [path for path in value.split(os.pathsep) if path] value_modified = False if subfolder: if subfolder.startswith(os.path.sep) or (os.path.altsep and subfolder.startswith(os.path.altsep)): subfolder = subfolder[1:] if subfolder.endswith(os.path.sep) or (os.path.altsep and subfolder.endswith(os.path.altsep)): subfolder = subfolder[:-1] for ws_path in _get_workspaces(environ, include_fuerte=True, include_non_existing=True): path_to_find = os.path.join(ws_path, subfolder) if subfolder else ws_path path_to_remove = None for env_path in env_paths: env_path_clean = env_path[:-1] if env_path and env_path[-1] in [os.path.sep, os.path.altsep] else env_path if env_path_clean == path_to_find: path_to_remove = env_path break if path_to_remove: env_paths.remove(path_to_remove) value_modified = True new_value = os.pathsep.join(env_paths) return new_value if value_modified else None def _get_workspaces(environ, include_fuerte=False, include_non_existing=False): ''' Based on CMAKE_PREFIX_PATH return all catkin workspaces. :param include_fuerte: The flag if paths starting with '/opt/ros/fuerte' should be considered workspaces, ``bool`` ''' # get all cmake prefix paths env_name = 'CMAKE_PREFIX_PATH' value = environ[env_name] if env_name in environ else '' paths = [path for path in value.split(os.pathsep) if path] # remove non-workspace paths workspaces = [path for path in paths if os.path.isfile(os.path.join(path, CATKIN_MARKER_FILE)) or (include_fuerte and path.startswith('/opt/ros/fuerte')) or (include_non_existing and not os.path.exists(path))] return workspaces def prepend_env_variables(environ, env_var_subfolders, workspaces): ''' Generate shell code to prepend environment variables for the all workspaces. ''' lines = [] lines.append(comment('prepend folders of workspaces to environment variables')) paths = [path for path in workspaces.split(os.pathsep) if path] prefix = _prefix_env_variable(environ, 'CMAKE_PREFIX_PATH', paths, '') lines.append(prepend(environ, 'CMAKE_PREFIX_PATH', prefix)) for key in sorted([key for key in env_var_subfolders.keys() if key != 'CMAKE_PREFIX_PATH']): subfolder = env_var_subfolders[key] prefix = _prefix_env_variable(environ, key, paths, subfolder) lines.append(prepend(environ, key, prefix)) return lines def _prefix_env_variable(environ, name, paths, subfolders): ''' Return the prefix to prepend to the environment variable NAME, adding any path in NEW_PATHS_STR without creating duplicate or empty items. ''' value = environ[name] if name in environ else '' environ_paths = [path for path in value.split(os.pathsep) if path] checked_paths = [] for path in paths: if not isinstance(subfolders, list): subfolders = [subfolders] for subfolder in subfolders: path_tmp = path if subfolder: path_tmp = os.path.join(path_tmp, subfolder) # exclude any path already in env and any path we already added if path_tmp not in environ_paths and path_tmp not in checked_paths: checked_paths.append(path_tmp) prefix_str = os.pathsep.join(checked_paths) if prefix_str != '' and environ_paths: prefix_str += os.pathsep return prefix_str def assignment(key, value): if not IS_WINDOWS: return 'export %s="%s"' % (key, value) else: return 'set %s=%s' % (key, value) def comment(msg): if not IS_WINDOWS: return '# %s' % msg else: return 'REM %s' % msg def prepend(environ, key, prefix): if key not in environ or not environ[key]: return assignment(key, prefix) if not IS_WINDOWS: return 'export %s="%s$%s"' % (key, prefix, key) else: return 'set %s=%s%%%s%%' % (key, prefix, key) def find_env_hooks(environ, cmake_prefix_path): ''' Generate shell code with found environment hooks for the all workspaces. ''' lines = [] lines.append(comment('found environment hooks in workspaces')) generic_env_hooks = [] generic_env_hooks_workspace = [] specific_env_hooks = [] specific_env_hooks_workspace = [] generic_env_hooks_by_filename = {} specific_env_hooks_by_filename = {} generic_env_hook_ext = 'bat' if IS_WINDOWS else 'sh' specific_env_hook_ext = environ['CATKIN_SHELL'] if not IS_WINDOWS and 'CATKIN_SHELL' in environ and environ['CATKIN_SHELL'] else None # remove non-workspace paths workspaces = [path for path in cmake_prefix_path.split(os.pathsep) if path and os.path.isfile(os.path.join(path, CATKIN_MARKER_FILE))] for workspace in reversed(workspaces): env_hook_dir = os.path.join(workspace, 'etc', 'catkin', 'profile.d') if os.path.isdir(env_hook_dir): for filename in sorted(os.listdir(env_hook_dir)): if filename.endswith('.%s' % generic_env_hook_ext): # remove previous env hook with same name if present if filename in generic_env_hooks_by_filename: i = generic_env_hooks.index(generic_env_hooks_by_filename[filename]) generic_env_hooks.pop(i) generic_env_hooks_workspace.pop(i) # append env hook generic_env_hooks.append(os.path.join(env_hook_dir, filename)) generic_env_hooks_workspace.append(workspace) generic_env_hooks_by_filename[filename] = generic_env_hooks[-1] elif specific_env_hook_ext is not None and filename.endswith('.%s' % specific_env_hook_ext): # remove previous env hook with same name if present if filename in specific_env_hooks_by_filename: i = specific_env_hooks.index(specific_env_hooks_by_filename[filename]) specific_env_hooks.pop(i) specific_env_hooks_workspace.pop(i) # append env hook specific_env_hooks.append(os.path.join(env_hook_dir, filename)) specific_env_hooks_workspace.append(workspace) specific_env_hooks_by_filename[filename] = specific_env_hooks[-1] env_hooks = generic_env_hooks + specific_env_hooks env_hooks_workspace = generic_env_hooks_workspace + specific_env_hooks_workspace count = len(env_hooks) lines.append(assignment('_CATKIN_ENVIRONMENT_HOOKS_COUNT', count)) for i in range(count): lines.append(assignment('_CATKIN_ENVIRONMENT_HOOKS_%d' % i, env_hooks[i])) lines.append(assignment('_CATKIN_ENVIRONMENT_HOOKS_%d_WORKSPACE' % i, env_hooks_workspace[i])) return lines def _parse_arguments(args=None): parser = argparse.ArgumentParser(description='Generates code blocks for the setup.SHELL script.') parser.add_argument('--extend', action='store_true', help='Skip unsetting previous environment variables to extend context') return parser.parse_known_args(args=args)[0] if __name__ == '__main__': try: try: args = _parse_arguments() except Exception as e: print(e, file=sys.stderr) sys.exit(1) # environment at generation time CMAKE_PREFIX_PATH = '/home/yuhan/catkin_ws/devel;/opt/ros/indigo'.split(';') # prepend current workspace if not already part of CPP base_path = os.path.dirname(__file__) if base_path not in CMAKE_PREFIX_PATH: CMAKE_PREFIX_PATH.insert(0, base_path) CMAKE_PREFIX_PATH = os.pathsep.join(CMAKE_PREFIX_PATH) environ = dict(os.environ) lines = [] if not args.extend: lines += rollback_env_variables(environ, ENV_VAR_SUBFOLDERS) lines += prepend_env_variables(environ, ENV_VAR_SUBFOLDERS, CMAKE_PREFIX_PATH) lines += find_env_hooks(environ, CMAKE_PREFIX_PATH) print('\n'.join(lines)) # need to explicitly flush the output sys.stdout.flush() except IOError as e: # and catch potantial "broken pipe" if stdout is not writable # which can happen when piping the output to a file but the disk is full if e.errno == errno.EPIPE: print(e, file=sys.stderr) sys.exit(2) raise sys.exit(0)
gpl-2.0
robhudson/kuma
vendor/packages/translate/storage/wordfast.py
25
16043
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Copyright 2007-2010 Zuza Software Foundation # # This file is part of the Translate Toolkit. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, see <http://www.gnu.org/licenses/>. """Manage the Wordfast Translation Memory format Wordfast TM format is the Translation Memory format used by the `Wordfast <http://www.wordfast.net/>`_ computer aided translation tool. It is a bilingual base class derived format with :class:`WordfastTMFile` and :class:`WordfastUnit` providing file and unit level access. Wordfast is a computer aided translation tool. It is an application built on top of Microsoft Word and is implemented as a rather sophisticated set of macros. Understanding that helps us understand many of the seemingly strange choices around this format including: encoding, escaping and file naming. Implementation The implementation covers the full requirements of a Wordfast TM file. The files are simple Tab Separated Value (TSV) files that can be read by Microsoft Excel and other spreadsheet programs. They use the .txt extension which does make it more difficult to automatically identify such files. The dialect of the TSV files is specified by :class:`WordfastDialect`. Encoding The files are UTF-16 or ISO-8859-1 (Latin1) encoded. These choices are most likely because Microsoft Word is the base editing tool for Wordfast. The format is tab separated so we are able to detect UTF-16 vs Latin-1 by searching for the occurance of a UTF-16 tab character and then continuing with the parsing. Timestamps :class:`WordfastTime` allows for the correct management of the Wordfast YYYYMMDD~HHMMSS timestamps. However, timestamps on individual units are not updated when edited. Header :class:`WordfastHeader` provides header management support. The header functionality is fully implemented through observing the behaviour of the files in real use cases, input from the Wordfast programmers and public documentation. Escaping Wordfast TM implements a form of escaping that covers two aspects: 1. Placeable: bold, formating, etc. These are left as is and ignored. It is up to the editor and future placeable implementation to manage these. 2. Escapes: items that may confuse Excel or translators are escaped as ``&'XX;``. These are fully implemented and are converted to and from Unicode. By observing behaviour and reading documentation we where able to observe all possible escapes. Unfortunately the escaping differs slightly between Windows and Mac version. This might cause errors in future. Functions allow for ``<_wf_to_char>`` and back to Wordfast escape (``<_char_to_wf>``). Extended Attributes The last 4 columns allow users to define and manage extended attributes. These are left as is and are not directly managed byour implemenation. """ import csv import time from translate.storage import base WF_TIMEFORMAT = "%Y%m%d~%H%M%S" """Time format used by Wordfast""" WF_FIELDNAMES_HEADER = ["date", "userlist", "tucount", "src-lang", "version", "target-lang", "license", "attr1list", "attr2list", "attr3list", "attr4list", "attr5list"] """Field names for the Wordfast header""" WF_FIELDNAMES = ["date", "user", "reuse", "src-lang", "source", "target-lang", "target", "attr1", "attr2", "attr3", "attr4"] """Field names for a Wordfast TU""" WF_FIELDNAMES_HEADER_DEFAULTS = { "date": "%19000101~121212", "userlist": "%User ID,TT,TT Translate-Toolkit", "tucount": "%TU=00000001", "src-lang": "%EN-US", "version": "%Wordfast TM v.5.51w9/00", "target-lang": "", "license": "%---00000001", "attr1list": "", "attr2list": "", "attr3list": "", "attr4list": "", } """Default or minimum header entries for a Wordfast file""" # TODO Needs validation. The following need to be checked against a WF TM file # to ensure that the correct Unicode values have been chosen for the characters. # For now these look correct and have been taken from Windows CP1252 and # Macintosh code points found for the respective character sets on Linux. WF_ESCAPE_MAP = ( ("&'26;", u"\u0026"), # & - Ampersand (must be first to prevent # escaping of escapes) ("&'82;", u"\u201A"), # ‚ - Single low-9 quotation mark ("&'85;", u"\u2026"), # … - Elippsis ("&'91;", u"\u2018"), # ‘ - left single quotation mark ("&'92;", u"\u2019"), # ’ - right single quotation mark ("&'93;", u"\u201C"), # “ - left double quotation mark ("&'94;", u"\u201D"), # ” - right double quotation mark ("&'96;", u"\u2013"), # – - en dash (validate) ("&'97;", u"\u2014"), # — - em dash (validate) ("&'99;", u"\u2122"), # ™ - Trade mark # Windows only ("&'A0;", u"\u00A0"), #   - Non breaking space ("&'A9;", u"\u00A9"), # © - Copyright ("&'AE;", u"\u00AE"), # ® - Registered ("&'BC;", u"\u00BC"), # ¼ ("&'BD;", u"\u00BD"), # ½ ("&'BE;", u"\u00BE"), # ¾ # Mac only ("&'A8;", u"\u00AE"), # ® - Registered ("&'AA;", u"\u2122"), # ™ - Trade mark ("&'C7;", u"\u00AB"), # « - Left-pointing double angle quotation mark ("&'C8;", u"\u00BB"), # » - Right-pointing double angle quotation mark ("&'C9;", u"\u2026"), # … - Horizontal Elippsis ("&'CA;", u"\u00A0"), #   - Non breaking space ("&'D0;", u"\u2013"), # – - en dash (validate) ("&'D1;", u"\u2014"), # — - em dash (validate) ("&'D2;", u"\u201C"), # “ - left double quotation mark ("&'D3;", u"\u201D"), # ” - right double quotation mark ("&'D4;", u"\u2018"), # ‘ - left single quotation mark ("&'D5;", u"\u2019"), # ’ - right single quotation mark ("&'E2;", u"\u201A"), # ‚ - Single low-9 quotation mark ("&'E3;", u"\u201E"), # „ - Double low-9 quotation mark # Other markers #("&'B;", u"\n"), # Soft-break - XXX creates a problem with # roundtripping could also be represented # by \u2028 ) """Mapping of Wordfast &'XX; escapes to correct Unicode characters""" TAB_UTF16 = "\x00\x09" """The tab \\t character as it would appear in UTF-16 encoding""" def _char_to_wf(string): """Char -> Wordfast &'XX; escapes Full roundtripping is not possible because of the escaping of NEWLINE \\n and TAB \\t""" # FIXME there is no platform check to ensure that we use Mac encodings # when running on a Mac if string: for code, char in WF_ESCAPE_MAP: string = string.replace(char.encode('utf-8'), code) string = string.replace("\n", "\\n").replace("\t", "\\t") return string def _wf_to_char(string): """Wordfast &'XX; escapes -> Char""" if string: for code, char in WF_ESCAPE_MAP: string = string.replace(code, char.encode('utf-8')) string = string.replace("\\n", "\n").replace("\\t", "\t") return string class WordfastDialect(csv.Dialect): """Describe the properties of a Wordfast generated TAB-delimited file.""" delimiter = "\t" lineterminator = "\r\n" quoting = csv.QUOTE_NONE csv.register_dialect("wordfast", WordfastDialect) class WordfastTime(object): """Manages time stamps in the Wordfast format of YYYYMMDD~hhmmss""" def __init__(self, newtime=None): self._time = None if not newtime: self.time = None elif isinstance(newtime, basestring): self.timestring = newtime elif isinstance(newtime, time.struct_time): self.time = newtime def get_timestring(self): """Get the time in the Wordfast time format""" if not self._time: return None else: return time.strftime(WF_TIMEFORMAT, self._time) def set_timestring(self, timestring): """Set the time_sturct object using a Wordfast time formated string :param timestring: A Wordfast time string (YYYMMDD~hhmmss) :type timestring: String """ self._time = time.strptime(timestring, WF_TIMEFORMAT) timestring = property(get_timestring, set_timestring) def get_time(self): """Get the time_struct object""" return self._time def set_time(self, newtime): """Set the time_struct object :param newtime: a new time object :type newtime: time.time_struct """ if newtime and isinstance(newtime, time.struct_time): self._time = newtime else: self._time = None time = property(get_time, set_time) def __str__(self): if not self.timestring: return "" else: return self.timestring class WordfastHeader(object): """A wordfast translation memory header""" def __init__(self, header=None): self._header_dict = [] if not header: self.header = self._create_default_header() elif isinstance(header, dict): self.header = header def _create_default_header(self): """Create a default Wordfast header with the date set to the current time""" defaultheader = WF_FIELDNAMES_HEADER_DEFAULTS defaultheader['date'] = '%%%s' % WordfastTime(time.localtime()).timestring return defaultheader def getheader(self): """Get the header dictionary""" return self._header_dict def setheader(self, newheader): self._header_dict = newheader header = property(getheader, setheader) def settargetlang(self, newlang): self._header_dict['target-lang'] = '%%%s' % newlang targetlang = property(None, settargetlang) def settucount(self, count): self._header_dict['tucount'] = '%%TU=%08d' % count tucount = property(None, settucount) class WordfastUnit(base.TranslationUnit): """A Wordfast translation memory unit""" def __init__(self, source=None): self._dict = {} if source: self.source = source super(WordfastUnit, self).__init__(source) def _update_timestamp(self): """Refresh the timestamp for the unit""" self._dict['date'] = WordfastTime(time.localtime()).timestring def getdict(self): """Get the dictionary of values for a Wordfast line""" return self._dict def setdict(self, newdict): """Set the dictionary of values for a Wordfast line :param newdict: a new dictionary with Wordfast line elements :type newdict: Dict """ # TODO First check that the values are OK self._dict = newdict dict = property(getdict, setdict) def _get_source_or_target(self, key): if self._dict.get(key, None) is None: return None elif self._dict[key]: return _wf_to_char(self._dict[key]).decode('utf-8') else: return "" def _set_source_or_target(self, key, newvalue): if newvalue is None: self._dict[key] = None if isinstance(newvalue, unicode): newvalue = newvalue.encode('utf-8') newvalue = _char_to_wf(newvalue) if not key in self._dict or newvalue != self._dict[key]: self._dict[key] = newvalue self._update_timestamp() def getsource(self): return self._get_source_or_target('source') def setsource(self, newsource): self._rich_source = None return self._set_source_or_target('source', newsource) source = property(getsource, setsource) def gettarget(self): return self._get_source_or_target('target') def settarget(self, newtarget): self._rich_target = None return self._set_source_or_target('target', newtarget) target = property(gettarget, settarget) def settargetlang(self, newlang): self._dict['target-lang'] = newlang targetlang = property(None, settargetlang) def __str__(self): return str(self._dict) def istranslated(self): if not self._dict.get('source', None): return False return bool(self._dict.get('target', None)) class WordfastTMFile(base.TranslationStore): """A Wordfast translation memory file""" Name = "Wordfast Translation Memory" Mimetypes = ["application/x-wordfast"] Extensions = ["txt"] def __init__(self, inputfile=None, unitclass=WordfastUnit): """construct a Wordfast TM, optionally reading in from inputfile.""" self.UnitClass = unitclass base.TranslationStore.__init__(self, unitclass=unitclass) self.filename = '' self.header = WordfastHeader() self._encoding = 'iso-8859-1' if inputfile is not None: self.parse(inputfile) def parse(self, input): """parsese the given file or file source string""" if hasattr(input, 'name'): self.filename = input.name elif not getattr(self, 'filename', ''): self.filename = '' if hasattr(input, "read"): tmsrc = input.read() input.close() input = tmsrc if TAB_UTF16 in input.split("\n")[0]: self._encoding = 'utf-16' else: self._encoding = 'iso-8859-1' try: input = input.decode(self._encoding).encode('utf-8') except: raise ValueError("Wordfast files are either UTF-16 (UCS2) or ISO-8859-1 encoded") for header in csv.DictReader(input.split("\n")[:1], fieldnames=WF_FIELDNAMES_HEADER, dialect="wordfast"): self.header = WordfastHeader(header) lines = csv.DictReader(input.split("\n")[1:], fieldnames=WF_FIELDNAMES, dialect="wordfast") for line in lines: newunit = WordfastUnit() newunit.dict = line self.addunit(newunit) def __str__(self): output = csv.StringIO() header_output = csv.StringIO() writer = csv.DictWriter(output, fieldnames=WF_FIELDNAMES, dialect="wordfast") unit_count = 0 for unit in self.units: if unit.istranslated(): unit_count += 1 writer.writerow(unit.dict) if unit_count == 0: return "" output.reset() self.header.tucount = unit_count outheader = csv.DictWriter(header_output, fieldnames=WF_FIELDNAMES_HEADER, dialect="wordfast") outheader.writerow(self.header.header) header_output.reset() decoded = "".join(header_output.readlines() + output.readlines()).decode('utf-8') try: return decoded.encode(self._encoding) except UnicodeEncodeError: return decoded.encode('utf-16')
mpl-2.0
cpanelli/-git-clone-https-chromium.googlesource.com-chromium-tools-depot_tools
third_party/pylint/checkers/similar.py
64
14174
# pylint: disable=W0622 # Copyright (c) 2004-2013 LOGILAB S.A. (Paris, FRANCE). # http://www.logilab.fr/ -- mailto:contact@logilab.fr # # This program is free software; you can redistribute it and/or modify it under # the terms of the GNU General Public License as published by the Free Software # Foundation; either version 2 of the License, or (at your option) any later # version. # # This program is distributed in the hope that it will be useful, but WITHOUT # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS # FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details # # You should have received a copy of the GNU General Public License along with # this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. """a similarities / code duplication command line tool and pylint checker """ from __future__ import print_function import sys from collections import defaultdict from logilab.common.ureports import Table from pylint.interfaces import IRawChecker from pylint.checkers import BaseChecker, table_lines_from_stats import six from six.moves import zip class Similar(object): """finds copy-pasted lines of code in a project""" def __init__(self, min_lines=4, ignore_comments=False, ignore_docstrings=False, ignore_imports=False): self.min_lines = min_lines self.ignore_comments = ignore_comments self.ignore_docstrings = ignore_docstrings self.ignore_imports = ignore_imports self.linesets = [] def append_stream(self, streamid, stream, encoding=None): """append a file to search for similarities""" if encoding is None: readlines = stream.readlines else: readlines = lambda: [line.decode(encoding) for line in stream] try: self.linesets.append(LineSet(streamid, readlines(), self.ignore_comments, self.ignore_docstrings, self.ignore_imports)) except UnicodeDecodeError: pass def run(self): """start looking for similarities and display results on stdout""" self._display_sims(self._compute_sims()) def _compute_sims(self): """compute similarities in appended files""" no_duplicates = defaultdict(list) for num, lineset1, idx1, lineset2, idx2 in self._iter_sims(): duplicate = no_duplicates[num] for couples in duplicate: if (lineset1, idx1) in couples or (lineset2, idx2) in couples: couples.add((lineset1, idx1)) couples.add((lineset2, idx2)) break else: duplicate.append(set([(lineset1, idx1), (lineset2, idx2)])) sims = [] for num, ensembles in six.iteritems(no_duplicates): for couples in ensembles: sims.append((num, couples)) sims.sort() sims.reverse() return sims def _display_sims(self, sims): """display computed similarities on stdout""" nb_lignes_dupliquees = 0 for num, couples in sims: print() print(num, "similar lines in", len(couples), "files") couples = sorted(couples) for lineset, idx in couples: print("==%s:%s" % (lineset.name, idx)) # pylint: disable=W0631 for line in lineset._real_lines[idx:idx+num]: print(" ", line.rstrip()) nb_lignes_dupliquees += num * (len(couples)-1) nb_total_lignes = sum([len(lineset) for lineset in self.linesets]) print("TOTAL lines=%s duplicates=%s percent=%.2f" \ % (nb_total_lignes, nb_lignes_dupliquees, nb_lignes_dupliquees*100. / nb_total_lignes)) def _find_common(self, lineset1, lineset2): """find similarities in the two given linesets""" lines1 = lineset1.enumerate_stripped lines2 = lineset2.enumerate_stripped find = lineset2.find index1 = 0 min_lines = self.min_lines while index1 < len(lineset1): skip = 1 num = 0 for index2 in find(lineset1[index1]): non_blank = 0 for num, ((_, line1), (_, line2)) in enumerate( zip(lines1(index1), lines2(index2))): if line1 != line2: if non_blank > min_lines: yield num, lineset1, index1, lineset2, index2 skip = max(skip, num) break if line1: non_blank += 1 else: # we may have reach the end num += 1 if non_blank > min_lines: yield num, lineset1, index1, lineset2, index2 skip = max(skip, num) index1 += skip def _iter_sims(self): """iterate on similarities among all files, by making a cartesian product """ for idx, lineset in enumerate(self.linesets[:-1]): for lineset2 in self.linesets[idx+1:]: for sim in self._find_common(lineset, lineset2): yield sim def stripped_lines(lines, ignore_comments, ignore_docstrings, ignore_imports): """return lines with leading/trailing whitespace and any ignored code features removed """ strippedlines = [] docstring = None for line in lines: line = line.strip() if ignore_docstrings: if not docstring and \ (line.startswith('"""') or line.startswith("'''")): docstring = line[:3] line = line[3:] if docstring: if line.endswith(docstring): docstring = None line = '' if ignore_imports: if line.startswith("import ") or line.startswith("from "): line = '' if ignore_comments: # XXX should use regex in checkers/format to avoid cutting # at a "#" in a string line = line.split('#', 1)[0].strip() strippedlines.append(line) return strippedlines class LineSet(object): """Holds and indexes all the lines of a single source file""" def __init__(self, name, lines, ignore_comments=False, ignore_docstrings=False, ignore_imports=False): self.name = name self._real_lines = lines self._stripped_lines = stripped_lines(lines, ignore_comments, ignore_docstrings, ignore_imports) self._index = self._mk_index() def __str__(self): return '<Lineset for %s>' % self.name def __len__(self): return len(self._real_lines) def __getitem__(self, index): return self._stripped_lines[index] def __lt__(self, other): return self.name < other.name def __hash__(self): return id(self) def enumerate_stripped(self, start_at=0): """return an iterator on stripped lines, starting from a given index if specified, else 0 """ idx = start_at if start_at: lines = self._stripped_lines[start_at:] else: lines = self._stripped_lines for line in lines: #if line: yield idx, line idx += 1 def find(self, stripped_line): """return positions of the given stripped line in this set""" return self._index.get(stripped_line, ()) def _mk_index(self): """create the index for this set""" index = defaultdict(list) for line_no, line in enumerate(self._stripped_lines): if line: index[line].append(line_no) return index MSGS = {'R0801': ('Similar lines in %s files\n%s', 'duplicate-code', 'Indicates that a set of similar lines has been detected \ among multiple file. This usually means that the code should \ be refactored to avoid this duplication.')} def report_similarities(sect, stats, old_stats): """make a layout with some stats about duplication""" lines = ['', 'now', 'previous', 'difference'] lines += table_lines_from_stats(stats, old_stats, ('nb_duplicated_lines', 'percent_duplicated_lines')) sect.append(Table(children=lines, cols=4, rheaders=1, cheaders=1)) # wrapper to get a pylint checker from the similar class class SimilarChecker(BaseChecker, Similar): """checks for similarities and duplicated code. This computation may be memory / CPU intensive, so you should disable it if you experiment some problems. """ __implements__ = (IRawChecker,) # configuration section name name = 'similarities' # messages msgs = MSGS # configuration options # for available dict keys/values see the optik parser 'add_option' method options = (('min-similarity-lines', {'default' : 4, 'type' : "int", 'metavar' : '<int>', 'help' : 'Minimum lines number of a similarity.'}), ('ignore-comments', {'default' : True, 'type' : 'yn', 'metavar' : '<y or n>', 'help': 'Ignore comments when computing similarities.'} ), ('ignore-docstrings', {'default' : True, 'type' : 'yn', 'metavar' : '<y or n>', 'help': 'Ignore docstrings when computing similarities.'} ), ('ignore-imports', {'default' : False, 'type' : 'yn', 'metavar' : '<y or n>', 'help': 'Ignore imports when computing similarities.'} ), ) # reports reports = (('RP0801', 'Duplication', report_similarities),) def __init__(self, linter=None): BaseChecker.__init__(self, linter) Similar.__init__(self, min_lines=4, ignore_comments=True, ignore_docstrings=True) self.stats = None def set_option(self, optname, value, action=None, optdict=None): """method called to set an option (registered in the options list) overridden to report options setting to Similar """ BaseChecker.set_option(self, optname, value, action, optdict) if optname == 'min-similarity-lines': self.min_lines = self.config.min_similarity_lines elif optname == 'ignore-comments': self.ignore_comments = self.config.ignore_comments elif optname == 'ignore-docstrings': self.ignore_docstrings = self.config.ignore_docstrings elif optname == 'ignore-imports': self.ignore_imports = self.config.ignore_imports def open(self): """init the checkers: reset linesets and statistics information""" self.linesets = [] self.stats = self.linter.add_stats(nb_duplicated_lines=0, percent_duplicated_lines=0) def process_module(self, node): """process a module the module's content is accessible via the stream object stream must implement the readlines method """ with node.stream() as stream: self.append_stream(self.linter.current_name, stream, node.file_encoding) def close(self): """compute and display similarities on closing (i.e. end of parsing)""" total = sum([len(lineset) for lineset in self.linesets]) duplicated = 0 stats = self.stats for num, couples in self._compute_sims(): msg = [] for lineset, idx in couples: msg.append("==%s:%s" % (lineset.name, idx)) msg.sort() # pylint: disable=W0631 for line in lineset._real_lines[idx:idx+num]: msg.append(line.rstrip()) self.add_message('R0801', args=(len(couples), '\n'.join(msg))) duplicated += num * (len(couples) - 1) stats['nb_duplicated_lines'] = duplicated stats['percent_duplicated_lines'] = total and duplicated * 100. / total def register(linter): """required method to auto register this checker """ linter.register_checker(SimilarChecker(linter)) def usage(status=0): """display command line usage information""" print("finds copy pasted blocks in a set of files") print() print('Usage: symilar [-d|--duplicates min_duplicated_lines] \ [-i|--ignore-comments] [--ignore-docstrings] [--ignore-imports] file1...') sys.exit(status) def Run(argv=None): """standalone command line access point""" if argv is None: argv = sys.argv[1:] from getopt import getopt s_opts = 'hdi' l_opts = ('help', 'duplicates=', 'ignore-comments', 'ignore-imports', 'ignore-docstrings') min_lines = 4 ignore_comments = False ignore_docstrings = False ignore_imports = False opts, args = getopt(argv, s_opts, l_opts) for opt, val in opts: if opt in ('-d', '--duplicates'): min_lines = int(val) elif opt in ('-h', '--help'): usage() elif opt in ('-i', '--ignore-comments'): ignore_comments = True elif opt in ('--ignore-docstrings',): ignore_docstrings = True elif opt in ('--ignore-imports',): ignore_imports = True if not args: usage(1) sim = Similar(min_lines, ignore_comments, ignore_docstrings, ignore_imports) for filename in args: with open(filename) as stream: sim.append_stream(filename, stream) sim.run() sys.exit(0) if __name__ == '__main__': Run()
bsd-3-clause
wnesl/gnuradio-IA
gnuradio-core/src/python/gnuradio/gr/qa_hier_block2.py
18
13570
#!/usr/bin/env python from gnuradio import gr, gr_unittest class test_hier_block2(gr_unittest.TestCase): def setUp(self): pass def tearDown(self): pass def test_001_make(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) self.assertEqual("test_block", hblock.name()) self.assertEqual(1, hblock.input_signature().max_streams()) self.assertEqual(1, hblock.output_signature().min_streams()) self.assertEqual(1, hblock.output_signature().max_streams()) self.assertEqual(gr.sizeof_int, hblock.output_signature().sizeof_stream_item(0)) def test_002_connect_input(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) hblock.connect(hblock, nop1) def test_004_connect_output(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) hblock.connect(nop1, hblock) def test_005_connect_output_in_use(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) nop2 = gr.nop(gr.sizeof_int) hblock.connect(nop1, hblock) self.assertRaises(ValueError, lambda: hblock.connect(nop2, hblock)) def test_006_connect_invalid_src_port_neg(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) self.assertRaises(ValueError, lambda: hblock.connect((hblock, -1), nop1)) def test_005_connect_invalid_src_port_exceeds(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) self.assertRaises(ValueError, lambda: hblock.connect((hblock, 1), nop1)) def test_007_connect_invalid_dst_port_neg(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) nop2 = gr.nop(gr.sizeof_int) self.assertRaises(ValueError, lambda: hblock.connect(nop1, (nop2, -1))) def test_008_connect_invalid_dst_port_exceeds(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.null_sink(gr.sizeof_int) nop2 = gr.null_sink(gr.sizeof_int) self.assertRaises(ValueError, lambda: hblock.connect(nop1, (nop2, 1))) def test_009_check_topology(self): hblock = gr.top_block("test_block") hblock.check_topology(0, 0) def test_010_run(self): expected = (1.0, 2.0, 3.0, 4.0) hblock = gr.top_block("test_block") src = gr.vector_source_f(expected, False) sink1 = gr.vector_sink_f() sink2 = gr.vector_sink_f() hblock.connect(src, sink1) hblock.connect(src, sink2) hblock.run() actual1 = sink1.data() actual2 = sink2.data() self.assertEquals(expected, actual1) self.assertEquals(expected, actual2) def test_012_disconnect_input(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) hblock.connect(hblock, nop1) hblock.disconnect(hblock, nop1) def test_013_disconnect_input_not_connected(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) nop2 = gr.nop(gr.sizeof_int) hblock.connect(hblock, nop1) self.assertRaises(ValueError, lambda: hblock.disconnect(hblock, nop2)) def test_014_disconnect_input_neg(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) hblock.connect(hblock, nop1) self.assertRaises(ValueError, lambda: hblock.disconnect((hblock, -1), nop1)) def test_015_disconnect_input_exceeds(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) hblock.connect(hblock, nop1) self.assertRaises(ValueError, lambda: hblock.disconnect((hblock, 1), nop1)) def test_016_disconnect_output(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) hblock.connect(nop1, hblock) hblock.disconnect(nop1, hblock) def test_017_disconnect_output_not_connected(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) nop2 = gr.nop(gr.sizeof_int) hblock.connect(nop1, hblock) self.assertRaises(ValueError, lambda: hblock.disconnect(nop2, hblock)) def test_018_disconnect_output_neg(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) hblock.connect(hblock, nop1) self.assertRaises(ValueError, lambda: hblock.disconnect(nop1, (hblock, -1))) def test_019_disconnect_output_exceeds(self): hblock = gr.hier_block2("test_block", gr.io_signature(1,1,gr.sizeof_int), gr.io_signature(1,1,gr.sizeof_int)) nop1 = gr.nop(gr.sizeof_int) hblock.connect(nop1, hblock) self.assertRaises(ValueError, lambda: hblock.disconnect(nop1, (hblock, 1))) def test_020_run(self): hblock = gr.top_block("test_block") data = (1.0, 2.0, 3.0, 4.0) src = gr.vector_source_f(data, False) dst = gr.vector_sink_f() hblock.connect(src, dst) hblock.run() self.assertEquals(data, dst.data()) def test_021_connect_single(self): hblock = gr.top_block("test_block") blk = gr.hier_block2("block", gr.io_signature(0, 0, 0), gr.io_signature(0, 0, 0)) hblock.connect(blk) def test_022_connect_single_with_ports(self): hblock = gr.top_block("test_block") blk = gr.hier_block2("block", gr.io_signature(1, 1, 1), gr.io_signature(1, 1, 1)) self.assertRaises(ValueError, lambda: hblock.connect(blk)) def test_023_connect_single_twice(self): hblock = gr.top_block("test_block") blk = gr.hier_block2("block", gr.io_signature(0, 0, 0), gr.io_signature(0, 0, 0)) hblock.connect(blk) self.assertRaises(ValueError, lambda: hblock.connect(blk)) def test_024_disconnect_single(self): hblock = gr.top_block("test_block") blk = gr.hier_block2("block", gr.io_signature(0, 0, 0), gr.io_signature(0, 0, 0)) hblock.connect(blk) hblock.disconnect(blk) def test_025_disconnect_single_not_connected(self): hblock = gr.top_block("test_block") blk = gr.hier_block2("block", gr.io_signature(0, 0, 0), gr.io_signature(0, 0, 0)) self.assertRaises(ValueError, lambda: hblock.disconnect(blk)) def test_026_run_single(self): expected_data = (1.0,) tb = gr.top_block("top_block") hb = gr.hier_block2("block", gr.io_signature(0, 0, 0), gr.io_signature(0, 0, 0)) src = gr.vector_source_f(expected_data) dst = gr.vector_sink_f() hb.connect(src, dst) tb.connect(hb) tb.run() self.assertEquals(expected_data, dst.data()) def test_027a_internally_unconnected_input(self): tb = gr.top_block() hb = gr.hier_block2("block", gr.io_signature(1, 1, 1), gr.io_signature(1, 1, 1)) hsrc = gr.vector_source_b([1,]) hb.connect(hsrc, hb) # wire output internally src = gr.vector_source_b([1, ]) dst = gr.vector_sink_b() tb.connect(src, hb, dst) # hb's input is not connected internally self.assertRaises(RuntimeError, lambda: tb.run()) def test_027b_internally_unconnected_output(self): tb = gr.top_block() hb = gr.hier_block2("block", gr.io_signature(1, 1, 1), gr.io_signature(1, 1, 1)) hdst = gr.vector_sink_b() hb.connect(hb, hdst) # wire input internally src = gr.vector_source_b([1, ]) dst = gr.vector_sink_b() tb.connect(src, hb, dst) # hb's output is not connected internally self.assertRaises(RuntimeError, lambda: tb.run()) def test_027c_fully_unconnected_output(self): tb = gr.top_block() hb = gr.hier_block2("block", gr.io_signature(1, 1, 1), gr.io_signature(1, 1, 1)) hsrc = gr.vector_sink_b() hb.connect(hb, hsrc) # wire input internally src = gr.vector_source_b([1, ]) dst = gr.vector_sink_b() tb.connect(src, hb) # hb's output is not connected internally or externally self.assertRaises(RuntimeError, lambda: tb.run()) def test_027d_fully_unconnected_input(self): tb = gr.top_block() hb = gr.hier_block2("block", gr.io_signature(1, 1, 1), gr.io_signature(1, 1, 1)) hdst = gr.vector_source_b([1,]) hb.connect(hdst, hb) # wire output internally dst = gr.vector_sink_b() tb.connect(hb, dst) # hb's input is not connected internally or externally self.assertRaises(RuntimeError, lambda: tb.run()) def test_028_singleton_reconfigure(self): tb = gr.top_block() hb = gr.hier_block2("block", gr.io_signature(0, 0, 0), gr.io_signature(0, 0, 0)) src = gr.vector_source_b([1, ]) dst = gr.vector_sink_b() hb.connect(src, dst) tb.connect(hb) # Singleton connect tb.lock() tb.disconnect_all() tb.connect(src, dst) tb.unlock() def test_029_singleton_disconnect(self): tb = gr.top_block() src = gr.vector_source_b([1, ]) dst = gr.vector_sink_b() tb.connect(src, dst) tb.disconnect(src) # Singleton disconnect tb.connect(src, dst) tb.run() self.assertEquals(dst.data(), (1,)) def test_030_nested_input(self): tb = gr.top_block() src = gr.vector_source_b([1,]) hb1 = gr.hier_block2("hb1", gr.io_signature(1, 1, gr.sizeof_char), gr.io_signature(0, 0, 0)) hb2 = gr.hier_block2("hb2", gr.io_signature(1, 1, gr.sizeof_char), gr.io_signature(0, 0, 0)) dst = gr.vector_sink_b() tb.connect(src, hb1) hb1.connect(hb1, hb2) hb2.connect(hb2, gr.kludge_copy(gr.sizeof_char), dst) tb.run() self.assertEquals(dst.data(), (1,)) def test_031_multiple_internal_inputs(self): tb = gr.top_block() src = gr.vector_source_f([1.0,]) hb = gr.hier_block2("hb", gr.io_signature(1, 1, gr.sizeof_float), gr.io_signature(1, 1, gr.sizeof_float)) m1 = gr.multiply_const_ff(1.0) m2 = gr.multiply_const_ff(2.0) add = gr.add_ff() hb.connect(hb, m1) # m1 is connected to hb external input #0 hb.connect(hb, m2) # m2 is also connected to hb external input #0 hb.connect(m1, (add, 0)) hb.connect(m2, (add, 1)) hb.connect(add, hb) # add is connected to hb external output #0 dst = gr.vector_sink_f() tb.connect(src, hb, dst) tb.run() self.assertEquals(dst.data(), (3.0,)) def test_032_nested_multiple_internal_inputs(self): tb = gr.top_block() src = gr.vector_source_f([1.0,]) hb = gr.hier_block2("hb", gr.io_signature(1, 1, gr.sizeof_float), gr.io_signature(1, 1, gr.sizeof_float)) hb2 = gr.hier_block2("hb", gr.io_signature(1, 1, gr.sizeof_float), gr.io_signature(1, 1, gr.sizeof_float)) m1 = gr.multiply_const_ff(1.0) m2 = gr.multiply_const_ff(2.0) add = gr.add_ff() hb2.connect(hb2, m1) # m1 is connected to hb2 external input #0 hb2.connect(hb2, m2) # m2 is also connected to hb2 external input #0 hb2.connect(m1, (add, 0)) hb2.connect(m2, (add, 1)) hb2.connect(add, hb2) # add is connected to hb2 external output #0 hb.connect(hb, hb2, hb) # hb as hb2 as nested internal block dst = gr.vector_sink_f() tb.connect(src, hb, dst) tb.run() self.assertEquals(dst.data(), (3.0,)) if __name__ == "__main__": gr_unittest.run(test_hier_block2, "test_hier_block2.xml")
gpl-3.0
chubbymaggie/claripy
claripy/vsa/bool_result.py
2
5012
from ..backend_object import BackendObject class BoolResult(BackendObject): def __init__(self, op=None, args=None): self._op = op self._args = args def value(self): raise NotImplementedError() def __len__(self): return BackendError() def __eq__(self, other): raise NotImplementedError() def __and__(self, other): raise NotImplementedError() def __invert__(self): raise NotImplementedError() def __or__(self, other): raise NotImplementedError() def identical(self, other): if self.value != other.value: return False if self._op != other._op: return False if self._args != other._args: return False return True def union(self, other): raise NotImplementedError() def size(self): #pylint:disable=no-self-use return None @staticmethod def is_maybe(o): if isinstance(o, Base): raise ClaripyValueError("BoolResult can't handle AST objects directly") return isinstance(o, MaybeResult) @staticmethod def has_true(o): if isinstance(o, Base): raise ClaripyValueError("BoolResult can't handle AST objects directly") return o is True or (isinstance(o, BoolResult) and True in o.value) @staticmethod def has_false(o): if isinstance(o, Base): raise ClaripyValueError("BoolResult can't handle AST objects directly") return o is False or (isinstance(o, BoolResult) and False in o.value) @staticmethod def is_true(o): if isinstance(o, Base): raise ClaripyValueError("BoolResult can't handle AST objects directly") return o is True or (isinstance(o, TrueResult)) @staticmethod def is_false(o): if isinstance(o, Base): raise ClaripyValueError("BoolResult can't handle AST objects directly") return o is False or (isinstance(o, FalseResult)) class TrueResult(BoolResult): cardinality = 1 @property def value(self): return (True, ) def identical(self, other): return isinstance(other, TrueResult) def __eq__(self, other): if isinstance(other, FalseResult): return FalseResult() elif isinstance(other, TrueResult): return TrueResult() else: return MaybeResult() def __invert__(self): return FalseResult() def __or__(self, other): return TrueResult() def __and__(self, other): if BoolResult.is_maybe(other): return MaybeResult() elif BoolResult.is_false(other): return FalseResult() else: return TrueResult() def union(self, other): if other is True or type(other) is TrueResult: return TrueResult() elif other is False or type(other) is FalseResult: return MaybeResult() elif type(other) is MaybeResult: return MaybeResult() else: return NotImplemented def __repr__(self): return '<True>' class FalseResult(BoolResult): cardinality = 1 @property def value(self): return (False, ) def identical(self, other): return isinstance(other, FalseResult) def __eq__(self, other): if isinstance(other, FalseResult): return TrueResult() elif isinstance(other, TrueResult): return FalseResult() else: return MaybeResult() def __invert__(self): return TrueResult() def __and__(self, other): return FalseResult() def __or__(self, other): return other def __repr__(self): return '<False>' def union(self, other): if other is True or type(other) is TrueResult: return MaybeResult() elif other is False or type(other) is FalseResult: return FalseResult() elif type(other) is MaybeResult: return MaybeResult() else: return NotImplemented class MaybeResult(BoolResult): cardinality = 2 @property def value(self): return (True, False) def identical(self, other): return isinstance(other, MaybeResult) def __eq__(self, other): return MaybeResult() def __invert__(self): return MaybeResult() def __and__(self, other): if BoolResult.is_false(other): return FalseResult() else: return MaybeResult() def union(self, other): return MaybeResult() def __or__(self, other): if BoolResult.is_true(other): return TrueResult() else: return self def __repr__(self): if self._op is None: return '<Maybe>' else: return '<Maybe(%s, %s)>' % (self._op, self._args) from ..errors import BackendError, ClaripyValueError from ..ast.base import Base
bsd-2-clause
emilyvon/titanium_mobile
support/android/android.py
33
15397
#!/usr/bin/env python # -*- coding: utf-8 -*- # # Appcelerator Titanium Mobile # Copyright (c) 2011-2013 by Appcelerator, Inc. All Rights Reserved. # Licensed under the terms of the Apache Public License # Please see the LICENSE included with this distribution for details. # # Android Application Script # import os, sys, shutil, platform, zipfile import string, subprocess, re from xml.etree.ElementTree import ElementTree from StringIO import StringIO from os.path import join, splitext, split, exists from shutil import copyfile from androidsdk import AndroidSDK from compiler import Compiler import bindings template_dir = os.path.abspath(os.path.dirname(sys._getframe(0).f_code.co_filename)) module_dir = os.path.join(os.path.dirname(template_dir), 'module') common_dir = os.path.join(os.path.dirname(template_dir), 'common') sys.path.extend([os.path.dirname(template_dir), module_dir, common_dir]) from mako.template import Template from tiapp import TiAppXML, touch_tiapp_xml from manifest import Manifest from module import ModuleDetector import simplejson ignoreFiles = ['.gitignore', '.cvsignore', '.DS_Store']; ignoreDirs = ['.git','.svn','_svn', 'CVS']; def run(args): return subprocess.Popen(args, stderr=subprocess.STDOUT, stdout=subprocess.PIPE).communicate()[0] def pipe(args1,args2): p1 = subprocess.Popen(args1, stdout=subprocess.PIPE) p2 = subprocess.Popen(args2, stdin=p1.stdout, stdout=subprocess.PIPE) return p2.communicate()[0] def copy_resources(source, target): if not os.path.exists(os.path.expanduser(target)): os.mkdir(os.path.expanduser(target)) for root, dirs, files in os.walk(source, True, None, True): for name in ignoreDirs: if name in dirs: dirs.remove(name) # don't visit ignored directories for file in files: if file in ignoreFiles: continue from_ = join(root, file) to_ = os.path.expanduser(from_.replace(source, target, 1)) to_directory = os.path.expanduser(split(to_)[0]) if not exists(to_directory): os.makedirs(to_directory) print "[TRACE] copying: %s to: %s" % (from_,to_) copyfile(from_, to_) class Android(object): def __init__(self, name, myid, sdk, deploy_type, java): self.name = name # android requires at least one dot in packageid if len(re.findall(r'\.',myid))==0: myid = 'com.%s' % myid self.id = myid self.sdk = sdk # Used in templating self.config = { 'appid': self.id, 'appname' : self.name, 'appversion' : '1', 'apiversion' : '7', #Android 2.1 'deploy_type': deploy_type, 'compile_js': False } self.config['classname'] = Android.strip_classname(self.name) self.deploy_type = deploy_type self.java = java @classmethod def strip_classname(cls, name): classname = ''.join([str.capitalize() for str in re.split('[^A-Za-z0-9_]', name)]) if re.search("^[0-9]", classname) != None: classname = "_" + classname return classname def newdir(self, *segments): path = os.path.join(*segments) if not os.path.exists(path): os.makedirs(path) return path def copyfile(self, file, src, dest): shutil.copy(os.path.join(src, file), os.path.join(dest, file)) def load_template(self, template): return Template(filename=template, output_encoding='utf-8', encoding_errors='replace') def render_android_manifest(self): template_dir = os.path.dirname(sys._getframe(0).f_code.co_filename) tmpl = self.load_template(os.path.join(template_dir, 'templates', 'AndroidManifest.xml')) return tmpl.render(config = self.config) def render(self, template_dir, template_file, dest, dest_file, **kwargs): tmpl = self.load_template(os.path.join(template_dir, 'templates', template_file)) f = None try: print "[TRACE] Generating %s" % os.path.join(dest, dest_file) f = open(os.path.join(dest, dest_file), "w") f.write(tmpl.render(config = self.config, **kwargs)) finally: if f!=None: f.close def build_app_info(self, project_dir): tiapp = ElementTree() assets_tiappxml = os.path.join(project_dir, 'build', 'android', 'bin', 'assets', 'tiapp.xml') self.app_info = {'fullscreen':'false','navbar-hidden':'false','deploy-type':self.deploy_type} self.app_properties = {} if not os.path.exists(assets_tiappxml): shutil.copy(os.path.join(project_dir, 'tiapp.xml'), assets_tiappxml) tiapp.parse(open(assets_tiappxml, 'r')) for key in ['id', 'name', 'version', 'publisher', 'url', 'copyright', 'description', 'icon', 'analytics', 'guid', 'navbar-hidden', 'fullscreen']: el = tiapp.find(key) if el != None: self.app_info[key] = el.text for property_el in tiapp.findall("property"): name = property_el.get("name") type = property_el.get("type") value = property_el.text if name == None: continue if type == None: type = "string" if value == None: value = "" self.app_properties[name] = {"type": type, "value": value} def generate_activities(self, app_package_dir): if not 'activities' in self.tiapp.android: return for key in self.tiapp.android['activities'].keys(): activity = self.tiapp.android['activities'][key] print '[DEBUG] generating activity class: ' + activity['classname'] self.render(template_dir, 'JSActivity.java', app_package_dir, activity['classname']+'.java', activity=activity) def generate_services(self, app_package_dir): if not 'services' in self.tiapp.android: return for key in self.tiapp.android['services'].keys(): service = self.tiapp.android['services'][key] service_type = service['service_type'] print '[DEBUG] generating service type "%s", class "%s"' %(service_type, service['classname']) if service_type == 'interval': self.render(template_dir, 'JSIntervalService.java', app_package_dir, service['classname']+'.java', service=service) else: self.render(template_dir, 'JSService.java', app_package_dir, service['classname']+'.java', service=service) def build_modules_info(self, resources_dir, app_bin_dir, include_all_ti_modules=False): self.app_modules = [] (modules, external_child_modules) = bindings.get_all_module_bindings() compiler = Compiler(self.tiapp, resources_dir, self.java, app_bin_dir, None, os.path.dirname(app_bin_dir), include_all_modules=include_all_ti_modules) compiler.compile(compile_bytecode=False, info_message=None) for module in compiler.modules: module_bindings = [] # TODO: we should also detect module properties for method in compiler.module_methods: if method.lower().startswith(module+'.') and '.' not in method: module_bindings.append(method[len(module)+1:]) module_onAppCreate = None module_class = None module_apiName = None for m in modules.keys(): if modules[m]['fullAPIName'].lower() == module: module_class = m module_apiName = modules[m]['fullAPIName'] if 'onAppCreate' in modules[m]: module_onAppCreate = modules[m]['onAppCreate'] break if module_apiName == None: continue # module wasn't found ext_modules = [] if module_class in external_child_modules: for child_module in external_child_modules[module_class]: if child_module['fullAPIName'].lower() in compiler.modules: ext_modules.append(child_module) self.app_modules.append({ 'api_name': module_apiName, 'class_name': module_class, 'bindings': module_bindings, 'external_child_modules': ext_modules, 'on_app_create': module_onAppCreate }) # discover app modules detector = ModuleDetector(self.project_dir) missing, detected_modules = detector.find_app_modules(self.tiapp, 'android', self.deploy_type) for missing_module in missing: print '[WARN] Couldn\'t find app module: %s' % missing_module['id'] self.custom_modules = [] for module in detected_modules: if module.jar == None: continue module_jar = zipfile.ZipFile(module.jar) module_bindings = bindings.get_module_bindings(module_jar) if module_bindings is None: continue for module_class in module_bindings['modules'].keys(): module_apiName = module_bindings['modules'][module_class]['apiName'] module_proxy = module_bindings['proxies'][module_class] module_id = module_proxy['proxyAttrs']['id'] module_proxy_class_name = module_proxy['proxyClassName'] module_onAppCreate = None if 'onAppCreate' in module_proxy: module_onAppCreate = module_proxy['onAppCreate'] print '[DEBUG] module_id = %s' % module_id if module_id == module.manifest.moduleid: # make sure that the module was not built before 1.8.0.1 try: module_api_version = int(module.manifest.apiversion) if module_api_version < 2: print "[ERROR] The 'apiversion' for '%s' in the module manifest is less than version 2. The module was likely built against a Titanium SDK pre 1.8.0.1. Please use a version of the module that has 'apiversion' 2 or greater" % module_id touch_tiapp_xml(os.path.join(self.project_dir, 'tiapp.xml')) sys.exit(1) except(TypeError, ValueError): print "[ERROR] The 'apiversion' for '%s' in the module manifest is not a valid value. Please use a version of the module that has an 'apiversion' value of 2 or greater set in it's manifest file" % module_id touch_tiapp_xml(os.path.join(self.project_dir, 'tiapp.xml')) sys.exit(1) is_native_js_module = (hasattr(module.manifest, 'commonjs') and module.manifest.commonjs) print '[DEBUG] appending module: %s' % module_class self.custom_modules.append({ 'module_id': module_id, 'module_apiName': module_apiName, 'proxy_name': module_proxy_class_name, 'class_name': module_class, 'manifest': module.manifest, 'on_app_create': module_onAppCreate, 'is_native_js_module': is_native_js_module }) if is_native_js_module: # Need to look at the app modules used in this external js module metadata_file = os.path.join(module.path, "metadata.json") metadata = None try: f = open(metadata_file, "r") metadata = f.read() finally: f.close() if metadata: metadata = simplejson.loads(metadata) if metadata.has_key("exports"): exported_module_ids = metadata["exports"] already_included_module_ids = [m["api_name"].lower() for m in self.app_modules] need_to_add = [m for m in exported_module_ids if m not in already_included_module_ids] if need_to_add: for to_add in need_to_add: module_onAppCreate = None module_class = None module_apiName = None for m in modules.keys(): if modules[m]['fullAPIName'].lower() == to_add: module_class = m module_apiName = modules[m]['fullAPIName'] if 'onAppCreate' in modules[m]: module_onAppCreate = modules[m]['onAppCreate'] break if module_apiName == None: continue # module wasn't found ext_modules = [] if module_class in external_child_modules: for child_module in external_child_modules[module_class]: if child_module['fullAPIName'].lower() in compiler.modules: ext_modules.append(child_module) self.app_modules.append({ 'api_name': module_apiName, 'class_name': module_class, 'bindings': [], 'external_child_modules': ext_modules, 'on_app_create': module_onAppCreate }) def create(self, dir, build_time=False, project_dir=None, include_all_ti_modules=False): template_dir = os.path.dirname(sys._getframe(0).f_code.co_filename) # Build up output directory tree if project_dir is None: project_dir = self.newdir(dir, self.name) self.project_dir = project_dir # Paths to Titanium assets that need to be linked into eclipse structure self.config['ti_tiapp_xml'] = os.path.join(project_dir, 'tiapp.xml') self.tiapp = TiAppXML(self.config['ti_tiapp_xml']) resource_dir = os.path.join(project_dir, 'Resources') self.config['ti_resources_dir'] = resource_dir json_contents = open(os.path.join(template_dir,'dependency.json')).read() depends_map = simplejson.loads(json_contents) app_build_dir = self.newdir(project_dir, 'build') app_dir = self.newdir(app_build_dir, 'android') #if os.path.exists(os.path.join(app_dir,'bin')): # shutil.rmtree(os.path.join(app_dir,'bin')) if os.path.exists(os.path.join(app_dir,'src')): shutil.rmtree(os.path.join(app_dir,'src')) if os.path.exists(os.path.join(app_dir,'res')): shutil.rmtree(os.path.join(app_dir,'res')) app_bin_dir = self.newdir(app_dir, 'bin') app_lib_dir = self.newdir(app_dir, 'lib') app_src_dir = self.newdir(app_dir, 'src') app_res_dir = self.newdir(app_dir, 'res') app_gen_dir = self.newdir(app_dir, 'gen') app_bin_classes_dir = self.newdir(app_bin_dir, 'classes') app_res_drawable_dir = self.newdir(app_res_dir, 'drawable') app_assets_dir = self.newdir(app_dir, 'assets') app_package_dir = self.newdir(app_gen_dir, *self.id.split('.')) app_bin_assets_dir = self.newdir(app_bin_dir, 'assets') app_bin_assets_resources_dir = self.newdir(app_bin_assets_dir, 'Resources') self.build_app_info(project_dir) self.build_modules_info(app_bin_assets_resources_dir, app_bin_dir, include_all_ti_modules=include_all_ti_modules) # Create android source self.render(template_dir, 'AppInfo.java', app_package_dir, self.config['classname'] + 'AppInfo.java', app_properties = self.app_properties, app_info = self.app_info) self.render(template_dir, 'AndroidManifest.xml', app_dir, 'AndroidManifest.xml') self.render(template_dir, 'App.java', app_package_dir, self.config['classname'] + 'Application.java', app_modules = self.app_modules, custom_modules = self.custom_modules) self.render(template_dir, 'Activity.java', app_package_dir, self.config['classname'] + 'Activity.java') self.generate_activities(app_package_dir) self.generate_services(app_package_dir) self.render(template_dir, 'classpath', app_dir, '.classpath') self.render(template_dir, 'project', app_dir, '.project') self.render(template_dir, 'default.properties', app_dir, 'default.properties') print "[TRACE] Generating app.json" f = None try: f = open(os.path.join(app_bin_assets_dir, "app.json"), "w") f.write(simplejson.dumps({"app_modules":self.app_modules})) finally: if f is not None: f.close() # Don't override a pre-existing .gitignore in case users have their own preferences # for what should be in it. (LH #2446) if not os.path.exists(os.path.join(app_dir, '.gitignore')): self.render(template_dir, 'gitignore', app_dir, '.gitignore') else: print "[TRACE] Skipping copying gitignore -> .gitignore because already exists" android_project_resources = os.path.join(project_dir,'Resources','android') if build_time==False and os.path.exists(android_project_resources): shutil.rmtree(android_project_resources) if not os.path.exists(android_project_resources): copy_resources(os.path.join(template_dir,'resources'),android_project_resources) if __name__ == '__main__': # this is for testing only for the time being if len(sys.argv) != 5 or sys.argv[1]=='--help': print "Usage: %s <name> <id> <directory> <sdk>" % os.path.basename(sys.argv[0]) sys.exit(1) sdk = AndroidSDK(sys.argv[4]) android = Android(sys.argv[1], sys.argv[2], sdk, None, 'java') android.create(sys.argv[3])
apache-2.0
will4906/PatentCrawler
config/crawler_config.py
1
8899
# -*- coding: utf-8 -*- """ Created on 2018/3/14 @author: will4906 采集的内容、方式定义 """ from bs4 import BeautifulSoup from controller.url_config import url_search, url_detail, url_related_info, url_full_text from crawler.items import DataItem from entity.crawler_item import BaseItem, ResultItem class PatentId(BaseItem): is_required = True crawler_id = url_search.get('crawler_id') english = 'patent_id' chinese = ['专利标志', '专利id', '专利ID', '专利Id'] @classmethod def parse(cls, raw, item, process=None): if process is not None: patent_id = process.find(attrs={'name': 'idHidden'}).get('value') item.patent_id = ResultItem(title=cls.title, value=str(patent_id)) return item class PatentName(BaseItem): is_required = True crawler_id = url_detail.get('crawler_id') english = ['patent_name', 'invention_name'] chinese = '专利名称' @classmethod def parse(cls, raw, item, process=None): if process is not None: patent_name = process.get('abstractInfoDTO').get('tioIndex').get('value') item.patent_name = ResultItem(title=cls.title, value=str(patent_name)) return item class Abstract(BaseItem): crawler_id = url_detail.get('crawler_id') english = 'abstract' chinese = '摘要' @classmethod def parse(cls, raw, item, process=None): if process is not None: abstract = BeautifulSoup(process.get('abstractInfoDTO').get('abIndexList')[0].get('value'), 'lxml').text.replace('\n', '').strip() item.abstract = ResultItem(title=cls.title, value=abstract) return item def push_item(json_list, item: DataItem, title, name): """ detail页面的解析函数 :param json_list: :param item: :param title: :param name: :return: """ if json_list is not None: aitem_list = json_list.get('abstractInfoDTO').get('abstractItemList') for a_item in aitem_list: if a_item.get('indexCnName').find(name) != -1: item.__setattr__(title, ResultItem(title=name, value=a_item.get('value'))) break if not hasattr(item, title): item.__setattr__(title, ResultItem(title=name, value="")) return item class RequestNumber(BaseItem): crawler_id = url_detail.get('crawler_id') english = ['request_number', 'application_number'] chinese = '申请号' @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'request_number', '申请号') class RequestDate(BaseItem): crawler_id = url_detail.get('crawler_id') english = 'request_date' chinese = '申请日' @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'request_date', '申请日') class PublishNumber(BaseItem): crawler_id = url_detail.get('crawler_id') english = 'publish_number' chinese = ['公开号', '公布号', '公开(公告)号'] @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'publish_number', '公开(公告)号') class PublishDate(BaseItem): crawler_id = url_detail.get('crawler_id') english = 'publish_date' chinese = ['公开日', '公布日', '公开(公告)日'] @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'publish_date', '公开(公告)日') class IpcClassificationNumber(BaseItem): crawler_id = url_detail.get('crawler_id') english = ['ipc_class_number', 'IPC', 'ipc', 'Ipc'] chinese = 'IPC分类号' @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'ipc_class_number', 'IPC分类号') class Applicant(BaseItem): crawler_id = url_detail.get('crawler_id') english = ['Applicant', 'applicant', 'assignee', 'Assignee', 'proposer'] chinese = ['申请人', '专利权人', '专利人', '申请(专利权)人'] @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'applicant', '申请(专利权)人') class Inventor(BaseItem): crawler_id = url_detail.get('crawler_id') english = ['Inventor', 'inventor'] chinese = '发明人' @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'inventor', '发明人') class PriorityNumber(BaseItem): crawler_id = url_detail.get('crawler_id') english = 'priority_number' chinese = '优先权号' @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'priority_number', '优先权号') class PriorityDate(BaseItem): crawler_id = url_detail.get('crawler_id') english = 'priority_date' chinese = '优先权日' @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'priority_date', '优先权日') class AddressOfApplicant(BaseItem): crawler_id = url_detail.get('crawler_id') english = ['proposer_address', 'address_of_the_Applicant', 'applicant_address'] chinese = '申请人地址' @classmethod def parse(cls, raw, item, process=None): if process is not None: item = push_item(process, item, 'proposer_address', '申请人地址') return item class ZipCodeOfTheApplicant(BaseItem): crawler_id = url_detail.get('crawler_id') english = ['proposer_post_code', 'zip_code_of_the_applicant', 'proposer_zip_code'] chinese = '申请人邮编' @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'proposer_zip_code', '申请人邮编') class CountryOfTheApplicant(BaseItem): crawler_id = url_detail.get('crawler_id') english = ['proposer_location', 'country_of_the_applicant', 'country_of_the_assignee'] chinese = ['申请人所在国(省)', '申请人所在地'] @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'proposer_location', '申请人所在国(省)') class CpcClassificationNumber(BaseItem): crawler_id = url_detail.get('crawler_id') english = ['cpc_class_number', 'cpc', 'CPC', 'Cpc'] chinese = 'CPC分类号' @classmethod def parse(cls, raw, item, process=None): return push_item(process, item, 'cpc_class_number', 'CPC分类号') class Cognation(BaseItem): crawler_id = url_related_info.get('crawler_id') table_name = 'cognation' english = 'cognation_list' chinese = '同族表' title = '同族' @classmethod def parse(cls, raw, item, process=None): if process is not None: cognation_list = process.get('cognationList') # print('cognation', cognation_list) if cognation_list is not None: pn_list = [] for cog in cognation_list: pn_list.append(cog.get('pn')) item.cognation_list = ResultItem(table=cls.table_name, title=cls.title, value=pn_list) return item class LawStateList(BaseItem): crawler_id = url_related_info.get('crawler_id') table_name = 'law_state' english = 'law_state_list' chinese = '法律状态表' title = ['法律状态', '法律状态时间'] @classmethod def set_title(cls, title): if title == cls.english: cls.title = ['law_status', 'law_status_date'] elif title == cls.chinese: cls.title = ['法律状态', '法律状态时间'] @classmethod def parse(cls, raw, item, process=None): if process is not None: law_state_list = process.get('lawStateList') if law_state_list is not None: tmp_list = [] for law in law_state_list: mean = law.get('lawStateCNMeaning') law_date = law.get('prsDate') part = (ResultItem(table=cls.table_name, title=cls.title[0], value=mean), ResultItem(table=cls.table_name, title=cls.title[1], value=law_date)) tmp_list.append(part) item.law_state_list = tmp_list return item class FullText(BaseItem): crawler_id = url_full_text.get('crawler_id') english = ['full_text', 'whole_text'] chinese = ['全文文本', '全文'] @classmethod def parse(cls, raw, item, process=None): if process is not None: item.full_text = ResultItem(table=cls.table_name, title=cls.title, value=BeautifulSoup(str(process.get('fullTextDTO').get('literaInfohtml')), 'lxml') .get_text().replace("'", '"').replace(';', ',')) return item
apache-2.0
w1ll1am23/home-assistant
tests/components/minio/common.py
27
1804
"""Minio Test event.""" TEST_EVENT = { "Records": [ { "eventVersion": "2.0", "eventSource": "minio:s3", "awsRegion": "", "eventTime": "2019-05-02T11:05:07Z", "eventName": "s3:ObjectCreated:Put", "userIdentity": {"principalId": "SO9KNO6YT9OGE39PQCZW"}, "requestParameters": { "accessKey": "SO9KNO6YT9OGE39PQCZW", "region": "", "sourceIPAddress": "172.27.0.1", }, "responseElements": { "x-amz-request-id": "159AD8E6F6805783", "x-minio-deployment-id": "90b265b8-bac5-413a-b12a-8915469fd769", "x-minio-origin-endpoint": "http://172.27.0.2:9000", }, "s3": { "s3SchemaVersion": "1.0", "configurationId": "Config", "bucket": { "name": "test", "ownerIdentity": {"principalId": "SO9KNO6YT9OGE39PQCZW"}, "arn": "arn:aws:s3:::test", }, "object": { "key": "5jJkTAo.jpg", "size": 108368, "eTag": "1af324731637228cbbb0b2e8c07d4e50", "contentType": "image/jpeg", "userMetadata": {"content-type": "image/jpeg"}, "versionId": "1", "sequencer": "159AD8E6F76DD9C4", }, }, "source": { "host": "", "port": "", "userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_3) " "AppleWebKit/605.1.15 (KHTML, like Gecko) " "Version/12.0.3 Safari/605.1.15", }, } ] }
apache-2.0
sunlianqiang/kbengine
kbe/src/lib/python/Lib/distutils/tests/test_config.py
90
3148
"""Tests for distutils.pypirc.pypirc.""" import sys import os import unittest import tempfile from distutils.core import PyPIRCCommand from distutils.core import Distribution from distutils.log import set_threshold from distutils.log import WARN from distutils.tests import support from test.support import run_unittest PYPIRC = """\ [distutils] index-servers = server1 server2 [server1] username:me password:secret [server2] username:meagain password: secret realm:acme repository:http://another.pypi/ """ PYPIRC_OLD = """\ [server-login] username:tarek password:secret """ WANTED = """\ [distutils] index-servers = pypi [pypi] username:tarek password:xxx """ class PyPIRCCommandTestCase(support.TempdirManager, support.LoggingSilencer, support.EnvironGuard, unittest.TestCase): def setUp(self): """Patches the environment.""" super(PyPIRCCommandTestCase, self).setUp() self.tmp_dir = self.mkdtemp() os.environ['HOME'] = self.tmp_dir self.rc = os.path.join(self.tmp_dir, '.pypirc') self.dist = Distribution() class command(PyPIRCCommand): def __init__(self, dist): PyPIRCCommand.__init__(self, dist) def initialize_options(self): pass finalize_options = initialize_options self._cmd = command self.old_threshold = set_threshold(WARN) def tearDown(self): """Removes the patch.""" set_threshold(self.old_threshold) super(PyPIRCCommandTestCase, self).tearDown() def test_server_registration(self): # This test makes sure PyPIRCCommand knows how to: # 1. handle several sections in .pypirc # 2. handle the old format # new format self.write_file(self.rc, PYPIRC) cmd = self._cmd(self.dist) config = cmd._read_pypirc() config = list(sorted(config.items())) waited = [('password', 'secret'), ('realm', 'pypi'), ('repository', 'https://pypi.python.org/pypi'), ('server', 'server1'), ('username', 'me')] self.assertEqual(config, waited) # old format self.write_file(self.rc, PYPIRC_OLD) config = cmd._read_pypirc() config = list(sorted(config.items())) waited = [('password', 'secret'), ('realm', 'pypi'), ('repository', 'https://pypi.python.org/pypi'), ('server', 'server-login'), ('username', 'tarek')] self.assertEqual(config, waited) def test_server_empty_registration(self): cmd = self._cmd(self.dist) rc = cmd._get_rc_file() self.assertFalse(os.path.exists(rc)) cmd._store_pypirc('tarek', 'xxx') self.assertTrue(os.path.exists(rc)) f = open(rc) try: content = f.read() self.assertEqual(content, WANTED) finally: f.close() def test_suite(): return unittest.makeSuite(PyPIRCCommandTestCase) if __name__ == "__main__": run_unittest(test_suite())
lgpl-3.0
moccu/django-cruditor
tests/test_views.py
1
15414
import pytest from django.contrib.auth.models import AnonymousUser from django.contrib.messages import SUCCESS as SUCCESS_LEVEL from django.contrib.messages import get_messages from django.core.exceptions import ImproperlyConfigured, PermissionDenied from django.urls import reverse from cruditor.views import Cruditor403View, Cruditor404View, CruditorListView from examples.minimal.views import DemoView from examples.store.models import Person from .factories import PersonFactory, RelatedPersonFactory @pytest.mark.django_db class TestBasicView: @pytest.fixture(autouse=True) def setup(self, rf, admin_user): self.request = rf.get('/') self.request.user = admin_user self.view = DemoView() self.view.request = self.request self.view.required_permission = 'accounts.some_permission' def test_not_logged_in(self, rf): self.request.user = AnonymousUser() response = self.view.dispatch(self.request) assert response.status_code == 200 assert 'breadcrumb' not in response.context_data['cruditor'] assert response.template_name[0] == DemoView.login_template_name def test_no_permission(self, admin_user): self.request.user.is_superuser = False with pytest.raises(PermissionDenied): self.view.dispatch(self.request) def test_no_permission_required(self, admin_user): self.view.required_permission = None self.request.user.is_superuser = False response = self.view.dispatch(self.request) assert response.status_code == 200 def test_permission_granted(self, admin_user): response = self.view.dispatch(self.request) assert response.status_code == 200 assert response.template_name[0] == 'minimal/demo.html' def test_cruditor_context(self): assert self.view.get_cruditor_context() == { 'breadcrumb': [ {'title': 'Additional breadcrumb', 'url': '/'}, {'title': 'Disabled item'}, {'title': 'Demo view', 'url': None} ], 'constants': { 'change_password_url': '/change-password/', 'extrahead_template_name': 'cruditor/includes/extrahead.html', 'index_url': '/', 'logout_url': '/logout/', 'menu_template_name': 'menu.html', 'menu_title': 'Examples Demo' }, 'title': 'Demo view', 'titlebuttons': None } def test_title(self): assert self.view.get_title() == 'Demo view' def test_model_verbose_name_explicit(self): self.view.model_verbose_name = 'Foo' assert self.view.get_model_verbose_name() == 'Foo' def test_model_verbose_name_from_meta(self): self.view.model = Person assert self.view.get_model_verbose_name() == 'Person' def test_model_verbose_name_fallback(self): assert self.view.get_model_verbose_name() == 'Item' def test_not_found_view(rf): response = Cruditor404View.as_view()(rf.get('/')) assert response.status_code == 404 assert response.template_name[0] == 'cruditor/404.html' def test_forbidden_view(rf): response = Cruditor403View.as_view()(rf.get('/')) assert response.status_code == 403 assert response.template_name[0] == 'cruditor/403.html' @pytest.mark.django_db class TestListView: def setup(self): self.person1 = PersonFactory.create(approved=True) self.person2 = PersonFactory.create(approved=False) def test_get_without_filter(self, admin_client): response = admin_client.get(reverse('collection:list')) assert response.status_code == 200 assert response.context['table'].data.data.count() == 2 assert response.context['filter_form'] is None def test_get_with_filter(self, admin_client): response = admin_client.get(reverse('collection:filter')) assert response.status_code == 200 assert response.context['table'].data.data.count() == 2 assert response.context['filter_form'] is not None assert not response.context['filter_form'].data def test_get_with_filter_active(self, admin_client): response = admin_client.get( reverse('collection:filter'), data={'approved': '2'}) assert response.status_code == 200 assert response.context['table'].data.data.count() == 1 assert response.context['filter_form'] is not None assert response.context['filter_form'].data def test_get_queryset_model(self): class DummyListView(CruditorListView): model = Person assert list(DummyListView().get_queryset()) == list(Person.objects.all()) def test_get_queryset_qset(self): class DummyListView(CruditorListView): queryset = Person.objects.filter(approved=True) assert DummyListView().get_queryset().get() == self.person1 def test_get_queryset_fallback(self): class DummyListView(CruditorListView): pass assert DummyListView().get_queryset() == [] def test_get_table_class(self): class DummyListView(CruditorListView): table_class = object assert DummyListView().get_table_class() def test_get_table_class_invalid(self): class DummyListView(CruditorListView): pass with pytest.raises(ImproperlyConfigured): DummyListView().get_table_class() @pytest.mark.django_db class TestAddView: def test_get(self, admin_client): response = admin_client.get(reverse('collection:add')) assert response.status_code == 200 assert response.context['cruditor']['title'] == 'Add Person' def test_post_valid(self, admin_client): response = admin_client.post( reverse('collection:add'), data={ 'first_name': 'John', 'last_name': 'Doe', 'country': 'Germany', 'reminder_0': '2018-05-25', 'reminder_1': '09:00:00', 'stars': '2', } ) assert response.status_code == 302 assert response['Location'] == reverse('collection:list') messages = list(get_messages(response.wsgi_request)) assert len(messages) == 1 assert messages[0].level == SUCCESS_LEVEL assert Person.objects.get().first_name == 'John' def test_post_invalid(self, admin_client): response = admin_client.post(reverse('collection:add'), data={}) assert response.status_code == 200 assert response.context['form'].is_valid() is False assert Person.objects.exists() is False @pytest.mark.django_db class TestChangeView: def setup(self): self.person = PersonFactory.create(first_name='Sally') def test_get(self, admin_client): response = admin_client.get( reverse('collection:change', args=(self.person.pk,))) assert response.status_code == 200 assert response.context['form'].instance == self.person def test_post_valid(self, admin_client): response = admin_client.post( reverse('collection:change', args=(self.person.pk,)), data={ 'first_name': 'John', 'last_name': 'Doe', 'country': 'Germany', 'reminder_0': '2018-05-25', 'reminder_1': '09:00:00', 'stars': '2', } ) assert response.status_code == 302 assert response['Location'] == reverse('collection:list') messages = list(get_messages(response.wsgi_request)) assert len(messages) == 0 self.person.refresh_from_db() assert self.person.first_name == 'John' def test_post_invalid(self, admin_client): response = admin_client.post( reverse('collection:change', args=(self.person.pk,)), data={}) assert response.status_code == 200 assert response.context['form'].is_valid() is False assert Person.objects.get().first_name == 'Sally' @pytest.mark.django_db class TestDeleteView: def setup(self): self.person = PersonFactory.create(first_name='Sally') def test_get(self, admin_client): response = admin_client.get( reverse('collection:delete', args=(self.person.pk,))) assert response.status_code == 200 assert Person.objects.exists() is True def test_post(self, admin_client): response = admin_client.post( reverse('collection:delete', args=(self.person.pk,))) assert response.status_code == 302 assert response['Location'] == reverse('collection:list') messages = list(get_messages(response.wsgi_request)) assert len(messages) == 1 assert messages[0].level == SUCCESS_LEVEL assert Person.objects.exists() is False def test_post_protected(self, admin_client): related = RelatedPersonFactory(person=self.person) response = admin_client.post( reverse('collection:delete', args=(self.person.pk,))) assert response.status_code == 200 assert response.context['linked_objects'] == [ 'Related person: {}'.format(str(related)), ] assert Person.objects.exists() is True def test_custom_button_label(self, admin_client): response = admin_client.get( reverse('collection:delete', args=(self.person.pk,))) assert response.context['form_save_button_label'] == 'Delete this person' assert 'Delete this person' in response.content.decode(response.charset) @pytest.mark.django_db class TestFormsetView: def setup(self): self.person = PersonFactory.create() self.related_persons = RelatedPersonFactory.create_batch(2, person=self.person) def test_get(self, admin_client): response = admin_client.get(reverse('formset:change', args=(self.person.pk,))) assert response.status_code == 200 assert response.context['form'].instance == self.person instances = [ form.instance for form in response.context['formsets']['related_persons'].forms ] assert self.related_persons[0] in instances assert self.related_persons[1] in instances def test_post_valid(self, admin_client): response = admin_client.post( reverse('formset:change', args=(self.person.pk,)), data={ 'first_name': 'John', 'last_name': 'Doe', 'country': 'Germany', 'reminder_0': '2018-05-25', 'reminder_1': '09:00:00', 'stars': '2', 'relatedperson_set-INITIAL_FORMS': '1', 'relatedperson_set-MAX_NUM_FORMS': '1000', 'relatedperson_set-MIN_NUM_FORMS': '0', 'relatedperson_set-TOTAL_FORMS': '1', 'relatedperson_set-0-DELETE': '', 'relatedperson_set-0-id': '1', 'relatedperson_set-0-first_name': 'Sally', 'relatedperson_set-0-last_name': 'Mary', 'relatedperson_set-0-person': '1', } ) assert response.status_code == 302 assert response['Location'] == reverse('formset:list') self.person.refresh_from_db() assert self.person.first_name == 'John' self.related_persons[0].refresh_from_db() assert self.related_persons[0].first_name == 'Sally' def test_post_invalid_formset(self, admin_client): response = admin_client.post( reverse('formset:change', args=(self.person.pk,)), data={ 'first_name': 'John', 'last_name': 'Doe', 'country': 'Germany', 'reminder_0': '2018-05-25', 'reminder_1': '09:00:00', 'stars': '2', 'relatedperson_set-INITIAL_FORMS': '1', 'relatedperson_set-MAX_NUM_FORMS': '1000', 'relatedperson_set-MIN_NUM_FORMS': '0', 'relatedperson_set-TOTAL_FORMS': '1', 'relatedperson_set-0-DELETE': '', 'relatedperson_set-0-id': '1', 'relatedperson_set-0-first_name': '', 'relatedperson_set-0-last_name': '', 'relatedperson_set-0-person': '1', } ) assert response.status_code == 200 assert response.context['form'].is_valid() is True assert response.context['formsets']['related_persons'].is_valid() is False def test_post_invalid_form(self, admin_client): response = admin_client.post( reverse('formset:change', args=(self.person.pk,)), data={ 'first_name': '', 'last_name': '', 'country': 'Germany', 'reminder_0': '2018-05-25', 'reminder_1': '09:00:00', 'stars': '2', 'relatedperson_set-INITIAL_FORMS': '1', 'relatedperson_set-MAX_NUM_FORMS': '1000', 'relatedperson_set-MIN_NUM_FORMS': '0', 'relatedperson_set-TOTAL_FORMS': '1', 'relatedperson_set-0-DELETE': '', 'relatedperson_set-0-id': '1', 'relatedperson_set-0-first_name': 'Sally', 'relatedperson_set-0-last_name': 'Mary', 'relatedperson_set-0-person': '1', } ) assert response.status_code == 200 assert response.context['form'].is_valid() is False assert response.context['formsets']['related_persons'].is_valid() is True class TestChangePasswordView: def test_get(self, admin_client): response = admin_client.get(reverse('change-password')) assert response.status_code == 200 assert list(response.context['form'].fields) == ['new_password1', 'new_password2'] def test_post_invalid(self, admin_user, admin_client): response = admin_client.post(reverse('change-password'), data={ 'new_password1': 'Secret', 'new_password2': 'Secret2' }) assert response.status_code == 200 assert response.context['form'].is_valid() is False admin_user.refresh_from_db() assert admin_user.check_password('password') is True def test_post_valid(self, admin_user, admin_client): response = admin_client.post(reverse('change-password'), data={ 'new_password1': 'Secret', 'new_password2': 'Secret' }) assert response.status_code == 302 assert response['Location'] == reverse('change-password') admin_user.refresh_from_db() assert admin_user.check_password('Secret') is True def test_logout_view(admin_client): response = admin_client.get(reverse('minimal:demo')) assert response.status_code == 200 assert response.template_name[0] == 'minimal/demo.html' response = admin_client.get(reverse('logout')) assert response.status_code == 200 assert response.template_name[0] == 'cruditor/logout.html' response = admin_client.get(reverse('minimal:demo')) assert response.status_code == 200 assert response.template_name[0] == 'cruditor/login.html'
mit
MoritzS/django
tests/urlpatterns_reverse/included_namespace_urls.py
72
1059
from django.conf.urls import include, url from .utils import URLObject from .views import empty_view, view_class_instance testobj3 = URLObject('testapp', 'test-ns3') testobj4 = URLObject('testapp', 'test-ns4') app_name = 'included_namespace_urls' urlpatterns = [ url(r'^normal/$', empty_view, name='inc-normal-view'), url(r'^normal/(?P<arg1>[0-9]+)/(?P<arg2>[0-9]+)/$', empty_view, name='inc-normal-view'), url(r'^\+\\\$\*/$', empty_view, name='inc-special-view'), url(r'^mixed_args/([0-9]+)/(?P<arg2>[0-9]+)/$', empty_view, name='inc-mixed-args'), url(r'^no_kwargs/([0-9]+)/([0-9]+)/$', empty_view, name='inc-no-kwargs'), url(r'^view_class/(?P<arg1>[0-9]+)/(?P<arg2>[0-9]+)/$', view_class_instance, name='inc-view-class'), url(r'^test3/', include(*testobj3.urls)), url(r'^test4/', include(*testobj4.urls)), url(r'^ns-included3/', include(('urlpatterns_reverse.included_urls', 'included_urls'), namespace='inc-ns3')), url(r'^ns-included4/', include('urlpatterns_reverse.namespace_urls', namespace='inc-ns4')), ]
bsd-3-clause
iwconfig/svtplay-dl
lib/svtplay_dl/fetcher/rtmp.py
4
1391
# ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- from __future__ import absolute_import import subprocess import shlex from svtplay_dl.log import log from svtplay_dl.utils import is_py2 from svtplay_dl.fetcher import VideoRetriever from svtplay_dl.output import output class RTMP(VideoRetriever): def name(self): return "rtmp" def download(self): """ Get the stream from RTMP """ args = [] if self.options.live: args.append("-v") if self.options.resume: args.append("-e") file_d = output(self.options, "flv", False) if file_d is None: return args += ["-o", self.options.output] if self.options.silent or self.options.output == "-": args.append("-q") if self.options.other: if is_py2: args += shlex.split(self.options.other.encode("utf-8")) else: args += shlex.split(self.options.other) if self.options.verbose: args.append("-V") command = ["rtmpdump", "-r", self.url] + args log.debug("Running: %s", " ".join(command)) try: subprocess.call(command) except OSError as e: log.error("Could not execute rtmpdump: " + e.strerror) return self.finished = True
mit
ryfx/modrana
modules/mod_options/mod_options.py
1
66705
# -*- coding: utf-8 -*- #---------------------------------------------------------------------------- # Handle option menus #---------------------------------------------------------------------------- # Copyright 2008, Oliver White # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. #--------------------------------------------------------------------------- from modules.base_module import RanaModule from core import utils from core.backports import six from core import constants # identifies item as a group GROUP_IDENTIFIER = "groupIdentifier" def getModule(*args, **kwargs): return Options(*args, **kwargs) class Options(RanaModule): """Handle options""" def __init__(self, *args, **kwargs): RanaModule.__init__(self, *args, **kwargs) self.options = {} # for fast searching defaults for corresponding keys self.keyDefault = {} # profile folder self.profileFolderPath = self.modrana.paths.getProfilePath() # check the profile path and create the folders if necessary utils.createFolderPath(self.profileFolderPath) # items menu cache self.itemMenus = {} # item tools special menu name self.keyStateListGroupID = None # option content variables self.monavPackList = [] def _getCategoryID(self, catId): return "opt_cat_%s" % catId # get a standardized id def addCategory(self, name, inId, icon, actionPrefix="", actionSuffix=""): """this method should be run only after menu module instance is available in self.menuModule and the options menu is cleared, eq. has at least the escape button""" # first we add the category button to the options catId = self._getCategoryID(inId) # get a standardized id action = "%sset:menu:%s%s" % (actionPrefix, catId, actionSuffix) self.menuModule.addItem('options', name, icon, action) # initialize menu for the new category menu self.menuModule.clearMenu(catId, "set:menu:options") # as a convenience, return the id return inId def _getGroupId(self, catId, groupId): parentId = self._getCategoryID(catId) return "%s_opt_group_%s" % (parentId, groupId) def addGroup(self, name, groupId, parentId, icon, actionPrefix="", actionSuffix="", registerToMenu=True, backButtonAction=None): """this method ads a new (empty) options group to category specified by catId, as a convenience feature, the id of the new group is returned""" catId = self._getCategoryID(parentId) groupId = self._getGroupId(parentId, groupId) # get a standardized id # handle empty parent ids - such ids can be valid because the menu switching is handled # handled entirely by the pre and post actions if not parentId: action = "%s%s" % (actionPrefix, actionSuffix) else: action = "%sset:menu:options#%s%s" % (actionPrefix, groupId, actionSuffix) if registerToMenu: # add to options menu structure ? self.menuModule.addItem(catId, name, icon, action) if backButtonAction is not None: self.options[groupId] = [backButtonAction, 0, []] else: self.options[groupId] = ["set:menu:%s" % catId, 0, []] return groupId def setGroupParent(self, groupID, parentID): """set the parent id of a given group id""" if groupID in self.options: self.options[groupID][0] = "set:menu:options#%s" % parentID else: self.log.error("can't set group parent - group %s not found", groupID) def getGroupParent(self, groupID): """set the parent id of a given group id""" if groupID in self.options: return self.options[groupID][0] else: self.log.error("can't get group parent - group %s not found", groupID) def clearGroup(self, groupID): """clear a given group from any options, preserving its parent setting""" self.options[groupID][1] = 0 self.options[groupID][2] = [] def addBoolOption(self, title, variable, group, default=None, action=None): on = constants.PANGO_ON off = constants.PANGO_OFF if action: states = ((False, off, action), (True, on, action)) else: states = ((False, off), (True, on)) data = {"type": "toggle", "states": states} self.addOption(title, variable, data, group, default) def addToggleOption(self, title, variable, choices, group, default=None): data = {"type": "toggle", "states": choices} self.addOption(title, variable, data, group, default) def addEditOption(self, title, variable, group, label="Edit variable", description=None): choices = {"type": "showAndEditVariable", "label": label, "description": description } self.addOption(title, variable, choices, group, None) def _generateNestedItems(self, inputList, variable, backAction, fakeMode=None): """generate item tuples for nested item menus""" menuItems = [] # an ordered list of all the menu items itemDict = {} # for easily assigning keys to labels index = 1 # id 0 is the escape button for value, name, icon, action in inputList: if value == "groupIdentifier": # this item is a group item = self.menuModule.generateItem("#%s" % name, "generic", action) else: # this item is a button that sets a value if fakeMode is None: # just use the current mode item = self.menuModule.generateItem("#%s" % name, "generic", "setWithCurrentMode:%s:%s|%s" % (variable, value, backAction)) else:# use a fake mode (used for per mode option state list) item = self.menuModule.generateItem("#%s" % name, "generic", "setWithMode:%s:%s:%s|%s" % ( fakeMode, variable, value, backAction)) menuItems.append(item) itemDict[value] = (name, index) index += 1 return menuItems, itemDict def _generateItems(self, valueNameList, variable, backAction, fakeMode=None): menuItems = [] # an ordered list of all the menu items itemDict = {} # for easily assigning keys to labels index = 1 # id 0 is the escape button for value, name in valueNameList: if fakeMode is None: # just use the current mode item = self.menuModule.generateItem("#%s" % name, "generic", "setWithCurrentMode:%s:%s|%s" % (variable, value, backAction)) else:# use a fake mode (used for per mode option state list) item = self.menuModule.generateItem("#%s" % name, "generic", "setWithMode:%s:%s:%s|%s" % (fakeMode, variable, value, backAction)) menuItems.append(item) itemDict[value] = (name, index) index += 1 return menuItems, itemDict def addItemsOption(self, title, variable, items, group, default=None, fakeMode=None, preAction=None): """add an option, that opens and item selection menu""" #NOTE: for the value - name mapping to work correctly, the value must be a string # -> this is caused by the value being sent as a string once one of the items is clicked # -> if the value, is not string, just the raw value value be shown # Example: # (100, "100%") will show 100 # ('100', "100%") will show 100% # the back action returns back to the group backAction = "set:menu:options#%s" % group # create and add the menu menu = self.menuModule.getClearedMenu(backAction) menuItems, itemDict = self._generateItems(items, variable, backAction, fakeMode=fakeMode) # load all items to the menu menu = self.menuModule.addItemsToThisMenu(menu, menuItems) # store the menu in the menu module # NOTE: for the returning back to the group to work correctly, # the menu is stored under a key combined from the variable and group names storageKey = self._getItemsOptionStorageKey(group, variable, fakeMode=fakeMode) # add the Item menu entry button self.menuModule.addItemMenu(storageKey, menu, wideButtons=True) # also store in the local options structure choices = {"type": "selectOneItem", 'label': "", 'description': "", 'default': default, 'items': items, 'itemDict': itemDict, 'storageKey': storageKey, 'preAction': preAction # send this message before entering the menu } # this means we are probably showing the option in the per mode state list if fakeMode is not None: choices['mode'] = fakeMode choices['noToolsIcon'] = True # disable the tools in the per mode state list self.addOption(title, variable, choices, group, default) def addNestedItemsOption(self, title, variable, items, group, default=None, fakeMode=None, preAction=None): """add an option, that opens and item selection menu with groups""" #NOTE: for the value - name mapping to work correctly, the value must be a string # -> this is caused by the value being sent as a string once one of the items is clicked # -> if the value, is not string, just the raw value value be shown # Example: # (100, "100%") will show 100 # ('100', "100%") will show 100% # the back action returns back to the group from which the menu was opened backAction = "set:menu:options#%s" % group # create submenus for the groups & a toplevel menu # NOTE: for the returning back to the group to work correctly, # the menu is stored under a key combined from the variable and group names storageKey = self._getItemsOptionStorageKey(group, variable, fakeMode=fakeMode) groupIndex = 1 # 0 is the back button topLevel = [] itemDict = {} tempItemDict = {} for item in items: value, name, icon, action = item["item"] if item.get('group', None): # this is a group subMenuItems = item.get('group', []) # create per-group submenu menuItems, tempItemDict = self._generateNestedItems(subMenuItems, variable, backAction, fakeMode=fakeMode) groupStorageKey = "%s_%d" % (storageKey, groupIndex) groupBackAction = "set:menu:%s" % storageKey self._setItemsAsMenu(groupStorageKey, menuItems, groupBackAction) # update value -> name mapping with correct group and subgroup IDs for layerKey in tempItemDict.keys(): label = tempItemDict[layerKey][0] subId = tempItemDict[layerKey][1] # label, toplevel id, group id # NOTE: group level highlighting is not yet implemented tempItemDict[layerKey] = (label, groupIndex, subId) # TODO: highlighting inside groups itemDict.update(tempItemDict) # override the action for the toplevel group button # to point to the group menu action = "set:menu:%s" % groupStorageKey groupIndex += 1 # add the toplevel button topLevel.append((value, name, icon, action)) # add the toplevel menu menuItems, tempItemDict = self._generateNestedItems(topLevel, variable, backAction, fakeMode=fakeMode) self._setItemsAsMenu(storageKey, menuItems, backAction) # update value -> name mapping for tomplevel buttons itemDict.update(tempItemDict) # also store in the local options structure choices = {"type": "selectOneItem", 'label': "", 'description': "", 'default': default, 'items': items, 'itemDict': itemDict, 'storageKey': storageKey, 'preAction': preAction # send this message before entering the menu } # this means we are probably showing the option in the per mode state list if fakeMode is not None: choices['mode'] = fakeMode choices['noToolsIcon'] = True # disable the tools in the per mode state list self.addOption(title, variable, choices, group, default) def _setItemsAsMenu(self, storageKey, menuItems, backAction, wideButtons=True): """create a new item menu (or overwrite an existing one) and register it in the menu Module""" menu = self.menuModule.getClearedMenu(backAction) menu = self.menuModule.addItemsToThisMenu(menu, menuItems) self.menuModule.addItemMenu(storageKey, menu, wideButtons=wideButtons) def _getItemsOptionStorageKey(self, group, variable, fakeMode=None): """return menu name for the special item selection itemized menu """ if fakeMode is None: return "options1Item*%s*%s" % (group, variable) else: return "options1Item*%s*%s*%s" % (group, variable, fakeMode) def _highlightActiveItem(self, menu, variable): """highlight currently active item in the item selection menu""" # test if the key was initialized if self.optionsKeyExists(variable): pass # text, icon, action, type, timedAction else: # not initialized, no need to highlight anything return menu def addOption(self, title, variable, choices, group, default=None): """add an option item""" # add group name to choices, # this is needed for the item tools menu to know where to return choices['groupName'] = group newOption = [title, variable, choices, group, default] if group in self.options: self.options[group][2].append(newOption) self.keyDefault[variable] = default else: self.log.error("group %s does not exist, call addGroup to create it first", group) def addRawOption(self, optionData): """add a raw option to options NOTE: the options contains its group ID""" (title, variable, choices, group, default) = optionData # as some options have have side effects when they are created, # we need to check option the type and replicate those effect as needed optionType = choices['type'] choices = dict(choices) if optionType == 'selectOneItem': if 'mode' in choices: fakeMode = choices['mode'] else: fakeMode = None items = choices['items'] self.addItemsOption(title, variable, items, group, default, fakeMode=fakeMode) else: # no side effects, just add the raw option if group in self.options: self.options[group][2].append(optionData) self.keyDefault[variable] = default else: self.log.error("group %s does not exist, can't add a raw option to it", group) def removeOption(self, categoryId, groupId, variable): """remove an option given by group and variable name""" group = self._getGroupId(categoryId, groupId) if group in self.options: remove = lambda x: x[1] == variable self.options[group][2][:] = [x for x in self.options[group][2] if not remove(x)] if variable in self.keyDefault: del self.keyDefault[variable] else: self.log.error("group %s does not exist, so option with variable %s can not be removed", group, variable) def getOption(self, groupID, index): """get a options item from a given group by its index""" if groupID in self.options: try: return self.options[groupID][2][index] except IndexError: self.log.error("group %s has no index %d, so this option can not be returned", groupID, index) return False else: self.log.error("group %s does not exist, so option with index %d can not be returned", groupID, index) return False def getKeyDefault(self, key, default=None): """get default value for a given key""" return self.keyDefault.get(key, default) def firstTime(self): # initialize the options menu self.initOptionsMenu() def initOptionsMenu(self): """Create the options menu structure. You can add your own options in here, or alternatively create them at runtime from your module's firstTime() function by calling addOption. That would be best if your module is only occasionally used, this function is best if the option is likely to be needed in all installations""" self.menuModule = self.m.get("menu", None) self.menuModule.clearMenu("options") # shortcuts addCat = self.addCategory addGroup = self.addGroup addOpt = self.addToggleOption addBoolOpt = self.addBoolOption addItems = self.addItemsOption # useful notifications notifyRestartNeeded = "ml:notification:m:restart modRana to apply this change;3" # * the Map category * catMap = addCat("Map", "map", "map") # ** map layers optionGroup = addGroup("Map layers", "map_layers", catMap, "generic") defaultBA = "set:menu:options" # default back action mapLayers = self.m.get('mapLayers', None) layerStructure = [] if mapLayers: groups = mapLayers.getGroupList() # sort the groups in alphabetical order by label groups = sorted(groups, key=lambda group: group.label) # assign layers to groups for group in groups: name = group.label icon = group.icon # if no icon is specified, use the generic icon if icon is None: icon = "generic" # list all layers for this group #groupLayers = filter(lambda x: layers[x].get("group", None) == key, layers.keys()) groupLayers = group.layers # layer keys to list of layers # groupLayers = map(lambda x: # (x, layers[x]['label'], layers[x]['icon'], defaultBA) # ,groupLayers) # sort them alphabetically by label groupLayers.sort(key=lambda layer: layer.label) # create (layerId, label, icon) tuples, # reuse the variable: groupLayers = list(map(lambda x: (x.id, x.label, x.icon, defaultBA), groupLayers)) # append their counter to the group name name = "%s (%d)" % (name, len(groupLayers)) layerStructure.append({"item": (GROUP_IDENTIFIER, name, icon, defaultBA), "group": groupLayers}) # append layers without group right after groups in the list nonGroupLayers = mapLayers.getLayersWithoutGroup() # sort the groups in alphabetical order by label nonGroupLayers = sorted(nonGroupLayers, key=lambda group: group.label) # convert to option format nonGroupLayers = map( lambda x: {'item': (x, x.label, x.icon, defaultBA)}, nonGroupLayers) layerStructure.extend(nonGroupLayers) # add empty layer layerStructure.append({'item': (None, "Empty layer", "generic", defaultBA)}) # add the option self.addNestedItemsOption("Main map", "layer", layerStructure, optionGroup, "mapnik") # ** Overlay group = addGroup("Map overlay", "map_overlay", catMap, "generic") addBoolOpt("Map as overlay", "overlay", group, False) self.addNestedItemsOption("Main map", "layer", layerStructure, group, "mapnik") self.addNestedItemsOption("Background map", "layer2", layerStructure, group, "cycle") addOpt("Transparency ratio", "transpRatio", [("0.25,1", "overlay:25%"), ("0.5,1", "overlay:50%"), ("0.75,1", "overlay:75%"), ("1,1", "overlay:100%")], group, "0.5,1") # ** Rotation group = addGroup("Rotation", "map_rotation", catMap, "generic") addBoolOpt("Rotate map in direction of travel", "rotateMap", group, False) # ** Scaling group = addGroup("Scaling", "map_scaling", catMap, "generic") addOpt("Map scale", "mapScale", [(1, "1X"), (2, "2X"), (4, "4X")], group, 1) # ** centering group = addGroup("Centering", "centering", catMap, "generic") addBoolOpt("Centre map", "centred", group, True) addOpt("Centering shift", "posShiftDirection", [("down", "shift down"), ("up", "shift up"), ("left", "shift left"), ("right", "shift right"), (False, "don't shift")], group, "down") addOpt("Centering shift amount", "posShiftAmount", [(0.25, "25%"), (0.5, "50%"), (0.75, "75%"), (1.0, "edge of the screen")], group, 0.75) changedMsg = "mapView:centeringDisableThresholdChanged" addOpt("Disable by dragging", "centeringDisableThreshold", [(2048, "normal drag - <i>default</i>", changedMsg), (15000, "long drag", changedMsg), (40000, "really long drag", changedMsg), (80000, "extremely long drag", changedMsg), (False, constants.PANGO_OFF, changedMsg)], group, 2048) # ** dragging group = addGroup("Dragging", "dragging", catMap, "generic") # check if we are on a powerful device or not and set the default accordingly if self.dmod.simpleMapDragging(): defaultMode = "staticMapDrag" else: defaultMode = "default" addOpt("Map dragging", "mapDraggingMode", [("default", "full redraw - <i>default</i>", "mapView:dragModeChanged"), ("staticMapDrag", "drag visible map - <i>fastest</i>", "mapView:dragModeChanged")], group, defaultMode) # ** map filtering group = addGroup("Filters", "map_filtering", catMap, "generic") addOpt("Negative", "invertMapTiles", [(False, "disabled"), (True, "enabled"), ('withNightTheme', "with night theme")], group, False) # ** map grid group = addGroup("Grid", "map_grid", catMap, "generic") addBoolOpt("Show grid", "drawMapGrid", group, False) addOpt("Grid color", "mapGridColor", [("white", '<span color="white">white</span>'), ("black", '<span color="black">black</span>'), ("red", '<span color="red">red</span>'), ("green", '<span color="green">green</span>'), ("blue", '<span color="blue">blue</span>')], group, "blue") addBoolOpt("Labels", "mapGridLabels", group, True) # ** tile storage group = addGroup("Tile storage", "tile_storage", catMap, "generic") addOpt("Tile storage", "tileStorageType", [('files', "files (default, more space used)"), ('sqlite', "sqlite (new, less space used)")], group, self.modrana.dmod.defaultTileStorageType) addBoolOpt("Store downloaded tiles", "storeDownloadedTiles", group, True) addOpt("Sqlite tile db commit interval", "sqliteTileDatabaseCommitInterval", [(1, "1 second", notifyRestartNeeded), (2, "2 seconds", notifyRestartNeeded), (5, "5 seconds (default)", notifyRestartNeeded), (10, "10 seconds", notifyRestartNeeded), (30, "30 seconds", notifyRestartNeeded), (60, "1 minute", notifyRestartNeeded)], group, 150) # * the view category * catView = addCat("View", "view", "view") # ** GUI group = addGroup("GUI", "gui", catView, "generic") addOpt("Hide main buttons", "hideDelay", [("never", "never hide buttons"), ("5", "hide buttons after 5 seconds"), ("10", "hide buttons after 10 seconds"), ("15", "hide buttons after 15 seconds"), ("30", "hide buttons after 30 seconds"), ("60", "hide buttons after 1 minute"), ("120", "hide buttons after 2 minutes")], group, "10") addOpt("GUI Rotation", "rotationMode", [("auto", "automatic", "device:modeChanged"), ("landscape", "landscape", "device:modeChanged"), ("portrait", "portrait", "device:modeChanged")], group, "auto") # if non-fullscreen mode is supported on the device, add a button # that enables starting in fullscreen # (requested by N900 users) if not self.modrana.dmod.fullscreenOnly(): addBoolOpt("Start in fullscreen", "startInFullscreen", group, False) # if the quit button should be shown even if it is not needed on the # given platform if not self.modrana.gui.showQuitButton(): addBoolOpt("Show Quit button", "showQuitButton", group, False) # ** screen # only add if supported on device display = self.m.get('display', None) if display: if display.screenBlankingControlSupported(): group = addGroup("Screen", "screen", catView, "generic") addOpt("Keep display ON", "screenBlankingMode", [("always", "always", "display:blankingModeChanged"), ("centred", "while centred", "display:blankingModeChanged"), ("moving", "while moving", "display:blankingModeChanged"), ("movingInFullscreen", "while moving in fullscreen", "display:blankingModeChanged"), ("fullscreen", "while in fullscreen", "display:blankingModeChanged"), ("gpsFix", "while there is a GPS fix", "display:blankingModeChanged"), #TODO: while there is actually a GPS lock ("never", "never", "display:blankingModeChanged")], group, "always") if display.usesDashboard(): addBoolOpt("Redraw when on dashboard", "redrawOnDashboard", group, False) # ** themes theme = self.m.get('theme', None) if theme: group = addGroup("Themes", "themes", catView, "generic") defaultTheme = constants.DEFAULT_THEME_ID themeList = theme.getAvailableThemeIds() # check if current theme as set in options exists currentTheme = self.get('currentTheme', None) if currentTheme is not None: if currentTheme not in themeList: self.log.error("theme with id %s is not available,\nswitching " "back to default theme", currentTheme) # theme not valid, reset to default self.set('currentTheme', defaultTheme) themeChangedMessage = "icons:themeChanged" nameValueList = map(lambda x: (x, x, themeChangedMessage), themeList) addOpt("Current theme", "currentTheme", nameValueList, group, defaultTheme ) # ** units group = addGroup("formats#Units and", "units", catView, "generic") addOpt("Units", "unitType", [("km", "use kilometers"), ("mile", "use miles")], group, "km") addOpt("Time format", "currentTimeFormat", [("24h", "24 hours"), ("12h", "12 hours")], group, "24h") addOpt("Small imperial units", "unitTypeImperialSmall", [("yards", "yards"), ("feet", "feet")], group, "yards") # ** menus group = addGroup("Menus", "menus", catView, "generic") addOpt("Listable menu rows", "listableMenuRows", [(2, "2 rows"), (3, "3 rows"), (4, "4 rows"), (5, "5 rows"), (6, "6 rows")], group, 4) if self.dmod.hasButtons(): # TODO: change this once there are more options for key shortcuts # * the Keys category * catKeys = addCat("Keys", "keys", "keys") # * Device buttons group = addGroup("Device buttons", "device_buttons", catKeys, "n900") if self.dmod.hasVolumeKeys(): addBoolOpt("Use volume keys for zooming", "useVolumeKeys", group, True, "device:updateKeys") # * the Navigation category catNavigation = addCat("Navigation", "navigation", "navigation") # * navigation language group = addGroup("Language", "tbt_language", catNavigation, "generic") # in the first string: first one goes to espeak, the second part goes to Google directionsLanguages = [('ca ca', 'Catalan'), ('zh-yue zh-TW', 'Chinese(Cantonese)'), ('zh zh-CN', 'Chinese(Mandarin)'), ('hr hr', 'Croatian'), ('cs cs', 'Czech'), ('nl nl', 'Dutch'), ('en en', 'English'), ('fi fi', 'Finnish'), ('fr fr', 'French'), ('de de', 'German'), ('el el', 'Greek'), ('hi hi', 'Hindi'), ('hu hu', 'Hungarian'), ('id id', 'Indonesian'), ('it it', 'Italian'), ('lv lv', 'Latvian'), ('no no', 'Norwegian'), ('pl pl', 'Polish'), ('pt pt-BR', 'Portuguese(Brazil)'), ('pt-pt pt-PT', 'Portuguese(European)'), ('ro ro', 'Romanian'), ('ru ru', 'Russian'), ('sr sr', 'Serbian'), ('sk sk', 'Slovak'), ('es es', 'Spanish'), ('ta ta', 'Tamil'), ('tr tr', 'Turkish'), ('vi vi', 'Vietnamese')] addItems("Language for directions", 'directionsLanguage', directionsLanguages, group, "en en") # TODO: use locale for default language ? addOpt("read Cyrillic with", "voiceNavigationCyrillicVoice", [('ru', "Russian voice"), (False, "current voice")], group, 'ru') # ** online routing submenu group = addGroup("Routing", "routing", catNavigation, "generic") addOpt("Routing provider", "routingProvider", [(constants.ROUTING_PROVIDER_GOOGLE, "Google - <b>online</b>"), (constants.ROUTING_PROVIDER_MONAV, "Monav - <b>on device</b>")], group, constants.DEFAULT_ROUTING_PROVIDER) addBoolOpt("Avoid major highways", "routingAvoidHighways", group, False) addBoolOpt("Avoid toll roads", "routingAvoidToll", group, False) # ** routing data submenu group = addGroup("data#Routing", "routing_data", catNavigation, "generic") self._reloadMonavPackList() #TODO: on demand reloading addItems("Monav data pack", 'preferredMonavDataPack', self.monavPackList, group, "no preferred pack" ) # * turn by turn navigation group = addGroup("Turn by turn", "turn_by_turn", catNavigation, "generic") addOpt("Autostart navigation", "autostartNavigationDefaultOnAutoselectTurn", [('disabled', constants.PANGO_OFF), ('enabled', constants.PANGO_ON)], group, 'enabled') addOpt("Make final turn announcement at", "pointReachedDistance", [(10, "10 m"), (20, "20 m"), (30, "30 m"), (60, "60 m"), (100, "100 m"), (200, "200 m"), (300, "300 m"), (400, "400 m"), (500, "500 m")], group, 30) addOpt("Announce turns at least this far ahead", "minAnnounceDistance", [(10, "10 m"), (20, "20 m"), (30, "30 m"), (60, "60 m"), (100, "100 m"), (200, "200 m"), (300, "300 m"), (500, "500 m")], group, 100) addOpt("Announce turns at least this long ahead", "minAnnounceTime", [(5, "5 s"), (10, "10 s"), (20, "20 s"), (30, "30 s"), (45, "45 s"), (60, "60 s"), (90, "90 s")], group, 10) # Note: actual values are in m/s, accurate to 2 decimal places. We # store them as strings so lookup will work reliably. addOpt("Increase turn announcement time above", "minAnnounceSpeed", [("5.56", "20 km/h (12 mph)"), ("8.33", "30 km/h (20 mph)"), ("11.11", "40 km/h (25 mph)"), ("13.89", "50 km/h (30 mph)"), ("22.22", "80 km/h (50 mph)"), ("27.78", "100 km/h (60 mph)")], group, "13.89") # Note: actual values are in m/s, accurate to 2 decimal places. We # store them as strings so lookup will work reliably. addOpt("Constant turn announcement time above", "maxAnnounceSpeed", [("13.89", "50 km/h (30 mph)"), ("22.22", "80 km/h (50 mph)"), ("27.78", "100 km/h (60 mph)"), ("33.33", "120 km/h (75 mph)"), ("44.44", "160 km/h (100 mph)")], group, "27.78") addOpt("Maximum turn announcement time", "maxAnnounceTime", [(20, "20 s"), (30, "30 s"), (45, "45 s"), (60, "60 s"), (90, "90 s"), (120, "120 s")], group, 60) # Note: these are exponents, stored as strings so lookup will work reliably. addOpt("Announcement time increase type", "announcePower", [("1.0", "Linear with speed"), ("0.5", "Very quickly, then linear"), ("0.75", "Quickly, then linear"), ("1.5", "Slowly, then linear"), ("2.0", "Quite slowly, then linear"), ("4.0", "Very slowly, then quite fast")], group, "2.0") # ** rerouting submenu group = addGroup("Rerouting", "rerouting", catNavigation, "generic") addItems("Rerouting trigger distance", "reroutingThreshold", [(None, "<b>disabled</b>"), ("10", "10 m"), ("20", "20 m"), ("30", "30 m (default)"), ("40", "40 m"), ("50", "50 m"), ("75", "75 m"), ("100", "100 m"), ("200", "200 m"), ("500", "500 m"), ("1000", "1000 m")], group, "30") # for some reason, the items menu doesn't work correctly for # non-string values (eq. 10 won't work correctly but "10" would # * the POI category catPOI = addCat("POI", "poi", "poi") # ** online POI search group = addGroup("Markers", "poi_markers", catPOI, "generic") addOpt("Show captions", "hideMarkerCaptionsBelowZl", [(-1, "always"), (5, "below zoomlevel 5"), (7, "below zoomlevel 7"), (10, "below zoomlevel 10"), (11, "below zoomlevel 11"), (12, "below zoomlevel 12"), (13, "below zoomlevel 13"), (14, "below zoomlevel 14"), (15, "below zoomlevel 15"), (16, "below zoomlevel 16"), (17, "below zoomlevel 17"), (18, "below zoomlevel 18"), (65535, "never"), ], group, 13) # ** POI storage group = addGroup("POI storage", "poi_storage", catPOI, "generic") addOpt("POI database", "POIDBFilename", [("poi.db", "shared with Mappero (EXPERIMENTAL)", "storePOI:reconnectToDb"), ("modrana_poi.db", "modRana only (default)", "storePOI:reconnectToDb")], group, "modrana_poi.db") # ExportPOIDatabaseToCSV is just a dummy value, # we just need to send a dump message to storePOI addOpt("Export POI Database to CSV", "EportPOIDatabaseToCSV", [("dump", "click to export", "storePOI:dumpToCSV"), ("dump", "click to export", "storePOI:dumpToCSV")], group, "dump") # ** online POI search group = addGroup("Local search", "poi_online", catPOI, "generic") addOpt("Local search ordering", "GLSOrdering", [("default", "ordering from provider"), ("distance", "order by distance") ], group, "default") addOpt("Local search radius", "localSearchRadius", [(500, "500 m"), (1000, "1 km"), (5000, "5 km"), (10000, "10 km"), (25000, "25 km"), (50000, "50 km")], group, 10000) # addOpt("Google local search results", "GLSResults", # [("8", "max 8 results"), # ("16", "max 16 results"), # ("32", "max 32 results")], # group, # "8") addOpt("Local search captions", "drawGLSResultCaptions", [("True", "draw captions"), ("False", "dont draw captions")], group, "True") # * the Location category * catLocation = addCat("Location", "location", "gps_satellite") # ** GPS group = addGroup("GPS", "gps", catLocation, "generic") addBoolOpt("GPS", "GPSEnabled", group, True, "location:checkGPSEnabled") if self.dmod.getLocationType() == 'gpsd': knots = "knots per second" meters = "meters per second" if self.dmod.getDeviceIDString() == 'neo': knots = "knots per second (old SHR)" meters = "meters per second (new SHR)" addOpt("GPSD reports speed in", "gpsdSpeedUnit", [('knotsPerSecond', knots), ('metersPerSecond', meters)], group, 'knotsPerSecond') # * the Network category * catNetwork = addCat("Network", "network", "network") # * network * group = addGroup("Network usage", "network_usage", catNetwork, "generic") addOpt("Network", "network", # [("off","No use of network"), #TODO: implement this :) [("minimal", "Don't Download Map Tiles"), ("full", "Unlimited use of network")], group, "full") addOpt("Auto tile download thread count", "maxAutoDownloadThreads2", [(1, "1 (serial)", notifyRestartNeeded), (5, "5", notifyRestartNeeded), (10, "10 (default)", notifyRestartNeeded), (20, "20", notifyRestartNeeded), (30, "30", notifyRestartNeeded), (40, "40", notifyRestartNeeded), (50, "50", notifyRestartNeeded)], group, 10) addOpt("Batch tile download thread count", "maxDlThreads", [(1, "1 (serial)"), (2, "2"), (3, "3"), (5, "5 (default)"), (10, "10"), (20, "20")], group, 10) # * the Sound category * catSound = addCat("Sound", "sound", "sound") # * sound output group = addGroup("Sound output", "sound_output", catSound, "sound") addBoolOpt("Application wide sound output", "soundEnabled", group, True) # * espeak group group = addGroup("Voice", "voice_out", catSound, "espeak") addOpt("Test voice output", "voiceTest", [("test", "<b>press to start test</b>", "voice:voiceTest"), ("test", "<b>press to start test</b>", "voice:voiceTest")], group, "test") addItems("Voice volume", "voiceVolume", [('0', "0% - silent"), ('20', "20%"), ('50', "50%"), ('100', "100% (default)"), ('200', "200%"), ('300', "300%"), ('400', "400%"), ('500', "500%"), ('600', "600%"), ('700', "700%"), ('1000', "1000%"), ('1100', "1100%"), ('1200', "1200%"), ('1300', "1300% (might be distorted)"), ('1400', "1400% (might be distorted)"), ('1500', "1500% (might be distorted)")], group, 100) addOpt("Voice parameters", "voiceParameters", [("auto", "<b>automatic</b>", "ms:options:espeakParams:auto"), ("manual", "<b>manual</b>", "ms:options:espeakParams:manual")], group, "auto") if self.get('voiceParameters', None) == "manual": self._updateVoiceManual('add') # *** special group for a list per mode item states *** self.keyStateListGroupID = addGroup("specialTools", 'specialGroup', 'specialParent', "generic", registerToMenu=False, backButtonAction="set:menu:optionsItemTools") # addOpt("Network", "threadedDownload", ## [("off","No use of network"), # [("True", "Use threads for download"), # ("False", "Don't use threads for download")], # "network", # "on") # addBoolOpt("Logging", "logging", "logging", True) # options = [] # for i in (1,2,5,10,20,40,60): # options.append((i, "%d sec" % i)) # addOpt("Frequency", "log_period", options, "logging", 2) # addBoolOpt("Vector maps", "vmap", "map", True) # [("0.5,0.5","over:50%,back:50%"), # ("0.25,0.75","over:25%,back:75%"), # ("0.75,0.25","over:75%,back:50%")], # "map", # "0.5,0.5") # addBoolOpt("Old tracklogs", "old_tracklogs", "map", False) # addBoolOpt("Latest tracklog", "tracklog", "map", True) # * the Debug category catDebug = addCat("Debug", "debug", "debug") # ** redraw group = addGroup("Redrawing", "redrawing", catDebug, "generic") addBoolOpt("Print redraw time to terminal", "showRedrawTime", group, False, "display:checkShowRedrawTime") # ** logging group = addGroup("Logging", "logging", catDebug, "generic") addBoolOpt("Log to file", "loggingStatus", group, False) addBoolOpt("Log file compression", "compressLogFile", group, False) # ** tiles group = addGroup("Tiles", "tiles", catDebug, "generic") addBoolOpt("Print tile cache status to terminal", "reportTileCacheStatus", group, False) addBoolOpt("Tile loading status to terminal", "tileLoadingDebug", group, False) addBoolOpt("Redraw screen once a new tile is loaded", "tileLoadedRedraw", group, True) addOpt("In memory tile cache size", "memoryTileCacheSize", [(20, "20", notifyRestartNeeded), (50, "50", notifyRestartNeeded), (100, "100", notifyRestartNeeded), (150, "150 (default)", notifyRestartNeeded), (200, "200", notifyRestartNeeded), (300, "300", notifyRestartNeeded), (500, "500", notifyRestartNeeded), (1000, "1000", notifyRestartNeeded)], group, 150) addOpt("Auto tile download queue size", "autoDownloadQueueSize", [(1, "1", notifyRestartNeeded), (10, "10", notifyRestartNeeded), (20, "20", notifyRestartNeeded), (50, "50", notifyRestartNeeded), (100, "100 (default)", notifyRestartNeeded), (200, "200", notifyRestartNeeded), (500, "500", notifyRestartNeeded), (1000, "1000", notifyRestartNeeded)], group, 100) addBoolOpt("Remove dups before batch dl", "checkTiles", group, False) # ** tracklog drawing group = addGroup("Tracklogs", "tracklogs", catDebug, "generic") addBoolOpt("Debug circles", "debugCircles", group, False) addBoolOpt("Debug squares", "debugSquares", group, False) # ** navigation group = self.addGroup("Navigation", "navigation", catDebug, "generic") addBoolOpt("Print Turn-By-Turn triggers", "debugTbT", group, False) # ** gps group = self.addGroup("GPS", "gps", catDebug, "generic") # only show relevant locationType = self.dmod.getLocationType() if locationType == 'liblocation': addBoolOpt("Show N900 GPS-fix", "n900GPSDebug", group, False) addBoolOpt("GPS debug messages", "gpsDebugEnabled", group, False) # addOpt("Tracklogs", "showTracklog", # [(False, "Don't draw tracklogs"), # ("simple", "Draw simple tracklogs")], # "view", # False) self._setUndefinedToDefault() def _setUndefinedToDefault(self): # Set all undefined options to default values for category, options in self.options.items(): for option in options[2]: (title, variable, choices, category, default) = option if default is not None: if not variable in self.d: self.set(variable, default) def _removeNonPersistent(self, inputDict): """keys that begin with # are not saved (as they mostly contain data that is either time sensitive or is reloaded on startup) ASSUMPTION: keys are strings of length>=1""" try: return dict((k, v) for k, v in six.iteritems(inputDict) if k[0] != '#') except Exception: self.log.exception("error while filtering options\nsome nonpersistent keys might have been left in\nNOTE: keys should be strings of length>=1") return self.d def _reloadKeyStateList(self, groupID, index, key): """reload the key state list to represent currently selected option""" # clear the group self.clearGroup(self.keyStateListGroupID) # for each mode show the current key state modes = self.modrana.getModes().keys() modes.sort() # get data for the given option optionData = self.getOption(groupID, index) # modify the option for mode in modes: optionD = list(optionData) # make a copy # modify the choices dictionary # NOTE: somehow, it is needed to do a copy not just to modify it in the option # or else the mode value is for all options added the same d = dict(optionD[2]) d['noToolsIcon'] = True # disable the tools icon d['mode'] = mode # force a mode optionD[2] = d optionD[3] = self.keyStateListGroupID # set the group to the state list self.addRawOption(optionD) def handleMessage(self, message, messageType, args): if messageType == "ml" and message == "scroll": (direction, menuName) = args index = self.options[menuName][1] maxIndex = len(self.options[menuName][2]) - 1 if direction == "up" and index > 0: newIndex = index - 1 self.options[menuName][1] = newIndex elif direction == "down" and index < maxIndex: newIndex = index + 1 self.options[menuName][1] = newIndex elif message == "save": self.modrana._saveOptions() elif messageType == 'ml' and message == "go2ItemToolsMenu": (groupID, index, key) = args index = int(index) # reload the tools menu menus = self.m.get('menu', None) if menus: menuName = 'optionsItemTools' reset = 'ms:options:resetKey:%s' % key notify = "ml:notification:m:Item has been reset to default;3" resetAction = "%s|%s|set:menu:options#%s" % (reset, notify, groupID) menus.clearMenu(menuName, 'set:menu:options#%s' % groupID) menus.addItem(menuName, 'state list#per mode', 'generic', 'ml:options:go2ItemStateListMenu:%s;%d;%s' % (groupID, index, key) ) menus.addItem(menuName, 'default#reset to', 'generic', resetAction) self.set('menu', menuName) elif messageType == 'ml' and message == "go2ItemStateListMenu": (groupID, index, key) = args index = int(index) # reload the option key state list for the given key self._reloadKeyStateList(groupID, index, key) # go to the menu self.set('menu', 'options#%s' % self.keyStateListGroupID) elif messageType == 'ms' and message == 'resetKey': # reset a given options item to default, including any key modifiers key = args self.modrana.purgeKey(key) default = self.getKeyDefault(key) self.set(key, default) elif messageType == 'ml' and message == 'addKeyModifier': # make the value of a key mode specific (key, mode) = args self.modrana.addKeyModifier(key, mode=mode) elif messageType == 'ml' and message == 'removeKeyModifier': (key, mode) = args # make the value of a key mode unspecific self.modrana.removeKeyModifier(key, mode=mode) elif messageType == "ms" and message == "espeakParams": # switch between espeak parameter modes if args == "manual": self._updateVoiceManual("remove") elif args == "auto": self._updateVoiceManual("add") elif messageType == "ml" and message == "editVariable": (variable, label, description) = args initialText = self.get(variable, "") entry = self.m.get('textEntry', None) if entry: key = "editVariable_%s" % variable entry.entryBox(self, key, label, initialText, description) #messages to toggles a key to be mode un/specific #Mode specific keys: #this means that the options key can have a different value #depending on the current mode, thus enabling better customization""" elif messageType == "ms" and message == "makeKeyModeSpecific": self.modrana.addKeyModifier(args) elif messageType == "ms" and message == "makeKeyModeUnSpecific": self.modrana.removeKeyModifier(args) elif messageType == 'ml' and message == 'update': if len(args) >= 1: target = args[0] if target == 'packListMonav': self._reloadMonavPackList() else: self.log.error('update target not specified') def _reloadMonavPackList(self): route = self.m.get('route', None) if route: self.log.info('reloading Monav data pack list') # wee need a list of (name, key) tuples self.monavPackList = map(lambda x: (x, x), route.getAvailableMonavDataPacks()) def _updateVoiceManual(self, action): """add or remove custom voice parameters option items""" if action == "add": groupId = self._getGroupId("sound", "voice_out") description = "<b>Note:</b> <tt>%language%</tt> will be replaced by current language code, <tt>%volume%</tt> with current voice volume, <tt>%message%</tt> with the message and <tt>%qmessage%</tt> will be replaced by the message in quotes" self.addEditOption("Edit voice string", "voiceString", groupId, "Edit voice string", description=description) message = "ms:voice:resetStringToDefault:espeak" self.addToggleOption("Reset voice string with <b>Espeak</b> default", "placeholder", [("foo", "<i>click to use this default</i>", message)], groupId, "foo") elif action == "remove": self.removeOption("sound", "voice_out", "voiceString") self.removeOption("sound", "voice_out", "placeholder") def handleTextEntryResult(self, key, result): (optionType, variable) = key.split("_", 1) if optionType == "editVariable": self.log.info("editing variable: %s with: %s", variable, result) self.set(variable, result) def drawMenu(self, cr, menuName, args=None): # custom options list drawing clickHandler = self.m.get('clickHandler', None) if self.menuModule and clickHandler: # elements allocation (e1, e2, e3, e4, alloc) = self.menuModule.threePlusOneMenuCoords() (x1, y1) = e1 (x2, y2) = e2 (x3, y3) = e3 (x4, y4) = e4 (w1, h1, dx, dy) = alloc (cancelButtonAction, firstItemIndex, options) = self.options[menuName] # Top row: # * parent menu timeout = self.modrana.gui.msLongPress self.menuModule.drawButton(cr, x1, y1, dx, dy, "", "back", cancelButtonAction, timedAction=(timeout, "set:menu:None")) # * scroll up self.menuModule.drawButton(cr, x2, y2, dx, dy, "", "up_list", "ml:options:scroll:up;%s" % menuName) # * scroll down self.menuModule.drawButton(cr, x3, y3, dx, dy, "", "down_list", "ml:options:scroll:down;%s" % menuName) # One option per row for row in (0, 1, 2): index = firstItemIndex + row numItems = len(options) cAction = None if 0 <= index < numItems: (title, variable, choices, group, default) = options[index] # What's it set to currently? if 'mode' in choices: mode = choices['mode'] fakeMode = mode else: mode = self.get('mode', 'car') fakeMode = None value = self.get(variable, None, mode=mode) # if the key has a modifier in this mode, # append the mode label to the title if 'mode' in choices: # this currently means we are in the option state list if self.get('mode', 'car') == choices['mode']: # current mode title = "%s: <small><sup><b>[%s]</b></sup></small>" % ( title, self.modrana.getModeLabel(mode)) else: title = "%s: <small><sup>[%s]</sup></small>" % ( title, self.modrana.getModeLabel(mode)) else: # normal option display if self.modrana.hasKeyModifierInMode(variable, mode): title = "%s: <small><sup><b>[%s]</b></sup></small>" % ( title, self.modrana.getModeLabel(mode)) else: title = "%s:" % title # Lookup the description of the currently-selected choice. # (if any, use str(value) if it doesn't match any defined options) # Also lookup the _next_ choice in the list, because that's what # we will set the option to if it's clicked optionType = choices["type"] if optionType == "showAndEditVariable": label = choices["label"] description = choices["description"] # show and edit the exact value of a variable manually valueDescription = self.get(variable, "variable is not set yet") valueDescription = "<tt><b>%s</b></tt>" % valueDescription payload = "%s;%s;%s" % (variable, label, description) onClick = "ml:options:editVariable:%s|set:needRedraw:True" % payload elif optionType == "selectOneItem": #show multiple items and make it possible to select one of them # get current value default = choices['default'] value = self.get(variable, default, mode=mode) # show label for the given value highlight = choices['itemDict'].get(value, (value, None)) valueDescription, highlightId = highlight[0], highlight[1] # if no description is found, just display the value valueDescription = "<tt><b>%s</b></tt>" % valueDescription preAction = "" # add any pre-actions (actions that should run before the # menu is entered, eq updating data, etc.) pre = choices.get('preAction', "") if pre: preAction += "%s|" % pre #assure highlighting if highlightId is not None: # add an action before switching to the next menu that # assures that items in the next menu are properly highlighted # according to the state of the corresponding variable preAction += "ml:menu:highlightItem:%s;%d|" % (choices['storageKey'], highlightId) if fakeMode is None: onClick = "%sset:menu:options1Item*%s*%s" % (preAction, group, variable) else: onClick = "%sset:menu:options1Item*%s*%s*%s" % (preAction, group, variable, fakeMode) # the fake mode is used for listing and setting options for other mode than the current one elif optionType == 'toggle': states = choices['states'] nextChoice = states[0] valueDescription = str(value) useNext = False for c in states: (cVal, cName) = (c[0], c[1]) if useNext: nextChoice = c useNext = False if str(value) == str(cVal): valueDescription = cName useNext = True if len(c) == 3: cAction = c[2] # What should happen if this option is clicked - # set the associated option to the next value in sequence onClick = "setWithMode:%s:%s:%s" % (mode, variable, str(nextChoice[0])) if cAction: onClick += "|%s" % cAction onClick += "|options:save" onClick += "|set:needRedraw:1" y = y4 + row * dy if w1 > h1: # landscape dx = (x4 - x1) w = w1 - dx else: # portrait dx = x2 w = w1 smallButtonW = dx / 2.0 smallButtonH = dy / 2.0 # Draw the option button and make it clickable self.menuModule.drawButton(cr, x4, y, w, dy, None, "generic", # background for a 3x1 icon onClick) # draw mode specific combined toggle & indicator if self.modrana.hasKeyModifierInMode(variable, mode): toggleText = constants.PANGO_ON + '#per Mode' modeSpecToggleAction = "ml:options:removeKeyModifier:%s;%s" % (variable, mode) else: toggleText = "OFF#per Mode" modeSpecToggleAction = "ml:options:addKeyModifier:%s;%s" % (variable, mode) self.menuModule.drawButton(cr, x4 + w - smallButtonW, y, smallButtonW, smallButtonH, toggleText, "generic", modeSpecToggleAction, layer=1) groupName = choices['groupName'] if 'noToolsIcon' not in choices: # draw tools button self.menuModule.drawButton(cr, x4 + w - smallButtonW, y + smallButtonH, smallButtonW, smallButtonH, None, "tools", # tools icon "ml:options:go2ItemToolsMenu:%s;%d;%s" % (groupName, index, variable), layer=1) border = 20 # 1st line: option name self.menuModule.showText(cr, title, x4 + border, y + border, w * 0.95 - smallButtonW - border) # 2nd line: current value self.menuModule.showText(cr, valueDescription, x4 + 0.15 * w, y + 0.6 * dy, w * 0.85 - smallButtonW - border) # in corner: row number indexX = x4 + w * 0.90 - smallButtonW self.menuModule.showText(cr, "%d/%d" % (index + 1, numItems), indexX, y + dy * 0.07, w * 0.10 - border, 20)
gpl-3.0
wetek-enigma/enigma2
lib/python/Screens/ChannelSelection.py
2
110011
# -*- coding: utf-8 -*- from boxbranding import getMachineBuild, getMachineBrand, getMachineName import os from Tools.Profile import profile from Screen import Screen import Screens.InfoBar import Components.ParentalControl from Components.Button import Button from Components.ServiceList import ServiceList, refreshServiceList from Components.ActionMap import NumberActionMap, ActionMap, HelpableActionMap from Components.MenuList import MenuList from Components.ServiceEventTracker import ServiceEventTracker, InfoBarBase from Components.Sources.List import List from Components.SystemInfo import SystemInfo from Components.UsageConfig import preferredTimerPath from Components.Renderer.Picon import getPiconName from Screens.TimerEdit import TimerSanityConflict profile("ChannelSelection.py 1") from EpgSelection import EPGSelection from enigma import eActionMap, eServiceReference, eEPGCache, eServiceCenter, eRCInput, eTimer, ePoint, eDVBDB, iPlayableService, iServiceInformation, getPrevAsciiCode, eEnv, loadPNG from Components.config import config, configfile, ConfigSubsection, ConfigText, ConfigYesNo from Tools.NumericalTextInput import NumericalTextInput profile("ChannelSelection.py 2") from Components.NimManager import nimmanager profile("ChannelSelection.py 2.1") from Components.Sources.RdsDecoder import RdsDecoder profile("ChannelSelection.py 2.2") from Components.Sources.ServiceEvent import ServiceEvent from Components.Sources.Event import Event profile("ChannelSelection.py 2.3") from Components.Input import Input profile("ChannelSelection.py 3") from Components.ChoiceList import ChoiceList, ChoiceEntryComponent from RecordTimer import RecordTimerEntry, AFTEREVENT from TimerEntry import TimerEntry, InstantRecordTimerEntry from Screens.InputBox import InputBox, PinInput from Screens.VirtualKeyBoard import VirtualKeyBoard from Screens.ChoiceBox import ChoiceBox from Screens.MessageBox import MessageBox from Screens.ServiceInfo import ServiceInfo from Screens.ButtonSetup import InfoBarButtonSetup, ButtonSetupActionMap, getButtonSetupFunctions profile("ChannelSelection.py 4") from Screens.PictureInPicture import PictureInPicture from Screens.RdsDisplay import RassInteractive from ServiceReference import ServiceReference from Tools.BoundFunction import boundFunction from Tools import Notifications from Plugins.Plugin import PluginDescriptor from Components.PluginComponent import plugins from time import localtime, time try: from Plugins.SystemPlugins.PiPServiceRelation.plugin import getRelationDict plugin_PiPServiceRelation_installed = True except: plugin_PiPServiceRelation_installed = False profile("ChannelSelection.py after imports") FLAG_SERVICE_NEW_FOUND = 64 FLAG_IS_DEDICATED_3D = 128 #define in lib/dvb/idvb.h as dxNewFound = 64 and dxIsDedicated3D = 128 class BouquetSelector(Screen): def __init__(self, session, bouquets, selectedFunc, enableWrapAround=True): Screen.__init__(self, session) Screen.setTitle(self, _("Choose Bouquet")) self.selectedFunc=selectedFunc self["actions"] = ActionMap(["OkCancelActions"], { "ok": self.okbuttonClick, "cancel": self.cancelClick }) entrys = [ (x[0], x[1]) for x in bouquets ] self["menu"] = MenuList(entrys, enableWrapAround) def getCurrent(self): cur = self["menu"].getCurrent() return cur and cur[1] def okbuttonClick(self): self.selectedFunc(self.getCurrent()) def up(self): self["menu"].up() def down(self): self["menu"].down() def cancelClick(self): self.close(False) class EpgBouquetSelector(BouquetSelector): def __init__(self, session, bouquets, selectedFunc, enableWrapAround=False): BouquetSelector.__init__(self, session, bouquets, selectedFunc, enableWrapAround=False) self.skinName = "BouquetSelector" self.bouquets=bouquets def okbuttonClick(self): self.selectedFunc(self.getCurrent(),self.bouquets) class SilentBouquetSelector: def __init__(self, bouquets, enableWrapAround=False, current=0): self.bouquets = [b[1] for b in bouquets] self.pos = current self.count = len(bouquets) self.enableWrapAround = enableWrapAround def up(self): if self.pos > 0 or self.enableWrapAround: self.pos = (self.pos - 1) % self.count def down(self): if self.pos < (self.count - 1) or self.enableWrapAround: self.pos = (self.pos + 1) % self.count def getCurrent(self): return self.bouquets[self.pos] # csel.bouquet_mark_edit values OFF = 0 EDIT_BOUQUET = 1 EDIT_ALTERNATIVES = 2 def append_when_current_valid(current, menu, args, level=0, key=""): if current and current.valid() and level <= config.usage.setup_level.index: menu.append(ChoiceEntryComponent(key, args)) def removed_userbouquets_available(): for file in os.listdir("/etc/enigma2/"): if file.startswith("userbouquet") and file.endswith(".del"): return True return False class ChannelContextMenu(Screen): def __init__(self, session, csel): Screen.__init__(self, session) Screen.setTitle(self, _("Channel list context menu")) #raise Exception("we need a better summary screen here") self.csel = csel self.bsel = None if self.isProtected(): self.onFirstExecBegin.append(boundFunction(self.session.openWithCallback, self.protectResult, PinInput, pinList=[x.value for x in config.ParentalControl.servicepin], triesEntry=config.ParentalControl.retries.servicepin, title=_("Please enter the correct pin code"), windowTitle=_("Enter pin code"))) self["actions"] = ActionMap(["OkCancelActions", "ColorActions", "NumberActions", "MenuActions"], { "ok": self.okbuttonClick, "cancel": self.cancelClick, "blue": self.showServiceInPiP, "red": self.playMain, "menu": self.openSetup, "2": self.renameEntry, "3": self.findCurrentlyPlayed, "5": self.addServiceToBouquetOrAlternative, "6": self.toggleMoveModeSelect, "8": self.removeEntry }) menu = [ ] self.removeFunction = False self.addFunction = False self.pipAvailable = False current = csel.getCurrentSelection() current_root = csel.getRoot() current_sel_path = current.getPath() current_sel_flags = current.flags inBouquetRootList = current_root and 'FROM BOUQUET "bouquets.' in current_root.getPath() #FIXME HACK inAlternativeList = current_root and 'FROM BOUQUET "alternatives' in current_root.getPath() self.inBouquet = csel.getMutableList() is not None haveBouquets = config.usage.multibouquet.value from Components.ParentalControl import parentalControl self.parentalControl = parentalControl self.parentalControlEnabled = config.ParentalControl.servicepinactive.value menu.append(ChoiceEntryComponent(text = (_("Settings..."), boundFunction(self.openSetup)))) if not (current_sel_path or current_sel_flags & (eServiceReference.isDirectory|eServiceReference.isMarker)): append_when_current_valid(current, menu, (_("show transponder info"), self.showServiceInformations), level=2) if csel.bouquet_mark_edit == OFF and not csel.entry_marked: if not inBouquetRootList: isPlayable = not (current_sel_flags & (eServiceReference.isMarker|eServiceReference.isDirectory)) if isPlayable: for p in plugins.getPlugins(PluginDescriptor.WHERE_CHANNEL_CONTEXT_MENU): append_when_current_valid(current, menu, (p.name, boundFunction(self.runPlugin, p)), key="bullet") if config.servicelist.startupservice.value == self.csel.getCurrentSelection().toString(): append_when_current_valid(current, menu, (_("stop using as startup service"), self.unsetStartupService), level=0) else: append_when_current_valid(current, menu, (_("set as startup service"), self.setStartupService), level=0) if self.parentalControlEnabled: if self.parentalControl.getProtectionLevel(csel.getCurrentSelection().toCompareString()) == -1: append_when_current_valid(current, menu, (_("add to parental protection"), boundFunction(self.addParentalProtection, csel.getCurrentSelection())), level=0) else: append_when_current_valid(current, menu, (_("remove from parental protection"), boundFunction(self.removeParentalProtection, csel.getCurrentSelection())), level=0) if config.ParentalControl.hideBlacklist.value and not parentalControl.sessionPinCached and config.ParentalControl.storeservicepin.value != "never": append_when_current_valid(current, menu, (_("Unhide parental control services"), boundFunction(self.unhideParentalServices)), level=0) if SystemInfo["3DMode"]: if eDVBDB.getInstance().getFlag(eServiceReference(current.toString())) & FLAG_IS_DEDICATED_3D: append_when_current_valid(current, menu, (_("Unmark service as dedicated 3D service"), self.removeDedicated3DFlag), level=0) else: append_when_current_valid(current, menu, (_("Mark service as dedicated 3D service"), self.addDedicated3DFlag), level=0) if haveBouquets: bouquets = self.csel.getBouquetList() if bouquets is None: bouquetCnt = 0 else: bouquetCnt = len(bouquets) if not self.inBouquet or bouquetCnt > 1: append_when_current_valid(current, menu, (_("add service to bouquet"), self.addServiceToBouquetSelected), level=0, key="5") self.addFunction = self.addServiceToBouquetSelected if not self.inBouquet: append_when_current_valid(current, menu, (_("remove entry"), self.removeEntry), level = 0, key="8") self.removeFunction = self.removeSatelliteService else: if not self.inBouquet: append_when_current_valid(current, menu, (_("add service to favourites"), self.addServiceToBouquetSelected), level=0, key="5") self.addFunction = self.addServiceToBouquetSelected if SystemInfo["PIPAvailable"]: if not self.parentalControlEnabled or self.parentalControl.getProtectionLevel(csel.getCurrentSelection().toCompareString()) == -1: if self.csel.dopipzap: append_when_current_valid(current, menu, (_("play in mainwindow"), self.playMain), level=0, key="red") else: append_when_current_valid(current, menu, (_("play as picture in picture"), self.showServiceInPiP), level=0, key="blue") append_when_current_valid(current, menu, (_("find currently played service"), self.findCurrentlyPlayed), level=0, key="3") else: if 'FROM SATELLITES' in current_root.getPath() and current and _("Services") in eServiceCenter.getInstance().info(current).getName(current): unsigned_orbpos = current.getUnsignedData(4) >> 16 if unsigned_orbpos == 0xFFFF: append_when_current_valid(current, menu, (_("remove cable services"), self.removeSatelliteServices), level = 0) elif unsigned_orbpos == 0xEEEE: append_when_current_valid(current, menu, (_("remove terrestrial services"), self.removeSatelliteServices), level = 0) else: append_when_current_valid(current, menu, (_("remove selected satellite"), self.removeSatelliteServices), level = 0) if haveBouquets: if not self.inBouquet and not "PROVIDERS" in current_sel_path: append_when_current_valid(current, menu, (_("copy to bouquets"), self.copyCurrentToBouquetList), level=0) if ("flags == %d" %(FLAG_SERVICE_NEW_FOUND)) in current_sel_path: append_when_current_valid(current, menu, (_("remove all new found flags"), self.removeAllNewFoundFlags), level=0) if self.inBouquet: append_when_current_valid(current, menu, (_("rename entry"), self.renameEntry), level=0, key="2") if not inAlternativeList: append_when_current_valid(current, menu, (_("remove entry"), self.removeEntry), level=0, key="8") self.removeFunction = self.removeCurrentService if current_root and ("flags == %d" %(FLAG_SERVICE_NEW_FOUND)) in current_root.getPath(): append_when_current_valid(current, menu, (_("remove new found flag"), self.removeNewFoundFlag), level=0) else: if self.parentalControlEnabled: if self.parentalControl.getProtectionLevel(csel.getCurrentSelection().toCompareString()) == -1: append_when_current_valid(current, menu, (_("add bouquet to parental protection"), boundFunction(self.addParentalProtection, csel.getCurrentSelection())), level=0) else: append_when_current_valid(current, menu, (_("remove bouquet from parental protection"), boundFunction(self.removeParentalProtection, csel.getCurrentSelection())), level=0) menu.append(ChoiceEntryComponent(text=(_("add bouquet"), self.showBouquetInputBox))) append_when_current_valid(current, menu, (_("rename entry"), self.renameEntry), level=0, key="2") append_when_current_valid(current, menu, (_("remove entry"), self.removeEntry), level=0, key="8") self.removeFunction = self.removeBouquet if removed_userbouquets_available(): append_when_current_valid(current, menu, (_("purge deleted userbouquets"), self.purgeDeletedBouquets), level=0) append_when_current_valid(current, menu, (_("restore deleted userbouquets"), self.restoreDeletedBouquets), level=0) if self.inBouquet: # current list is editable? if csel.bouquet_mark_edit == OFF: if csel.movemode: append_when_current_valid(current, menu, (_("disable move mode"), self.toggleMoveMode), level=0, key="6") else: append_when_current_valid(current, menu, (_("enable move mode"), self.toggleMoveMode), level=1, key="6") if not csel.entry_marked and not inBouquetRootList and current_root and not (current_root.flags & eServiceReference.isGroup): if current.type != -1: menu.append(ChoiceEntryComponent(text=(_("add marker"), self.showMarkerInputBox))) if not csel.movemode: if haveBouquets: append_when_current_valid(current, menu, (_("enable bouquet edit"), self.bouquetMarkStart), level=0) else: append_when_current_valid(current, menu, (_("enable favourite edit"), self.bouquetMarkStart), level=0) if current_sel_flags & eServiceReference.isGroup: append_when_current_valid(current, menu, (_("edit alternatives"), self.editAlternativeServices), level=2) append_when_current_valid(current, menu, (_("show alternatives"), self.showAlternativeServices), level=2) append_when_current_valid(current, menu, (_("remove all alternatives"), self.removeAlternativeServices), level=2) elif not current_sel_flags & eServiceReference.isMarker: append_when_current_valid(current, menu, (_("add alternatives"), self.addAlternativeServices), level=2) else: if csel.bouquet_mark_edit == EDIT_BOUQUET: if haveBouquets: append_when_current_valid(current, menu, (_("end bouquet edit"), self.bouquetMarkEnd), level=0) append_when_current_valid(current, menu, (_("abort bouquet edit"), self.bouquetMarkAbort), level=0) else: append_when_current_valid(current, menu, (_("end favourites edit"), self.bouquetMarkEnd), level=0) append_when_current_valid(current, menu, (_("abort favourites edit"), self.bouquetMarkAbort), level=0) if current_sel_flags & eServiceReference.isMarker: append_when_current_valid(current, menu, (_("rename entry"), self.renameEntry), level=0, key="2") append_when_current_valid(current, menu, (_("remove entry"), self.removeEntry), level=0, key="8") self.removeFunction = self.removeCurrentService else: append_when_current_valid(current, menu, (_("end alternatives edit"), self.bouquetMarkEnd), level=0) append_when_current_valid(current, menu, (_("abort alternatives edit"), self.bouquetMarkAbort), level=0) menu.append(ChoiceEntryComponent(text = (_("Reload Services"), self.reloadServices))) self["menu"] = ChoiceList(menu) def set3DMode(self, value): if config.osd.threeDmode.value == "auto" and self.session.nav.currentlyPlayingServiceReference == self.csel.getCurrentSelection(): from Screens.VideoMode import applySettings applySettings(value and "sidebyside" or config.osd.threeDmode.value) def addDedicated3DFlag(self): eDVBDB.getInstance().addFlag(eServiceReference(self.csel.getCurrentSelection().toString()), FLAG_IS_DEDICATED_3D) eDVBDB.getInstance().reloadBouquets() self.set3DMode(True) self.close() def removeDedicated3DFlag(self): eDVBDB.getInstance().removeFlag(eServiceReference(self.csel.getCurrentSelection().toString()), FLAG_IS_DEDICATED_3D) eDVBDB.getInstance().reloadBouquets() self.set3DMode(False) self.close() def isProtected(self): return self.csel.protectContextMenu and config.ParentalControl.setuppinactive.value and config.ParentalControl.config_sections.context_menus.value def protectResult(self, answer): if answer: self.csel.protectContextMenu = False elif answer is not None: self.session.openWithCallback(self.close, MessageBox, _("The pin code you entered is wrong."), MessageBox.TYPE_ERROR) else: self.close() def addServiceToBouquetOrAlternative(self): if self.addFunction: self.addFunction() else: return 0 def getCurrentSelectionName(self): cur = self.csel.getCurrentSelection() if cur and cur.valid(): name = eServiceCenter.getInstance().info(cur).getName(cur) or ServiceReference(cur).getServiceName() or "" name = name.replace('\xc2\x86', '').replace('\xc2\x87', '') return name return "" def removeEntry(self): ref = self.csel.servicelist.getCurrent() if self.removeFunction and ref and ref.valid(): if self.csel.confirmRemove: list = [(_("yes"), True), (_("no"), False), (_("yes") + ", " + _("and never ask again this session again"), "never")] self.session.openWithCallback(self.removeFunction, MessageBox, _("Are you sure to remove this entry?") + "\n%s" % self.getCurrentSelectionName(), list=list) else: self.removeFunction(True) else: return 0 def removeCurrentService(self, answer): if answer: if answer == "never": self.csel.confirmRemove = False self.csel.removeCurrentService() self.close() def removeSatelliteService(self, answer): if answer: if answer == "never": self.csel.confirmRemove = False self.csel.removeSatelliteService() self.close() def removeBouquet(self, answer): if answer: self.csel.removeBouquet() eDVBDB.getInstance().reloadBouquets() self.close() def purgeDeletedBouquets(self): self.session.openWithCallback(self.purgeDeletedBouquetsCallback, MessageBox, _("Are you sure to purge all deleted userbouquets?")) def purgeDeletedBouquetsCallback(self, answer): if answer: for file in os.listdir("/etc/enigma2/"): if file.startswith("userbouquet") and file.endswith(".del"): file = "/etc/enigma2/" + file print "permantly remove file ", file os.remove(file) self.close() def restoreDeletedBouquets(self): for file in os.listdir("/etc/enigma2/"): if file.startswith("userbouquet") and file.endswith(".del"): file = "/etc/enigma2/" + file print "restore file ", file[:-4] os.rename(file, file[:-4]) eDVBDBInstance = eDVBDB.getInstance() eDVBDBInstance.setLoadUnlinkedUserbouquets(True) eDVBDBInstance.reloadBouquets() eDVBDBInstance.setLoadUnlinkedUserbouquets(config.misc.load_unlinked_userbouquets.value) refreshServiceList() self.csel.showFavourites() self.close() def playMain(self): sel = self.csel.getCurrentSelection() if sel and sel.valid() and self.csel.dopipzap and (not self.parentalControlEnabled or self.parentalControl.getProtectionLevel(self.csel.getCurrentSelection().toCompareString()) == -1): self.csel.zap() self.csel.setCurrentSelection(sel) self.close(True) else: return 0 def okbuttonClick(self): self["menu"].getCurrent()[0][1]() def openSetup(self): from Screens.Setup import Setup self.session.openWithCallback(self.cancelClick, Setup, "channelselection") def cancelClick(self, dummy=False): self.close(False) def reloadServices(self): eDVBDB.getInstance().reloadBouquets() eDVBDB.getInstance().reloadServicelist() self.session.openWithCallback(self.close, MessageBox, _("The servicelist is reloaded."), MessageBox.TYPE_INFO, timeout = 5) def showServiceInformations(self): self.session.open( ServiceInfo, self.csel.getCurrentSelection() ) def setStartupService(self): self.session.openWithCallback(self.setStartupServiceCallback, MessageBox, _("Set startup service"), list = [(_("Only on startup"), "startup"), (_("Also on standby"), "standby")]) def setStartupServiceCallback(self, answer): if answer: config.servicelist.startupservice.value = self.csel.getCurrentSelection().toString() path = ';'.join([i.toString() for i in self.csel.servicePath]) config.servicelist.startuproot.value = path config.servicelist.startupmode.value = config.servicelist.lastmode.value config.servicelist.startupservice_onstandby.value = answer == "standby" config.servicelist.save() configfile.save() self.close() def unsetStartupService(self): config.servicelist.startupservice.value = '' config.servicelist.startupservice_onstandby.value = False config.servicelist.save() configfile.save() self.close() def setStartupServiceStandby(self): config.servicelist.startupservice_standby.value = self.csel.getCurrentSelection().toString() config.servicelist.save() configfile.save() self.close() def unsetStartupServiceStandby(self): config.servicelist.startupservice_standby.value = '' config.servicelist.save() configfile.save() self.close() def showBouquetInputBox(self): self.session.openWithCallback(self.bouquetInputCallback, VirtualKeyBoard, title=_("Please enter a name for the new bouquet"), text="bouquetname", maxSize=False, visible_width=56, type=Input.TEXT) def bouquetInputCallback(self, bouquet): if bouquet is not None: self.csel.addBouquet(bouquet, None) self.close() def addParentalProtection(self, service): self.parentalControl.protectService(service.toCompareString()) if config.ParentalControl.hideBlacklist.value and not self.parentalControl.sessionPinCached: self.csel.servicelist.resetRoot() self.close() def removeParentalProtection(self, service): self.session.openWithCallback(boundFunction(self.pinEntered, service.toCompareString()), PinInput, pinList=[config.ParentalControl.servicepin[0].value], triesEntry=config.ParentalControl.retries.servicepin, title=_("Enter the service pin"), windowTitle=_("Enter pin code")) def pinEntered(self, service, answer): if answer: self.parentalControl.unProtectService(service) self.close() elif answer is not None: self.session.openWithCallback(self.close, MessageBox, _("The pin code you entered is wrong."), MessageBox.TYPE_ERROR) else: self.close() def unhideParentalServices(self): if self.csel.protectContextMenu: self.session.openWithCallback(self.unhideParentalServicesCallback, PinInput, pinList=[config.ParentalControl.servicepin[0].value], triesEntry=config.ParentalControl.retries.servicepin, title=_("Enter the service pin"), windowTitle=_("Enter pin code")) else: self.unhideParentalServicesCallback(True) def unhideParentalServicesCallback(self, answer): if answer: service = self.csel.servicelist.getCurrent() self.parentalControl.setSessionPinCached() self.parentalControl.hideBlacklist() self.csel.servicelist.resetRoot() self.csel.servicelist.setCurrent(service) self.close() elif answer is not None: self.session.openWithCallback(self.close, MessageBox, _("The pin code you entered is wrong."), MessageBox.TYPE_ERROR) else: self.close() def showServiceInPiP(self): if self.csel.dopipzap or (self.parentalControlEnabled and not self.parentalControl.getProtectionLevel(self.csel.getCurrentSelection().toCompareString()) == -1): return 0 service = self.session.nav.getCurrentService() info = service and service.info() xres = str(info.getInfo(iServiceInformation.sVideoWidth)) if int(xres) <= 720 or not getMachineBuild() == 'blackbox7405': if self.session.pipshown: del self.session.pip if SystemInfo["LCDMiniTV"] and int(config.lcd.modepip.value) >= 1: print '[LCDMiniTV] disable PIP' f = open("/proc/stb/lcd/mode", "w") f.write(config.lcd.minitvmode.value) f.close() self.session.pip = self.session.instantiateDialog(PictureInPicture) self.session.pip.setAnimationMode(0) self.session.pip.show() newservice = self.csel.servicelist.getCurrent() currentBouquet = self.csel.servicelist and self.csel.servicelist.getRoot() if newservice and newservice.valid(): if self.session.pip.playService(newservice): self.session.pipshown = True self.session.pip.servicePath = self.csel.getCurrentServicePath() self.session.pip.servicePath[1] = currentBouquet if SystemInfo["LCDMiniTV"] and int(config.lcd.modepip.value) >= 1: print '[LCDMiniTV] enable PIP' f = open("/proc/stb/lcd/mode", "w") f.write(config.lcd.minitvpipmode.value) f.close() f = open("/proc/stb/vmpeg/1/dst_width", "w") f.write("0") f.close() f = open("/proc/stb/vmpeg/1/dst_height", "w") f.write("0") f.close() f = open("/proc/stb/vmpeg/1/dst_apply", "w") f.write("1") f.close() self.close(True) else: self.session.pipshown = False del self.session.pip if SystemInfo["LCDMiniTV"] and int(config.lcd.modepip.value) >= 1: print '[LCDMiniTV] disable PIP' f = open("/proc/stb/lcd/mode", "w") f.write(config.lcd.minitvmode.value) f.close() self.session.openWithCallback(self.close, MessageBox, _("Could not open Picture in Picture"), MessageBox.TYPE_ERROR) else: self.session.open(MessageBox, _("Your %s %s does not support PiP HD") % (getMachineBrand(), getMachineName()), type = MessageBox.TYPE_INFO,timeout = 5 ) def addServiceToBouquetSelected(self): bouquets = self.csel.getBouquetList() if bouquets is None: cnt = 0 else: cnt = len(bouquets) if cnt > 1: # show bouquet list self.bsel = self.session.openWithCallback(self.bouquetSelClosed, BouquetSelector, bouquets, self.addCurrentServiceToBouquet) elif cnt == 1: # add to only one existing bouquet self.addCurrentServiceToBouquet(bouquets[0][1], closeBouquetSelection=False) def bouquetSelClosed(self, recursive): self.bsel = None if recursive: self.close(False) def removeSatelliteServices(self): self.csel.removeSatelliteServices() self.close() def copyCurrentToBouquetList(self): self.csel.copyCurrentToBouquetList() self.close() def showMarkerInputBox(self): self.session.openWithCallback(self.markerInputCallback, VirtualKeyBoard, title=_("Please enter a name for the new marker"), text="markername", maxSize=False, visible_width=56, type=Input.TEXT) def markerInputCallback(self, marker): if marker is not None: self.csel.addMarker(marker) self.close() def addCurrentServiceToBouquet(self, dest, closeBouquetSelection=True): self.csel.addServiceToBouquet(dest) if self.bsel is not None: self.bsel.close(True) else: self.close(closeBouquetSelection) # close bouquet selection def renameEntry(self): if self.inBouquet and self.csel.servicelist.getCurrent() and self.csel.servicelist.getCurrent().valid() and not self.csel.entry_marked: self.csel.renameEntry() self.close() else: return 0 def toggleMoveMode(self): if self.inBouquet and self.csel.servicelist.getCurrent() and self.csel.servicelist.getCurrent().valid(): self.csel.toggleMoveMode() self.close() else: return 0 def toggleMoveModeSelect(self): if self.inBouquet and self.csel.servicelist.getCurrent() and self.csel.servicelist.getCurrent().valid(): self.csel.toggleMoveMode(True) self.close() else: return 0 def bouquetMarkStart(self): self.csel.startMarkedEdit(EDIT_BOUQUET) self.close() def bouquetMarkEnd(self): self.csel.endMarkedEdit(abort=False) self.close() def bouquetMarkAbort(self): self.csel.endMarkedEdit(abort=True) self.close() def removeNewFoundFlag(self): eDVBDB.getInstance().removeFlag(self.csel.getCurrentSelection(), FLAG_SERVICE_NEW_FOUND) self.close() def removeAllNewFoundFlags(self): curpath = self.csel.getCurrentSelection().getPath() idx = curpath.find("satellitePosition == ") if idx != -1: tmp = curpath[idx+21:] idx = tmp.find(')') if idx != -1: satpos = int(tmp[:idx]) eDVBDB.getInstance().removeFlags(FLAG_SERVICE_NEW_FOUND, -1, -1, -1, satpos) self.close() def editAlternativeServices(self): self.csel.startMarkedEdit(EDIT_ALTERNATIVES) self.close() def showAlternativeServices(self): self.csel["Service"].editmode = True self.csel.enterPath(self.csel.getCurrentSelection()) self.close() def removeAlternativeServices(self): self.csel.removeAlternativeServices() self.close() def addAlternativeServices(self): self.csel.addAlternativeServices() self.csel.startMarkedEdit(EDIT_ALTERNATIVES) self.close() def findCurrentlyPlayed(self): sel = self.csel.getCurrentSelection() if sel and sel.valid() and not self.csel.entry_marked: currentPlayingService = (hasattr(self.csel, "dopipzap") and self.csel.dopipzap) and self.session.pip.getCurrentService() or self.session.nav.getCurrentlyPlayingServiceOrGroup() self.csel.servicelist.setCurrent(currentPlayingService, adjust=False) if self.csel.getCurrentSelection() != currentPlayingService: self.csel.setCurrentSelection(sel) self.close() else: return 0 def runPlugin(self, plugin): plugin(session=self.session, service=self.csel.getCurrentSelection()) self.close() class SelectionEventInfo: def __init__(self): self["Service"] = self["ServiceEvent"] = ServiceEvent() self["Event"] = Event() self.servicelist.connectSelChanged(self.__selectionChanged) self.timer = eTimer() self.timer.callback.append(self.updateEventInfo) self.onShown.append(self.__selectionChanged) def __selectionChanged(self): if self.execing: self.timer.start(100, True) def updateEventInfo(self): cur = self.getCurrentSelection() service = self["Service"] service.newService(cur) self["Event"].newEvent(service.event) def parseCurentEvent(list): if len(list) >= 0: list = list[0] begin = list[2] - (config.recording.margin_before.value * 60) end = list[2] + list[3] + (config.recording.margin_after.value * 60) name = list[1] description = list[5] eit = list[0] return begin, end, name, description, eit return False def parseNextEvent(list): if len(list) > 0: list = list[1] begin = list[2] - (config.recording.margin_before.value * 60) end = list[2] + list[3] + (config.recording.margin_after.value * 60) name = list[1] description = list[5] eit = list[0] return begin, end, name, description, eit return False class ChannelSelectionEPG(InfoBarButtonSetup): def __init__(self): self.ChoiceBoxDialog = None self.RemoveTimerDialog = None self.hotkeys = [("Info (EPG)", "info", "Infobar/openEventView"), ("Info (EPG)" + " " + _("long"), "info_long", "Infobar/showEventInfoPlugins"), ("Epg/Guide", "epg", "Infobar/EPGPressed/1"), ("Epg/Guide" + " " + _("long"), "epg_long", "Infobar/showEventInfoPlugins")] self["ChannelSelectEPGActions"] = ButtonSetupActionMap(["ChannelSelectEPGActions"], dict((x[1], self.ButtonSetupGlobal) for x in self.hotkeys)) self.currentSavedPath = [] self.onExecBegin.append(self.clearLongkeyPressed) self["ChannelSelectEPGActions"] = ActionMap(["ChannelSelectEPGActions"], { "showEPGList": self.showEPGList, }) self["recordingactions"] = HelpableActionMap(self, "InfobarInstantRecord", { "ShortRecord": (self.RecordTimerQuestion, _("Add a record timer")), 'LongRecord': (self.doZapTimer, _('Add a zap timer for next event')) },-1) self['dialogactions'] = ActionMap(['SetupActions'], { 'cancel': self.closeChoiceBoxDialog, }) self['dialogactions'].execEnd() def getKeyFunctions(self, key): selection = eval("config.misc.ButtonSetup." + key + ".value.split(',')") selected = [] for x in selection: function = list(function for function in getButtonSetupFunctions() if function[1] == x and function[2] == "EPG") if function: selected.append(function[0]) return selected def RecordTimerQuestion(self): serviceref = ServiceReference(self.getCurrentSelection()) refstr = ':'.join(serviceref.ref.toString().split(':')[:11]) self.epgcache = eEPGCache.getInstance() test = [ 'ITBDSECX', (refstr, 1, -1, 12*60) ] # search next 12 hours self.list = [] if self.epgcache is None else self.epgcache.lookupEvent(test) if len(self.list) < 1: return eventid = self.list[0][0] if len(self.list) == 1: eventidnext = None else: eventidnext = self.list[1][0] eventname = str(self.list[0][1]) if eventid is None: return indx = int(self.servicelist.getCurrentIndex()) selx = self.servicelist.instance.size().width() while indx+1 > config.usage.serviceitems_per_page.value: indx = indx - config.usage.serviceitems_per_page.value pos = self.servicelist.instance.position().y() sely = int(pos)+(int(self.servicelist.ItemHeight)*int(indx)) temp = int(self.servicelist.instance.position().y())+int(self.servicelist.instance.size().height()) if int(sely) >= temp: sely = int(sely) - int(self.listHeight) menu1 = _("Record now") menu2 = _("Record next") for timer in self.session.nav.RecordTimer.timer_list: if timer.eit == eventid and ':'.join(timer.service_ref.ref.toString().split(':')[:11]) == refstr: menu1 = _("Stop recording now") elif eventidnext is not None: if timer.eit == eventidnext and ':'.join(timer.service_ref.ref.toString().split(':')[:11]) == refstr: menu2 = _("Change next timer") if eventidnext is not None: menu = [(menu1, 'CALLFUNC', self.ChoiceBoxCB, self.doRecordCurrentTimer), (menu2, 'CALLFUNC', self.ChoiceBoxCB, self.doRecordNextTimer)] else: menu = [(menu1, 'CALLFUNC', self.ChoiceBoxCB, self.doRecordCurrentTimer)] self.ChoiceBoxDialog = self.session.instantiateDialog(ChoiceBox, list=menu, keys=['red', 'green'], skin_name="RecordTimerQuestion") self.ChoiceBoxDialog.instance.move(ePoint(selx-self.ChoiceBoxDialog.instance.size().width(),self.instance.position().y()+sely)) self.showChoiceBoxDialog() def ChoiceBoxCB(self, choice): self.closeChoiceBoxDialog() if choice: try: choice() except: choice def RemoveTimerDialogCB(self, choice): self.closeChoiceBoxDialog() if choice: choice(self) def showChoiceBoxDialog(self): self['actions'].setEnabled(False) self['recordingactions'].setEnabled(False) self['ChannelSelectEPGActions'].setEnabled(False) self["ChannelSelectBaseActions"].setEnabled(False) self['dialogactions'].execBegin() self.ChoiceBoxDialog['actions'].execBegin() self.ChoiceBoxDialog.show() def closeChoiceBoxDialog(self): self['dialogactions'].execEnd() if self.ChoiceBoxDialog: self.ChoiceBoxDialog['actions'].execEnd() self.session.deleteDialog(self.ChoiceBoxDialog) self['actions'].setEnabled(True) self['recordingactions'].setEnabled(True) self['ChannelSelectEPGActions'].setEnabled(True) self["ChannelSelectBaseActions"].setEnabled(True) def doRecordCurrentTimer(self): self.doInstantTimer(0, parseCurentEvent) def doRecordNextTimer(self): self.doInstantTimer(0, parseNextEvent, True) def doZapTimer(self): self.doInstantTimer(1, parseNextEvent) def editTimer(self, timer): self.session.open(TimerEntry, timer) def doInstantTimer(self, zap, parseEvent, next=False): serviceref = ServiceReference(self.getCurrentSelection()) refstr = ':'.join(serviceref.ref.toString().split(':')[:11]) self.epgcache = eEPGCache.getInstance() test = [ 'ITBDSECX', (refstr, 1, -1, 12*60) ] # search next 12 hours self.list = [] if self.epgcache is None else self.epgcache.lookupEvent(test) if self.list is None: return if not next: eventid = self.list[0][0] eventname = str(self.list[0][1]) else: if len(self.list) < 2: return eventid = self.list[1][0] eventname = str(self.list[1][1]) if eventid is None: return indx = int(self.servicelist.getCurrentIndex()) selx = self.servicelist.instance.size().width() while indx+1 > config.usage.serviceitems_per_page.value: indx = indx - config.usage.serviceitems_per_page.value pos = self.servicelist.instance.position().y() sely = int(pos)+(int(self.servicelist.ItemHeight)*int(indx)) temp = int(self.servicelist.instance.position().y())+int(self.servicelist.instance.size().height()) if int(sely) >= temp: sely = int(sely) - int(self.listHeight) for timer in self.session.nav.RecordTimer.timer_list: if timer.eit == eventid and ':'.join(timer.service_ref.ref.toString().split(':')[:11]) == refstr: if not next: cb_func = lambda ret: self.removeTimer(timer) menu = [(_("Yes"), 'CALLFUNC', cb_func), (_("No"), 'CALLFUNC', self.ChoiceBoxCB)] self.ChoiceBoxDialog = self.session.instantiateDialog(MessageBox, text=_('Do you really want to remove the timer for %s?') % eventname, list=menu, skin_name="RemoveTimerQuestion", picon=False) else: cb_func1 = lambda ret: self.removeTimer(timer) cb_func2 = lambda ret: self.editTimer(timer) menu = [(_("Delete timer"), 'CALLFUNC', self.RemoveTimerDialogCB, cb_func1), (_("Edit timer"), 'CALLFUNC', self.RemoveTimerDialogCB, cb_func2)] self.ChoiceBoxDialog = self.session.instantiateDialog(ChoiceBox, title=_("Select action for timer %s:") % eventname, list=menu, keys=['green', 'blue'], skin_name="RecordTimerQuestion") self.ChoiceBoxDialog.instance.move(ePoint(selx-self.ChoiceBoxDialog.instance.size().width(),self.instance.position().y()+sely)) self.showChoiceBoxDialog() break else: newEntry = RecordTimerEntry(serviceref, checkOldTimers = True, dirname = preferredTimerPath(), *parseEvent(self.list)) if not newEntry: return self.InstantRecordDialog = self.session.instantiateDialog(InstantRecordTimerEntry, newEntry, zap) retval = [True, self.InstantRecordDialog.retval()] self.session.deleteDialogWithCallback(self.finishedAdd, self.InstantRecordDialog, retval) def finishedAdd(self, answer): # print "finished add" if answer[0]: entry = answer[1] simulTimerList = self.session.nav.RecordTimer.record(entry) if simulTimerList is not None: for x in simulTimerList: if x.setAutoincreaseEnd(entry): self.session.nav.RecordTimer.timeChanged(x) simulTimerList = self.session.nav.RecordTimer.record(entry) if simulTimerList is not None: if not entry.repeated and not config.recording.margin_before.value and not config.recording.margin_after.value and len(simulTimerList) > 1: change_time = False conflict_begin = simulTimerList[1].begin conflict_end = simulTimerList[1].end if conflict_begin == entry.end: entry.end -= 30 change_time = True elif entry.begin == conflict_end: entry.begin += 30 change_time = True if change_time: simulTimerList = self.session.nav.RecordTimer.record(entry) if simulTimerList is not None: self.session.openWithCallback(self.finishSanityCorrection, TimerSanityConflict, simulTimerList) def finishSanityCorrection(self, answer): self.finishedAdd(answer) def removeTimer(self, timer): timer.afterEvent = AFTEREVENT.NONE self.session.nav.RecordTimer.removeEntry(timer) self.closeChoiceBoxDialog() def showEPGList(self): ref=self.getCurrentSelection() if ref: self.savedService = ref self.session.openWithCallback(self.SingleServiceEPGClosed, EPGSelection, ref, serviceChangeCB=self.changeServiceCB, EPGtype="single") def SingleServiceEPGClosed(self, ret=False): if ret: service = self.getCurrentSelection() if service is not None: self.saveChannel(service) self.addToHistory(service) self.close() else: self.setCurrentSelection(self.savedService) def changeServiceCB(self, direction, epg): beg = self.getCurrentSelection() while True: if direction > 0: self.moveDown() else: self.moveUp() cur = self.getCurrentSelection() if cur == beg or not (cur.flags & eServiceReference.isMarker): break epg.setService(ServiceReference(self.getCurrentSelection())) def zapToService(self, service, preview=False, zapback=False): if self.startServiceRef is None: self.startServiceRef = self.session.nav.getCurrentlyPlayingServiceOrGroup() if service is not None: if self.servicelist.getRoot() != self.epg_bouquet: self.servicelist.clearPath() if self.servicelist.bouquet_root != self.epg_bouquet: self.servicelist.enterPath(self.servicelist.bouquet_root) self.servicelist.enterPath(self.epg_bouquet) self.servicelist.setCurrent(service) if not zapback or preview: self.zap(enable_pipzap=True) if (self.dopipzap or zapback) and not preview: self.zapBack() if not preview: self.startServiceRef = None self.startRoot = None class ChannelSelectionEdit: def __init__(self): self.entry_marked = False self.bouquet_mark_edit = OFF self.mutableList = None self.__marked = [ ] self.saved_title = None self.saved_root = None self.current_ref = None self.editMode = False self.confirmRemove = True class ChannelSelectionEditActionMap(ActionMap): def __init__(self, csel, contexts=None, actions=None, prio=0): if not contexts: contexts = [] if not actions: actions = {} ActionMap.__init__(self, contexts, actions, prio) self.csel = csel def action(self, contexts, action): if action == "cancel": self.csel.handleEditCancel() return 0 # fall-trough elif action == "ok": return 0 # fall-trough else: return ActionMap.action(self, contexts, action) self["ChannelSelectEditActions"] = ChannelSelectionEditActionMap(self, ["ChannelSelectEditActions", "OkCancelActions"], { "contextMenu": self.doContext, }) def getMutableList(self, root=eServiceReference()): if not self.mutableList is None: return self.mutableList serviceHandler = eServiceCenter.getInstance() if not root.valid(): root=self.getRoot() list = root and serviceHandler.list(root) if list is not None: return list.startEdit() return None def buildBouquetID(self, str): tmp = str.lower() name = '' for c in tmp: if ('a' <= c <= 'z') or ('0' <= c <= '9'): name += c else: name += '_' return name def renameEntry(self): self.editMode = True cur = self.getCurrentSelection() if cur and cur.valid(): name = eServiceCenter.getInstance().info(cur).getName(cur) or ServiceReference(cur).getServiceName() or "" name = name.replace('\xc2\x86', '').replace('\xc2\x87', '') if name: self.session.openWithCallback(self.renameEntryCallback, VirtualKeyBoard, title=_("Please enter new name:"), text=name) else: return 0 def renameEntryCallback(self, name): if name: mutableList = self.getMutableList() if mutableList: current = self.servicelist.getCurrent() current.setName(name) index = self.servicelist.getCurrentIndex() mutableList.removeService(current, False) mutableList.addService(current) mutableList.moveService(current, index) mutableList.flushChanges() self.servicelist.addService(current, True) self.servicelist.removeCurrent() if not self.servicelist.atEnd(): self.servicelist.moveUp() def addMarker(self, name): current = self.servicelist.getCurrent() mutableList = self.getMutableList() cnt = 0 while mutableList: str = '1:64:%d:0:0:0:0:0:0:0::%s'%(cnt, name) ref = eServiceReference(str) if current and current.valid(): if not mutableList.addService(ref, current): self.servicelist.addService(ref, True) mutableList.flushChanges() break elif not mutableList.addService(ref): self.servicelist.addService(ref, True) mutableList.flushChanges() break cnt+=1 def addAlternativeServices(self): cur_service = ServiceReference(self.getCurrentSelection()) root = self.getRoot() cur_root = root and ServiceReference(root) mutableBouquet = cur_root.list().startEdit() if mutableBouquet: name = cur_service.getServiceName() if self.mode == MODE_TV: str = '1:134:1:0:0:0:0:0:0:0:FROM BOUQUET \"alternatives.%s.tv\" ORDER BY bouquet'%(self.buildBouquetID(name)) else: str = '1:134:2:0:0:0:0:0:0:0:FROM BOUQUET \"alternatives.%s.radio\" ORDER BY bouquet'%(self.buildBouquetID(name)) new_ref = ServiceReference(str) if not mutableBouquet.addService(new_ref.ref, cur_service.ref): mutableBouquet.removeService(cur_service.ref) mutableBouquet.flushChanges() eDVBDB.getInstance().reloadBouquets() mutableAlternatives = new_ref.list().startEdit() if mutableAlternatives: mutableAlternatives.setListName(name) if mutableAlternatives.addService(cur_service.ref): print "add", cur_service.ref.toString(), "to new alternatives failed" mutableAlternatives.flushChanges() self.servicelist.addService(new_ref.ref, True) self.servicelist.removeCurrent() if not self.atEnd(): self.servicelist.moveUp() if cur_service.ref.toString() == self.lastservice.value: self.saveChannel(new_ref.ref) if self.startServiceRef and cur_service.ref == self.startServiceRef: self.startServiceRef = new_ref.ref else: print "get mutable list for new created alternatives failed" else: print "add", str, "to", cur_root.getServiceName(), "failed" else: print "bouquetlist is not editable" def addBouquet(self, bName, services): serviceHandler = eServiceCenter.getInstance() mutableBouquetList = serviceHandler.list(self.bouquet_root).startEdit() if mutableBouquetList: if self.mode == MODE_TV: bName += _(" (TV)") str = '1:7:1:0:0:0:0:0:0:0:FROM BOUQUET \"userbouquet.%s.tv\" ORDER BY bouquet'%(self.buildBouquetID(bName)) else: bName += _(" (Radio)") str = '1:7:2:0:0:0:0:0:0:0:FROM BOUQUET \"userbouquet.%s.radio\" ORDER BY bouquet'%(self.buildBouquetID(bName)) new_bouquet_ref = eServiceReference(str) if not mutableBouquetList.addService(new_bouquet_ref): mutableBouquetList.flushChanges() eDVBDB.getInstance().reloadBouquets() mutableBouquet = serviceHandler.list(new_bouquet_ref).startEdit() if mutableBouquet: mutableBouquet.setListName(bName) if services is not None: for service in services: if mutableBouquet.addService(service): print "add", service.toString(), "to new bouquet failed" mutableBouquet.flushChanges() else: print "get mutable list for new created bouquet failed" # do some voodoo to check if current_root is equal to bouquet_root cur_root = self.getRoot() str1 = cur_root and cur_root.toString() pos1 = str1 and str1.find("FROM BOUQUET") or -1 pos2 = self.bouquet_rootstr.find("FROM BOUQUET") if pos1 != -1 and pos2 != -1 and str1[pos1:] == self.bouquet_rootstr[pos2:]: self.servicelist.addService(new_bouquet_ref) self.servicelist.resetRoot() else: print "add", str, "to bouquets failed" else: print "bouquetlist is not editable" def copyCurrentToBouquetList(self): provider = ServiceReference(self.getCurrentSelection()) providerName = provider.getServiceName() serviceHandler = eServiceCenter.getInstance() services = serviceHandler.list(provider.ref) self.addBouquet(providerName, services and services.getContent('R', True)) def removeAlternativeServices(self): cur_service = ServiceReference(self.getCurrentSelection()) end = self.atEnd() root = self.getRoot() cur_root = root and ServiceReference(root) list = cur_service.list() first_in_alternative = list and list.getNext() if first_in_alternative: edit_root = cur_root and cur_root.list().startEdit() if edit_root: if not edit_root.addService(first_in_alternative, cur_service.ref): self.servicelist.addService(first_in_alternative, True) if cur_service.ref.toString() == self.lastservice.value: self.saveChannel(first_in_alternative) if self.startServiceRef and cur_service.ref == self.startServiceRef: self.startServiceRef = first_in_alternative else: print "couldn't add first alternative service to current root" else: print "couldn't edit current root!!" else: print "remove empty alternative list !!" self.removeBouquet() if not end: self.servicelist.moveUp() def removeBouquet(self): refstr = self.getCurrentSelection().toString() pos = refstr.find('FROM BOUQUET "') filename = None self.removeCurrentService(bouquet=True) def removeSatelliteService(self): current = self.getCurrentSelection() eDVBDB.getInstance().removeService(current) refreshServiceList() if not self.atEnd(): self.servicelist.moveUp() def removeSatelliteServices(self): current = self.getCurrentSelection() unsigned_orbpos = current.getUnsignedData(4) >> 16 if unsigned_orbpos == 0xFFFF: messageText = _("Are you sure to remove all cable services?") elif unsigned_orbpos == 0xEEEE: messageText = _("Are you sure to remove all terrestrial services?") else: if unsigned_orbpos > 1800: unsigned_orbpos = 3600 - unsigned_orbpos direction = _("W") else: direction = _("E") messageText = _("Are you sure to remove all %d.%d%s%s services?") % (unsigned_orbpos/10, unsigned_orbpos%10, "\xc2\xb0", direction) self.session.openWithCallback(self.removeSatelliteServicesCallback, MessageBox, messageText) def removeSatelliteServicesCallback(self, answer): if answer: currentIndex = self.servicelist.getCurrentIndex() current = self.getCurrentSelection() unsigned_orbpos = current.getUnsignedData(4) >> 16 if unsigned_orbpos == 0xFFFF: eDVBDB.getInstance().removeServices(int("0xFFFF0000", 16) - 0x100000000) elif unsigned_orbpos == 0xEEEE: eDVBDB.getInstance().removeServices(int("0xEEEE0000", 16) - 0x100000000) else: curpath = current.getPath() idx = curpath.find("satellitePosition == ") if idx != -1: tmp = curpath[idx + 21:] idx = tmp.find(')') if idx != -1: satpos = int(tmp[:idx]) eDVBDB.getInstance().removeServices(-1, -1, -1, satpos) refreshServiceList() if hasattr(self, 'showSatellites'): self.showSatellites() self.servicelist.moveToIndex(currentIndex) if currentIndex != self.servicelist.getCurrentIndex(): self.servicelist.instance.moveSelection(self.servicelist.instance.moveEnd) # multiple marked entry stuff ( edit mode, later multiepg selection ) def startMarkedEdit(self, type): self.savedPath = self.servicePath[:] if type == EDIT_ALTERNATIVES: self.current_ref = self.getCurrentSelection() self.enterPath(self.current_ref) self.mutableList = self.getMutableList() # add all services from the current list to internal marked set in listboxservicecontent self.clearMarks() # this clears the internal marked set in the listboxservicecontent self.saved_title = self.getTitle() pos = self.saved_title.find(')') new_title = self.saved_title[:pos+1] if type == EDIT_ALTERNATIVES: self.bouquet_mark_edit = EDIT_ALTERNATIVES new_title += ' ' + _("[alternative edit]") else: self.bouquet_mark_edit = EDIT_BOUQUET if config.usage.multibouquet.value: new_title += ' ' + _("[bouquet edit]") else: new_title += ' ' + _("[favourite edit]") self.setTitle(new_title) self.__marked = self.servicelist.getRootServices() for x in self.__marked: self.servicelist.addMarked(eServiceReference(x)) self["Service"].editmode = True def endMarkedEdit(self, abort): if not abort and self.mutableList is not None: new_marked = set(self.servicelist.getMarked()) old_marked = set(self.__marked) removed = old_marked - new_marked added = new_marked - old_marked changed = False for x in removed: changed = True self.mutableList.removeService(eServiceReference(x)) for x in added: changed = True self.mutableList.addService(eServiceReference(x)) if changed: if self.bouquet_mark_edit == EDIT_ALTERNATIVES and not new_marked and self.__marked: self.mutableList.addService(eServiceReference(self.__marked[0])) self.mutableList.flushChanges() self.__marked = [] self.clearMarks() self.bouquet_mark_edit = OFF self.mutableList = None self.setTitle(self.saved_title) self.saved_title = None # self.servicePath is just a reference to servicePathTv or Radio... # so we never ever do use the asignment operator in self.servicePath del self.servicePath[:] # remove all elements self.servicePath += self.savedPath # add saved elements del self.savedPath self.setRoot(self.servicePath[-1]) if self.current_ref: self.setCurrentSelection(self.current_ref) self.current_ref = None def clearMarks(self): self.servicelist.clearMarks() def doMark(self): ref = self.servicelist.getCurrent() if self.servicelist.isMarked(ref): self.servicelist.removeMarked(ref) else: self.servicelist.addMarked(ref) def removeCurrentEntry(self, bouquet=False): if self.confirmRemove: list = [(_("yes"), True), (_("no"), False), (_("yes") + ", " + _("and never ask again this session again"), "never")] self.session.openWithCallback(boundFunction(self.removeCurrentEntryCallback, bouquet), MessageBox, _("Are you sure to remove this entry?"), list=list) else: self.removeCurrentEntryCallback(bouquet, True) def removeCurrentEntryCallback(self, bouquet, answer): if answer: if answer == "never": self.confirmRemove = False if bouquet: self.removeBouquet() else: self.removeCurrentService() def removeCurrentService(self, bouquet=False): self.editMode = True ref = self.servicelist.getCurrent() mutableList = self.getMutableList() if ref.valid() and mutableList is not None: if not mutableList.removeService(ref): mutableList.flushChanges() #FIXME dont flush on each single removed service self.servicelist.removeCurrent() self.servicelist.resetRoot() playingref = self.session.nav.getCurrentlyPlayingServiceOrGroup() if not bouquet and playingref and ref == playingref: try: doClose = not config.usage.servicelistpreview_mode.value or ref == self.session.nav.getCurrentlyPlayingServiceOrGroup() except: doClose = False if self.startServiceRef is None and not doClose: self.startServiceRef = self.session.nav.getCurrentlyPlayingServiceOrGroup() ref = self.getCurrentSelection() if self.movemode and (self.isBasePathEqual(self.bouquet_root) or "userbouquet." in ref.toString()): self.toggleMoveMarked() elif (ref.flags & eServiceReference.flagDirectory) == eServiceReference.flagDirectory: if Components.ParentalControl.parentalControl.isServicePlayable(ref, self.bouquetParentalControlCallback, self.session): self.enterPath(ref) self.gotoCurrentServiceOrProvider(ref) elif self.bouquet_mark_edit != OFF: if not (self.bouquet_mark_edit == EDIT_ALTERNATIVES and ref.flags & eServiceReference.isGroup): self.doMark() elif not (ref.flags & eServiceReference.isMarker or ref.type == -1): root = self.getRoot() if not root or not (root.flags & eServiceReference.isGroup): self.zap(enable_pipzap=doClose, preview_zap=not doClose) self.asciiOff() def addServiceToBouquet(self, dest, service=None): mutableList = self.getMutableList(dest) if not mutableList is None: if service is None: #use current selected service service = self.servicelist.getCurrent() if not mutableList.addService(service): mutableList.flushChanges() # do some voodoo to check if current_root is equal to dest cur_root = self.getRoot() str1 = cur_root and cur_root.toString() or -1 str2 = dest.toString() pos1 = str1.find("FROM BOUQUET") pos2 = str2.find("FROM BOUQUET") if pos1 != -1 and pos2 != -1 and str1[pos1:] == str2[pos2:]: self.servicelist.addService(service) self.servicelist.resetRoot() def toggleMoveMode(self, select=False): self.editMode = True if self.movemode: if self.entry_marked: self.toggleMoveMarked() # unmark current entry self.movemode = False self.mutableList.flushChanges() # FIXME add check if changes was made self.mutableList = None self.setTitle(self.saved_title) self.saved_title = None self.servicelist.resetRoot() self.servicelist.setCurrent(self.servicelist.getCurrent()) else: self.mutableList = self.getMutableList() self.movemode = True select and self.toggleMoveMarked() self.saved_title = self.getTitle() pos = self.saved_title.find(')') self.setTitle(self.saved_title[:pos+1] + ' ' + _("[move mode]") + self.saved_title[pos+1:]); self.servicelist.setCurrent(self.servicelist.getCurrent()) self["Service"].editmode = True def handleEditCancel(self): if self.movemode: #movemode active? self.toggleMoveMode() # disable move mode elif self.bouquet_mark_edit != OFF: self.endMarkedEdit(True) # abort edit mode def toggleMoveMarked(self): if self.entry_marked: self.servicelist.setCurrentMarked(False) self.entry_marked = False self.pathChangeDisabled = False # re-enable path change else: self.servicelist.setCurrentMarked(True) self.entry_marked = True self.pathChangeDisabled = True # no path change allowed in movemod def doContext(self): self.session.openWithCallback(self.exitContext, ChannelContextMenu, self) def exitContext(self, close = False): l = self["list"] l.setFontsize() l.setItemsPerPage() l.setMode('MODE_TV') if close: self.cancel() MODE_TV = 0 MODE_RADIO = 1 # type 1 = digital television service # type 4 = nvod reference service (NYI) # type 17 = MPEG-2 HD digital television service # type 22 = advanced codec SD digital television # type 24 = advanced codec SD NVOD reference service (NYI) # type 25 = advanced codec HD digital television # type 27 = advanced codec HD NVOD reference service (NYI) # type 2 = digital radio sound service # type 10 = advanced codec digital radio sound service # type 31 = High Efficiency Video Coing digital television service_types_tv = '1:7:1:0:0:0:0:0:0:0:(type == 1) || (type == 17) || (type == 22) || (type == 25) || (type == 31) || (type == 134) || (type == 195)' service_types_radio = '1:7:2:0:0:0:0:0:0:0:(type == 2) || (type == 10)' class ChannelSelectionBase(Screen): def __init__(self, session): Screen.__init__(self, session) self["key_red"] = Button(_("All")) self["key_green"] = Button(_("Reception lists")) self["key_yellow"] = Button(_("Providers")) self["key_blue"] = Button(_("Favourites")) self["list"] = ServiceList(self) self.servicelist = self["list"] self.numericalTextInput = NumericalTextInput(handleTimeout=False) self.servicePathTV = [ ] self.servicePathRadio = [ ] self.servicePath = [ ] self.history = [ ] self.rootChanged = False self.startRoot = None self.selectionNumber = "" self.clearNumberSelectionNumberTimer = eTimer() self.clearNumberSelectionNumberTimer.callback.append(self.clearNumberSelectionNumber) self.protectContextMenu = True self.mode = MODE_TV self.dopipzap = False self.pathChangeDisabled = False self.movemode = False self.showSatDetails = False self["ChannelSelectBaseActions"] = NumberActionMap(["ChannelSelectBaseActions", "NumberActions", "InputAsciiActions"], { "showFavourites": self.showFavourites, "showAllServices": self.showAllServices, "showProviders": self.showProviders, "showSatellites": boundFunction(self.showSatellites, changeMode=True), "nextBouquet": self.nextBouquet, "prevBouquet": self.prevBouquet, "nextMarker": self.nextMarker, "prevMarker": self.prevMarker, "gotAsciiCode": self.keyAsciiCode, "1": self.keyNumberGlobal, "2": self.keyNumberGlobal, "3": self.keyNumberGlobal, "4": self.keyNumberGlobal, "5": self.keyNumberGlobal, "6": self.keyNumberGlobal, "7": self.keyNumberGlobal, "8": self.keyNumberGlobal, "9": self.keyNumberGlobal, "0": self.keyNumberGlobal }) self.maintitle = _("Channel selection") self.recallBouquetMode() self.onShown.append(self.applyKeyMap) def applyKeyMap(self): if config.usage.show_channel_jump_in_servicelist.value == "alpha": self.numericalTextInput.setUseableChars(u'abcdefghijklmnopqrstuvwxyz1234567890') else: self.numericalTextInput.setUseableChars(u'1234567890ABCDEFGHIJKLMNOPQRSTUVWXYZ') def getBouquetNumOffset(self, bouquet): if not config.usage.multibouquet.value: return 0 str = bouquet.toString() offset = 0 if 'userbouquet.' in bouquet.toCompareString(): serviceHandler = eServiceCenter.getInstance() servicelist = serviceHandler.list(bouquet) if servicelist is not None: while True: serviceIterator = servicelist.getNext() if not serviceIterator.valid(): #check if end of list break number = serviceIterator.getChannelNum() if number > 0: offset = number - 1 break return offset def recallBouquetMode(self): if self.mode == MODE_TV: self.service_types = service_types_tv if config.usage.multibouquet.value: self.bouquet_rootstr = '1:7:1:0:0:0:0:0:0:0:FROM BOUQUET "bouquets.tv" ORDER BY bouquet' else: self.bouquet_rootstr = '%s FROM BOUQUET "userbouquet.favourites.tv" ORDER BY bouquet'% self.service_types else: self.service_types = service_types_radio if config.usage.multibouquet.value: self.bouquet_rootstr = '1:7:1:0:0:0:0:0:0:0:FROM BOUQUET "bouquets.radio" ORDER BY bouquet' else: self.bouquet_rootstr = '%s FROM BOUQUET "userbouquet.favourites.radio" ORDER BY bouquet'% self.service_types self.bouquet_root = eServiceReference(self.bouquet_rootstr) def setTvMode(self): self.mode = MODE_TV self.servicePath = self.servicePathTV self.recallBouquetMode() # title = self.maintitle # pos = title.find(" (") # if pos != -1: # title = title[:pos] title = _(' (TV)') self.setTitle(title) def setRadioMode(self): self.mode = MODE_RADIO self.servicePath = self.servicePathRadio self.recallBouquetMode() # title = self.maintitle # pos = title.find(" (") # if pos != -1: # title = title[:pos] title = _(' (Radio)') self.setTitle(title) def setRoot(self, root, justSet=False): if self.startRoot is None: self.startRoot = self.getRoot() path = root.getPath() isBouquet = 'FROM BOUQUET' in path and (root.flags & eServiceReference.isDirectory) inBouquetRootList = 'FROM BOUQUET "bouquets.' in path #FIXME HACK if not inBouquetRootList and isBouquet: self.servicelist.setMode(ServiceList.MODE_FAVOURITES) else: self.servicelist.setMode(ServiceList.MODE_NORMAL) self.servicelist.setRoot(root, justSet) self.rootChanged = True self.buildTitleString() def removeModeStr(self, str): if self.mode == MODE_TV: pos = str.find(_(' (TV)')) else: pos = str.find(_(' (Radio)')) if pos != -1: return str[:pos] return str def getServiceName(self, ref): str = self.removeModeStr(ServiceReference(ref).getServiceName()) if 'User - bouquets' in str: return _('User - bouquets') if not str: pathstr = ref.getPath() if 'FROM PROVIDERS' in pathstr: return _('Provider') if 'FROM SATELLITES' in pathstr: return _('Reception lists') if ') ORDER BY name' in pathstr: return _('All') return str def buildTitleString(self): titleStr = self.getTitle() nameStr = '' pos = titleStr.find(']') if pos == -1: pos = titleStr.find(')') if pos != -1: titleStr = titleStr[:pos+1] if titleStr.find(' (TV)') != -1: titleStr = titleStr[-5:] elif titleStr.find(' (Radio)') != -1: titleStr = titleStr[-8:] Len = len(self.servicePath) if Len > 0: base_ref = self.servicePath[0] if Len > 1: end_ref = self.servicePath[Len - 1] else: end_ref = None nameStr = self.getServiceName(base_ref) # titleStr += ' - ' + nameStr if end_ref is not None: # if Len > 2: # titleStr += '/../' # else: # titleStr += '/' nameStr = self.getServiceName(end_ref) titleStr += nameStr self.setTitle(titleStr) def moveUp(self): self.servicelist.moveUp() def moveDown(self): self.servicelist.moveDown() def moveTop(self): self.servicelist.moveTop() def moveEnd(self): self.servicelist.moveEnd() def clearPath(self): del self.servicePath[:] def enterPath(self, ref, justSet=False): self.servicePath.append(ref) self.setRoot(ref, justSet) def enterUserbouquet(self, root, save_root=True): self.clearPath() self.recallBouquetMode() if self.bouquet_root: self.enterPath(self.bouquet_root) self.enterPath(root) self.startRoot = None if save_root: self.saveRoot() def pathUp(self, justSet=False): prev = self.servicePath.pop() if self.servicePath: current = self.servicePath[-1] self.setRoot(current, justSet) if not justSet: self.setCurrentSelection(prev) return prev def isBasePathEqual(self, ref): if len(self.servicePath) > 1 and self.servicePath[0] == ref: return True return False def isPrevPathEqual(self, ref): length = len(self.servicePath) if length > 1 and self.servicePath[length - 2] == ref: return True return False def preEnterPath(self, refstr): return False def showAllServices(self): self["key_green"].setText(_("Reception lists")) if not self.pathChangeDisabled: refstr = '%s ORDER BY name'%(self.service_types) if not self.preEnterPath(refstr): ref = eServiceReference(refstr) currentRoot = self.getRoot() if currentRoot is None or currentRoot != ref: self.clearPath() self.enterPath(ref) playingref = self.session.nav.getCurrentlyPlayingServiceOrGroup() if playingref: self.setCurrentSelectionAlternative(playingref) def showSatellites(self, changeMode=False): if not self.pathChangeDisabled: refstr = '%s FROM SATELLITES ORDER BY satellitePosition'%(self.service_types) if self.showSatDetails: self["key_green"].setText(_("Simple")) else: self["key_green"].setText(_("Extended")) if not self.preEnterPath(refstr): ref = eServiceReference(refstr) justSet = False prev = None if self.isBasePathEqual(ref): if self.isPrevPathEqual(ref): justSet = True prev = self.pathUp(justSet) else: currentRoot = self.getRoot() if currentRoot is None or currentRoot != ref: justSet = True self.clearPath() self.enterPath(ref, True) if changeMode and currentRoot and currentRoot == ref: self.showSatDetails = not self.showSatDetails justSet = True self.clearPath() self.enterPath(ref, True) if self.showSatDetails: self["key_green"].setText(_("Simple")) else: self["key_green"].setText(_("Extended")) if justSet: addCableAndTerrestrialLater = [] serviceHandler = eServiceCenter.getInstance() servicelist = serviceHandler.list(ref) if servicelist is not None: while True: service = servicelist.getNext() if not service.valid(): #check if end of list break unsigned_orbpos = service.getUnsignedData(4) >> 16 orbpos = service.getData(4) >> 16 if orbpos < 0: orbpos += 3600 if "FROM PROVIDER" in service.getPath(): service_type = self.showSatDetails and _("Providers") elif ("flags == %d" %(FLAG_SERVICE_NEW_FOUND)) in service.getPath(): service_type = self.showSatDetails and _("New") else: service_type = _("Services") if service_type: if unsigned_orbpos == 0xFFFF: #Cable service_name = _("Cable") addCableAndTerrestrialLater.append(("%s - %s" % (service_name, service_type), service.toString())) elif unsigned_orbpos == 0xEEEE: #Terrestrial service_name = _("Terrestrial") addCableAndTerrestrialLater.append(("%s - %s" % (service_name, service_type), service.toString())) else: try: service_name = str(nimmanager.getSatDescription(orbpos)) except: if orbpos > 1800: # west orbpos = 3600 - orbpos h = _("W") else: h = _("E") service_name = ("%d.%d" + h) % (orbpos / 10, orbpos % 10) service.setName("%s - %s" % (service_name, service_type)) self.servicelist.addService(service) cur_ref = self.session.nav.getCurrentlyPlayingServiceReference() self.servicelist.l.sort() if cur_ref: pos = self.service_types.rfind(':') refstr = '%s (channelID == %08x%04x%04x) && %s ORDER BY name' %(self.service_types[:pos+1], cur_ref.getUnsignedData(4), # NAMESPACE cur_ref.getUnsignedData(2), # TSID cur_ref.getUnsignedData(3), # ONID self.service_types[pos+1:]) ref = eServiceReference(refstr) ref.setName(_("Current transponder")) self.servicelist.addService(ref, beforeCurrent=True) for (service_name, service_ref) in addCableAndTerrestrialLater: ref = eServiceReference(service_ref) ref.setName(service_name) self.servicelist.addService(ref, beforeCurrent=True) self.servicelist.l.FillFinished() if prev is not None: self.setCurrentSelection(prev) elif cur_ref: refstr = cur_ref.toString() op = "".join(refstr.split(':', 10)[6:7]) if len(op) >= 4: hop = int(op[:-4],16) if len(op) >= 7 and not op.endswith('0000'): op = op[:-4] + '0000' refstr = '1:7:0:0:0:0:%s:0:0:0:(satellitePosition == %s) && %s ORDER BY name' % (op, hop, self.service_types[self.service_types.rfind(':')+1:]) self.setCurrentSelectionAlternative(eServiceReference(refstr)) def showProviders(self): self["key_green"].setText(_("Reception lists")) if not self.pathChangeDisabled: refstr = '%s FROM PROVIDERS ORDER BY name'%(self.service_types) if not self.preEnterPath(refstr): ref = eServiceReference(refstr) if self.isBasePathEqual(ref): self.pathUp() else: currentRoot = self.getRoot() if currentRoot is None or currentRoot != ref: self.clearPath() self.enterPath(ref) service = self.session.nav.getCurrentService() if service: info = service.info() if info: provider = info.getInfoString(iServiceInformation.sProvider) refstr = '1:7:0:0:0:0:0:0:0:0:(provider == \"%s\") && %s ORDER BY name:%s' % (provider, self.service_types[self.service_types.rfind(':')+1:],provider) self.setCurrentSelectionAlternative(eServiceReference(refstr)) def changeBouquet(self, direction): if not self.pathChangeDisabled: if len(self.servicePath) > 1: ref = eServiceReference('%s FROM SATELLITES ORDER BY satellitePosition' % self.service_types) if self.isBasePathEqual(ref): self.showSatellites() else: self.pathUp() if direction < 0: self.moveUp() else: self.moveDown() ref = self.getCurrentSelection() self.enterPath(ref) prev = None root = self.getRoot() for path in self.history: if len(path) > 2 and path[1] == root: prev = path[2] if prev is not None: self.setCurrentSelection(prev) def inBouquet(self): if self.servicePath and self.servicePath[0] == self.bouquet_root: return True return False def atBegin(self): return self.servicelist.atBegin() def atEnd(self): return self.servicelist.atEnd() def nextBouquet(self): if "reverseB" in config.usage.servicelist_cursor_behavior.value: if config.usage.channelbutton_mode.value == '0' or config.usage.channelbutton_mode.value == '3': self.changeBouquet(-1) else: self.servicelist.moveDown() else: if config.usage.channelbutton_mode.value == '0' or config.usage.channelbutton_mode.value == '3': self.changeBouquet(+1) else: self.servicelist.moveUp() def prevBouquet(self): if "reverseB" in config.usage.servicelist_cursor_behavior.value: if config.usage.channelbutton_mode.value == '0' or config.usage.channelbutton_mode.value == '3': self.changeBouquet(+1) else: self.servicelist.moveUp() else: if config.usage.channelbutton_mode.value == '0' or config.usage.channelbutton_mode.value == '3': self.changeBouquet(-1) else: self.servicelist.moveDown() def showFavourites(self): self["key_green"].setText(_("Reception lists")) if not self.pathChangeDisabled: if not self.preEnterPath(self.bouquet_rootstr): if self.isBasePathEqual(self.bouquet_root): self.pathUp() else: currentRoot = self.getRoot() if currentRoot is None or currentRoot != self.bouquet_root: self.clearPath() self.enterPath(self.bouquet_root) def keyNumberGlobal(self, number): if config.usage.show_channel_jump_in_servicelist.value == "quick": if self.isBasePathEqual(self.bouquet_root): if hasattr(self, "editMode") and self.editMode: if number == 2: self.renameEntry() if number == 6: self.toggleMoveMode(select=True) if number == 8: self.removeCurrentEntry(bouquet=False) else: self.numberSelectionActions(number) else: current_root = self.getRoot() if current_root and 'FROM BOUQUET "bouquets.' in current_root.getPath(): if hasattr(self, "editMode") and self.editMode: if number == 2: self.renameEntry() if number == 6: self.toggleMoveMode(select=True) if number == 8: self.removeCurrentEntry(bouquet=True) else: self.numberSelectionActions(number) else: unichar = self.numericalTextInput.getKey(number) charstr = unichar.encode("utf-8") if len(charstr) == 1: self.servicelist.moveToChar(charstr[0]) else: unichar = self.numericalTextInput.getKey(number) charstr = unichar.encode("utf-8") if len(charstr) == 1: self.servicelist.moveToChar(charstr[0]) def numberSelectionActions(self, number): if not(hasattr(self, "movemode") and self.movemode): if len(self.selectionNumber)>4: self.clearNumberSelectionNumber() self.selectionNumber = self.selectionNumber + str(number) ref, bouquet = Screens.InfoBar.InfoBar.instance.searchNumber(int(self.selectionNumber), bouquet=self.getRoot()) if ref: if not ref.flags & eServiceReference.isMarker: self.enterUserbouquet(bouquet, save_root=False) self.servicelist.setCurrent(ref) self.clearNumberSelectionNumberTimer.start(1000, True) else: self.clearNumberSelectionNumber() def clearNumberSelectionNumber(self): self.clearNumberSelectionNumberTimer.stop() self.selectionNumber = "" def keyAsciiCode(self): unichar = unichr(getPrevAsciiCode()) charstr = unichar.encode('utf-8') if len(charstr) == 1: self.servicelist.moveToChar(charstr[0]) def getRoot(self): return self.servicelist.getRoot() def getCurrentSelection(self): return self.servicelist.getCurrent() def setCurrentSelection(self, service): if service: self.servicelist.setCurrent(service, adjust=False) def setCurrentSelectionAlternative(self, ref): if self.bouquet_mark_edit == EDIT_ALTERNATIVES and not (ref.flags & eServiceReference.isDirectory): for markedService in self.servicelist.getMarked(): markedService = eServiceReference(markedService) self.setCurrentSelection(markedService) if markedService == self.getCurrentSelection(): return self.setCurrentSelection(ref) def getBouquetList(self): bouquets = [] serviceHandler = eServiceCenter.getInstance() if config.usage.multibouquet.value: list = serviceHandler.list(self.bouquet_root) if list: while True: s = list.getNext() if not s.valid(): break if s.flags & eServiceReference.isDirectory and not s.flags & eServiceReference.isInvisible: info = serviceHandler.info(s) if info: bouquets.append((info.getName(s), s)) return bouquets else: info = serviceHandler.info(self.bouquet_root) if info: bouquets.append((info.getName(self.bouquet_root), self.bouquet_root)) return bouquets return None def keyGoUp(self): if len(self.servicePath) > 1: if self.isBasePathEqual(self.bouquet_root): self.showFavourites() else: ref = eServiceReference('%s FROM SATELLITES ORDER BY satellitePosition'% self.service_types) if self.isBasePathEqual(ref): self.showSatellites() else: ref = eServiceReference('%s FROM PROVIDERS ORDER BY name'% self.service_types) if self.isBasePathEqual(ref): self.showProviders() else: self.showAllServices() def nextMarker(self): self.servicelist.moveToNextMarker() def prevMarker(self): self.servicelist.moveToPrevMarker() def gotoCurrentServiceOrProvider(self, ref): str = ref.toString() if _("Providers") in str: service = self.session.nav.getCurrentService() if service: info = service.info() if info: provider = info.getInfoString(iServiceInformation.sProvider) op = int(self.session.nav.getCurrentlyPlayingServiceOrGroup().toString().split(':')[6][:-4] or "0",16) refstr = '1:7:0:0:0:0:0:0:0:0:(provider == \"%s\") && (satellitePosition == %s) && %s ORDER BY name:%s' % (provider, op, self.service_types[self.service_types.rfind(':')+1:],provider) self.servicelist.setCurrent(eServiceReference(refstr)) elif not self.isBasePathEqual(self.bouquet_root) or self.bouquet_mark_edit == EDIT_ALTERNATIVES: playingref = self.session.nav.getCurrentlyPlayingServiceOrGroup() if playingref: self.setCurrentSelectionAlternative(playingref) HISTORYSIZE = 20 #config for lastservice config.tv = ConfigSubsection() config.tv.lastservice = ConfigText() config.tv.lastroot = ConfigText() config.radio = ConfigSubsection() config.radio.lastservice = ConfigText() config.radio.lastroot = ConfigText() config.servicelist = ConfigSubsection() config.servicelist.lastmode = ConfigText(default='tv') config.servicelist.startupservice = ConfigText() config.servicelist.startupservice_standby = ConfigText() config.servicelist.startupservice_onstandby = ConfigYesNo(default = False) config.servicelist.startuproot = ConfigText() config.servicelist.startupmode = ConfigText(default='tv') class ChannelSelection(ChannelSelectionBase, ChannelSelectionEdit, ChannelSelectionEPG, SelectionEventInfo): instance = None def __init__(self, session): ChannelSelectionBase.__init__(self, session) ChannelSelectionEdit.__init__(self) ChannelSelectionEPG.__init__(self) SelectionEventInfo.__init__(self) if config.usage.servicelist_mode.value == 'simple': self.skinName = ["SlimChannelSelection","SimpleChannelSelection","ChannelSelection"] else: self.skinName = "ChannelSelection" self["actions"] = ActionMap(["OkCancelActions", "TvRadioActions"], { "cancel": self.cancel, "ok": self.channelSelected, "keyRadio": self.toogleTvRadio, "keyTV": self.toogleTvRadio, }) self.radioTV = 0 self.__event_tracker = ServiceEventTracker(screen=self, eventmap= { iPlayableService.evStart: self.__evServiceStart, iPlayableService.evEnd: self.__evServiceEnd }) assert ChannelSelection.instance is None, "class InfoBar is a singleton class and just one instance of this class is allowed!" ChannelSelection.instance = self self.startServiceRef = None self.history_tv = [] self.history_radio = [] self.history = self.history_tv self.history_pos = 0 self.delhistpoint = None if config.servicelist.startupservice.value and config.servicelist.startuproot.value: config.servicelist.lastmode.value = config.servicelist.startupmode.value if config.servicelist.lastmode.value == 'tv': config.tv.lastservice.value = config.servicelist.startupservice.value config.tv.lastroot.value = config.servicelist.startuproot.value elif config.servicelist.lastmode.value == 'radio': config.radio.lastservice.value = config.servicelist.startupservice.value config.radio.lastroot.value = config.servicelist.startuproot.value self.lastservice = config.tv.lastservice self.lastroot = config.tv.lastroot self.revertMode = None config.usage.multibouquet.addNotifier(self.multibouquet_config_changed) self.new_service_played = False self.dopipzap = False if config.misc.remotecontrol_text_support.value: self.onExecBegin.append(self.asciiOff) else: self.onExecBegin.append(self.asciiOn) self.mainScreenMode = None self.mainScreenRoot = None self.lastChannelRootTimer = eTimer() self.lastChannelRootTimer.callback.append(self.__onCreate) self.lastChannelRootTimer.start(100, True) self.pipzaptimer = eTimer() def asciiOn(self): rcinput = eRCInput.getInstance() rcinput.setKeyboardMode(rcinput.kmAscii) def asciiOff(self): rcinput = eRCInput.getInstance() rcinput.setKeyboardMode(rcinput.kmNone) def multibouquet_config_changed(self, val): self.recallBouquetMode() def __evServiceStart(self): if self.dopipzap and hasattr(self.session, 'pip'): self.servicelist.setPlayableIgnoreService(self.session.pip.getCurrentServiceReference() or eServiceReference()) else: service = self.session.nav.getCurrentService() if service: info = service.info() if info: refstr = info.getInfoString(iServiceInformation.sServiceref) self.servicelist.setPlayableIgnoreService(eServiceReference(refstr)) def __evServiceEnd(self): self.servicelist.setPlayableIgnoreService(eServiceReference()) def setMode(self): self.rootChanged = True self.restoreRoot() lastservice = eServiceReference(self.lastservice.value) if lastservice.valid(): self.setCurrentSelection(lastservice) def toogleTvRadio(self): if self.radioTV == 1: self.radioTV = 0 self.setModeTv() else: self.radioTV = 1 self.setModeRadio() def setModeTv(self): if self.revertMode is None and config.servicelist.lastmode.value == 'radio': self.revertMode = MODE_RADIO self.history = self.history_tv self.lastservice = config.tv.lastservice self.lastroot = config.tv.lastroot config.servicelist.lastmode.value = 'tv' self.setTvMode() self.setMode() def setModeRadio(self): if self.revertMode is None and config.servicelist.lastmode.value == 'tv': self.revertMode = MODE_TV if config.usage.e1like_radio_mode.value: self.history = self.history_radio self.lastservice = config.radio.lastservice self.lastroot = config.radio.lastroot config.servicelist.lastmode.value = 'radio' self.setRadioMode() self.setMode() def __onCreate(self): if config.usage.e1like_radio_mode.value: if config.servicelist.lastmode.value == 'tv': self.setModeTv() else: self.setModeRadio() else: self.setModeTv() lastservice = eServiceReference(self.lastservice.value) if lastservice.valid(): self.zap() def channelSelected(self): ref = self.getCurrentSelection() try: doClose = not config.usage.servicelistpreview_mode.value or ref == self.session.nav.getCurrentlyPlayingServiceOrGroup() except: doClose = False if self.startServiceRef is None and not doClose: self.startServiceRef = self.session.nav.getCurrentlyPlayingServiceOrGroup() ref = self.getCurrentSelection() if self.movemode and (self.isBasePathEqual(self.bouquet_root) or "userbouquet." in ref.toString()): self.toggleMoveMarked() elif (ref.flags & eServiceReference.flagDirectory) == eServiceReference.flagDirectory: if Components.ParentalControl.parentalControl.isServicePlayable(ref, self.bouquetParentalControlCallback, self.session): self.enterPath(ref) self.gotoCurrentServiceOrProvider(ref) elif self.bouquet_mark_edit != OFF: if not (self.bouquet_mark_edit == EDIT_ALTERNATIVES and ref.flags & eServiceReference.isGroup): self.doMark() elif not (ref.flags & eServiceReference.isMarker or ref.type == -1): root = self.getRoot() if not root or not (root.flags & eServiceReference.isGroup): self.zap(enable_pipzap=doClose, preview_zap=not doClose) self.asciiOff() if doClose: if self.dopipzap: self.zapBack() self.startServiceRef = None self.startRoot = None self.correctChannelNumber() self.movemode and self.toggleMoveMode() self.editMode = False self.protectContextMenu = True self["key_green"].setText(_("Reception lists")) self.close(ref) def bouquetParentalControlCallback(self, ref): self.enterPath(ref) self.gotoCurrentServiceOrProvider(ref) def togglePipzap(self): assert self.session.pip title = self.instance.getTitle() pos = title.find(' (') if pos != -1: title = title[:pos] if self.dopipzap: # Mark PiP as inactive and effectively deactivate pipzap self.hidePipzapMessage() self.dopipzap = False # Disable PiP if not playing a service if self.session.pip.pipservice is None: self.session.pipshown = False del self.session.pip self.__evServiceStart() # Move to playing service lastservice = eServiceReference(self.lastservice.value) if lastservice.valid() and self.getCurrentSelection() != lastservice: self.setCurrentSelection(lastservice) title += _(' (TV)') else: # Mark PiP as active and effectively active pipzap self.showPipzapMessage() self.dopipzap = True self.__evServiceStart() # Move to service playing in pip (will not work with subservices) self.setCurrentSelection(self.session.pip.getCurrentService()) title += _(' (PiP)') self.setTitle(title) self.buildTitleString() def showPipzapMessage(self): time = config.usage.infobar_timeout.index if time: self.pipzaptimer.callback.append(self.hidePipzapMessage) self.pipzaptimer.startLongTimer(time) self.session.pip.active() def hidePipzapMessage(self): if self.pipzaptimer.isActive(): self.pipzaptimer.callback.remove(self.hidePipzapMessage) self.pipzaptimer.stop() self.session.pip.inactive() #called from infoBar and channelSelected def zap(self, enable_pipzap=False, preview_zap=False, checkParentalControl=True, ref=None): self.curRoot = self.startRoot nref = ref or self.getCurrentSelection() ref = self.session.nav.getCurrentlyPlayingServiceOrGroup() if enable_pipzap and self.dopipzap: ref = self.session.pip.getCurrentService() if ref is None or ref != nref: nref = self.session.pip.resolveAlternatePipService(nref) if nref and (not checkParentalControl or Components.ParentalControl.parentalControl.isServicePlayable(nref, boundFunction(self.zap, enable_pipzap=True, checkParentalControl=False))): self.session.pip.playService(nref) self.__evServiceStart() self.showPipzapMessage() else: self.setStartRoot(self.curRoot) self.setCurrentSelection(ref) elif ref is None or ref != nref: Screens.InfoBar.InfoBar.instance.checkTimeshiftRunning(boundFunction(self.zapCheckTimeshiftCallback, enable_pipzap, preview_zap, nref)) elif not preview_zap: self.saveRoot() self.saveChannel(nref) config.servicelist.lastmode.save() self.setCurrentSelection(nref) if self.startServiceRef is None or nref != self.startServiceRef: self.addToHistory(nref) self.rootChanged = False self.revertMode = None def zapCheckTimeshiftCallback(self, enable_pipzap, preview_zap, nref, answer): if answer: self.new_service_played = True self.session.nav.playService(nref) if not preview_zap: self.saveRoot() self.saveChannel(nref) config.servicelist.lastmode.save() if self.startServiceRef is None or nref != self.startServiceRef: self.addToHistory(nref) if self.dopipzap: self.setCurrentSelection(self.session.pip.getCurrentService()) else: self.mainScreenMode = config.servicelist.lastmode.value self.mainScreenRoot = self.getRoot() self.revertMode = None else: Notifications.RemovePopup("Parental control") self.setCurrentSelection(nref) else: self.setStartRoot(self.curRoot) self.setCurrentSelection(self.session.nav.getCurrentlyPlayingServiceOrGroup()) if not preview_zap: self.hide() def newServicePlayed(self): ret = self.new_service_played self.new_service_played = False return ret def addToHistory(self, ref): if self.delhistpoint is not None: x = self.delhistpoint while x <= len(self.history)-1: del self.history[x] self.delhistpoint = None if self.servicePath is not None: tmp = self.servicePath[:] tmp.append(ref) self.history.append(tmp) hlen = len(self.history) x = 0 while x < hlen - 1: if self.history[x][-1] == ref: del self.history[x] hlen -= 1 else: x += 1 if hlen > HISTORYSIZE: del self.history[0] hlen -= 1 self.history_pos = hlen - 1 def historyBack(self): hlen = len(self.history) currentPlayedRef = self.session.nav.getCurrentlyPlayingServiceOrGroup() if hlen > 0 and currentPlayedRef and self.history[self.history_pos][-1] != currentPlayedRef: self.addToHistory(currentPlayedRef) hlen = len(self.history) if hlen > 1 and self.history_pos > 0: self.history_pos -= 1 self.setHistoryPath() self.delhistpoint = self.history_pos+1 def historyNext(self): self.delhistpoint = None hlen = len(self.history) if hlen > 1 and self.history_pos < hlen - 1: self.history_pos += 1 self.setHistoryPath() def setHistoryPath(self, doZap=True): path = self.history[self.history_pos][:] ref = path.pop() del self.servicePath[:] self.servicePath += path self.saveRoot() root = path[-1] cur_root = self.getRoot() if cur_root and cur_root != root: self.setRoot(root) self.servicelist.setCurrent(ref) if doZap: self.session.nav.playService(ref) if self.dopipzap: self.setCurrentSelection(self.session.pip.getCurrentService()) else: self.setCurrentSelection(ref) self.saveChannel(ref) def historyClear(self): if self and self.servicelist: for i in range(0, len(self.history)-1): del self.history[0] self.history_pos = len(self.history)-1 return True return False def historyZap(self, direction): hlen = len(self.history) if hlen < 1: return mark = self.history_pos selpos = self.history_pos + direction if selpos < 0: selpos = 0 if selpos > hlen-1: selpos = hlen-1 serviceHandler = eServiceCenter.getInstance() historylist = [ ] for x in self.history: info = serviceHandler.info(x[-1]) if info: historylist.append((info.getName(x[-1]), x[-1])) self.session.openWithCallback(self.historyMenuClosed, HistoryZapSelector, historylist, selpos, mark, invert_items=True, redirect_buttons=True, wrap_around=True) def historyMenuClosed(self, retval): if not retval: return hlen = len(self.history) pos = 0 for x in self.history: if x[-1] == retval: break pos += 1 self.delhistpoint = pos+1 if pos < hlen and pos != self.history_pos: tmp = self.history[pos] # self.history.append(tmp) # del self.history[pos] self.history_pos = pos self.setHistoryPath() def saveRoot(self): path = '' for i in self.servicePath: path += i.toString() path += ';' if path and path != self.lastroot.value: if self.mode == MODE_RADIO and 'FROM BOUQUET "bouquets.tv"' in path: self.setModeTv() elif self.mode == MODE_TV and 'FROM BOUQUET "bouquets.radio"' in path: self.setModeRadio() self.lastroot.value = path self.lastroot.save() def restoreRoot(self): tmp = [ x for x in self.lastroot.value.split(';') if x != '' ] current = [ x.toString() for x in self.servicePath ] if tmp != current or self.rootChanged: self.clearPath() cnt = 0 for i in tmp: self.servicePath.append(eServiceReference(i)) cnt += 1 if cnt: path = self.servicePath.pop() self.enterPath(path) else: self.showFavourites() self.saveRoot() self.rootChanged = False def preEnterPath(self, refstr): if self.servicePath and self.servicePath[0] != eServiceReference(refstr): pathstr = self.lastroot.value if pathstr is not None and refstr in pathstr: self.restoreRoot() lastservice = eServiceReference(self.lastservice.value) if lastservice.valid(): self.setCurrentSelection(lastservice) return True return False def saveChannel(self, ref): if ref is not None: refstr = ref.toString() else: refstr = '' if refstr != self.lastservice.value: self.lastservice.value = refstr self.lastservice.save() def setCurrentServicePath(self, path, doZap=True): if self.history: self.history[self.history_pos] = path else: self.history.append(path) self.setHistoryPath(doZap) def getCurrentServicePath(self): if self.history: return self.history[self.history_pos] return None def recallPrevService(self): hlen = len(self.history) currentPlayedRef = self.session.nav.getCurrentlyPlayingServiceOrGroup() if hlen > 0 and currentPlayedRef and self.history[self.history_pos][-1] != currentPlayedRef: self.addToHistory(currentPlayedRef) hlen = len(self.history) if hlen > 1: if self.history_pos == hlen - 1: tmp = self.history[self.history_pos] self.history[self.history_pos] = self.history[self.history_pos - 1] self.history[self.history_pos - 1] = tmp else: tmp = self.history[self.history_pos + 1] self.history[self.history_pos + 1] = self.history[self.history_pos] self.history[self.history_pos] = tmp self.setHistoryPath() def cancel(self): if self.revertMode is None: self.restoreRoot() if self.dopipzap: # This unfortunately won't work with subservices self.setCurrentSelection(self.session.pip.getCurrentService()) else: lastservice = eServiceReference(self.lastservice.value) if lastservice.valid() and self.getCurrentSelection() != lastservice: self.setCurrentSelection(lastservice) self.asciiOff() if config.usage.servicelistpreview_mode.value: self.zapBack() self.correctChannelNumber() self.editMode = False self.protectContextMenu = True self.close(None) def zapBack(self): currentPlayedRef = self.session.nav.getCurrentlyPlayingServiceOrGroup() if self.startServiceRef and (currentPlayedRef is None or currentPlayedRef != self.startServiceRef): self.setStartRoot(self.startRoot) self.new_service_played = True self.session.nav.playService(self.startServiceRef) self.saveChannel(self.startServiceRef) else: self.restoreMode() self.startServiceRef = None self.startRoot = None if self.dopipzap: # This unfortunately won't work with subservices self.setCurrentSelection(self.session.pip.getCurrentService()) else: lastservice = eServiceReference(self.lastservice.value) if lastservice.valid() and self.getCurrentSelection() == lastservice: pass # keep current selection else: self.setCurrentSelection(currentPlayedRef) def setStartRoot(self, root): if root: if self.revertMode == MODE_TV: self.setModeTv() elif self.revertMode == MODE_RADIO: self.setModeRadio() self.revertMode = None self.enterUserbouquet(root) def restoreMode(self): if self.revertMode == MODE_TV: self.setModeTv() elif self.revertMode == MODE_RADIO: self.setModeRadio() self.revertMode = None def correctChannelNumber(self): current_ref = self.session.nav.getCurrentlyPlayingServiceOrGroup() if self.dopipzap: tmp_mode = config.servicelist.lastmode.value tmp_root = self.getRoot() tmp_ref = self.getCurrentSelection() pip_ref = self.session.pip.getCurrentService() if tmp_ref and pip_ref and tmp_ref != pip_ref: self.revertMode = None return if self.mainScreenMode == "tv": self.setModeTv() elif self.mainScreenMode == "radio": self.setModeRadio() if self.mainScreenRoot: self.setRoot(self.mainScreenRoot) self.setCurrentSelection(current_ref) selected_ref = self.getCurrentSelection() if selected_ref and current_ref and selected_ref.getChannelNum() != current_ref.getChannelNum(): oldref = self.session.nav.currentlyPlayingServiceReference if oldref and selected_ref == oldref or (oldref != current_ref and selected_ref == current_ref): self.session.nav.currentlyPlayingServiceOrGroup = selected_ref self.session.nav.pnav.navEvent(iPlayableService.evStart) if self.dopipzap: if tmp_mode == "tv": self.setModeTv() elif tmp_mode == "radio": self.setModeRadio() self.enterUserbouquet(tmp_root) title = self.instance.getTitle() pos = title.find(" (") if pos != -1: title = title[:pos] title += _(" (PiP)") self.setTitle(title) self.buildTitleString() if tmp_ref and pip_ref and tmp_ref.getChannelNum() != pip_ref.getChannelNum(): self.session.pip.currentService = tmp_ref self.setCurrentSelection(tmp_ref) self.revertMode = None class PiPZapSelection(ChannelSelection): def __init__(self, session): ChannelSelection.__init__(self, session) self.skinName = ["SlimChannelSelection","SimpleChannelSelection","ChannelSelection"] self.startservice = None self.pipzapfailed = None if plugin_PiPServiceRelation_installed: self.pipServiceRelation = getRelationDict() else: self.pipServiceRelation = {} self.keymaptimer = eTimer() self.keymaptimer.callback.append(self.enableKeyMap) self.onShown.append(self.disableKeyMap) def disableKeyMap(self): if not hasattr(self.session, 'pip'): if not self.pipzapfailed: self.startservice = self.session.nav.getCurrentlyPlayingServiceReference() or self.servicelist.getCurrent() else: self.startservice = self.startservice self.setCurrentSelection(self.startservice) self.session.pip = self.session.instantiateDialog(PictureInPicture) self.session.pip.show() self.session.pip.playService(self.startservice) eActionMap.getInstance().unbindNativeKey("ListboxActions", 0) eActionMap.getInstance().unbindNativeKey("ListboxActions", 1) self.keymaptimer.start(1000, True) def enableKeyMap(self): eActionMap.getInstance().bindKey("keymap.xml", "generic", 103, 5, "ListboxActions", "moveUp") eActionMap.getInstance().bindKey("keymap.xml", "generic", 108, 5, "ListboxActions", "moveDown") def channelSelected(self): ref = self.servicelist.getCurrent() if (ref.flags & eServiceReference.flagDirectory) == eServiceReference.flagDirectory: self.enterPath(ref) self.gotoCurrentServiceOrProvider(ref) elif not (ref.flags & eServiceReference.isMarker or ref.toString().startswith("-1")): root = self.getRoot() if not root or not (root.flags & eServiceReference.isGroup): n_service = self.pipServiceRelation.get(str(ref), None) if n_service is not None: newservice = eServiceReference(n_service) else: newservice = ref if not hasattr(self.session, 'pip'): self.session.pip = self.session.instantiateDialog(PictureInPicture) self.session.pip.show() if self.session.pip.playService(newservice): self.pipzapfailed = False self.session.pipshown = True self.session.pip.servicePath = self.getCurrentServicePath() self.setStartRoot(self.curRoot) self.saveRoot() self.saveChannel(ref) self.setCurrentSelection(ref) if SystemInfo["LCDMiniTVPiP"] and int(config.lcd.minitvpipmode.value) >= 1: print '[LCDMiniTV] enable PIP' f = open("/proc/stb/lcd/mode", "w") f.write(config.lcd.minitvpipmode.value) f.close() f = open("/proc/stb/vmpeg/1/dst_width", "w") f.write("0") f.close() f = open("/proc/stb/vmpeg/1/dst_height", "w") f.write("0") f.close() f = open("/proc/stb/vmpeg/1/dst_apply", "w") f.write("1") f.close() self.close(True) else: self.pipzapfailed = True self.session.pipshown = False del self.session.pip if SystemInfo["LCDMiniTVPiP"] and int(config.lcd.minitvpipmode.value) >= 1: print '[LCDMiniTV] disable PIP' f = open("/proc/stb/lcd/mode", "w") f.write(config.lcd.minitvmode.value) f.close() self.close(None) def cancel(self): self.asciiOff() if self.startservice and hasattr(self.session, 'pip') and self.session.pip.getCurrentService() and self.startservice == self.session.pip.getCurrentService(): self.session.pipshown = False del self.session.pip if SystemInfo["LCDMiniTVPiP"] and int(config.lcd.minitvpipmode.value) >= 1: print '[LCDMiniTV] disable PIP' f = open("/proc/stb/lcd/mode", "w") f.write(config.lcd.minitvmode.value) f.close() self.correctChannelNumber() self.close(None) class RadioInfoBar(Screen): def __init__(self, session): Screen.__init__(self, session) Screen.setTitle(self, _("Radio Channel Selection")) self['RdsDecoder'] = RdsDecoder(self.session.nav) class ChannelSelectionRadio(ChannelSelectionBase, ChannelSelectionEdit, ChannelSelectionEPG, InfoBarBase, SelectionEventInfo): ALLOW_SUSPEND = True def __init__(self, session, infobar): ChannelSelectionBase.__init__(self, session) ChannelSelectionEdit.__init__(self) ChannelSelectionEPG.__init__(self) InfoBarBase.__init__(self) SelectionEventInfo.__init__(self) self.infobar = infobar self.startServiceRef = None self.onLayoutFinish.append(self.onCreate) self.info = session.instantiateDialog(RadioInfoBar) # our simple infobar self.info.setAnimationMode(0) self["actions"] = ActionMap(["OkCancelActions", "TvRadioActions"], { "keyTV": self.cancel, "keyRadio": self.cancel, "cancel": self.cancel, "ok": self.channelSelected, }) self.__event_tracker = ServiceEventTracker(screen=self, eventmap= { iPlayableService.evStart: self.__evServiceStart, iPlayableService.evEnd: self.__evServiceEnd }) ########## RDS Radiotext / Rass Support BEGIN self.infobar = infobar # reference to real infobar (the one and only) self["RdsDecoder"] = self.info["RdsDecoder"] self["RdsActions"] = HelpableActionMap(self, "InfobarRdsActions", { "startRassInteractive": (self.startRassInteractive, _("View Rass interactive...")) },-1) self["RdsActions"].setEnabled(False) infobar.rds_display.onRassInteractivePossibilityChanged.append(self.RassInteractivePossibilityChanged) self.onClose.append(self.__onClose) self.onExecBegin.append(self.__onExecBegin) self.onExecEnd.append(self.__onExecEnd) def __onClose(self): del self.info["RdsDecoder"] self.session.deleteDialog(self.info) self.infobar.rds_display.onRassInteractivePossibilityChanged.remove(self.RassInteractivePossibilityChanged) lastservice=eServiceReference(config.tv.lastservice.value) self.session.nav.playService(lastservice) def startRassInteractive(self): self.info.hide() self.infobar.rass_interactive = self.session.openWithCallback(self.RassInteractiveClosed, RassInteractive) def RassInteractiveClosed(self): self.info.show() self.infobar.rass_interactive = None self.infobar.RassSlidePicChanged() def RassInteractivePossibilityChanged(self, state): self['RdsActions'].setEnabled(state) def __onExecBegin(self): self.info.show() def __onExecEnd(self): self.info.hide() def cancel(self): self.info.hide() self.close(None) def __evServiceStart(self): service = self.session.nav.getCurrentService() if service: info = service.info() if info: refstr = info.getInfoString(iServiceInformation.sServiceref) self.servicelist.setPlayableIgnoreService(eServiceReference(refstr)) def __evServiceEnd(self): self.servicelist.setPlayableIgnoreService(eServiceReference()) def saveRoot(self): path = '' for i in self.servicePathRadio: path += i.toString() path += ';' if path and path != config.radio.lastroot.value: config.radio.lastroot.value = path config.radio.lastroot.save() def restoreRoot(self): tmp = [ x for x in config.radio.lastroot.value.split(';') if x != '' ] current = [ x.toString() for x in self.servicePath ] if tmp != current or self.rootChanged: cnt = 0 for i in tmp: self.servicePathRadio.append(eServiceReference(i)) cnt += 1 if cnt: path = self.servicePathRadio.pop() self.enterPath(path) else: self.showFavourites() self.saveRoot() self.rootChanged = False def preEnterPath(self, refstr): if self.servicePathRadio and self.servicePathRadio[0] != eServiceReference(refstr): pathstr = config.radio.lastroot.value if pathstr is not None and refstr in pathstr: self.restoreRoot() lastservice = eServiceReference(config.radio.lastservice.value) if lastservice.valid(): self.setCurrentSelection(lastservice) return True return False def onCreate(self): self.setRadioMode() self.restoreRoot() lastservice = eServiceReference(config.radio.lastservice.value) if lastservice.valid(): self.servicelist.setCurrent(lastservice) self.session.nav.playService(lastservice) else: self.session.nav.stopService() self.info.show() def channelSelected(self, doClose=False): # just return selected service ref = self.getCurrentSelection() if self.movemode: self.toggleMoveMarked() elif (ref.flags & eServiceReference.flagDirectory) == eServiceReference.flagDirectory: self.enterPath(ref) self.gotoCurrentServiceOrProvider(ref) elif self.bouquet_mark_edit != OFF: if not (self.bouquet_mark_edit == EDIT_ALTERNATIVES and ref.flags & eServiceReference.isGroup): self.doMark() elif not (ref.flags & eServiceReference.isMarker): # no marker cur_root = self.getRoot() if not cur_root or not (cur_root.flags & eServiceReference.isGroup): playingref = self.session.nav.getCurrentlyPlayingServiceOrGroup() if playingref is None or playingref != ref: self.session.nav.playService(ref) config.radio.lastservice.value = ref.toString() config.radio.lastservice.save() self.saveRoot() def zapBack(self): self.channelSelected() class SimpleChannelSelection(ChannelSelectionBase): def __init__(self, session, title, currentBouquet=False): ChannelSelectionBase.__init__(self, session) self["actions"] = ActionMap(["OkCancelActions", "TvRadioActions"], { "cancel": self.close, "ok": self.channelSelected, "keyRadio": self.setModeRadio, "keyTV": self.setModeTv, }) self.bouquet_mark_edit = OFF self.title = title self.currentBouquet = currentBouquet self.onLayoutFinish.append(self.layoutFinished) def layoutFinished(self): self.setModeTv() if self.currentBouquet: ref = Screens.InfoBar.InfoBar.instance.servicelist.getRoot() if ref: self.enterPath(ref) self.gotoCurrentServiceOrProvider(ref) def saveRoot(self): pass def keyRecord(self): return 0 def channelSelected(self): # just return selected service ref = self.getCurrentSelection() if (ref.flags & eServiceReference.flagDirectory) == eServiceReference.flagDirectory: self.enterPath(ref) self.gotoCurrentServiceOrProvider(ref) elif not (ref.flags & eServiceReference.isMarker): ref = self.getCurrentSelection() self.close(ref) def setModeTv(self): self.setTvMode() self.showFavourites() def setModeRadio(self): self.setRadioMode() self.showFavourites() class HistoryZapSelector(Screen): def __init__(self, session, items=None, sel_item=0, mark_item=0, invert_items=False, redirect_buttons=False, wrap_around=True): if not items: items = [] Screen.__init__(self, session) self.redirectButton = redirect_buttons self.invertItems = invert_items if self.invertItems: self.currentPos = len(items) - sel_item - 1 else: self.currentPos = sel_item self["actions"] = ActionMap(["OkCancelActions", "InfobarCueSheetActions"], { "ok": self.okbuttonClick, "cancel": self.cancelClick, "jumpPreviousMark": self.prev, "jumpNextMark": self.next, "toggleMark": self.okbuttonClick, }) self.setTitle(_("History zap...")) self.list = [] cnt = 0 serviceHandler = eServiceCenter.getInstance() for x in items: info = serviceHandler.info(x[-1]) if info: serviceName = info.getName(x[-1]) if serviceName is None: serviceName = "" eventName = "" descriptionName = "" durationTime = "" # if config.plugins.SetupZapSelector.event.value != "0": event = info.getEvent(x[-1]) if event: eventName = event.getEventName() if eventName is None: eventName = "" else: eventName = eventName.replace('(18+)', '').replace('18+', '').replace('(16+)', '').replace('16+', '').replace('(12+)', '').replace('12+', '').replace('(7+)', '').replace('7+', '').replace('(6+)', '').replace('6+', '').replace('(0+)', '').replace('0+', '') # if config.plugins.SetupZapSelector.event.value == "2": descriptionName = event.getShortDescription() if descriptionName is None or descriptionName == "": descriptionName = event.getExtendedDescription() if descriptionName is None: descriptionName = "" # if config.plugins.SetupZapSelector.duration.value: begin = event.getBeginTime() if begin is not None: end = begin + event.getDuration() remaining = (end - int(time())) / 60 prefix = "" if remaining > 0: prefix = "+" local_begin = localtime(begin) local_end = localtime(end) durationTime = _("%02d.%02d - %02d.%02d (%s%d min)") % (local_begin[3],local_begin[4],local_end[3],local_end[4],prefix, remaining) png = "" picon = getPiconName(str(ServiceReference(x[1]))) if picon != "": png = loadPNG(picon) if self.invertItems: self.list.insert(0, (x[1], cnt == mark_item and "»" or "", x[0], eventName, descriptionName, durationTime, png)) else: self.list.append((x[1], cnt == mark_item and "»" or "", x[0], eventName, descriptionName, durationTime, png)) cnt += 1 self["menu"] = List(self.list, enableWrapAround=wrap_around) self.onShown.append(self.__onShown) def __onShown(self): self["menu"].index = self.currentPos def prev(self): if self.redirectButton: self.down() else: self.up() def next(self): if self.redirectButton: self.up() else: self.down() def up(self): self["menu"].selectPrevious() def down(self): self["menu"].selectNext() def getCurrent(self): cur = self["menu"].current return cur and cur[0] def okbuttonClick(self): self.close(self.getCurrent()) def cancelClick(self): self.close(None)
gpl-2.0
justinlulejian/fah-gae
lib/requests/packages/chardet/langcyrillicmodel.py
2762
17725
######################## BEGIN LICENSE BLOCK ######################## # The Original Code is Mozilla Communicator client code. # # The Initial Developer of the Original Code is # Netscape Communications Corporation. # Portions created by the Initial Developer are Copyright (C) 1998 # the Initial Developer. All Rights Reserved. # # Contributor(s): # Mark Pilgrim - port to Python # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA # 02110-1301 USA ######################### END LICENSE BLOCK ######################### # KOI8-R language model # Character Mapping Table: KOI8R_CharToOrderMap = ( 255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, # 80 207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, # 90 223,224,225, 68,226,227,228,229,230,231,232,233,234,235,236,237, # a0 238,239,240,241,242,243,244,245,246,247,248,249,250,251,252,253, # b0 27, 3, 21, 28, 13, 2, 39, 19, 26, 4, 23, 11, 8, 12, 5, 1, # c0 15, 16, 9, 7, 6, 14, 24, 10, 17, 18, 20, 25, 30, 29, 22, 54, # d0 59, 37, 44, 58, 41, 48, 53, 46, 55, 42, 60, 36, 49, 38, 31, 34, # e0 35, 43, 45, 32, 40, 52, 56, 33, 61, 62, 51, 57, 47, 63, 50, 70, # f0 ) win1251_CharToOrderMap = ( 255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, 207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, 223,224,225,226,227,228,229,230,231,232,233,234,235,236,237,238, 239,240,241,242,243,244,245,246, 68,247,248,249,250,251,252,253, 37, 44, 33, 46, 41, 48, 56, 51, 42, 60, 36, 49, 38, 31, 34, 35, 45, 32, 40, 52, 53, 55, 58, 50, 57, 63, 70, 62, 61, 47, 59, 43, 3, 21, 10, 19, 13, 2, 24, 20, 4, 23, 11, 8, 12, 5, 1, 15, 9, 7, 6, 14, 39, 26, 28, 22, 25, 29, 54, 18, 17, 30, 27, 16, ) latin5_CharToOrderMap = ( 255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, 207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, 223,224,225,226,227,228,229,230,231,232,233,234,235,236,237,238, 37, 44, 33, 46, 41, 48, 56, 51, 42, 60, 36, 49, 38, 31, 34, 35, 45, 32, 40, 52, 53, 55, 58, 50, 57, 63, 70, 62, 61, 47, 59, 43, 3, 21, 10, 19, 13, 2, 24, 20, 4, 23, 11, 8, 12, 5, 1, 15, 9, 7, 6, 14, 39, 26, 28, 22, 25, 29, 54, 18, 17, 30, 27, 16, 239, 68,240,241,242,243,244,245,246,247,248,249,250,251,252,255, ) macCyrillic_CharToOrderMap = ( 255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 37, 44, 33, 46, 41, 48, 56, 51, 42, 60, 36, 49, 38, 31, 34, 35, 45, 32, 40, 52, 53, 55, 58, 50, 57, 63, 70, 62, 61, 47, 59, 43, 191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, 207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, 223,224,225,226,227,228,229,230,231,232,233,234,235,236,237,238, 239,240,241,242,243,244,245,246,247,248,249,250,251,252, 68, 16, 3, 21, 10, 19, 13, 2, 24, 20, 4, 23, 11, 8, 12, 5, 1, 15, 9, 7, 6, 14, 39, 26, 28, 22, 25, 29, 54, 18, 17, 30, 27,255, ) IBM855_CharToOrderMap = ( 255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 191,192,193,194, 68,195,196,197,198,199,200,201,202,203,204,205, 206,207,208,209,210,211,212,213,214,215,216,217, 27, 59, 54, 70, 3, 37, 21, 44, 28, 58, 13, 41, 2, 48, 39, 53, 19, 46,218,219, 220,221,222,223,224, 26, 55, 4, 42,225,226,227,228, 23, 60,229, 230,231,232,233,234,235, 11, 36,236,237,238,239,240,241,242,243, 8, 49, 12, 38, 5, 31, 1, 34, 15,244,245,246,247, 35, 16,248, 43, 9, 45, 7, 32, 6, 40, 14, 52, 24, 56, 10, 33, 17, 61,249, 250, 18, 62, 20, 51, 25, 57, 30, 47, 29, 63, 22, 50,251,252,255, ) IBM866_CharToOrderMap = ( 255,255,255,255,255,255,255,255,255,255,254,255,255,254,255,255, # 00 255,255,255,255,255,255,255,255,255,255,255,255,255,255,255,255, # 10 253,253,253,253,253,253,253,253,253,253,253,253,253,253,253,253, # 20 252,252,252,252,252,252,252,252,252,252,253,253,253,253,253,253, # 30 253,142,143,144,145,146,147,148,149,150,151,152, 74,153, 75,154, # 40 155,156,157,158,159,160,161,162,163,164,165,253,253,253,253,253, # 50 253, 71,172, 66,173, 65,174, 76,175, 64,176,177, 77, 72,178, 69, # 60 67,179, 78, 73,180,181, 79,182,183,184,185,253,253,253,253,253, # 70 37, 44, 33, 46, 41, 48, 56, 51, 42, 60, 36, 49, 38, 31, 34, 35, 45, 32, 40, 52, 53, 55, 58, 50, 57, 63, 70, 62, 61, 47, 59, 43, 3, 21, 10, 19, 13, 2, 24, 20, 4, 23, 11, 8, 12, 5, 1, 15, 191,192,193,194,195,196,197,198,199,200,201,202,203,204,205,206, 207,208,209,210,211,212,213,214,215,216,217,218,219,220,221,222, 223,224,225,226,227,228,229,230,231,232,233,234,235,236,237,238, 9, 7, 6, 14, 39, 26, 28, 22, 25, 29, 54, 18, 17, 30, 27, 16, 239, 68,240,241,242,243,244,245,246,247,248,249,250,251,252,255, ) # Model Table: # total sequences: 100% # first 512 sequences: 97.6601% # first 1024 sequences: 2.3389% # rest sequences: 0.1237% # negative sequences: 0.0009% RussianLangModel = ( 0,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,1,1,3,3,3,3,1,3,3,3,2,3,2,3,3, 3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,0,3,2,2,2,2,2,0,0,2, 3,3,3,2,3,3,3,3,3,3,3,3,3,3,2,3,3,0,0,3,3,3,3,3,3,3,3,3,2,3,2,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,2,2,3,3,3,3,3,3,3,3,3,2,3,3,0,0,3,3,3,3,3,3,3,3,2,3,3,1,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,2,3,2,3,3,3,3,3,3,3,3,3,3,3,3,3,0,0,3,3,3,3,3,3,3,3,3,3,3,2,1, 0,0,0,0,0,0,0,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,3,3,0,0,3,3,3,3,3,3,3,3,3,3,3,2,1, 0,0,0,0,0,1,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,3,3,2,2,2,3,1,3,3,1,3,3,3,3,2,2,3,0,2,2,2,3,3,2,1,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,2,3,3,3,3,3,2,2,3,2,3,3,3,2,1,2,2,0,1,2,2,2,2,2,2,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,3,2,2,2,3,0,2,2,3,3,2,1,2,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,1,0,0,2,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,2,3,3,1,2,3,2,2,3,2,3,3,3,3,2,2,3,0,3,2,2,3,1,1,1,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,3,3,2,2,3,3,3,3,3,2,3,3,3,3,2,2,2,0,3,3,3,2,2,2,2,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,3,3,3,3,2,3,2,3,3,3,3,3,3,2,3,2,2,0,1,3,2,1,2,2,1,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,3,3,3,3,3,2,1,1,3,0,1,1,1,1,2,1,1,0,2,2,2,1,2,0,1,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,2,3,3,2,2,2,2,1,3,2,3,2,3,2,1,2,2,0,1,1,2,1,2,1,2,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,3,3,3,3,3,3,2,2,3,2,3,3,3,2,2,2,2,0,2,2,2,2,3,1,1,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, 3,2,3,2,2,3,3,3,3,3,3,3,3,3,1,3,2,0,0,3,3,3,3,2,3,3,3,3,2,3,2,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 2,3,3,3,3,3,2,2,3,3,0,2,1,0,3,2,3,2,3,0,0,1,2,0,0,1,0,1,2,1,1,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,0,3,0,2,3,3,3,3,2,3,3,3,3,1,2,2,0,0,2,3,2,2,2,3,2,3,2,2,3,0,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,2,3,0,2,3,2,3,0,1,2,3,3,2,0,2,3,0,0,2,3,2,2,0,1,3,1,3,2,2,1,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,1,3,0,2,3,3,3,3,3,3,3,3,2,1,3,2,0,0,2,2,3,3,3,2,3,3,0,2,2,0,0, 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,2,2,3,3,2,2,2,3,3,0,0,1,1,1,1,1,2,0,0,1,1,1,1,0,1,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,2,2,3,3,3,3,3,3,3,0,3,2,3,3,2,3,2,0,2,1,0,1,1,0,1,0, 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,3,2,3,3,3,2,2,2,2,3,1,3,2,3,1,1,2,1,0,2,2,2,2,1,3,1,0, 0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0, 2,2,3,3,3,3,3,1,2,2,1,3,1,0,3,0,0,3,0,0,0,1,1,0,1,2,1,0,0,0,0,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,2,2,1,1,3,3,3,2,2,1,2,2,3,1,1,2,0,0,2,2,1,3,0,0,2,1,1,2,1,1,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,2,3,3,3,3,1,2,2,2,1,2,1,3,3,1,1,2,1,2,1,2,2,0,2,0,0,1,1,0,1,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 2,3,3,3,3,3,2,1,3,2,2,3,2,0,3,2,0,3,0,1,0,1,1,0,0,1,1,1,1,0,1,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,2,3,3,3,2,2,2,3,3,1,2,1,2,1,0,1,0,1,1,0,1,0,0,2,1,1,1,0,1,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0, 3,1,1,2,1,2,3,3,2,2,1,2,2,3,0,2,1,0,0,2,2,3,2,1,2,2,2,2,2,3,1,0, 0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 3,3,3,3,3,1,1,0,1,1,2,2,1,1,3,0,0,1,3,1,1,1,0,0,0,1,0,1,1,0,0,0, 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 2,1,3,3,3,2,0,0,0,2,1,0,1,0,2,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 2,0,1,0,0,2,3,2,2,2,1,2,2,2,1,2,1,0,0,1,1,1,0,2,0,1,1,1,0,0,1,1, 1,0,0,0,0,0,1,2,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0, 2,3,3,3,3,0,0,0,0,1,0,0,0,0,3,0,1,2,1,0,0,0,0,0,0,0,1,1,0,0,1,1, 1,0,1,0,1,2,0,0,1,1,2,1,0,1,1,1,1,0,1,1,1,1,0,1,0,0,1,0,0,1,1,0, 2,2,3,2,2,2,3,1,2,2,2,2,2,2,2,2,1,1,1,1,1,1,1,0,1,0,1,1,1,0,2,1, 1,1,1,1,1,1,1,1,2,1,1,1,1,1,1,1,1,1,1,0,1,0,1,1,0,1,1,1,0,1,1,0, 3,3,3,2,2,2,2,3,2,2,1,1,2,2,2,2,1,1,3,1,2,1,2,0,0,1,1,0,1,0,2,1, 1,1,1,1,1,2,1,0,1,1,1,1,0,1,0,0,1,1,0,0,1,0,1,0,0,1,0,0,0,1,1,0, 2,0,0,1,0,3,2,2,2,2,1,2,1,2,1,2,0,0,0,2,1,2,2,1,1,2,2,0,1,1,0,2, 1,1,1,1,1,0,1,1,1,2,1,1,1,2,1,0,1,2,1,1,1,1,0,1,1,1,0,0,1,0,0,1, 1,3,2,2,2,1,1,1,2,3,0,0,0,0,2,0,2,2,1,0,0,0,0,0,0,1,0,0,0,0,1,1, 1,0,1,1,0,1,0,1,1,0,1,1,0,2,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0, 2,3,2,3,2,1,2,2,2,2,1,0,0,0,2,0,0,1,1,0,0,0,0,0,0,0,1,1,0,0,2,1, 1,1,2,1,0,2,0,0,1,0,1,0,0,1,0,0,1,1,0,1,1,0,0,0,0,0,1,0,0,0,0,0, 3,0,0,1,0,2,2,2,3,2,2,2,2,2,2,2,0,0,0,2,1,2,1,1,1,2,2,0,0,0,1,2, 1,1,1,1,1,0,1,2,1,1,1,1,1,1,1,0,1,1,1,1,1,1,0,1,1,1,1,1,1,0,0,1, 2,3,2,3,3,2,0,1,1,1,0,0,1,0,2,0,1,1,3,1,0,0,0,0,0,0,0,1,0,0,2,1, 1,1,1,1,1,1,1,0,1,0,1,1,1,1,0,1,1,1,0,0,1,1,0,1,0,0,0,0,0,0,1,0, 2,3,3,3,3,1,2,2,2,2,0,1,1,0,2,1,1,1,2,1,0,1,1,0,0,1,0,1,0,0,2,0, 0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 2,3,3,3,2,0,0,1,1,2,2,1,0,0,2,0,1,1,3,0,0,1,0,0,0,0,0,1,0,1,2,1, 1,1,2,0,1,1,1,0,1,0,1,1,0,1,0,1,1,1,1,0,1,0,0,0,0,0,0,1,0,1,1,0, 1,3,2,3,2,1,0,0,2,2,2,0,1,0,2,0,1,1,1,0,1,0,0,0,3,0,1,1,0,0,2,1, 1,1,1,0,1,1,0,0,0,0,1,1,0,1,0,0,2,1,1,0,1,0,0,0,1,0,1,0,0,1,1,0, 3,1,2,1,1,2,2,2,2,2,2,1,2,2,1,1,0,0,0,2,2,2,0,0,0,1,2,1,0,1,0,1, 2,1,1,1,1,1,1,1,1,1,1,1,1,1,1,0,2,1,1,1,0,1,0,1,1,0,1,1,1,0,0,1, 3,0,0,0,0,2,0,1,1,1,1,1,1,1,0,1,0,0,0,1,1,1,0,1,0,1,1,0,0,1,0,1, 1,1,0,0,1,0,0,0,1,0,1,1,0,0,1,0,1,0,1,0,0,0,0,1,0,0,0,1,0,0,0,1, 1,3,3,2,2,0,0,0,2,2,0,0,0,1,2,0,1,1,2,0,0,0,0,0,0,0,0,1,0,0,2,1, 0,1,1,0,0,1,1,0,0,0,1,1,0,1,1,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,1,0, 2,3,2,3,2,0,0,0,0,1,1,0,0,0,2,0,2,0,2,0,0,0,0,0,1,0,0,1,0,0,1,1, 1,1,2,0,1,2,1,0,1,1,2,1,1,1,1,1,2,1,1,0,1,0,0,1,1,1,1,1,0,1,1,0, 1,3,2,2,2,1,0,0,2,2,1,0,1,2,2,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,1,1, 0,0,1,1,0,1,1,0,0,1,1,0,1,1,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0, 1,0,0,1,0,2,3,1,2,2,2,2,2,2,1,1,0,0,0,1,0,1,0,2,1,1,1,0,0,0,0,1, 1,1,0,1,1,0,1,1,1,1,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1,0,0,0, 2,0,2,0,0,1,0,3,2,1,2,1,2,2,0,1,0,0,0,2,1,0,0,2,1,1,1,1,0,2,0,2, 2,1,1,1,1,1,1,1,1,1,1,1,1,2,1,0,1,1,1,1,0,0,0,1,1,1,1,0,1,0,0,1, 1,2,2,2,2,1,0,0,1,0,0,0,0,0,2,0,1,1,1,1,0,0,0,0,1,0,1,2,0,0,2,0, 1,0,1,1,1,2,1,0,1,0,1,1,0,0,1,0,1,1,1,0,1,0,0,0,1,0,0,1,0,1,1,0, 2,1,2,2,2,0,3,0,1,1,0,0,0,0,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1, 0,0,0,1,1,1,0,0,1,0,1,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0, 1,2,2,3,2,2,0,0,1,1,2,0,1,2,1,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,1, 0,1,1,0,0,1,1,0,0,1,1,0,0,1,1,0,1,1,0,0,1,0,0,0,0,0,0,0,0,1,1,0, 2,2,1,1,2,1,2,2,2,2,2,1,2,2,0,1,0,0,0,1,2,2,2,1,2,1,1,1,1,1,2,1, 1,1,1,1,1,1,1,1,1,1,0,0,1,1,1,0,1,1,1,0,0,0,0,1,1,1,0,1,1,0,0,1, 1,2,2,2,2,0,1,0,2,2,0,0,0,0,2,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,2,0, 0,0,1,0,0,1,0,0,0,0,1,0,1,1,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0, 0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0, 0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 1,2,2,2,2,0,0,0,2,2,2,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,1, 0,1,1,0,0,1,1,0,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 1,2,2,2,2,0,0,0,0,1,0,0,1,1,2,0,0,0,0,1,0,1,0,0,1,0,0,2,0,0,0,1, 0,0,1,0,0,1,0,0,0,1,1,0,0,0,0,0,1,0,0,1,1,0,0,0,0,0,0,0,0,0,0,0, 1,2,2,2,1,1,2,0,2,1,1,1,1,0,2,2,0,0,0,0,0,0,0,0,0,1,1,0,0,0,1,1, 0,0,1,0,1,1,0,0,0,0,1,0,0,0,0,0,1,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0, 1,0,2,1,2,0,0,0,0,0,1,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,1,0,0,0,0, 0,0,1,0,1,1,0,0,0,0,1,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,1,0, 1,0,0,0,0,2,0,1,2,1,0,1,1,1,0,1,0,0,0,1,0,1,0,0,1,0,1,0,0,0,0,1, 0,0,0,0,0,1,0,0,1,1,0,0,1,1,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1, 2,2,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, 1,0,0,0,1,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,0,1,0,0,1,0,0,0,0,0, 2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, 1,1,1,0,1,0,1,0,0,1,1,1,1,0,0,0,1,0,0,0,0,1,0,0,0,1,0,1,0,0,0,0, 1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1, 1,1,0,1,1,0,1,0,1,0,0,0,0,1,1,0,1,1,0,0,0,0,0,1,0,1,1,0,1,0,0,0, 0,1,1,1,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0, 0,0,0,0,0,1,0,0,0,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0, ) Koi8rModel = { 'charToOrderMap': KOI8R_CharToOrderMap, 'precedenceMatrix': RussianLangModel, 'mTypicalPositiveRatio': 0.976601, 'keepEnglishLetter': False, 'charsetName': "KOI8-R" } Win1251CyrillicModel = { 'charToOrderMap': win1251_CharToOrderMap, 'precedenceMatrix': RussianLangModel, 'mTypicalPositiveRatio': 0.976601, 'keepEnglishLetter': False, 'charsetName': "windows-1251" } Latin5CyrillicModel = { 'charToOrderMap': latin5_CharToOrderMap, 'precedenceMatrix': RussianLangModel, 'mTypicalPositiveRatio': 0.976601, 'keepEnglishLetter': False, 'charsetName': "ISO-8859-5" } MacCyrillicModel = { 'charToOrderMap': macCyrillic_CharToOrderMap, 'precedenceMatrix': RussianLangModel, 'mTypicalPositiveRatio': 0.976601, 'keepEnglishLetter': False, 'charsetName': "MacCyrillic" }; Ibm866Model = { 'charToOrderMap': IBM866_CharToOrderMap, 'precedenceMatrix': RussianLangModel, 'mTypicalPositiveRatio': 0.976601, 'keepEnglishLetter': False, 'charsetName': "IBM866" } Ibm855Model = { 'charToOrderMap': IBM855_CharToOrderMap, 'precedenceMatrix': RussianLangModel, 'mTypicalPositiveRatio': 0.976601, 'keepEnglishLetter': False, 'charsetName': "IBM855" } # flake8: noqa
mit
Johnzero/OE7
openerp/report/render/rml2pdf/customfonts.py
57
6317
# -*- coding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution # Copyright (C) 2004-2009 P. Christeas, Tiny SPRL (<http://tiny.be>). # Copyright (C) 2010-2013 OpenERP SA. (http://www.openerp.com) # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import glob import logging import os import platform from reportlab import rl_config from openerp.tools import config #.apidoc title: TTF Font Table """This module allows the mapping of some system-available TTF fonts to the reportlab engine. This file could be customized per distro (although most Linux/Unix ones) should have the same filenames, only need the code below). Due to an awful configuration that ships with reportlab at many Linux and Ubuntu distros, we have to override the search path, too. """ _logger = logging.getLogger(__name__) CustomTTFonts = [ ('Helvetica',"DejaVu Sans", "DejaVuSans.ttf", 'normal'), ('Helvetica',"DejaVu Sans Bold", "DejaVuSans-Bold.ttf", 'bold'), ('Helvetica',"DejaVu Sans Oblique", "DejaVuSans-Oblique.ttf", 'italic'), ('Helvetica',"DejaVu Sans BoldOblique", "DejaVuSans-BoldOblique.ttf", 'bolditalic'), ('Times',"Liberation Serif", "LiberationSerif-Regular.ttf", 'normal'), ('Times',"Liberation Serif Bold", "LiberationSerif-Bold.ttf", 'bold'), ('Times',"Liberation Serif Italic", "LiberationSerif-Italic.ttf", 'italic'), ('Times',"Liberation Serif BoldItalic", "LiberationSerif-BoldItalic.ttf", 'bolditalic'), ('Times-Roman',"Liberation Serif", "LiberationSerif-Regular.ttf", 'normal'), ('Times-Roman',"Liberation Serif Bold", "LiberationSerif-Bold.ttf", 'bold'), ('Times-Roman',"Liberation Serif Italic", "LiberationSerif-Italic.ttf", 'italic'), ('Times-Roman',"Liberation Serif BoldItalic", "LiberationSerif-BoldItalic.ttf", 'bolditalic'), ('Courier',"FreeMono", "FreeMono.ttf", 'normal'), ('Courier',"FreeMono Bold", "FreeMonoBold.ttf", 'bold'), ('Courier',"FreeMono Oblique", "FreeMonoOblique.ttf", 'italic'), ('Courier',"FreeMono BoldOblique", "FreeMonoBoldOblique.ttf", 'bolditalic'), # Sun-ExtA can be downloaded from http://okuc.net/SunWb/ ('Sun-ExtA',"Sun-ExtA", "Sun-ExtA.ttf", 'normal'), ] TTFSearchPath_Linux = [ '/usr/share/fonts/truetype', # SuSE '/usr/share/fonts/dejavu', '/usr/share/fonts/liberation', # Fedora, RHEL '/usr/share/fonts/truetype/*', # Ubuntu, '/usr/share/fonts/TTF/*', # at Mandriva/Mageia '/usr/share/fonts/TTF', # Arch Linux ] TTFSearchPath_Windows = [ 'c:/winnt/fonts', 'c:/windows/fonts' ] TTFSearchPath_Darwin = [ #mac os X - from #http://developer.apple.com/technotes/tn/tn2024.html '~/Library/Fonts', '/Library/Fonts', '/Network/Library/Fonts', '/System/Library/Fonts', ] TTFSearchPathMap = { 'Darwin': TTFSearchPath_Darwin, 'Windows': TTFSearchPath_Windows, 'Linux': TTFSearchPath_Linux, } # ----- The code below is less distro-specific, please avoid editing! ------- __foundFonts = None def FindCustomFonts(): """Fill the __foundFonts list with those filenames, whose fonts can be found in the reportlab ttf font path. This process needs only be done once per loading of this module, it is cached. But, if the system admin adds some font in the meanwhile, the server must be restarted eventually. """ dirpath = [] global __foundFonts __foundFonts = {} searchpath = [] if config.get('fonts_search_path'): searchpath += map(str.strip, config.get('fonts_search_path').split(',')) local_platform = platform.system() if local_platform in TTFSearchPathMap: searchpath += TTFSearchPathMap[local_platform] # Append the original search path of reportlab (at the end) searchpath += rl_config.TTFSearchPath # Perform the search for font files ourselves, as reportlab's # TTFOpenFile is not very good at it. for dirglob in searchpath: dirglob = os.path.expanduser(dirglob) for dirname in glob.iglob(dirglob): abp = os.path.abspath(dirname) if os.path.isdir(abp): dirpath.append(abp) for k, (name, font, filename, mode) in enumerate(CustomTTFonts): if filename in __foundFonts: continue for d in dirpath: abs_filename = os.path.join(d, filename) if os.path.exists(abs_filename): _logger.debug("Found font %s at %s", filename, abs_filename) __foundFonts[filename] = abs_filename break def SetCustomFonts(rmldoc): """ Map some font names to the corresponding TTF fonts The ttf font may not even have the same name, as in Times -> Liberation Serif. This function is called once per report, so it should avoid system-wide processing (cache it, instead). """ global __foundFonts if __foundFonts is None: FindCustomFonts() for name, font, filename, mode in CustomTTFonts: if os.path.isabs(filename) and os.path.exists(filename): rmldoc.setTTFontMapping(name, font, filename, mode) elif filename in __foundFonts: rmldoc.setTTFontMapping(name, font, __foundFonts[filename], mode) return True #eof # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
agpl-3.0
yaqiyang/autorest
src/generator/AutoRest.Python.Azure.Tests/Expected/AcceptanceTests/Lro/autorestlongrunningoperationtestservice/operations/lr_os_custom_header_operations.py
4
15472
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is # regenerated. # -------------------------------------------------------------------------- from msrest.pipeline import ClientRawResponse from msrestazure.azure_exceptions import CloudError from msrestazure.azure_operation import AzureOperationPoller import uuid from .. import models class LROsCustomHeaderOperations(object): """LROsCustomHeaderOperations operations. :param client: Client for service requests. :param config: Configuration of service client. :param serializer: An object model serializer. :param deserializer: An objec model deserializer. """ def __init__(self, client, config, serializer, deserializer): self._client = client self._serialize = serializer self._deserialize = deserializer self.config = config def put_async_retry_succeeded( self, product=None, custom_headers=None, raw=False, **operation_config): """x-ms-client-request-id = 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0 is required message header for all requests. Long running put request, service returns a 200 to the initial request, with an entity that contains ProvisioningState=’Creating’. Poll the endpoint indicated in the Azure-AsyncOperation header for operation status. :param product: Product to put :type product: :class:`Product <fixtures.acceptancetestslro.models.Product>` :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :rtype: :class:`AzureOperationPoller<msrestazure.azure_operation.AzureOperationPoller>` instance that returns :class:`Product <fixtures.acceptancetestslro.models.Product>` :rtype: :class:`ClientRawResponse<msrest.pipeline.ClientRawResponse>` if raw=true """ # Construct URL url = '/lro/customheader/putasync/retry/succeeded' # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct body if product is not None: body_content = self._serialize.body(product, 'Product') else: body_content = None # Construct and send request def long_running_send(): request = self._client.put(url, query_parameters) return self._client.send( request, header_parameters, body_content, **operation_config) def get_long_running_status(status_link, headers=None): request = self._client.get(status_link) if headers: request.headers.update(headers) return self._client.send( request, header_parameters, **operation_config) def get_long_running_output(response): if response.status_code not in [200]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp deserialized = None header_dict = {} if response.status_code == 200: deserialized = self._deserialize('Product', response) header_dict = { 'Azure-AsyncOperation': 'str', 'Location': 'str', 'Retry-After': 'int', } if raw: client_raw_response = ClientRawResponse(deserialized, response) client_raw_response.add_headers(header_dict) return client_raw_response return deserialized if raw: response = long_running_send() return get_long_running_output(response) long_running_operation_timeout = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) return AzureOperationPoller( long_running_send, get_long_running_output, get_long_running_status, long_running_operation_timeout) def put201_creating_succeeded200( self, product=None, custom_headers=None, raw=False, **operation_config): """x-ms-client-request-id = 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0 is required message header for all requests. Long running put request, service returns a 201 to the initial request, with an entity that contains ProvisioningState=’Creating’. Polls return this value until the last poll returns a ‘200’ with ProvisioningState=’Succeeded’. :param product: Product to put :type product: :class:`Product <fixtures.acceptancetestslro.models.Product>` :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :rtype: :class:`AzureOperationPoller<msrestazure.azure_operation.AzureOperationPoller>` instance that returns :class:`Product <fixtures.acceptancetestslro.models.Product>` :rtype: :class:`ClientRawResponse<msrest.pipeline.ClientRawResponse>` if raw=true """ # Construct URL url = '/lro/customheader/put/201/creating/succeeded/200' # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct body if product is not None: body_content = self._serialize.body(product, 'Product') else: body_content = None # Construct and send request def long_running_send(): request = self._client.put(url, query_parameters) return self._client.send( request, header_parameters, body_content, **operation_config) def get_long_running_status(status_link, headers=None): request = self._client.get(status_link) if headers: request.headers.update(headers) return self._client.send( request, header_parameters, **operation_config) def get_long_running_output(response): if response.status_code not in [200, 201]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp deserialized = None if response.status_code == 200: deserialized = self._deserialize('Product', response) if response.status_code == 201: deserialized = self._deserialize('Product', response) if raw: client_raw_response = ClientRawResponse(deserialized, response) return client_raw_response return deserialized if raw: response = long_running_send() return get_long_running_output(response) long_running_operation_timeout = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) return AzureOperationPoller( long_running_send, get_long_running_output, get_long_running_status, long_running_operation_timeout) def post202_retry200( self, product=None, custom_headers=None, raw=False, **operation_config): """x-ms-client-request-id = 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0 is required message header for all requests. Long running post request, service returns a 202 to the initial request, with 'Location' and 'Retry-After' headers, Polls return a 200 with a response body after success. :param product: Product to put :type product: :class:`Product <fixtures.acceptancetestslro.models.Product>` :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :rtype: :class:`AzureOperationPoller<msrestazure.azure_operation.AzureOperationPoller>` instance that returns None :rtype: :class:`ClientRawResponse<msrest.pipeline.ClientRawResponse>` if raw=true """ # Construct URL url = '/lro/customheader/post/202/retry/200' # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct body if product is not None: body_content = self._serialize.body(product, 'Product') else: body_content = None # Construct and send request def long_running_send(): request = self._client.post(url, query_parameters) return self._client.send( request, header_parameters, body_content, **operation_config) def get_long_running_status(status_link, headers=None): request = self._client.get(status_link) if headers: request.headers.update(headers) return self._client.send( request, header_parameters, **operation_config) def get_long_running_output(response): if response.status_code not in [202]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp if raw: client_raw_response = ClientRawResponse(None, response) client_raw_response.add_headers({ 'Location': 'str', 'Retry-After': 'int', }) return client_raw_response if raw: response = long_running_send() return get_long_running_output(response) long_running_operation_timeout = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) return AzureOperationPoller( long_running_send, get_long_running_output, get_long_running_status, long_running_operation_timeout) def post_async_retry_succeeded( self, product=None, custom_headers=None, raw=False, **operation_config): """x-ms-client-request-id = 9C4D50EE-2D56-4CD3-8152-34347DC9F2B0 is required message header for all requests. Long running post request, service returns a 202 to the initial request, with an entity that contains ProvisioningState=’Creating’. Poll the endpoint indicated in the Azure-AsyncOperation header for operation status. :param product: Product to put :type product: :class:`Product <fixtures.acceptancetestslro.models.Product>` :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response :rtype: :class:`AzureOperationPoller<msrestazure.azure_operation.AzureOperationPoller>` instance that returns None :rtype: :class:`ClientRawResponse<msrest.pipeline.ClientRawResponse>` if raw=true """ # Construct URL url = '/lro/customheader/postasync/retry/succeeded' # Construct parameters query_parameters = {} # Construct headers header_parameters = {} header_parameters['Content-Type'] = 'application/json; charset=utf-8' if self.config.generate_client_request_id: header_parameters['x-ms-client-request-id'] = str(uuid.uuid1()) if custom_headers: header_parameters.update(custom_headers) if self.config.accept_language is not None: header_parameters['accept-language'] = self._serialize.header("self.config.accept_language", self.config.accept_language, 'str') # Construct body if product is not None: body_content = self._serialize.body(product, 'Product') else: body_content = None # Construct and send request def long_running_send(): request = self._client.post(url, query_parameters) return self._client.send( request, header_parameters, body_content, **operation_config) def get_long_running_status(status_link, headers=None): request = self._client.get(status_link) if headers: request.headers.update(headers) return self._client.send( request, header_parameters, **operation_config) def get_long_running_output(response): if response.status_code not in [202]: exp = CloudError(response) exp.request_id = response.headers.get('x-ms-request-id') raise exp if raw: client_raw_response = ClientRawResponse(None, response) client_raw_response.add_headers({ 'Azure-AsyncOperation': 'str', 'Location': 'str', 'Retry-After': 'int', }) return client_raw_response if raw: response = long_running_send() return get_long_running_output(response) long_running_operation_timeout = operation_config.get( 'long_running_operation_timeout', self.config.long_running_operation_timeout) return AzureOperationPoller( long_running_send, get_long_running_output, get_long_running_status, long_running_operation_timeout)
mit
fedorpatlin/ansible
lib/ansible/modules/windows/win_firewall_rule.py
41
3499
#!/usr/bin/env python # (c) 2014, Timothy Vandenbrande <timothy.vandenbrande@gmail.com> # # This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <http://www.gnu.org/licenses/>. ANSIBLE_METADATA = {'metadata_version': '1.0', 'status': ['preview'], 'supported_by': 'community'} DOCUMENTATION = r''' --- module: win_firewall_rule version_added: "2.0" author: Timothy Vandenbrande short_description: Windows firewall automation description: - allows you to create/remove/update firewall rules options: enable: description: - is this firewall rule enabled or disabled default: true required: false state: description: - should this rule be added or removed default: "present" required: true choices: ['present', 'absent'] name: description: - the rules name default: null required: true direction: description: - is this rule for inbound or outbound traffic default: null required: true choices: ['in', 'out'] action: description: - what to do with the items this rule is for default: null required: true choices: ['allow', 'block', 'bypass'] description: description: - description for the firewall rule default: null required: false localip: description: - the local ip address this rule applies to default: 'any' required: false remoteip: description: - the remote ip address/range this rule applies to default: 'any' required: false localport: description: - the local port this rule applies to default: 'any' required: false remoteport: description: - the remote port this rule applies to default: 'any' required: false program: description: - the program this rule applies to default: null required: false service: description: - the service this rule applies to default: 'any' required: false protocol: description: - the protocol this rule applies to default: 'any' required: false profile: description: - the profile this rule applies to, e.g. Domain,Private,Public default: 'any' required: false force: description: - Enforces the change if a rule with different values exists default: false required: false ''' EXAMPLES = r''' - name: Firewall rule to allow smtp on TCP port 25 action: win_firewall_rule args: name: smtp enable: yes state: present localport: 25 action: allow direction: In protocol: TCP '''
gpl-3.0
gannetson/django
django/core/serializers/xml_serializer.py
16
15345
""" XML serializer. """ from __future__ import unicode_literals from collections import OrderedDict from xml.dom import pulldom from xml.sax import handler from xml.sax.expatreader import ExpatParser as _ExpatParser from django.apps import apps from django.conf import settings from django.core.serializers import base from django.db import DEFAULT_DB_ALIAS, models from django.utils.encoding import smart_text from django.utils.xmlutils import SimplerXMLGenerator class Serializer(base.Serializer): """ Serializes a QuerySet to XML. """ def indent(self, level): if self.options.get('indent') is not None: self.xml.ignorableWhitespace('\n' + ' ' * self.options.get('indent') * level) def start_serialization(self): """ Start serialization -- open the XML document and the root element. """ self.xml = SimplerXMLGenerator(self.stream, self.options.get("encoding", settings.DEFAULT_CHARSET)) self.xml.startDocument() self.xml.startElement("django-objects", {"version": "1.0"}) def end_serialization(self): """ End serialization -- end the document. """ self.indent(0) self.xml.endElement("django-objects") self.xml.endDocument() def start_object(self, obj): """ Called as each object is handled. """ if not hasattr(obj, "_meta"): raise base.SerializationError("Non-model object (%s) encountered during serialization" % type(obj)) self.indent(1) attrs = OrderedDict([("model", smart_text(obj._meta))]) if not self.use_natural_primary_keys or not hasattr(obj, 'natural_key'): obj_pk = obj._get_pk_val() if obj_pk is not None: attrs['pk'] = smart_text(obj_pk) self.xml.startElement("object", attrs) def end_object(self, obj): """ Called after handling all fields for an object. """ self.indent(1) self.xml.endElement("object") def handle_field(self, obj, field): """ Called to handle each field on an object (except for ForeignKeys and ManyToManyFields) """ self.indent(2) self.xml.startElement("field", OrderedDict([ ("name", field.name), ("type", field.get_internal_type()), ])) # Get a "string version" of the object's data. if getattr(obj, field.name) is not None: self.xml.characters(field.value_to_string(obj)) else: self.xml.addQuickElement("None") self.xml.endElement("field") def handle_fk_field(self, obj, field): """ Called to handle a ForeignKey (we need to treat them slightly differently from regular fields). """ self._start_relational_field(field) related_att = getattr(obj, field.get_attname()) if related_att is not None: if self.use_natural_foreign_keys and hasattr(field.remote_field.model, 'natural_key'): related = getattr(obj, field.name) # If related object has a natural key, use it related = related.natural_key() # Iterable natural keys are rolled out as subelements for key_value in related: self.xml.startElement("natural", {}) self.xml.characters(smart_text(key_value)) self.xml.endElement("natural") else: self.xml.characters(smart_text(related_att)) else: self.xml.addQuickElement("None") self.xml.endElement("field") def handle_m2m_field(self, obj, field): """ Called to handle a ManyToManyField. Related objects are only serialized as references to the object's PK (i.e. the related *data* is not dumped, just the relation). """ if field.remote_field.through._meta.auto_created: self._start_relational_field(field) if self.use_natural_foreign_keys and hasattr(field.remote_field.model, 'natural_key'): # If the objects in the m2m have a natural key, use it def handle_m2m(value): natural = value.natural_key() # Iterable natural keys are rolled out as subelements self.xml.startElement("object", {}) for key_value in natural: self.xml.startElement("natural", {}) self.xml.characters(smart_text(key_value)) self.xml.endElement("natural") self.xml.endElement("object") else: def handle_m2m(value): self.xml.addQuickElement("object", attrs={ 'pk': smart_text(value._get_pk_val()) }) for relobj in getattr(obj, field.name).iterator(): handle_m2m(relobj) self.xml.endElement("field") def _start_relational_field(self, field): """ Helper to output the <field> element for relational fields """ self.indent(2) self.xml.startElement("field", OrderedDict([ ("name", field.name), ("rel", field.remote_field.__class__.__name__), ("to", smart_text(field.remote_field.model._meta)), ])) class Deserializer(base.Deserializer): """ Deserialize XML. """ def __init__(self, stream_or_string, **options): super(Deserializer, self).__init__(stream_or_string, **options) self.event_stream = pulldom.parse(self.stream, self._make_parser()) self.db = options.pop('using', DEFAULT_DB_ALIAS) self.ignore = options.pop('ignorenonexistent', False) def _make_parser(self): """Create a hardened XML parser (no custom/external entities).""" return DefusedExpatParser() def __next__(self): for event, node in self.event_stream: if event == "START_ELEMENT" and node.nodeName == "object": self.event_stream.expandNode(node) return self._handle_object(node) raise StopIteration def _handle_object(self, node): """ Convert an <object> node to a DeserializedObject. """ # Look up the model using the model loading mechanism. If this fails, # bail. Model = self._get_model_from_node(node, "model") # Start building a data dictionary from the object. data = {} if node.hasAttribute('pk'): data[Model._meta.pk.attname] = Model._meta.pk.to_python( node.getAttribute('pk')) # Also start building a dict of m2m data (this is saved as # {m2m_accessor_attribute : [list_of_related_objects]}) m2m_data = {} field_names = {f.name for f in Model._meta.get_fields()} # Deserialize each field. for field_node in node.getElementsByTagName("field"): # If the field is missing the name attribute, bail (are you # sensing a pattern here?) field_name = field_node.getAttribute("name") if not field_name: raise base.DeserializationError("<field> node is missing the 'name' attribute") # Get the field from the Model. This will raise a # FieldDoesNotExist if, well, the field doesn't exist, which will # be propagated correctly unless ignorenonexistent=True is used. if self.ignore and field_name not in field_names: continue field = Model._meta.get_field(field_name) # As is usually the case, relation fields get the special treatment. if field.remote_field and isinstance(field.remote_field, models.ManyToManyRel): m2m_data[field.name] = self._handle_m2m_field_node(field_node, field) elif field.remote_field and isinstance(field.remote_field, models.ManyToOneRel): data[field.attname] = self._handle_fk_field_node(field_node, field) else: if field_node.getElementsByTagName('None'): value = None else: value = field.to_python(getInnerText(field_node).strip()) data[field.name] = value obj = base.build_instance(Model, data, self.db) # Return a DeserializedObject so that the m2m data has a place to live. return base.DeserializedObject(obj, m2m_data) def _handle_fk_field_node(self, node, field): """ Handle a <field> node for a ForeignKey """ # Check if there is a child node named 'None', returning None if so. if node.getElementsByTagName('None'): return None else: if hasattr(field.remote_field.model._default_manager, 'get_by_natural_key'): keys = node.getElementsByTagName('natural') if keys: # If there are 'natural' subelements, it must be a natural key field_value = [getInnerText(k).strip() for k in keys] obj = field.remote_field.model._default_manager.db_manager(self.db).get_by_natural_key(*field_value) obj_pk = getattr(obj, field.remote_field.field_name) # If this is a natural foreign key to an object that # has a FK/O2O as the foreign key, use the FK value if field.remote_field.model._meta.pk.remote_field: obj_pk = obj_pk.pk else: # Otherwise, treat like a normal PK field_value = getInnerText(node).strip() obj_pk = field.remote_field.model._meta.get_field(field.remote_field.field_name).to_python(field_value) return obj_pk else: field_value = getInnerText(node).strip() return field.remote_field.model._meta.get_field(field.remote_field.field_name).to_python(field_value) def _handle_m2m_field_node(self, node, field): """ Handle a <field> node for a ManyToManyField. """ if hasattr(field.remote_field.model._default_manager, 'get_by_natural_key'): def m2m_convert(n): keys = n.getElementsByTagName('natural') if keys: # If there are 'natural' subelements, it must be a natural key field_value = [getInnerText(k).strip() for k in keys] obj_pk = field.remote_field.model._default_manager.db_manager(self.db).get_by_natural_key(*field_value).pk else: # Otherwise, treat like a normal PK value. obj_pk = field.remote_field.model._meta.pk.to_python(n.getAttribute('pk')) return obj_pk else: m2m_convert = lambda n: field.remote_field.model._meta.pk.to_python(n.getAttribute('pk')) return [m2m_convert(c) for c in node.getElementsByTagName("object")] def _get_model_from_node(self, node, attr): """ Helper to look up a model from a <object model=...> or a <field rel=... to=...> node. """ model_identifier = node.getAttribute(attr) if not model_identifier: raise base.DeserializationError( "<%s> node is missing the required '%s' attribute" % (node.nodeName, attr)) try: return apps.get_model(model_identifier) except (LookupError, TypeError): raise base.DeserializationError( "<%s> node has invalid model identifier: '%s'" % (node.nodeName, model_identifier)) def getInnerText(node): """ Get all the inner text of a DOM node (recursively). """ # inspired by http://mail.python.org/pipermail/xml-sig/2005-March/011022.html inner_text = [] for child in node.childNodes: if child.nodeType == child.TEXT_NODE or child.nodeType == child.CDATA_SECTION_NODE: inner_text.append(child.data) elif child.nodeType == child.ELEMENT_NODE: inner_text.extend(getInnerText(child)) else: pass return "".join(inner_text) # Below code based on Christian Heimes' defusedxml class DefusedExpatParser(_ExpatParser): """ An expat parser hardened against XML bomb attacks. Forbids DTDs, external entity references """ def __init__(self, *args, **kwargs): _ExpatParser.__init__(self, *args, **kwargs) self.setFeature(handler.feature_external_ges, False) self.setFeature(handler.feature_external_pes, False) def start_doctype_decl(self, name, sysid, pubid, has_internal_subset): raise DTDForbidden(name, sysid, pubid) def entity_decl(self, name, is_parameter_entity, value, base, sysid, pubid, notation_name): raise EntitiesForbidden(name, value, base, sysid, pubid, notation_name) def unparsed_entity_decl(self, name, base, sysid, pubid, notation_name): # expat 1.2 raise EntitiesForbidden(name, None, base, sysid, pubid, notation_name) def external_entity_ref_handler(self, context, base, sysid, pubid): raise ExternalReferenceForbidden(context, base, sysid, pubid) def reset(self): _ExpatParser.reset(self) parser = self._parser parser.StartDoctypeDeclHandler = self.start_doctype_decl parser.EntityDeclHandler = self.entity_decl parser.UnparsedEntityDeclHandler = self.unparsed_entity_decl parser.ExternalEntityRefHandler = self.external_entity_ref_handler class DefusedXmlException(ValueError): """Base exception.""" def __repr__(self): return str(self) class DTDForbidden(DefusedXmlException): """Document type definition is forbidden.""" def __init__(self, name, sysid, pubid): super(DTDForbidden, self).__init__() self.name = name self.sysid = sysid self.pubid = pubid def __str__(self): tpl = "DTDForbidden(name='{}', system_id={!r}, public_id={!r})" return tpl.format(self.name, self.sysid, self.pubid) class EntitiesForbidden(DefusedXmlException): """Entity definition is forbidden.""" def __init__(self, name, value, base, sysid, pubid, notation_name): super(EntitiesForbidden, self).__init__() self.name = name self.value = value self.base = base self.sysid = sysid self.pubid = pubid self.notation_name = notation_name def __str__(self): tpl = "EntitiesForbidden(name='{}', system_id={!r}, public_id={!r})" return tpl.format(self.name, self.sysid, self.pubid) class ExternalReferenceForbidden(DefusedXmlException): """Resolving an external reference is forbidden.""" def __init__(self, context, base, sysid, pubid): super(ExternalReferenceForbidden, self).__init__() self.context = context self.base = base self.sysid = sysid self.pubid = pubid def __str__(self): tpl = "ExternalReferenceForbidden(system_id='{}', public_id={})" return tpl.format(self.sysid, self.pubid)
bsd-3-clause
jgoclawski/django
django/contrib/auth/hashers.py
211
17463
from __future__ import unicode_literals import base64 import binascii import hashlib import importlib from collections import OrderedDict from django.conf import settings from django.core.exceptions import ImproperlyConfigured from django.core.signals import setting_changed from django.dispatch import receiver from django.utils import lru_cache from django.utils.crypto import ( constant_time_compare, get_random_string, pbkdf2, ) from django.utils.encoding import force_bytes, force_str, force_text from django.utils.module_loading import import_string from django.utils.translation import ugettext_noop as _ UNUSABLE_PASSWORD_PREFIX = '!' # This will never be a valid encoded hash UNUSABLE_PASSWORD_SUFFIX_LENGTH = 40 # number of random chars to add after UNUSABLE_PASSWORD_PREFIX def is_password_usable(encoded): if encoded is None or encoded.startswith(UNUSABLE_PASSWORD_PREFIX): return False try: identify_hasher(encoded) except ValueError: return False return True def check_password(password, encoded, setter=None, preferred='default'): """ Returns a boolean of whether the raw password matches the three part encoded digest. If setter is specified, it'll be called when you need to regenerate the password. """ if password is None or not is_password_usable(encoded): return False preferred = get_hasher(preferred) hasher = identify_hasher(encoded) must_update = hasher.algorithm != preferred.algorithm if not must_update: must_update = preferred.must_update(encoded) is_correct = hasher.verify(password, encoded) if setter and is_correct and must_update: setter(password) return is_correct def make_password(password, salt=None, hasher='default'): """ Turn a plain-text password into a hash for database storage Same as encode() but generates a new random salt. If password is None then a concatenation of UNUSABLE_PASSWORD_PREFIX and a random string will be returned which disallows logins. Additional random string reduces chances of gaining access to staff or superuser accounts. See ticket #20079 for more info. """ if password is None: return UNUSABLE_PASSWORD_PREFIX + get_random_string(UNUSABLE_PASSWORD_SUFFIX_LENGTH) hasher = get_hasher(hasher) if not salt: salt = hasher.salt() return hasher.encode(password, salt) @lru_cache.lru_cache() def get_hashers(): hashers = [] for hasher_path in settings.PASSWORD_HASHERS: hasher_cls = import_string(hasher_path) hasher = hasher_cls() if not getattr(hasher, 'algorithm'): raise ImproperlyConfigured("hasher doesn't specify an " "algorithm name: %s" % hasher_path) hashers.append(hasher) return hashers @lru_cache.lru_cache() def get_hashers_by_algorithm(): return {hasher.algorithm: hasher for hasher in get_hashers()} @receiver(setting_changed) def reset_hashers(**kwargs): if kwargs['setting'] == 'PASSWORD_HASHERS': get_hashers.cache_clear() get_hashers_by_algorithm.cache_clear() def get_hasher(algorithm='default'): """ Returns an instance of a loaded password hasher. If algorithm is 'default', the default hasher will be returned. This function will also lazy import hashers specified in your settings file if needed. """ if hasattr(algorithm, 'algorithm'): return algorithm elif algorithm == 'default': return get_hashers()[0] else: hashers = get_hashers_by_algorithm() try: return hashers[algorithm] except KeyError: raise ValueError("Unknown password hashing algorithm '%s'. " "Did you specify it in the PASSWORD_HASHERS " "setting?" % algorithm) def identify_hasher(encoded): """ Returns an instance of a loaded password hasher. Identifies hasher algorithm by examining encoded hash, and calls get_hasher() to return hasher. Raises ValueError if algorithm cannot be identified, or if hasher is not loaded. """ # Ancient versions of Django created plain MD5 passwords and accepted # MD5 passwords with an empty salt. if ((len(encoded) == 32 and '$' not in encoded) or (len(encoded) == 37 and encoded.startswith('md5$$'))): algorithm = 'unsalted_md5' # Ancient versions of Django accepted SHA1 passwords with an empty salt. elif len(encoded) == 46 and encoded.startswith('sha1$$'): algorithm = 'unsalted_sha1' else: algorithm = encoded.split('$', 1)[0] return get_hasher(algorithm) def mask_hash(hash, show=6, char="*"): """ Returns the given hash, with only the first ``show`` number shown. The rest are masked with ``char`` for security reasons. """ masked = hash[:show] masked += char * len(hash[show:]) return masked class BasePasswordHasher(object): """ Abstract base class for password hashers When creating your own hasher, you need to override algorithm, verify(), encode() and safe_summary(). PasswordHasher objects are immutable. """ algorithm = None library = None def _load_library(self): if self.library is not None: if isinstance(self.library, (tuple, list)): name, mod_path = self.library else: mod_path = self.library try: module = importlib.import_module(mod_path) except ImportError as e: raise ValueError("Couldn't load %r algorithm library: %s" % (self.__class__.__name__, e)) return module raise ValueError("Hasher %r doesn't specify a library attribute" % self.__class__.__name__) def salt(self): """ Generates a cryptographically secure nonce salt in ASCII """ return get_random_string() def verify(self, password, encoded): """ Checks if the given password is correct """ raise NotImplementedError('subclasses of BasePasswordHasher must provide a verify() method') def encode(self, password, salt): """ Creates an encoded database value The result is normally formatted as "algorithm$salt$hash" and must be fewer than 128 characters. """ raise NotImplementedError('subclasses of BasePasswordHasher must provide an encode() method') def safe_summary(self, encoded): """ Returns a summary of safe values The result is a dictionary and will be used where the password field must be displayed to construct a safe representation of the password. """ raise NotImplementedError('subclasses of BasePasswordHasher must provide a safe_summary() method') def must_update(self, encoded): return False class PBKDF2PasswordHasher(BasePasswordHasher): """ Secure password hashing using the PBKDF2 algorithm (recommended) Configured to use PBKDF2 + HMAC + SHA256. The result is a 64 byte binary string. Iterations may be changed safely but you must rename the algorithm if you change SHA256. """ algorithm = "pbkdf2_sha256" iterations = 24000 digest = hashlib.sha256 def encode(self, password, salt, iterations=None): assert password is not None assert salt and '$' not in salt if not iterations: iterations = self.iterations hash = pbkdf2(password, salt, iterations, digest=self.digest) hash = base64.b64encode(hash).decode('ascii').strip() return "%s$%d$%s$%s" % (self.algorithm, iterations, salt, hash) def verify(self, password, encoded): algorithm, iterations, salt, hash = encoded.split('$', 3) assert algorithm == self.algorithm encoded_2 = self.encode(password, salt, int(iterations)) return constant_time_compare(encoded, encoded_2) def safe_summary(self, encoded): algorithm, iterations, salt, hash = encoded.split('$', 3) assert algorithm == self.algorithm return OrderedDict([ (_('algorithm'), algorithm), (_('iterations'), iterations), (_('salt'), mask_hash(salt)), (_('hash'), mask_hash(hash)), ]) def must_update(self, encoded): algorithm, iterations, salt, hash = encoded.split('$', 3) return int(iterations) != self.iterations class PBKDF2SHA1PasswordHasher(PBKDF2PasswordHasher): """ Alternate PBKDF2 hasher which uses SHA1, the default PRF recommended by PKCS #5. This is compatible with other implementations of PBKDF2, such as openssl's PKCS5_PBKDF2_HMAC_SHA1(). """ algorithm = "pbkdf2_sha1" digest = hashlib.sha1 class BCryptSHA256PasswordHasher(BasePasswordHasher): """ Secure password hashing using the bcrypt algorithm (recommended) This is considered by many to be the most secure algorithm but you must first install the bcrypt library. Please be warned that this library depends on native C code and might cause portability issues. """ algorithm = "bcrypt_sha256" digest = hashlib.sha256 library = ("bcrypt", "bcrypt") rounds = 12 def salt(self): bcrypt = self._load_library() return bcrypt.gensalt(rounds=self.rounds) def encode(self, password, salt): bcrypt = self._load_library() # Need to reevaluate the force_bytes call once bcrypt is supported on # Python 3 # Hash the password prior to using bcrypt to prevent password truncation # See: https://code.djangoproject.com/ticket/20138 if self.digest is not None: # We use binascii.hexlify here because Python3 decided that a hex encoded # bytestring is somehow a unicode. password = binascii.hexlify(self.digest(force_bytes(password)).digest()) else: password = force_bytes(password) data = bcrypt.hashpw(password, salt) return "%s$%s" % (self.algorithm, force_text(data)) def verify(self, password, encoded): algorithm, data = encoded.split('$', 1) assert algorithm == self.algorithm bcrypt = self._load_library() # Hash the password prior to using bcrypt to prevent password truncation # See: https://code.djangoproject.com/ticket/20138 if self.digest is not None: # We use binascii.hexlify here because Python3 decided that a hex encoded # bytestring is somehow a unicode. password = binascii.hexlify(self.digest(force_bytes(password)).digest()) else: password = force_bytes(password) # Ensure that our data is a bytestring data = force_bytes(data) # force_bytes() necessary for py-bcrypt compatibility hashpw = force_bytes(bcrypt.hashpw(password, data)) return constant_time_compare(data, hashpw) def safe_summary(self, encoded): algorithm, empty, algostr, work_factor, data = encoded.split('$', 4) assert algorithm == self.algorithm salt, checksum = data[:22], data[22:] return OrderedDict([ (_('algorithm'), algorithm), (_('work factor'), work_factor), (_('salt'), mask_hash(salt)), (_('checksum'), mask_hash(checksum)), ]) def must_update(self, encoded): algorithm, empty, algostr, rounds, data = encoded.split('$', 4) return int(rounds) != self.rounds class BCryptPasswordHasher(BCryptSHA256PasswordHasher): """ Secure password hashing using the bcrypt algorithm This is considered by many to be the most secure algorithm but you must first install the bcrypt library. Please be warned that this library depends on native C code and might cause portability issues. This hasher does not first hash the password which means it is subject to the 72 character bcrypt password truncation, most use cases should prefer the BCryptSha512PasswordHasher. See: https://code.djangoproject.com/ticket/20138 """ algorithm = "bcrypt" digest = None class SHA1PasswordHasher(BasePasswordHasher): """ The SHA1 password hashing algorithm (not recommended) """ algorithm = "sha1" def encode(self, password, salt): assert password is not None assert salt and '$' not in salt hash = hashlib.sha1(force_bytes(salt + password)).hexdigest() return "%s$%s$%s" % (self.algorithm, salt, hash) def verify(self, password, encoded): algorithm, salt, hash = encoded.split('$', 2) assert algorithm == self.algorithm encoded_2 = self.encode(password, salt) return constant_time_compare(encoded, encoded_2) def safe_summary(self, encoded): algorithm, salt, hash = encoded.split('$', 2) assert algorithm == self.algorithm return OrderedDict([ (_('algorithm'), algorithm), (_('salt'), mask_hash(salt, show=2)), (_('hash'), mask_hash(hash)), ]) class MD5PasswordHasher(BasePasswordHasher): """ The Salted MD5 password hashing algorithm (not recommended) """ algorithm = "md5" def encode(self, password, salt): assert password is not None assert salt and '$' not in salt hash = hashlib.md5(force_bytes(salt + password)).hexdigest() return "%s$%s$%s" % (self.algorithm, salt, hash) def verify(self, password, encoded): algorithm, salt, hash = encoded.split('$', 2) assert algorithm == self.algorithm encoded_2 = self.encode(password, salt) return constant_time_compare(encoded, encoded_2) def safe_summary(self, encoded): algorithm, salt, hash = encoded.split('$', 2) assert algorithm == self.algorithm return OrderedDict([ (_('algorithm'), algorithm), (_('salt'), mask_hash(salt, show=2)), (_('hash'), mask_hash(hash)), ]) class UnsaltedSHA1PasswordHasher(BasePasswordHasher): """ Very insecure algorithm that you should *never* use; stores SHA1 hashes with an empty salt. This class is implemented because Django used to accept such password hashes. Some older Django installs still have these values lingering around so we need to handle and upgrade them properly. """ algorithm = "unsalted_sha1" def salt(self): return '' def encode(self, password, salt): assert salt == '' hash = hashlib.sha1(force_bytes(password)).hexdigest() return 'sha1$$%s' % hash def verify(self, password, encoded): encoded_2 = self.encode(password, '') return constant_time_compare(encoded, encoded_2) def safe_summary(self, encoded): assert encoded.startswith('sha1$$') hash = encoded[6:] return OrderedDict([ (_('algorithm'), self.algorithm), (_('hash'), mask_hash(hash)), ]) class UnsaltedMD5PasswordHasher(BasePasswordHasher): """ Incredibly insecure algorithm that you should *never* use; stores unsalted MD5 hashes without the algorithm prefix, also accepts MD5 hashes with an empty salt. This class is implemented because Django used to store passwords this way and to accept such password hashes. Some older Django installs still have these values lingering around so we need to handle and upgrade them properly. """ algorithm = "unsalted_md5" def salt(self): return '' def encode(self, password, salt): assert salt == '' return hashlib.md5(force_bytes(password)).hexdigest() def verify(self, password, encoded): if len(encoded) == 37 and encoded.startswith('md5$$'): encoded = encoded[5:] encoded_2 = self.encode(password, '') return constant_time_compare(encoded, encoded_2) def safe_summary(self, encoded): return OrderedDict([ (_('algorithm'), self.algorithm), (_('hash'), mask_hash(encoded, show=3)), ]) class CryptPasswordHasher(BasePasswordHasher): """ Password hashing using UNIX crypt (not recommended) The crypt module is not supported on all platforms. """ algorithm = "crypt" library = "crypt" def salt(self): return get_random_string(2) def encode(self, password, salt): crypt = self._load_library() assert len(salt) == 2 data = crypt.crypt(force_str(password), salt) # we don't need to store the salt, but Django used to do this return "%s$%s$%s" % (self.algorithm, '', data) def verify(self, password, encoded): crypt = self._load_library() algorithm, salt, data = encoded.split('$', 2) assert algorithm == self.algorithm return constant_time_compare(data, crypt.crypt(force_str(password), data)) def safe_summary(self, encoded): algorithm, salt, data = encoded.split('$', 2) assert algorithm == self.algorithm return OrderedDict([ (_('algorithm'), algorithm), (_('salt'), salt), (_('hash'), mask_hash(data, show=3)), ])
bsd-3-clause
dandan94/OpenGLTest
finalOpenGL/HelloGLFW/lib/boost_1_59_0/libs/python/test/shared_ptr.py
46
2017
# Copyright David Abrahams 2004. Distributed under the Boost # Software License, Version 1.0. (See accompanying # file LICENSE_1_0.txt or copy at http://www.boost.org/LICENSE_1_0.txt) ''' >>> from shared_ptr_ext import * Test that shared_ptr<Derived> can be converted to shared_ptr<Base> >>> Y.store(YYY(42)) >>> x = X(17) >>> null_x = null(x) >>> null_x # should be None >>> identity(null_x) # should also be None >>> a = New(1) >>> A.call_f(a) 1 >>> New(0) >>> type(factory(3)) <class 'shared_ptr_ext.Y'> >>> type(factory(42)) <class 'shared_ptr_ext.YY'> >>> class P(Z): ... def v(self): ... return -Z.v(self); ... def __del__(self): ... print 'bye' ... >>> p = P(12) >>> p.value() 12 >>> p.v() -12 >>> look(p) 12 >>> try: modify(p) ... except TypeError: pass ... else: 'print expected a TypeError' >>> look(None) -1 >>> store(p) >>> del p >>> Z.get().v() -12 >>> Z.count() 1 >>> Z.look_store() 12 >>> Z.release() bye >>> Z.count() 0 >>> z = Z(13) >>> z.value() 13 >>> z.v() 13 >>> try: modify(z) ... except TypeError: pass ... else: 'print expected a TypeError' >>> Z.get() # should be None >>> store(z) >>> assert Z.get() is z # show that deleter introspection works >>> del z >>> Z.get().value() 13 >>> Z.count() 1 >>> Z.look_store() 13 >>> Z.release() >>> Z.count() 0 >>> x = X(17) >>> x.value() 17 >>> look(x) 17 >>> try: modify(x) ... except TypeError: pass ... else: 'print expected a TypeError' >>> look(None) -1 >>> store(x) >>> del x >>> X.count() 1 >>> X.look_store() 17 >>> X.release() >>> X.count() 0 >>> y = Y(19) >>> y.value() 19 >>> modify(y) >>> look(y) -1 >>> store(Y(23)) >>> Y.count() 1 >>> Y.look_store() 23 >>> Y.release() >>> Y.count() 0 ''' def run(args = None): import sys import doctest if args is not None: sys.argv = args return doctest.testmod(sys.modules.get(__name__)) if __name__ == '__main__': print "running..." import sys status = run()[0] if (status == 0): print "Done." sys.exit(status)
gpl-3.0
Rogdham/pelican
pelican/tests/test_urlwrappers.py
7
3417
# -*- coding: utf-8 -*- from __future__ import unicode_literals from pelican.tests.support import unittest from pelican.urlwrappers import Author, Category, Tag, URLWrapper class TestURLWrapper(unittest.TestCase): def test_ordering(self): # URLWrappers are sorted by name wrapper_a = URLWrapper(name='first', settings={}) wrapper_b = URLWrapper(name='last', settings={}) self.assertFalse(wrapper_a > wrapper_b) self.assertFalse(wrapper_a >= wrapper_b) self.assertFalse(wrapper_a == wrapper_b) self.assertTrue(wrapper_a != wrapper_b) self.assertTrue(wrapper_a <= wrapper_b) self.assertTrue(wrapper_a < wrapper_b) wrapper_b.name = 'first' self.assertFalse(wrapper_a > wrapper_b) self.assertTrue(wrapper_a >= wrapper_b) self.assertTrue(wrapper_a == wrapper_b) self.assertFalse(wrapper_a != wrapper_b) self.assertTrue(wrapper_a <= wrapper_b) self.assertFalse(wrapper_a < wrapper_b) wrapper_a.name = 'last' self.assertTrue(wrapper_a > wrapper_b) self.assertTrue(wrapper_a >= wrapper_b) self.assertFalse(wrapper_a == wrapper_b) self.assertTrue(wrapper_a != wrapper_b) self.assertFalse(wrapper_a <= wrapper_b) self.assertFalse(wrapper_a < wrapper_b) def test_equality(self): tag = Tag('test', settings={}) cat = Category('test', settings={}) author = Author('test', settings={}) # same name, but different class self.assertNotEqual(tag, cat) self.assertNotEqual(tag, author) # should be equal vs text representing the same name self.assertEqual(tag, u'test') # should not be equal vs binary self.assertNotEqual(tag, b'test') # Tags describing the same should be equal tag_equal = Tag('Test', settings={}) self.assertEqual(tag, tag_equal) # Author describing the same should be equal author_equal = Author('Test', settings={}) self.assertEqual(author, author_equal) cat_ascii = Category('指導書', settings={}) self.assertEqual(cat_ascii, u'zhi-dao-shu') def test_slugify_with_substitutions_and_dots(self): tag = Tag('Tag Dot', settings={ 'TAG_SUBSTITUTIONS': [('Tag Dot', 'tag.dot', True)] }) cat = Category('Category Dot', settings={ 'CATEGORY_SUBSTITUTIONS': (('Category Dot', 'cat.dot', True),) }) self.assertEqual(tag.slug, 'tag.dot') self.assertEqual(cat.slug, 'cat.dot') def test_author_slug_substitutions(self): settings = { 'AUTHOR_SUBSTITUTIONS': [ ('Alexander Todorov', 'atodorov', False), ('Krasimir Tsonev', 'krasimir', False), ] } author1 = Author('Mr. Senko', settings=settings) author2 = Author('Alexander Todorov', settings=settings) author3 = Author('Krasimir Tsonev', settings=settings) self.assertEqual(author1.slug, 'mr-senko') self.assertEqual(author2.slug, 'atodorov') self.assertEqual(author3.slug, 'krasimir')
agpl-3.0
gklyne/annalist
src/annalist_root/annalist/views/form_utils/fieldchoice.py
1
5788
""" This module defines a class used to represent a choice for an enumerated-value field. """ from __future__ import unicode_literals from __future__ import absolute_import, division, print_function __author__ = "Graham Klyne (GK@ACM.ORG)" __copyright__ = "Copyright 2015, G. Klyne" __license__ = "MIT (http://opensource.org/licenses/MIT)" import re import logging log = logging.getLogger(__name__) from collections import namedtuple from utils.py3porting import to_unicode from django.utils.html import format_html, mark_safe, escape _FieldChoice_tuple = namedtuple("FieldChoice", ("id", "value", "label", "link", "choice_value")) class FieldChoice(_FieldChoice_tuple): """ Class representing a choice for an enumerated field. >>> c1 = FieldChoice('id1', 'value1', 'label1', 'link1', choice_value=True) >>> c1 == FieldChoice(id='id1', value='value1', label='label1', link='link1', choice_value=True) True >>> c1.id == 'id1' True >>> c1.value == 'value1' True >>> c1.label == 'label1' True >>> c1.link == 'link1' True >>> c1.choice_html() == u'label1&nbsp;&nbsp;&nbsp;(value1)' True >>> c2 = FieldChoice('id2', 'value2', 'label2', 'link2', choice_value=False) >>> c2 == FieldChoice(id='id2', value='value2', label='label2', link='link2', choice_value=False) True >>> c2.id == 'id2' True >>> c2.value == 'value2' True >>> c2.label == 'label2' True >>> c2.link == 'link2' True >>> c2.choice() == u'label2' True >>> c3 = FieldChoice(id='id3', value='value3', link='link3') >>> c3 == FieldChoice(id='id3', value='value3', label='value3', link='link3', choice_value=False) True >>> c3.id == 'id3' True >>> c3.value == 'value3' True >>> c3.label == 'value3' True >>> c3.link == 'link3' True >>> c4 = FieldChoice('id4', link='link4') >>> c4 == FieldChoice(id='id4', value='id4', label='id4', link='link4', choice_value=False) True >>> c4.id == 'id4' True >>> c4.value == 'id4' True >>> c4.label == 'id4' True >>> c4.link == 'link4' True >>> c5 = FieldChoice('') >>> c5 == FieldChoice(id='', value='', label='', link=None, choice_value=False) True """ def __new__(_cls, id=None, value=None, label=None, link=None, choice_value=False): if value is None: value = id if label is None: label = value result = super(FieldChoice, _cls).__new__(_cls, id, value, label, link, choice_value) return result def __eq__(self, other): """ Returns True if self == other for sorting and equivalence purposes """ return self.id.__eq__(other.id) def __ne__(self, other): """ Returns True if self != other for sorting and equivalence purposes Note: required for Python2. """ return self.id.__ne__(other.id) def __lt__(self, other): """ Returns True if self < other for sorting purposes """ return self.id.__lt__(other.id) def __hash__(self): """ pylint says this should be defined if __eq__ is defined. Something to do with sets? """ return hash(self.id) def choice(self, sep=u"\xa0\xa0\xa0"): """ Return choice string """ if self.choice_value: choice_text = self.option_label(sep=sep) else: choice_text = to_unicode(self.label) return choice_text def choice_html(self, sep=u"&nbsp;&nbsp;&nbsp;"): """ Return choice string HTML for option in drop-down list. """ return self.choice(sep=sep) def add_link(self, link=None): return FieldChoice(self.id, self.value, self.label, link) def option_label(self, sep=u"\xa0\xa0\xa0"): """ Returns string used for displayed option label. This function is used mainly for testing, to isolate details of option presentation from the majority of test cases. """ if self.label: return format_html(u"{}{}({})", self.label, mark_safe(sep), self.value) else: return escape(self.value) def option_label_html(self, sep=u"&nbsp;&nbsp;&nbsp;"): """ Variation of option_label returns HTML-encoded form of label text """ return self.option_label(sep=sep) def update_choice_labels(fieldchoices): """ Update choice labels in supplied list of FieldChoice values so that duplicate labels can be distinguished. Returns an updated list of options. """ # Detect non-unique labels labels = {} for o in fieldchoices: l = o.label labels[l] = labels.get(l, 0) + 1 # Generate updated choice values new_choices = [] for o in fieldchoices: if labels[o.label] > 1: new_choices.append( FieldChoice(id=o.id, value=o.value, label=o.label, link=o.link, choice_value=True) ) else: new_choices.append(o) return new_choices def get_choice_labels(fieldchoices): """ Return a list of choice labels based on the supplied list of FieldChoice values >>> c1 = FieldChoice('id1', 'value1', 'label1', 'link1') >>> c2a = FieldChoice('id2a', 'value2a', 'label2', 'link2') >>> c2b = FieldChoice('id2b', 'value2b', 'label2', 'link2') >>> labels = get_choice_labels([c1,c2a,c2b]) >>> labels == ['label1', u'label2\\xa0\\xa0\\xa0(value2a)', u'label2\\xa0\\xa0\\xa0(value2b)'] True """ return [ fc.choice() for fc in update_choice_labels(fieldchoices) ] if __name__ == "__main__": import doctest doctest.testmod() # End.
mit
kaczla/PJN
src/Przecinki/scikit.py
1
1048
#!/usr/bin/python2 # -*- coding: utf-8 -*- import sys import matplotlib.pyplot as plt import numpy as np from sklearn import datasets from sklearn.cross_validation import cross_val_predict from sklearn import linear_model from sklearn import datasets X = [] Y = [] for line in sys.stdin: line = line.rstrip() X.append([len(line.split())]) Y.append(line.count(",")) lr = linear_model.LinearRegression() predicted = cross_val_predict(lr, X, Y) FILE = open(sys.argv[1], "r") X_TEST = [] Y_TEST = [] for line in FILE: line = line.rstrip() Y_TEST.append(line.count(",")) line = line.replace(",", "") X_TEST.append([len(line.split())]) regr = linear_model.LinearRegression() regr.fit(X, Y) print "Coefficients: ", regr.coef_ print "Residual sum of squares: %.2f" % np.mean((regr.predict(X_TEST) - Y_TEST) ** 2) print "Variance score: %.2f" % regr.score(X_TEST, Y_TEST) plt.scatter(X_TEST, Y_TEST, color='black') plt.plot(X_TEST, regr.predict(X_TEST), color='green', linewidth=2) plt.xticks(()) plt.yticks(()) plt.show()
gpl-2.0
maoy/zknova
nova/tests/api/openstack/test_faults.py
1
7383
# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright 2010 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from xml.dom import minidom import webob import webob.dec import webob.exc from nova.api.openstack import common from nova.api.openstack import wsgi from nova.openstack.common import jsonutils from nova import test class TestFaults(test.TestCase): """Tests covering `nova.api.openstack.faults:Fault` class.""" def _prepare_xml(self, xml_string): """Remove characters from string which hinder XML equality testing.""" xml_string = xml_string.replace(" ", "") xml_string = xml_string.replace("\n", "") xml_string = xml_string.replace("\t", "") return xml_string def test_400_fault_json(self): # Test fault serialized to JSON via file-extension and/or header. requests = [ webob.Request.blank('/.json'), webob.Request.blank('/', headers={"Accept": "application/json"}), ] for request in requests: fault = wsgi.Fault(webob.exc.HTTPBadRequest(explanation='scram')) response = request.get_response(fault) expected = { "badRequest": { "message": "scram", "code": 400, }, } actual = jsonutils.loads(response.body) self.assertEqual(response.content_type, "application/json") self.assertEqual(expected, actual) def test_413_fault_json(self): # Test fault serialized to JSON via file-extension and/or header. requests = [ webob.Request.blank('/.json'), webob.Request.blank('/', headers={"Accept": "application/json"}), ] for request in requests: exc = webob.exc.HTTPRequestEntityTooLarge fault = wsgi.Fault(exc(explanation='sorry', headers={'Retry-After': 4})) response = request.get_response(fault) expected = { "overLimit": { "message": "sorry", "code": 413, "retryAfter": 4, }, } actual = jsonutils.loads(response.body) self.assertEqual(response.content_type, "application/json") self.assertEqual(expected, actual) def test_raise(self): # Ensure the ability to raise :class:`Fault` in WSGI-ified methods. @webob.dec.wsgify def raiser(req): raise wsgi.Fault(webob.exc.HTTPNotFound(explanation='whut?')) req = webob.Request.blank('/.xml') resp = req.get_response(raiser) self.assertEqual(resp.content_type, "application/xml") self.assertEqual(resp.status_int, 404) self.assertTrue('whut?' in resp.body) def test_raise_403(self): # Ensure the ability to raise :class:`Fault` in WSGI-ified methods. @webob.dec.wsgify def raiser(req): raise wsgi.Fault(webob.exc.HTTPForbidden(explanation='whut?')) req = webob.Request.blank('/.xml') resp = req.get_response(raiser) self.assertEqual(resp.content_type, "application/xml") self.assertEqual(resp.status_int, 403) self.assertTrue('resizeNotAllowed' not in resp.body) self.assertTrue('forbidden' in resp.body) def test_fault_has_status_int(self): # Ensure the status_int is set correctly on faults. fault = wsgi.Fault(webob.exc.HTTPBadRequest(explanation='what?')) self.assertEqual(fault.status_int, 400) def test_xml_serializer(self): # Ensure that a v1.1 request responds with a v1.1 xmlns. request = webob.Request.blank('/v1.1', headers={"Accept": "application/xml"}) fault = wsgi.Fault(webob.exc.HTTPBadRequest(explanation='scram')) response = request.get_response(fault) self.assertTrue(common.XML_NS_V11 in response.body) self.assertEqual(response.content_type, "application/xml") self.assertEqual(response.status_int, 400) class FaultsXMLSerializationTestV11(test.TestCase): """Tests covering `nova.api.openstack.faults:Fault` class.""" def _prepare_xml(self, xml_string): xml_string = xml_string.replace(" ", "") xml_string = xml_string.replace("\n", "") xml_string = xml_string.replace("\t", "") return xml_string def test_400_fault(self): metadata = {'attributes': {"badRequest": 'code'}} serializer = wsgi.XMLDictSerializer(metadata=metadata, xmlns=common.XML_NS_V11) fixture = { "badRequest": { "message": "scram", "code": 400, }, } output = serializer.serialize(fixture) actual = minidom.parseString(self._prepare_xml(output)) expected = minidom.parseString(self._prepare_xml(""" <badRequest code="400" xmlns="%s"> <message>scram</message> </badRequest> """) % common.XML_NS_V11) self.assertEqual(expected.toxml(), actual.toxml()) def test_413_fault(self): metadata = {'attributes': {"overLimit": 'code'}} serializer = wsgi.XMLDictSerializer(metadata=metadata, xmlns=common.XML_NS_V11) fixture = { "overLimit": { "message": "sorry", "code": 413, "retryAfter": 4, }, } output = serializer.serialize(fixture) actual = minidom.parseString(self._prepare_xml(output)) expected = minidom.parseString(self._prepare_xml(""" <overLimit code="413" xmlns="%s"> <message>sorry</message> <retryAfter>4</retryAfter> </overLimit> """) % common.XML_NS_V11) self.assertEqual(expected.toxml(), actual.toxml()) def test_404_fault(self): metadata = {'attributes': {"itemNotFound": 'code'}} serializer = wsgi.XMLDictSerializer(metadata=metadata, xmlns=common.XML_NS_V11) fixture = { "itemNotFound": { "message": "sorry", "code": 404, }, } output = serializer.serialize(fixture) actual = minidom.parseString(self._prepare_xml(output)) expected = minidom.parseString(self._prepare_xml(""" <itemNotFound code="404" xmlns="%s"> <message>sorry</message> </itemNotFound> """) % common.XML_NS_V11) self.assertEqual(expected.toxml(), actual.toxml())
apache-2.0
hfp/tensorflow-xsmm
tensorflow/python/training/basic_loops.py
13
2348
# Copyright 2016 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Basic loop for training.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow.python.framework import errors from tensorflow.python.util.tf_export import tf_export @tf_export(v1=["train.basic_train_loop"]) def basic_train_loop(supervisor, train_step_fn, args=None, kwargs=None, master=""): """Basic loop to train a model. Calls `train_step_fn` in a loop to train a model. The function is called as: ```python train_step_fn(session, *args, **kwargs) ``` It is passed a `tf.Session` in addition to `args` and `kwargs`. The function typically runs one training step in the session. Args: supervisor: `tf.train.Supervisor` to run the training services. train_step_fn: Callable to execute one training step. Called repeatedly as `train_step_fn(session, *args **kwargs)`. args: Optional positional arguments passed to `train_step_fn`. kwargs: Optional keyword arguments passed to `train_step_fn`. master: Master to use to create the training session. Defaults to `""` which causes the session to be created in the local process. """ if args is None: args = [] if kwargs is None: kwargs = {} should_retry = True while should_retry: try: should_retry = False with supervisor.managed_session(master) as sess: while not supervisor.should_stop(): train_step_fn(sess, *args, **kwargs) except errors.AbortedError: # Always re-run on AbortedError as it indicates a restart of one of the # distributed tensorflow servers. should_retry = True
apache-2.0
Sylrob434/CouchPotatoServer
couchpotato/core/media/movie/providers/trailer/youtube_dl/extractor/fourtube.py
19
3728
from __future__ import unicode_literals import re from .common import InfoExtractor from ..utils import ( compat_urllib_request, unified_strdate, str_to_int, parse_duration, clean_html, ) class FourTubeIE(InfoExtractor): IE_NAME = '4tube' _VALID_URL = r'https?://(?:www\.)?4tube\.com/videos/(?P<id>\d+)' _TEST = { 'url': 'http://www.4tube.com/videos/209733/hot-babe-holly-michaels-gets-her-ass-stuffed-by-black', 'md5': '6516c8ac63b03de06bc8eac14362db4f', 'info_dict': { 'id': '209733', 'ext': 'mp4', 'title': 'Hot Babe Holly Michaels gets her ass stuffed by black', 'uploader': 'WCP Club', 'uploader_id': 'wcp-club', 'upload_date': '20131031', 'duration': 583, } } def _real_extract(self, url): mobj = re.match(self._VALID_URL, url) video_id = mobj.group('id') webpage_url = 'http://www.4tube.com/videos/' + video_id webpage = self._download_webpage(webpage_url, video_id) self.report_extraction(video_id) playlist_json = self._html_search_regex(r'var playerConfigPlaylist\s+=\s+([^;]+)', webpage, 'Playlist') media_id = self._search_regex(r'idMedia:\s*(\d+)', playlist_json, 'Media Id') sources = self._search_regex(r'sources:\s*\[([^\]]*)\]', playlist_json, 'Sources').split(',') title = self._search_regex(r'title:\s*"([^"]*)', playlist_json, 'Title') thumbnail_url = self._search_regex(r'image:\s*"([^"]*)', playlist_json, 'Thumbnail', fatal=False) uploader_str = self._search_regex(r'<span>Uploaded by</span>(.*?)<span>', webpage, 'uploader', fatal=False) mobj = re.search(r'<a href="/sites/(?P<id>[^"]+)"><strong>(?P<name>[^<]+)</strong></a>', uploader_str) (uploader, uploader_id) = (mobj.group('name'), mobj.group('id')) if mobj else (clean_html(uploader_str), None) upload_date = None view_count = None duration = None description = self._html_search_meta('description', webpage, 'description') if description: upload_date = self._search_regex(r'Published Date: (\d{2} [a-zA-Z]{3} \d{4})', description, 'upload date', fatal=False) if upload_date: upload_date = unified_strdate(upload_date) view_count = self._search_regex(r'Views: ([\d,\.]+)', description, 'view count', fatal=False) if view_count: view_count = str_to_int(view_count) duration = parse_duration(self._search_regex(r'Length: (\d+m\d+s)', description, 'duration', fatal=False)) token_url = "http://tkn.4tube.com/{0}/desktop/{1}".format(media_id, "+".join(sources)) headers = { b'Content-Type': b'application/x-www-form-urlencoded', b'Origin': b'http://www.4tube.com', } token_req = compat_urllib_request.Request(token_url, b'{}', headers) tokens = self._download_json(token_req, video_id) formats = [{ 'url': tokens[format]['token'], 'format_id': format + 'p', 'resolution': format + 'p', 'quality': int(format), } for format in sources] self._sort_formats(formats) return { 'id': video_id, 'title': title, 'formats': formats, 'thumbnail': thumbnail_url, 'uploader': uploader, 'uploader_id': uploader_id, 'upload_date': upload_date, 'view_count': view_count, 'duration': duration, 'age_limit': 18, 'webpage_url': webpage_url, }
gpl-3.0
srluge/SickRage
lib/sqlalchemy/orm/mapper.py
75
108686
# orm/mapper.py # Copyright (C) 2005-2014 the SQLAlchemy authors and contributors <see AUTHORS file> # # This module is part of SQLAlchemy and is released under # the MIT License: http://www.opensource.org/licenses/mit-license.php """Logic to map Python classes to and from selectables. Defines the :class:`~sqlalchemy.orm.mapper.Mapper` class, the central configurational unit which associates a class with a database table. This is a semi-private module; the main configurational API of the ORM is available in :class:`~sqlalchemy.orm.`. """ from __future__ import absolute_import import types import weakref from itertools import chain from collections import deque from .. import sql, util, log, exc as sa_exc, event, schema, inspection from ..sql import expression, visitors, operators, util as sql_util from . import instrumentation, attributes, exc as orm_exc, loading from . import properties from .interfaces import MapperProperty, _InspectionAttr, _MappedAttribute from .base import _class_to_mapper, _state_mapper, class_mapper, \ state_str, _INSTRUMENTOR from .path_registry import PathRegistry import sys _mapper_registry = weakref.WeakKeyDictionary() _already_compiling = False _memoized_configured_property = util.group_expirable_memoized_property() # a constant returned by _get_attr_by_column to indicate # this mapper is not handling an attribute for a particular # column NO_ATTRIBUTE = util.symbol('NO_ATTRIBUTE') # lock used to synchronize the "mapper configure" step _CONFIGURE_MUTEX = util.threading.RLock() @inspection._self_inspects @log.class_logger class Mapper(_InspectionAttr): """Define the correlation of class attributes to database table columns. The :class:`.Mapper` object is instantiated using the :func:`~sqlalchemy.orm.mapper` function. For information about instantiating new :class:`.Mapper` objects, see that function's documentation. When :func:`.mapper` is used explicitly to link a user defined class with table metadata, this is referred to as *classical mapping*. Modern SQLAlchemy usage tends to favor the :mod:`sqlalchemy.ext.declarative` extension for class configuration, which makes usage of :func:`.mapper` behind the scenes. Given a particular class known to be mapped by the ORM, the :class:`.Mapper` which maintains it can be acquired using the :func:`.inspect` function:: from sqlalchemy import inspect mapper = inspect(MyClass) A class which was mapped by the :mod:`sqlalchemy.ext.declarative` extension will also have its mapper available via the ``__mapper__`` attribute. """ _new_mappers = False def __init__(self, class_, local_table=None, properties=None, primary_key=None, non_primary=False, inherits=None, inherit_condition=None, inherit_foreign_keys=None, extension=None, order_by=False, always_refresh=False, version_id_col=None, version_id_generator=None, polymorphic_on=None, _polymorphic_map=None, polymorphic_identity=None, concrete=False, with_polymorphic=None, allow_partial_pks=True, batch=True, column_prefix=None, include_properties=None, exclude_properties=None, passive_updates=True, confirm_deleted_rows=True, eager_defaults=False, legacy_is_orphan=False, _compiled_cache_size=100, ): """Return a new :class:`~.Mapper` object. This function is typically used behind the scenes via the Declarative extension. When using Declarative, many of the usual :func:`.mapper` arguments are handled by the Declarative extension itself, including ``class_``, ``local_table``, ``properties``, and ``inherits``. Other options are passed to :func:`.mapper` using the ``__mapper_args__`` class variable:: class MyClass(Base): __tablename__ = 'my_table' id = Column(Integer, primary_key=True) type = Column(String(50)) alt = Column("some_alt", Integer) __mapper_args__ = { 'polymorphic_on' : type } Explicit use of :func:`.mapper` is often referred to as *classical mapping*. The above declarative example is equivalent in classical form to:: my_table = Table("my_table", metadata, Column('id', Integer, primary_key=True), Column('type', String(50)), Column("some_alt", Integer) ) class MyClass(object): pass mapper(MyClass, my_table, polymorphic_on=my_table.c.type, properties={ 'alt':my_table.c.some_alt }) .. seealso:: :ref:`classical_mapping` - discussion of direct usage of :func:`.mapper` :param class\_: The class to be mapped. When using Declarative, this argument is automatically passed as the declared class itself. :param local_table: The :class:`.Table` or other selectable to which the class is mapped. May be ``None`` if this mapper inherits from another mapper using single-table inheritance. When using Declarative, this argument is automatically passed by the extension, based on what is configured via the ``__table__`` argument or via the :class:`.Table` produced as a result of the ``__tablename__`` and :class:`.Column` arguments present. :param always_refresh: If True, all query operations for this mapped class will overwrite all data within object instances that already exist within the session, erasing any in-memory changes with whatever information was loaded from the database. Usage of this flag is highly discouraged; as an alternative, see the method :meth:`.Query.populate_existing`. :param allow_partial_pks: Defaults to True. Indicates that a composite primary key with some NULL values should be considered as possibly existing within the database. This affects whether a mapper will assign an incoming row to an existing identity, as well as if :meth:`.Session.merge` will check the database first for a particular primary key value. A "partial primary key" can occur if one has mapped to an OUTER JOIN, for example. :param batch: Defaults to ``True``, indicating that save operations of multiple entities can be batched together for efficiency. Setting to False indicates that an instance will be fully saved before saving the next instance. This is used in the extremely rare case that a :class:`.MapperEvents` listener requires being called in between individual row persistence operations. :param column_prefix: A string which will be prepended to the mapped attribute name when :class:`.Column` objects are automatically assigned as attributes to the mapped class. Does not affect explicitly specified column-based properties. See the section :ref:`column_prefix` for an example. :param concrete: If True, indicates this mapper should use concrete table inheritance with its parent mapper. See the section :ref:`concrete_inheritance` for an example. :param confirm_deleted_rows: defaults to True; when a DELETE occurs of one more more rows based on specific primary keys, a warning is emitted when the number of rows matched does not equal the number of rows expected. This parameter may be set to False to handle the case where database ON DELETE CASCADE rules may be deleting some of those rows automatically. The warning may be changed to an exception in a future release. .. versionadded:: 0.9.4 - added :paramref:`.mapper.confirm_deleted_rows` as well as conditional matched row checking on delete. :param eager_defaults: if True, the ORM will immediately fetch the value of server-generated default values after an INSERT or UPDATE, rather than leaving them as expired to be fetched on next access. This can be used for event schemes where the server-generated values are needed immediately before the flush completes. By default, this scheme will emit an individual ``SELECT`` statement per row inserted or updated, which note can add significant performance overhead. However, if the target database supports :term:`RETURNING`, the default values will be returned inline with the INSERT or UPDATE statement, which can greatly enhance performance for an application that needs frequent access to just-generated server defaults. .. versionchanged:: 0.9.0 The ``eager_defaults`` option can now make use of :term:`RETURNING` for backends which support it. :param exclude_properties: A list or set of string column names to be excluded from mapping. See :ref:`include_exclude_cols` for an example. :param extension: A :class:`.MapperExtension` instance or list of :class:`.MapperExtension` instances which will be applied to all operations by this :class:`.Mapper`. **Deprecated.** Please see :class:`.MapperEvents`. :param include_properties: An inclusive list or set of string column names to map. See :ref:`include_exclude_cols` for an example. :param inherits: A mapped class or the corresponding :class:`.Mapper` of one indicating a superclass to which this :class:`.Mapper` should *inherit* from. The mapped class here must be a subclass of the other mapper's class. When using Declarative, this argument is passed automatically as a result of the natural class hierarchy of the declared classes. .. seealso:: :ref:`inheritance_toplevel` :param inherit_condition: For joined table inheritance, a SQL expression which will define how the two tables are joined; defaults to a natural join between the two tables. :param inherit_foreign_keys: When ``inherit_condition`` is used and the columns present are missing a :class:`.ForeignKey` configuration, this parameter can be used to specify which columns are "foreign". In most cases can be left as ``None``. :param legacy_is_orphan: Boolean, defaults to ``False``. When ``True``, specifies that "legacy" orphan consideration is to be applied to objects mapped by this mapper, which means that a pending (that is, not persistent) object is auto-expunged from an owning :class:`.Session` only when it is de-associated from *all* parents that specify a ``delete-orphan`` cascade towards this mapper. The new default behavior is that the object is auto-expunged when it is de-associated with *any* of its parents that specify ``delete-orphan`` cascade. This behavior is more consistent with that of a persistent object, and allows behavior to be consistent in more scenarios independently of whether or not an orphanable object has been flushed yet or not. See the change note and example at :ref:`legacy_is_orphan_addition` for more detail on this change. .. versionadded:: 0.8 - the consideration of a pending object as an "orphan" has been modified to more closely match the behavior as that of persistent objects, which is that the object is expunged from the :class:`.Session` as soon as it is de-associated from any of its orphan-enabled parents. Previously, the pending object would be expunged only if de-associated from all of its orphan-enabled parents. The new flag ``legacy_is_orphan`` is added to :func:`.orm.mapper` which re-establishes the legacy behavior. :param non_primary: Specify that this :class:`.Mapper` is in addition to the "primary" mapper, that is, the one used for persistence. The :class:`.Mapper` created here may be used for ad-hoc mapping of the class to an alternate selectable, for loading only. :paramref:`.Mapper.non_primary` is not an often used option, but is useful in some specific :func:`.relationship` cases. .. seealso:: :ref:`relationship_non_primary_mapper` :param order_by: A single :class:`.Column` or list of :class:`.Column` objects for which selection operations should use as the default ordering for entities. By default mappers have no pre-defined ordering. :param passive_updates: Indicates UPDATE behavior of foreign key columns when a primary key column changes on a joined-table inheritance mapping. Defaults to ``True``. When True, it is assumed that ON UPDATE CASCADE is configured on the foreign key in the database, and that the database will handle propagation of an UPDATE from a source column to dependent columns on joined-table rows. When False, it is assumed that the database does not enforce referential integrity and will not be issuing its own CASCADE operation for an update. The unit of work process will emit an UPDATE statement for the dependent columns during a primary key change. .. seealso:: :ref:`passive_updates` - description of a similar feature as used with :func:`.relationship` :param polymorphic_on: Specifies the column, attribute, or SQL expression used to determine the target class for an incoming row, when inheriting classes are present. This value is commonly a :class:`.Column` object that's present in the mapped :class:`.Table`:: class Employee(Base): __tablename__ = 'employee' id = Column(Integer, primary_key=True) discriminator = Column(String(50)) __mapper_args__ = { "polymorphic_on":discriminator, "polymorphic_identity":"employee" } It may also be specified as a SQL expression, as in this example where we use the :func:`.case` construct to provide a conditional approach:: class Employee(Base): __tablename__ = 'employee' id = Column(Integer, primary_key=True) discriminator = Column(String(50)) __mapper_args__ = { "polymorphic_on":case([ (discriminator == "EN", "engineer"), (discriminator == "MA", "manager"), ], else_="employee"), "polymorphic_identity":"employee" } It may also refer to any attribute configured with :func:`.column_property`, or to the string name of one:: class Employee(Base): __tablename__ = 'employee' id = Column(Integer, primary_key=True) discriminator = Column(String(50)) employee_type = column_property( case([ (discriminator == "EN", "engineer"), (discriminator == "MA", "manager"), ], else_="employee") ) __mapper_args__ = { "polymorphic_on":employee_type, "polymorphic_identity":"employee" } .. versionchanged:: 0.7.4 ``polymorphic_on`` may be specified as a SQL expression, or refer to any attribute configured with :func:`.column_property`, or to the string name of one. When setting ``polymorphic_on`` to reference an attribute or expression that's not present in the locally mapped :class:`.Table`, yet the value of the discriminator should be persisted to the database, the value of the discriminator is not automatically set on new instances; this must be handled by the user, either through manual means or via event listeners. A typical approach to establishing such a listener looks like:: from sqlalchemy import event from sqlalchemy.orm import object_mapper @event.listens_for(Employee, "init", propagate=True) def set_identity(instance, *arg, **kw): mapper = object_mapper(instance) instance.discriminator = mapper.polymorphic_identity Where above, we assign the value of ``polymorphic_identity`` for the mapped class to the ``discriminator`` attribute, thus persisting the value to the ``discriminator`` column in the database. .. seealso:: :ref:`inheritance_toplevel` :param polymorphic_identity: Specifies the value which identifies this particular class as returned by the column expression referred to by the ``polymorphic_on`` setting. As rows are received, the value corresponding to the ``polymorphic_on`` column expression is compared to this value, indicating which subclass should be used for the newly reconstructed object. :param properties: A dictionary mapping the string names of object attributes to :class:`.MapperProperty` instances, which define the persistence behavior of that attribute. Note that :class:`.Column` objects present in the mapped :class:`.Table` are automatically placed into ``ColumnProperty`` instances upon mapping, unless overridden. When using Declarative, this argument is passed automatically, based on all those :class:`.MapperProperty` instances declared in the declared class body. :param primary_key: A list of :class:`.Column` objects which define the primary key to be used against this mapper's selectable unit. This is normally simply the primary key of the ``local_table``, but can be overridden here. :param version_id_col: A :class:`.Column` that will be used to keep a running version id of rows in the table. This is used to detect concurrent updates or the presence of stale data in a flush. The methodology is to detect if an UPDATE statement does not match the last known version id, a :class:`~sqlalchemy.orm.exc.StaleDataError` exception is thrown. By default, the column must be of :class:`.Integer` type, unless ``version_id_generator`` specifies an alternative version generator. .. seealso:: :ref:`mapper_version_counter` - discussion of version counting and rationale. :param version_id_generator: Define how new version ids should be generated. Defaults to ``None``, which indicates that a simple integer counting scheme be employed. To provide a custom versioning scheme, provide a callable function of the form:: def generate_version(version): return next_version Alternatively, server-side versioning functions such as triggers, or programmatic versioning schemes outside of the version id generator may be used, by specifying the value ``False``. Please see :ref:`server_side_version_counter` for a discussion of important points when using this option. .. versionadded:: 0.9.0 ``version_id_generator`` supports server-side version number generation. .. seealso:: :ref:`custom_version_counter` :ref:`server_side_version_counter` :param with_polymorphic: A tuple in the form ``(<classes>, <selectable>)`` indicating the default style of "polymorphic" loading, that is, which tables are queried at once. <classes> is any single or list of mappers and/or classes indicating the inherited classes that should be loaded at once. The special value ``'*'`` may be used to indicate all descending classes should be loaded immediately. The second tuple argument <selectable> indicates a selectable that will be used to query for multiple classes. .. seealso:: :ref:`with_polymorphic` - discussion of polymorphic querying techniques. """ self.class_ = util.assert_arg_type(class_, type, 'class_') self.class_manager = None self._primary_key_argument = util.to_list(primary_key) self.non_primary = non_primary if order_by is not False: self.order_by = util.to_list(order_by) else: self.order_by = order_by self.always_refresh = always_refresh if isinstance(version_id_col, MapperProperty): self.version_id_prop = version_id_col self.version_id_col = None else: self.version_id_col = version_id_col if version_id_generator is False: self.version_id_generator = False elif version_id_generator is None: self.version_id_generator = lambda x: (x or 0) + 1 else: self.version_id_generator = version_id_generator self.concrete = concrete self.single = False self.inherits = inherits self.local_table = local_table self.inherit_condition = inherit_condition self.inherit_foreign_keys = inherit_foreign_keys self._init_properties = properties or {} self._delete_orphans = [] self.batch = batch self.eager_defaults = eager_defaults self.column_prefix = column_prefix self.polymorphic_on = expression._clause_element_as_expr( polymorphic_on) self._dependency_processors = [] self.validators = util.immutabledict() self.passive_updates = passive_updates self.legacy_is_orphan = legacy_is_orphan self._clause_adapter = None self._requires_row_aliasing = False self._inherits_equated_pairs = None self._memoized_values = {} self._compiled_cache_size = _compiled_cache_size self._reconstructor = None self._deprecated_extensions = util.to_list(extension or []) self.allow_partial_pks = allow_partial_pks if self.inherits and not self.concrete: self.confirm_deleted_rows = False else: self.confirm_deleted_rows = confirm_deleted_rows self._set_with_polymorphic(with_polymorphic) if isinstance(self.local_table, expression.SelectBase): raise sa_exc.InvalidRequestError( "When mapping against a select() construct, map against " "an alias() of the construct instead." "This because several databases don't allow a " "SELECT from a subquery that does not have an alias." ) if self.with_polymorphic and \ isinstance(self.with_polymorphic[1], expression.SelectBase): self.with_polymorphic = (self.with_polymorphic[0], self.with_polymorphic[1].alias()) # our 'polymorphic identity', a string name that when located in a # result set row indicates this Mapper should be used to construct # the object instance for that row. self.polymorphic_identity = polymorphic_identity # a dictionary of 'polymorphic identity' names, associating those # names with Mappers that will be used to construct object instances # upon a select operation. if _polymorphic_map is None: self.polymorphic_map = {} else: self.polymorphic_map = _polymorphic_map if include_properties is not None: self.include_properties = util.to_set(include_properties) else: self.include_properties = None if exclude_properties: self.exclude_properties = util.to_set(exclude_properties) else: self.exclude_properties = None self.configured = False # prevent this mapper from being constructed # while a configure_mappers() is occurring (and defer a # configure_mappers() until construction succeeds) _CONFIGURE_MUTEX.acquire() try: self.dispatch._events._new_mapper_instance(class_, self) self._configure_inheritance() self._configure_legacy_instrument_class() self._configure_class_instrumentation() self._configure_listeners() self._configure_properties() self._configure_polymorphic_setter() self._configure_pks() Mapper._new_mappers = True self._log("constructed") self._expire_memoizations() finally: _CONFIGURE_MUTEX.release() # major attributes initialized at the classlevel so that # they can be Sphinx-documented. is_mapper = True """Part of the inspection API.""" @property def mapper(self): """Part of the inspection API. Returns self. """ return self @property def entity(self): """Part of the inspection API. Returns self.class\_. """ return self.class_ local_table = None """The :class:`.Selectable` which this :class:`.Mapper` manages. Typically is an instance of :class:`.Table` or :class:`.Alias`. May also be ``None``. The "local" table is the selectable that the :class:`.Mapper` is directly responsible for managing from an attribute access and flush perspective. For non-inheriting mappers, the local table is the same as the "mapped" table. For joined-table inheritance mappers, local_table will be the particular sub-table of the overall "join" which this :class:`.Mapper` represents. If this mapper is a single-table inheriting mapper, local_table will be ``None``. .. seealso:: :attr:`~.Mapper.mapped_table`. """ mapped_table = None """The :class:`.Selectable` to which this :class:`.Mapper` is mapped. Typically an instance of :class:`.Table`, :class:`.Join`, or :class:`.Alias`. The "mapped" table is the selectable that the mapper selects from during queries. For non-inheriting mappers, the mapped table is the same as the "local" table. For joined-table inheritance mappers, mapped_table references the full :class:`.Join` representing full rows for this particular subclass. For single-table inheritance mappers, mapped_table references the base table. .. seealso:: :attr:`~.Mapper.local_table`. """ inherits = None """References the :class:`.Mapper` which this :class:`.Mapper` inherits from, if any. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ configured = None """Represent ``True`` if this :class:`.Mapper` has been configured. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. .. seealso:: :func:`.configure_mappers`. """ concrete = None """Represent ``True`` if this :class:`.Mapper` is a concrete inheritance mapper. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ tables = None """An iterable containing the collection of :class:`.Table` objects which this :class:`.Mapper` is aware of. If the mapper is mapped to a :class:`.Join`, or an :class:`.Alias` representing a :class:`.Select`, the individual :class:`.Table` objects that comprise the full construct will be represented here. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ primary_key = None """An iterable containing the collection of :class:`.Column` objects which comprise the 'primary key' of the mapped table, from the perspective of this :class:`.Mapper`. This list is against the selectable in :attr:`~.Mapper.mapped_table`. In the case of inheriting mappers, some columns may be managed by a superclass mapper. For example, in the case of a :class:`.Join`, the primary key is determined by all of the primary key columns across all tables referenced by the :class:`.Join`. The list is also not necessarily the same as the primary key column collection associated with the underlying tables; the :class:`.Mapper` features a ``primary_key`` argument that can override what the :class:`.Mapper` considers as primary key columns. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ class_ = None """The Python class which this :class:`.Mapper` maps. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ class_manager = None """The :class:`.ClassManager` which maintains event listeners and class-bound descriptors for this :class:`.Mapper`. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ single = None """Represent ``True`` if this :class:`.Mapper` is a single table inheritance mapper. :attr:`~.Mapper.local_table` will be ``None`` if this flag is set. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ non_primary = None """Represent ``True`` if this :class:`.Mapper` is a "non-primary" mapper, e.g. a mapper that is used only to selet rows but not for persistence management. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ polymorphic_on = None """The :class:`.Column` or SQL expression specified as the ``polymorphic_on`` argument for this :class:`.Mapper`, within an inheritance scenario. This attribute is normally a :class:`.Column` instance but may also be an expression, such as one derived from :func:`.cast`. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ polymorphic_map = None """A mapping of "polymorphic identity" identifiers mapped to :class:`.Mapper` instances, within an inheritance scenario. The identifiers can be of any type which is comparable to the type of column represented by :attr:`~.Mapper.polymorphic_on`. An inheritance chain of mappers will all reference the same polymorphic map object. The object is used to correlate incoming result rows to target mappers. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ polymorphic_identity = None """Represent an identifier which is matched against the :attr:`~.Mapper.polymorphic_on` column during result row loading. Used only with inheritance, this object can be of any type which is comparable to the type of column represented by :attr:`~.Mapper.polymorphic_on`. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ base_mapper = None """The base-most :class:`.Mapper` in an inheritance chain. In a non-inheriting scenario, this attribute will always be this :class:`.Mapper`. In an inheritance scenario, it references the :class:`.Mapper` which is parent to all other :class:`.Mapper` objects in the inheritance chain. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ columns = None """A collection of :class:`.Column` or other scalar expression objects maintained by this :class:`.Mapper`. The collection behaves the same as that of the ``c`` attribute on any :class:`.Table` object, except that only those columns included in this mapping are present, and are keyed based on the attribute name defined in the mapping, not necessarily the ``key`` attribute of the :class:`.Column` itself. Additionally, scalar expressions mapped by :func:`.column_property` are also present here. This is a *read only* attribute determined during mapper construction. Behavior is undefined if directly modified. """ validators = None """An immutable dictionary of attributes which have been decorated using the :func:`~.orm.validates` decorator. The dictionary contains string attribute names as keys mapped to the actual validation method. """ c = None """A synonym for :attr:`~.Mapper.columns`.""" @util.memoized_property def _path_registry(self): return PathRegistry.per_mapper(self) def _configure_inheritance(self): """Configure settings related to inherting and/or inherited mappers being present.""" # a set of all mappers which inherit from this one. self._inheriting_mappers = util.WeakSequence() if self.inherits: if isinstance(self.inherits, type): self.inherits = class_mapper(self.inherits, configure=False) if not issubclass(self.class_, self.inherits.class_): raise sa_exc.ArgumentError( "Class '%s' does not inherit from '%s'" % (self.class_.__name__, self.inherits.class_.__name__)) if self.non_primary != self.inherits.non_primary: np = not self.non_primary and "primary" or "non-primary" raise sa_exc.ArgumentError( "Inheritance of %s mapper for class '%s' is " "only allowed from a %s mapper" % (np, self.class_.__name__, np)) # inherit_condition is optional. if self.local_table is None: self.local_table = self.inherits.local_table self.mapped_table = self.inherits.mapped_table self.single = True elif not self.local_table is self.inherits.local_table: if self.concrete: self.mapped_table = self.local_table for mapper in self.iterate_to_root(): if mapper.polymorphic_on is not None: mapper._requires_row_aliasing = True else: if self.inherit_condition is None: # figure out inherit condition from our table to the # immediate table of the inherited mapper, not its # full table which could pull in other stuff we dont # want (allows test/inheritance.InheritTest4 to pass) self.inherit_condition = sql_util.join_condition( self.inherits.local_table, self.local_table) self.mapped_table = sql.join( self.inherits.mapped_table, self.local_table, self.inherit_condition) fks = util.to_set(self.inherit_foreign_keys) self._inherits_equated_pairs = sql_util.criterion_as_pairs( self.mapped_table.onclause, consider_as_foreign_keys=fks) else: self.mapped_table = self.local_table if self.polymorphic_identity is not None and not self.concrete: self._identity_class = self.inherits._identity_class else: self._identity_class = self.class_ if self.version_id_col is None: self.version_id_col = self.inherits.version_id_col self.version_id_generator = self.inherits.version_id_generator elif self.inherits.version_id_col is not None and \ self.version_id_col is not self.inherits.version_id_col: util.warn( "Inheriting version_id_col '%s' does not match inherited " "version_id_col '%s' and will not automatically populate " "the inherited versioning column. " "version_id_col should only be specified on " "the base-most mapper that includes versioning." % (self.version_id_col.description, self.inherits.version_id_col.description) ) if self.order_by is False and \ not self.concrete and \ self.inherits.order_by is not False: self.order_by = self.inherits.order_by self.polymorphic_map = self.inherits.polymorphic_map self.batch = self.inherits.batch self.inherits._inheriting_mappers.append(self) self.base_mapper = self.inherits.base_mapper self.passive_updates = self.inherits.passive_updates self._all_tables = self.inherits._all_tables if self.polymorphic_identity is not None: self.polymorphic_map[self.polymorphic_identity] = self else: self._all_tables = set() self.base_mapper = self self.mapped_table = self.local_table if self.polymorphic_identity is not None: self.polymorphic_map[self.polymorphic_identity] = self self._identity_class = self.class_ if self.mapped_table is None: raise sa_exc.ArgumentError( "Mapper '%s' does not have a mapped_table specified." % self) def _set_with_polymorphic(self, with_polymorphic): if with_polymorphic == '*': self.with_polymorphic = ('*', None) elif isinstance(with_polymorphic, (tuple, list)): if isinstance(with_polymorphic[0], util.string_types + (tuple, list)): self.with_polymorphic = with_polymorphic else: self.with_polymorphic = (with_polymorphic, None) elif with_polymorphic is not None: raise sa_exc.ArgumentError("Invalid setting for with_polymorphic") else: self.with_polymorphic = None if isinstance(self.local_table, expression.SelectBase): raise sa_exc.InvalidRequestError( "When mapping against a select() construct, map against " "an alias() of the construct instead." "This because several databases don't allow a " "SELECT from a subquery that does not have an alias." ) if self.with_polymorphic and \ isinstance(self.with_polymorphic[1], expression.SelectBase): self.with_polymorphic = (self.with_polymorphic[0], self.with_polymorphic[1].alias()) if self.configured: self._expire_memoizations() def _set_concrete_base(self, mapper): """Set the given :class:`.Mapper` as the 'inherits' for this :class:`.Mapper`, assuming this :class:`.Mapper` is concrete and does not already have an inherits.""" assert self.concrete assert not self.inherits assert isinstance(mapper, Mapper) self.inherits = mapper self.inherits.polymorphic_map.update(self.polymorphic_map) self.polymorphic_map = self.inherits.polymorphic_map for mapper in self.iterate_to_root(): if mapper.polymorphic_on is not None: mapper._requires_row_aliasing = True self.batch = self.inherits.batch for mp in self.self_and_descendants: mp.base_mapper = self.inherits.base_mapper self.inherits._inheriting_mappers.append(self) self.passive_updates = self.inherits.passive_updates self._all_tables = self.inherits._all_tables for key, prop in mapper._props.items(): if key not in self._props and \ not self._should_exclude(key, key, local=False, column=None): self._adapt_inherited_property(key, prop, False) def _set_polymorphic_on(self, polymorphic_on): self.polymorphic_on = polymorphic_on self._configure_polymorphic_setter(True) def _configure_legacy_instrument_class(self): if self.inherits: self.dispatch._update(self.inherits.dispatch) super_extensions = set( chain(*[m._deprecated_extensions for m in self.inherits.iterate_to_root()])) else: super_extensions = set() for ext in self._deprecated_extensions: if ext not in super_extensions: ext._adapt_instrument_class(self, ext) def _configure_listeners(self): if self.inherits: super_extensions = set( chain(*[m._deprecated_extensions for m in self.inherits.iterate_to_root()])) else: super_extensions = set() for ext in self._deprecated_extensions: if ext not in super_extensions: ext._adapt_listener(self, ext) def _configure_class_instrumentation(self): """If this mapper is to be a primary mapper (i.e. the non_primary flag is not set), associate this Mapper with the given class_ and entity name. Subsequent calls to ``class_mapper()`` for the class_/entity name combination will return this mapper. Also decorate the `__init__` method on the mapped class to include optional auto-session attachment logic. """ manager = attributes.manager_of_class(self.class_) if self.non_primary: if not manager or not manager.is_mapped: raise sa_exc.InvalidRequestError( "Class %s has no primary mapper configured. Configure " "a primary mapper first before setting up a non primary " "Mapper." % self.class_) self.class_manager = manager self._identity_class = manager.mapper._identity_class _mapper_registry[self] = True return if manager is not None: assert manager.class_ is self.class_ if manager.is_mapped: raise sa_exc.ArgumentError( "Class '%s' already has a primary mapper defined. " "Use non_primary=True to " "create a non primary Mapper. clear_mappers() will " "remove *all* current mappers from all classes." % self.class_) #else: # a ClassManager may already exist as # ClassManager.instrument_attribute() creates # new managers for each subclass if they don't yet exist. _mapper_registry[self] = True self.dispatch.instrument_class(self, self.class_) if manager is None: manager = instrumentation.register_class(self.class_) self.class_manager = manager manager.mapper = self manager.deferred_scalar_loader = util.partial( loading.load_scalar_attributes, self) # The remaining members can be added by any mapper, # e_name None or not. if manager.info.get(_INSTRUMENTOR, False): return event.listen(manager, 'first_init', _event_on_first_init, raw=True) event.listen(manager, 'init', _event_on_init, raw=True) event.listen(manager, 'resurrect', _event_on_resurrect, raw=True) for key, method in util.iterate_attributes(self.class_): if isinstance(method, types.FunctionType): if hasattr(method, '__sa_reconstructor__'): self._reconstructor = method event.listen(manager, 'load', _event_on_load, raw=True) elif hasattr(method, '__sa_validators__'): validation_opts = method.__sa_validation_opts__ for name in method.__sa_validators__: self.validators = self.validators.union( {name: (method, validation_opts)} ) manager.info[_INSTRUMENTOR] = self @classmethod def _configure_all(cls): """Class-level path to the :func:`.configure_mappers` call. """ configure_mappers() def dispose(self): # Disable any attribute-based compilation. self.configured = True if hasattr(self, '_configure_failed'): del self._configure_failed if not self.non_primary and \ self.class_manager is not None and \ self.class_manager.is_mapped and \ self.class_manager.mapper is self: instrumentation.unregister_class(self.class_) def _configure_pks(self): self.tables = sql_util.find_tables(self.mapped_table) self._pks_by_table = {} self._cols_by_table = {} all_cols = util.column_set(chain(*[ col.proxy_set for col in self._columntoproperty])) pk_cols = util.column_set(c for c in all_cols if c.primary_key) # identify primary key columns which are also mapped by this mapper. tables = set(self.tables + [self.mapped_table]) self._all_tables.update(tables) for t in tables: if t.primary_key and pk_cols.issuperset(t.primary_key): # ordering is important since it determines the ordering of # mapper.primary_key (and therefore query.get()) self._pks_by_table[t] = \ util.ordered_column_set(t.primary_key).\ intersection(pk_cols) self._cols_by_table[t] = \ util.ordered_column_set(t.c).\ intersection(all_cols) # determine cols that aren't expressed within our tables; mark these # as "read only" properties which are refreshed upon INSERT/UPDATE self._readonly_props = set( self._columntoproperty[col] for col in self._columntoproperty if not hasattr(col, 'table') or col.table not in self._cols_by_table) # if explicit PK argument sent, add those columns to the # primary key mappings if self._primary_key_argument: for k in self._primary_key_argument: if k.table not in self._pks_by_table: self._pks_by_table[k.table] = util.OrderedSet() self._pks_by_table[k.table].add(k) # otherwise, see that we got a full PK for the mapped table elif self.mapped_table not in self._pks_by_table or \ len(self._pks_by_table[self.mapped_table]) == 0: raise sa_exc.ArgumentError( "Mapper %s could not assemble any primary " "key columns for mapped table '%s'" % (self, self.mapped_table.description)) elif self.local_table not in self._pks_by_table and \ isinstance(self.local_table, schema.Table): util.warn("Could not assemble any primary " "keys for locally mapped table '%s' - " "no rows will be persisted in this Table." % self.local_table.description) if self.inherits and \ not self.concrete and \ not self._primary_key_argument: # if inheriting, the "primary key" for this mapper is # that of the inheriting (unless concrete or explicit) self.primary_key = self.inherits.primary_key else: # determine primary key from argument or mapped_table pks - # reduce to the minimal set of columns if self._primary_key_argument: primary_key = sql_util.reduce_columns( [self.mapped_table.corresponding_column(c) for c in self._primary_key_argument], ignore_nonexistent_tables=True) else: primary_key = sql_util.reduce_columns( self._pks_by_table[self.mapped_table], ignore_nonexistent_tables=True) if len(primary_key) == 0: raise sa_exc.ArgumentError( "Mapper %s could not assemble any primary " "key columns for mapped table '%s'" % (self, self.mapped_table.description)) self.primary_key = tuple(primary_key) self._log("Identified primary key columns: %s", primary_key) def _configure_properties(self): # Column and other ClauseElement objects which are mapped self.columns = self.c = util.OrderedProperties() # object attribute names mapped to MapperProperty objects self._props = util.OrderedDict() # table columns mapped to lists of MapperProperty objects # using a list allows a single column to be defined as # populating multiple object attributes self._columntoproperty = _ColumnMapping(self) # load custom properties if self._init_properties: for key, prop in self._init_properties.items(): self._configure_property(key, prop, False) # pull properties from the inherited mapper if any. if self.inherits: for key, prop in self.inherits._props.items(): if key not in self._props and \ not self._should_exclude(key, key, local=False, column=None): self._adapt_inherited_property(key, prop, False) # create properties for each column in the mapped table, # for those columns which don't already map to a property for column in self.mapped_table.columns: if column in self._columntoproperty: continue column_key = (self.column_prefix or '') + column.key if self._should_exclude( column.key, column_key, local=self.local_table.c.contains_column(column), column=column ): continue # adjust the "key" used for this column to that # of the inheriting mapper for mapper in self.iterate_to_root(): if column in mapper._columntoproperty: column_key = mapper._columntoproperty[column].key self._configure_property(column_key, column, init=False, setparent=True) def _configure_polymorphic_setter(self, init=False): """Configure an attribute on the mapper representing the 'polymorphic_on' column, if applicable, and not already generated by _configure_properties (which is typical). Also create a setter function which will assign this attribute to the value of the 'polymorphic_identity' upon instance construction, also if applicable. This routine will run when an instance is created. """ setter = False if self.polymorphic_on is not None: setter = True if isinstance(self.polymorphic_on, util.string_types): # polymorphic_on specified as as string - link # it to mapped ColumnProperty try: self.polymorphic_on = self._props[self.polymorphic_on] except KeyError: raise sa_exc.ArgumentError( "Can't determine polymorphic_on " "value '%s' - no attribute is " "mapped to this name." % self.polymorphic_on) if self.polymorphic_on in self._columntoproperty: # polymorphic_on is a column that is already mapped # to a ColumnProperty prop = self._columntoproperty[self.polymorphic_on] polymorphic_key = prop.key self.polymorphic_on = prop.columns[0] polymorphic_key = prop.key elif isinstance(self.polymorphic_on, MapperProperty): # polymorphic_on is directly a MapperProperty, # ensure it's a ColumnProperty if not isinstance(self.polymorphic_on, properties.ColumnProperty): raise sa_exc.ArgumentError( "Only direct column-mapped " "property or SQL expression " "can be passed for polymorphic_on") prop = self.polymorphic_on self.polymorphic_on = prop.columns[0] polymorphic_key = prop.key elif not expression._is_column(self.polymorphic_on): # polymorphic_on is not a Column and not a ColumnProperty; # not supported right now. raise sa_exc.ArgumentError( "Only direct column-mapped " "property or SQL expression " "can be passed for polymorphic_on" ) else: # polymorphic_on is a Column or SQL expression and # doesn't appear to be mapped. this means it can be 1. # only present in the with_polymorphic selectable or # 2. a totally standalone SQL expression which we'd # hope is compatible with this mapper's mapped_table col = self.mapped_table.corresponding_column( self.polymorphic_on) if col is None: # polymorphic_on doesn't derive from any # column/expression isn't present in the mapped # table. we will make a "hidden" ColumnProperty # for it. Just check that if it's directly a # schema.Column and we have with_polymorphic, it's # likely a user error if the schema.Column isn't # represented somehow in either mapped_table or # with_polymorphic. Otherwise as of 0.7.4 we # just go with it and assume the user wants it # that way (i.e. a CASE statement) setter = False instrument = False col = self.polymorphic_on if isinstance(col, schema.Column) and ( self.with_polymorphic is None or \ self.with_polymorphic[1].\ corresponding_column(col) is None ): raise sa_exc.InvalidRequestError( "Could not map polymorphic_on column " "'%s' to the mapped table - polymorphic " "loads will not function properly" % col.description) else: # column/expression that polymorphic_on derives from # is present in our mapped table # and is probably mapped, but polymorphic_on itself # is not. This happens when # the polymorphic_on is only directly present in the # with_polymorphic selectable, as when use # polymorphic_union. # we'll make a separate ColumnProperty for it. instrument = True key = getattr(col, 'key', None) if key: if self._should_exclude(col.key, col.key, False, col): raise sa_exc.InvalidRequestError( "Cannot exclude or override the " "discriminator column %r" % col.key) else: self.polymorphic_on = col = \ col.label("_sa_polymorphic_on") key = col.key self._configure_property( key, properties.ColumnProperty(col, _instrument=instrument), init=init, setparent=True) polymorphic_key = key else: # no polymorphic_on was set. # check inheriting mappers for one. for mapper in self.iterate_to_root(): # determine if polymorphic_on of the parent # should be propagated here. If the col # is present in our mapped table, or if our mapped # table is the same as the parent (i.e. single table # inheritance), we can use it if mapper.polymorphic_on is not None: if self.mapped_table is mapper.mapped_table: self.polymorphic_on = mapper.polymorphic_on else: self.polymorphic_on = \ self.mapped_table.corresponding_column( mapper.polymorphic_on) # we can use the parent mapper's _set_polymorphic_identity # directly; it ensures the polymorphic_identity of the # instance's mapper is used so is portable to subclasses. if self.polymorphic_on is not None: self._set_polymorphic_identity = \ mapper._set_polymorphic_identity self._validate_polymorphic_identity = \ mapper._validate_polymorphic_identity else: self._set_polymorphic_identity = None return if setter: def _set_polymorphic_identity(state): dict_ = state.dict state.get_impl(polymorphic_key).set(state, dict_, state.manager.mapper.polymorphic_identity, None) def _validate_polymorphic_identity(mapper, state, dict_): if polymorphic_key in dict_ and \ dict_[polymorphic_key] not in \ mapper._acceptable_polymorphic_identities: util.warn( "Flushing object %s with " "incompatible polymorphic identity %r; the " "object may not refresh and/or load correctly" % ( state_str(state), dict_[polymorphic_key] ) ) self._set_polymorphic_identity = _set_polymorphic_identity self._validate_polymorphic_identity = _validate_polymorphic_identity else: self._set_polymorphic_identity = None _validate_polymorphic_identity = None @_memoized_configured_property def _version_id_prop(self): if self.version_id_col is not None: return self._columntoproperty[self.version_id_col] else: return None @_memoized_configured_property def _acceptable_polymorphic_identities(self): identities = set() stack = deque([self]) while stack: item = stack.popleft() if item.mapped_table is self.mapped_table: identities.add(item.polymorphic_identity) stack.extend(item._inheriting_mappers) return identities def _adapt_inherited_property(self, key, prop, init): if not self.concrete: self._configure_property(key, prop, init=False, setparent=False) elif key not in self._props: self._configure_property( key, properties.ConcreteInheritedProperty(), init=init, setparent=True) def _configure_property(self, key, prop, init=True, setparent=True): self._log("_configure_property(%s, %s)", key, prop.__class__.__name__) if not isinstance(prop, MapperProperty): prop = self._property_from_column(key, prop) if isinstance(prop, properties.ColumnProperty): col = self.mapped_table.corresponding_column(prop.columns[0]) # if the column is not present in the mapped table, # test if a column has been added after the fact to the # parent table (or their parent, etc.) [ticket:1570] if col is None and self.inherits: path = [self] for m in self.inherits.iterate_to_root(): col = m.local_table.corresponding_column(prop.columns[0]) if col is not None: for m2 in path: m2.mapped_table._reset_exported() col = self.mapped_table.corresponding_column( prop.columns[0]) break path.append(m) # subquery expression, column not present in the mapped # selectable. if col is None: col = prop.columns[0] # column is coming in after _readonly_props was # initialized; check for 'readonly' if hasattr(self, '_readonly_props') and \ (not hasattr(col, 'table') or col.table not in self._cols_by_table): self._readonly_props.add(prop) else: # if column is coming in after _cols_by_table was # initialized, ensure the col is in the right set if hasattr(self, '_cols_by_table') and \ col.table in self._cols_by_table and \ col not in self._cols_by_table[col.table]: self._cols_by_table[col.table].add(col) # if this properties.ColumnProperty represents the "polymorphic # discriminator" column, mark it. We'll need this when rendering # columns in SELECT statements. if not hasattr(prop, '_is_polymorphic_discriminator'): prop._is_polymorphic_discriminator = \ (col is self.polymorphic_on or prop.columns[0] is self.polymorphic_on) self.columns[key] = col for col in prop.columns + prop._orig_columns: for col in col.proxy_set: self._columntoproperty[col] = prop prop.key = key if setparent: prop.set_parent(self, init) if key in self._props and \ getattr(self._props[key], '_mapped_by_synonym', False): syn = self._props[key]._mapped_by_synonym raise sa_exc.ArgumentError( "Can't call map_column=True for synonym %r=%r, " "a ColumnProperty already exists keyed to the name " "%r for column %r" % (syn, key, key, syn) ) if key in self._props and \ not isinstance(prop, properties.ColumnProperty) and \ not isinstance(self._props[key], properties.ColumnProperty): util.warn("Property %s on %s being replaced with new " "property %s; the old property will be discarded" % ( self._props[key], self, prop, )) self._props[key] = prop if not self.non_primary: prop.instrument_class(self) for mapper in self._inheriting_mappers: mapper._adapt_inherited_property(key, prop, init) if init: prop.init() prop.post_instrument_class(self) if self.configured: self._expire_memoizations() def _property_from_column(self, key, prop): """generate/update a :class:`.ColumnProprerty` given a :class:`.Column` object. """ # we were passed a Column or a list of Columns; # generate a properties.ColumnProperty columns = util.to_list(prop) column = columns[0] if not expression._is_column(column): raise sa_exc.ArgumentError( "%s=%r is not an instance of MapperProperty or Column" % (key, prop)) prop = self._props.get(key, None) if isinstance(prop, properties.ColumnProperty): if prop.parent is self: raise sa_exc.InvalidRequestError( "Implicitly combining column %s with column " "%s under attribute '%s'. Please configure one " "or more attributes for these same-named columns " "explicitly." % (prop.columns[-1], column, key)) # existing properties.ColumnProperty from an inheriting # mapper. make a copy and append our column to it prop = prop.copy() prop.columns.insert(0, column) self._log("inserting column to existing list " "in properties.ColumnProperty %s" % (key)) return prop elif prop is None or isinstance(prop, properties.ConcreteInheritedProperty): mapped_column = [] for c in columns: mc = self.mapped_table.corresponding_column(c) if mc is None: mc = self.local_table.corresponding_column(c) if mc is not None: # if the column is in the local table but not the # mapped table, this corresponds to adding a # column after the fact to the local table. # [ticket:1523] self.mapped_table._reset_exported() mc = self.mapped_table.corresponding_column(c) if mc is None: raise sa_exc.ArgumentError( "When configuring property '%s' on %s, " "column '%s' is not represented in the mapper's " "table. Use the `column_property()` function to " "force this column to be mapped as a read-only " "attribute." % (key, self, c)) mapped_column.append(mc) return properties.ColumnProperty(*mapped_column) else: raise sa_exc.ArgumentError( "WARNING: when configuring property '%s' on %s, " "column '%s' conflicts with property '%r'. " "To resolve this, map the column to the class under a " "different name in the 'properties' dictionary. Or, " "to remove all awareness of the column entirely " "(including its availability as a foreign key), " "use the 'include_properties' or 'exclude_properties' " "mapper arguments to control specifically which table " "columns get mapped." % (key, self, column.key, prop)) def _post_configure_properties(self): """Call the ``init()`` method on all ``MapperProperties`` attached to this mapper. This is a deferred configuration step which is intended to execute once all mappers have been constructed. """ self._log("_post_configure_properties() started") l = [(key, prop) for key, prop in self._props.items()] for key, prop in l: self._log("initialize prop %s", key) if prop.parent is self and not prop._configure_started: prop.init() if prop._configure_finished: prop.post_instrument_class(self) self._log("_post_configure_properties() complete") self.configured = True def add_properties(self, dict_of_properties): """Add the given dictionary of properties to this mapper, using `add_property`. """ for key, value in dict_of_properties.items(): self.add_property(key, value) def add_property(self, key, prop): """Add an individual MapperProperty to this mapper. If the mapper has not been configured yet, just adds the property to the initial properties dictionary sent to the constructor. If this Mapper has already been configured, then the given MapperProperty is configured immediately. """ self._init_properties[key] = prop self._configure_property(key, prop, init=self.configured) def _expire_memoizations(self): for mapper in self.iterate_to_root(): _memoized_configured_property.expire_instance(mapper) @property def _log_desc(self): return "(" + self.class_.__name__ + \ "|" + \ (self.local_table is not None and self.local_table.description or str(self.local_table)) +\ (self.non_primary and "|non-primary" or "") + ")" def _log(self, msg, *args): self.logger.info( "%s " + msg, *((self._log_desc,) + args) ) def _log_debug(self, msg, *args): self.logger.debug( "%s " + msg, *((self._log_desc,) + args) ) def __repr__(self): return '<Mapper at 0x%x; %s>' % ( id(self), self.class_.__name__) def __str__(self): return "Mapper|%s|%s%s" % ( self.class_.__name__, self.local_table is not None and self.local_table.description or None, self.non_primary and "|non-primary" or "" ) def _is_orphan(self, state): orphan_possible = False for mapper in self.iterate_to_root(): for (key, cls) in mapper._delete_orphans: orphan_possible = True has_parent = attributes.manager_of_class(cls).has_parent( state, key, optimistic=state.has_identity) if self.legacy_is_orphan and has_parent: return False elif not self.legacy_is_orphan and not has_parent: return True if self.legacy_is_orphan: return orphan_possible else: return False def has_property(self, key): return key in self._props def get_property(self, key, _configure_mappers=True): """return a MapperProperty associated with the given key. """ if _configure_mappers and Mapper._new_mappers: configure_mappers() try: return self._props[key] except KeyError: raise sa_exc.InvalidRequestError( "Mapper '%s' has no property '%s'" % (self, key)) def get_property_by_column(self, column): """Given a :class:`.Column` object, return the :class:`.MapperProperty` which maps this column.""" return self._columntoproperty[column] @property def iterate_properties(self): """return an iterator of all MapperProperty objects.""" if Mapper._new_mappers: configure_mappers() return iter(self._props.values()) def _mappers_from_spec(self, spec, selectable): """given a with_polymorphic() argument, return the set of mappers it represents. Trims the list of mappers to just those represented within the given selectable, if present. This helps some more legacy-ish mappings. """ if spec == '*': mappers = list(self.self_and_descendants) elif spec: mappers = set() for m in util.to_list(spec): m = _class_to_mapper(m) if not m.isa(self): raise sa_exc.InvalidRequestError( "%r does not inherit from %r" % (m, self)) if selectable is None: mappers.update(m.iterate_to_root()) else: mappers.add(m) mappers = [m for m in self.self_and_descendants if m in mappers] else: mappers = [] if selectable is not None: tables = set(sql_util.find_tables(selectable, include_aliases=True)) mappers = [m for m in mappers if m.local_table in tables] return mappers def _selectable_from_mappers(self, mappers, innerjoin): """given a list of mappers (assumed to be within this mapper's inheritance hierarchy), construct an outerjoin amongst those mapper's mapped tables. """ from_obj = self.mapped_table for m in mappers: if m is self: continue if m.concrete: raise sa_exc.InvalidRequestError( "'with_polymorphic()' requires 'selectable' argument " "when concrete-inheriting mappers are used.") elif not m.single: if innerjoin: from_obj = from_obj.join(m.local_table, m.inherit_condition) else: from_obj = from_obj.outerjoin(m.local_table, m.inherit_condition) return from_obj @_memoized_configured_property def _single_table_criterion(self): if self.single and \ self.inherits and \ self.polymorphic_on is not None: return self.polymorphic_on.in_( m.polymorphic_identity for m in self.self_and_descendants) else: return None @_memoized_configured_property def _with_polymorphic_mappers(self): if Mapper._new_mappers: configure_mappers() if not self.with_polymorphic: return [] return self._mappers_from_spec(*self.with_polymorphic) @_memoized_configured_property def _with_polymorphic_selectable(self): if not self.with_polymorphic: return self.mapped_table spec, selectable = self.with_polymorphic if selectable is not None: return selectable else: return self._selectable_from_mappers( self._mappers_from_spec(spec, selectable), False) with_polymorphic_mappers = _with_polymorphic_mappers """The list of :class:`.Mapper` objects included in the default "polymorphic" query. """ @property def selectable(self): """The :func:`.select` construct this :class:`.Mapper` selects from by default. Normally, this is equivalent to :attr:`.mapped_table`, unless the ``with_polymorphic`` feature is in use, in which case the full "polymorphic" selectable is returned. """ return self._with_polymorphic_selectable def _with_polymorphic_args(self, spec=None, selectable=False, innerjoin=False): if self.with_polymorphic: if not spec: spec = self.with_polymorphic[0] if selectable is False: selectable = self.with_polymorphic[1] elif selectable is False: selectable = None mappers = self._mappers_from_spec(spec, selectable) if selectable is not None: return mappers, selectable else: return mappers, self._selectable_from_mappers(mappers, innerjoin) @_memoized_configured_property def _polymorphic_properties(self): return list(self._iterate_polymorphic_properties( self._with_polymorphic_mappers)) def _iterate_polymorphic_properties(self, mappers=None): """Return an iterator of MapperProperty objects which will render into a SELECT.""" if mappers is None: mappers = self._with_polymorphic_mappers if not mappers: for c in self.iterate_properties: yield c else: # in the polymorphic case, filter out discriminator columns # from other mappers, as these are sometimes dependent on that # mapper's polymorphic selectable (which we don't want rendered) for c in util.unique_list( chain(*[ list(mapper.iterate_properties) for mapper in [self] + mappers ]) ): if getattr(c, '_is_polymorphic_discriminator', False) and \ (self.polymorphic_on is None or c.columns[0] is not self.polymorphic_on): continue yield c @util.memoized_property def attrs(self): """A namespace of all :class:`.MapperProperty` objects associated this mapper. This is an object that provides each property based on its key name. For instance, the mapper for a ``User`` class which has ``User.name`` attribute would provide ``mapper.attrs.name``, which would be the :class:`.ColumnProperty` representing the ``name`` column. The namespace object can also be iterated, which would yield each :class:`.MapperProperty`. :class:`.Mapper` has several pre-filtered views of this attribute which limit the types of properties returned, inclding :attr:`.synonyms`, :attr:`.column_attrs`, :attr:`.relationships`, and :attr:`.composites`. .. seealso:: :attr:`.Mapper.all_orm_descriptors` """ if Mapper._new_mappers: configure_mappers() return util.ImmutableProperties(self._props) @util.memoized_property def all_orm_descriptors(self): """A namespace of all :class:`._InspectionAttr` attributes associated with the mapped class. These attributes are in all cases Python :term:`descriptors` associated with the mapped class or its superclasses. This namespace includes attributes that are mapped to the class as well as attributes declared by extension modules. It includes any Python descriptor type that inherits from :class:`._InspectionAttr`. This includes :class:`.QueryableAttribute`, as well as extension types such as :class:`.hybrid_property`, :class:`.hybrid_method` and :class:`.AssociationProxy`. To distinguish between mapped attributes and extension attributes, the attribute :attr:`._InspectionAttr.extension_type` will refer to a constant that distinguishes between different extension types. When dealing with a :class:`.QueryableAttribute`, the :attr:`.QueryableAttribute.property` attribute refers to the :class:`.MapperProperty` property, which is what you get when referring to the collection of mapped properties via :attr:`.Mapper.attrs`. .. versionadded:: 0.8.0 .. seealso:: :attr:`.Mapper.attrs` """ return util.ImmutableProperties( dict(self.class_manager._all_sqla_attributes())) @_memoized_configured_property def synonyms(self): """Return a namespace of all :class:`.SynonymProperty` properties maintained by this :class:`.Mapper`. .. seealso:: :attr:`.Mapper.attrs` - namespace of all :class:`.MapperProperty` objects. """ return self._filter_properties(properties.SynonymProperty) @_memoized_configured_property def column_attrs(self): """Return a namespace of all :class:`.ColumnProperty` properties maintained by this :class:`.Mapper`. .. seealso:: :attr:`.Mapper.attrs` - namespace of all :class:`.MapperProperty` objects. """ return self._filter_properties(properties.ColumnProperty) @_memoized_configured_property def relationships(self): """Return a namespace of all :class:`.RelationshipProperty` properties maintained by this :class:`.Mapper`. .. seealso:: :attr:`.Mapper.attrs` - namespace of all :class:`.MapperProperty` objects. """ return self._filter_properties(properties.RelationshipProperty) @_memoized_configured_property def composites(self): """Return a namespace of all :class:`.CompositeProperty` properties maintained by this :class:`.Mapper`. .. seealso:: :attr:`.Mapper.attrs` - namespace of all :class:`.MapperProperty` objects. """ return self._filter_properties(properties.CompositeProperty) def _filter_properties(self, type_): if Mapper._new_mappers: configure_mappers() return util.ImmutableProperties(util.OrderedDict( (k, v) for k, v in self._props.items() if isinstance(v, type_) )) @_memoized_configured_property def _get_clause(self): """create a "get clause" based on the primary key. this is used by query.get() and many-to-one lazyloads to load this item by primary key. """ params = [(primary_key, sql.bindparam(None, type_=primary_key.type)) for primary_key in self.primary_key] return sql.and_(*[k == v for (k, v) in params]), \ util.column_dict(params) @_memoized_configured_property def _equivalent_columns(self): """Create a map of all *equivalent* columns, based on the determination of column pairs that are equated to one another based on inherit condition. This is designed to work with the queries that util.polymorphic_union comes up with, which often don't include the columns from the base table directly (including the subclass table columns only). The resulting structure is a dictionary of columns mapped to lists of equivalent columns, i.e. { tablea.col1: set([tableb.col1, tablec.col1]), tablea.col2: set([tabled.col2]) } """ result = util.column_dict() def visit_binary(binary): if binary.operator == operators.eq: if binary.left in result: result[binary.left].add(binary.right) else: result[binary.left] = util.column_set((binary.right,)) if binary.right in result: result[binary.right].add(binary.left) else: result[binary.right] = util.column_set((binary.left,)) for mapper in self.base_mapper.self_and_descendants: if mapper.inherit_condition is not None: visitors.traverse( mapper.inherit_condition, {}, {'binary': visit_binary}) return result def _is_userland_descriptor(self, obj): if isinstance(obj, (_MappedAttribute, instrumentation.ClassManager, expression.ColumnElement)): return False else: return True def _should_exclude(self, name, assigned_name, local, column): """determine whether a particular property should be implicitly present on the class. This occurs when properties are propagated from an inherited class, or are applied from the columns present in the mapped table. """ # check for class-bound attributes and/or descriptors, # either local or from an inherited class if local: if self.class_.__dict__.get(assigned_name, None) is not None \ and self._is_userland_descriptor( self.class_.__dict__[assigned_name]): return True else: if getattr(self.class_, assigned_name, None) is not None \ and self._is_userland_descriptor( getattr(self.class_, assigned_name)): return True if self.include_properties is not None and \ name not in self.include_properties and \ (column is None or column not in self.include_properties): self._log("not including property %s" % (name)) return True if self.exclude_properties is not None and \ ( name in self.exclude_properties or \ (column is not None and column in self.exclude_properties) ): self._log("excluding property %s" % (name)) return True return False def common_parent(self, other): """Return true if the given mapper shares a common inherited parent as this mapper.""" return self.base_mapper is other.base_mapper def _canload(self, state, allow_subtypes): s = self.primary_mapper() if self.polymorphic_on is not None or allow_subtypes: return _state_mapper(state).isa(s) else: return _state_mapper(state) is s def isa(self, other): """Return True if the this mapper inherits from the given mapper.""" m = self while m and m is not other: m = m.inherits return bool(m) def iterate_to_root(self): m = self while m: yield m m = m.inherits @_memoized_configured_property def self_and_descendants(self): """The collection including this mapper and all descendant mappers. This includes not just the immediately inheriting mappers but all their inheriting mappers as well. """ descendants = [] stack = deque([self]) while stack: item = stack.popleft() descendants.append(item) stack.extend(item._inheriting_mappers) return util.WeakSequence(descendants) def polymorphic_iterator(self): """Iterate through the collection including this mapper and all descendant mappers. This includes not just the immediately inheriting mappers but all their inheriting mappers as well. To iterate through an entire hierarchy, use ``mapper.base_mapper.polymorphic_iterator()``. """ return iter(self.self_and_descendants) def primary_mapper(self): """Return the primary mapper corresponding to this mapper's class key (class).""" return self.class_manager.mapper @property def primary_base_mapper(self): return self.class_manager.mapper.base_mapper def identity_key_from_row(self, row, adapter=None): """Return an identity-map key for use in storing/retrieving an item from the identity map. :param row: A :class:`.RowProxy` instance. The columns which are mapped by this :class:`.Mapper` should be locatable in the row, preferably via the :class:`.Column` object directly (as is the case when a :func:`.select` construct is executed), or via string names of the form ``<tablename>_<colname>``. """ pk_cols = self.primary_key if adapter: pk_cols = [adapter.columns[c] for c in pk_cols] return self._identity_class, \ tuple(row[column] for column in pk_cols) def identity_key_from_primary_key(self, primary_key): """Return an identity-map key for use in storing/retrieving an item from an identity map. :param primary_key: A list of values indicating the identifier. """ return self._identity_class, tuple(primary_key) def identity_key_from_instance(self, instance): """Return the identity key for the given instance, based on its primary key attributes. If the instance's state is expired, calling this method will result in a database check to see if the object has been deleted. If the row no longer exists, :class:`~sqlalchemy.orm.exc.ObjectDeletedError` is raised. This value is typically also found on the instance state under the attribute name `key`. """ return self.identity_key_from_primary_key( self.primary_key_from_instance(instance)) def _identity_key_from_state(self, state): dict_ = state.dict manager = state.manager return self._identity_class, tuple([ manager[self._columntoproperty[col].key].\ impl.get(state, dict_, attributes.PASSIVE_OFF) for col in self.primary_key ]) def primary_key_from_instance(self, instance): """Return the list of primary key values for the given instance. If the instance's state is expired, calling this method will result in a database check to see if the object has been deleted. If the row no longer exists, :class:`~sqlalchemy.orm.exc.ObjectDeletedError` is raised. """ state = attributes.instance_state(instance) return self._primary_key_from_state(state) def _primary_key_from_state(self, state): dict_ = state.dict manager = state.manager return [ manager[self._columntoproperty[col].key].\ impl.get(state, dict_, attributes.PASSIVE_OFF) for col in self.primary_key ] def _get_state_attr_by_column(self, state, dict_, column, passive=attributes.PASSIVE_OFF): prop = self._columntoproperty[column] return state.manager[prop.key].impl.get(state, dict_, passive=passive) def _set_state_attr_by_column(self, state, dict_, column, value): prop = self._columntoproperty[column] state.manager[prop.key].impl.set(state, dict_, value, None) def _get_committed_attr_by_column(self, obj, column): state = attributes.instance_state(obj) dict_ = attributes.instance_dict(obj) return self._get_committed_state_attr_by_column(state, dict_, column) def _get_committed_state_attr_by_column(self, state, dict_, column, passive=attributes.PASSIVE_OFF): prop = self._columntoproperty[column] return state.manager[prop.key].impl.\ get_committed_value(state, dict_, passive=passive) def _optimized_get_statement(self, state, attribute_names): """assemble a WHERE clause which retrieves a given state by primary key, using a minimized set of tables. Applies to a joined-table inheritance mapper where the requested attribute names are only present on joined tables, not the base table. The WHERE clause attempts to include only those tables to minimize joins. """ props = self._props tables = set(chain( *[sql_util.find_tables(c, check_columns=True) for key in attribute_names for c in props[key].columns] )) if self.base_mapper.local_table in tables: return None class ColumnsNotAvailable(Exception): pass def visit_binary(binary): leftcol = binary.left rightcol = binary.right if leftcol is None or rightcol is None: return if leftcol.table not in tables: leftval = self._get_committed_state_attr_by_column( state, state.dict, leftcol, passive=attributes.PASSIVE_NO_INITIALIZE) if leftval is attributes.PASSIVE_NO_RESULT or leftval is None: raise ColumnsNotAvailable() binary.left = sql.bindparam(None, leftval, type_=binary.right.type) elif rightcol.table not in tables: rightval = self._get_committed_state_attr_by_column( state, state.dict, rightcol, passive=attributes.PASSIVE_NO_INITIALIZE) if rightval is attributes.PASSIVE_NO_RESULT or \ rightval is None: raise ColumnsNotAvailable() binary.right = sql.bindparam(None, rightval, type_=binary.right.type) allconds = [] try: start = False for mapper in reversed(list(self.iterate_to_root())): if mapper.local_table in tables: start = True elif not isinstance(mapper.local_table, expression.TableClause): return None if start and not mapper.single: allconds.append(visitors.cloned_traverse( mapper.inherit_condition, {}, {'binary': visit_binary} ) ) except ColumnsNotAvailable: return None cond = sql.and_(*allconds) cols = [] for key in attribute_names: cols.extend(props[key].columns) return sql.select(cols, cond, use_labels=True) def cascade_iterator(self, type_, state, halt_on=None): """Iterate each element and its mapper in an object graph, for all relationships that meet the given cascade rule. :param type_: The name of the cascade rule (i.e. save-update, delete, etc.) :param state: The lead InstanceState. child items will be processed per the relationships defined for this object's mapper. the return value are object instances; this provides a strong reference so that they don't fall out of scope immediately. """ visited_states = set() prp, mpp = object(), object() visitables = deque([(deque(self._props.values()), prp, state, state.dict)]) while visitables: iterator, item_type, parent_state, parent_dict = visitables[-1] if not iterator: visitables.pop() continue if item_type is prp: prop = iterator.popleft() if type_ not in prop.cascade: continue queue = deque(prop.cascade_iterator(type_, parent_state, parent_dict, visited_states, halt_on)) if queue: visitables.append((queue, mpp, None, None)) elif item_type is mpp: instance, instance_mapper, corresponding_state, \ corresponding_dict = iterator.popleft() yield instance, instance_mapper, \ corresponding_state, corresponding_dict visitables.append((deque(instance_mapper._props.values()), prp, corresponding_state, corresponding_dict)) @_memoized_configured_property def _compiled_cache(self): return util.LRUCache(self._compiled_cache_size) @_memoized_configured_property def _sorted_tables(self): table_to_mapper = {} for mapper in self.base_mapper.self_and_descendants: for t in mapper.tables: table_to_mapper.setdefault(t, mapper) extra_dependencies = [] for table, mapper in table_to_mapper.items(): super_ = mapper.inherits if super_: extra_dependencies.extend([ (super_table, table) for super_table in super_.tables ]) def skip(fk): # attempt to skip dependencies that are not # significant to the inheritance chain # for two tables that are related by inheritance. # while that dependency may be important, it's techinically # not what we mean to sort on here. parent = table_to_mapper.get(fk.parent.table) dep = table_to_mapper.get(fk.column.table) if parent is not None and \ dep is not None and \ dep is not parent and \ dep.inherit_condition is not None: cols = set(sql_util._find_columns(dep.inherit_condition)) if parent.inherit_condition is not None: cols = cols.union(sql_util._find_columns( parent.inherit_condition)) return fk.parent not in cols and fk.column not in cols else: return fk.parent not in cols return False sorted_ = sql_util.sort_tables(table_to_mapper, skip_fn=skip, extra_dependencies=extra_dependencies) ret = util.OrderedDict() for t in sorted_: ret[t] = table_to_mapper[t] return ret def _memo(self, key, callable_): if key in self._memoized_values: return self._memoized_values[key] else: self._memoized_values[key] = value = callable_() return value @util.memoized_property def _table_to_equated(self): """memoized map of tables to collections of columns to be synchronized upwards to the base mapper.""" result = util.defaultdict(list) for table in self._sorted_tables: cols = set(table.c) for m in self.iterate_to_root(): if m._inherits_equated_pairs and \ cols.intersection( util.reduce(set.union, [l.proxy_set for l, r in m._inherits_equated_pairs]) ): result[table].append((m, m._inherits_equated_pairs)) return result def configure_mappers(): """Initialize the inter-mapper relationships of all mappers that have been constructed thus far. This function can be called any number of times, but in most cases is handled internally. """ if not Mapper._new_mappers: return _CONFIGURE_MUTEX.acquire() try: global _already_compiling if _already_compiling: return _already_compiling = True try: # double-check inside mutex if not Mapper._new_mappers: return Mapper.dispatch(Mapper).before_configured() # initialize properties on all mappers # note that _mapper_registry is unordered, which # may randomly conceal/reveal issues related to # the order of mapper compilation for mapper in list(_mapper_registry): if getattr(mapper, '_configure_failed', False): e = sa_exc.InvalidRequestError( "One or more mappers failed to initialize - " "can't proceed with initialization of other " "mappers. Original exception was: %s" % mapper._configure_failed) e._configure_failed = mapper._configure_failed raise e if not mapper.configured: try: mapper._post_configure_properties() mapper._expire_memoizations() mapper.dispatch.mapper_configured( mapper, mapper.class_) except: exc = sys.exc_info()[1] if not hasattr(exc, '_configure_failed'): mapper._configure_failed = exc raise Mapper._new_mappers = False finally: _already_compiling = False finally: _CONFIGURE_MUTEX.release() Mapper.dispatch(Mapper).after_configured() def reconstructor(fn): """Decorate a method as the 'reconstructor' hook. Designates a method as the "reconstructor", an ``__init__``-like method that will be called by the ORM after the instance has been loaded from the database or otherwise reconstituted. The reconstructor will be invoked with no arguments. Scalar (non-collection) database-mapped attributes of the instance will be available for use within the function. Eagerly-loaded collections are generally not yet available and will usually only contain the first element. ORM state changes made to objects at this stage will not be recorded for the next flush() operation, so the activity within a reconstructor should be conservative. """ fn.__sa_reconstructor__ = True return fn def validates(*names, **kw): """Decorate a method as a 'validator' for one or more named properties. Designates a method as a validator, a method which receives the name of the attribute as well as a value to be assigned, or in the case of a collection, the value to be added to the collection. The function can then raise validation exceptions to halt the process from continuing (where Python's built-in ``ValueError`` and ``AssertionError`` exceptions are reasonable choices), or can modify or replace the value before proceeding. The function should otherwise return the given value. Note that a validator for a collection **cannot** issue a load of that collection within the validation routine - this usage raises an assertion to avoid recursion overflows. This is a reentrant condition which is not supported. :param \*names: list of attribute names to be validated. :param include_removes: if True, "remove" events will be sent as well - the validation function must accept an additional argument "is_remove" which will be a boolean. .. versionadded:: 0.7.7 :param include_backrefs: defaults to ``True``; if ``False``, the validation function will not emit if the originator is an attribute event related via a backref. This can be used for bi-directional :func:`.validates` usage where only one validator should emit per attribute operation. .. versionadded:: 0.9.0 .. seealso:: :ref:`simple_validators` - usage examples for :func:`.validates` """ include_removes = kw.pop('include_removes', False) include_backrefs = kw.pop('include_backrefs', True) def wrap(fn): fn.__sa_validators__ = names fn.__sa_validation_opts__ = { "include_removes": include_removes, "include_backrefs": include_backrefs } return fn return wrap def _event_on_load(state, ctx): instrumenting_mapper = state.manager.info[_INSTRUMENTOR] if instrumenting_mapper._reconstructor: instrumenting_mapper._reconstructor(state.obj()) def _event_on_first_init(manager, cls): """Initial mapper compilation trigger. instrumentation calls this one when InstanceState is first generated, and is needed for legacy mutable attributes to work. """ instrumenting_mapper = manager.info.get(_INSTRUMENTOR) if instrumenting_mapper: if Mapper._new_mappers: configure_mappers() def _event_on_init(state, args, kwargs): """Run init_instance hooks. This also includes mapper compilation, normally not needed here but helps with some piecemeal configuration scenarios (such as in the ORM tutorial). """ instrumenting_mapper = state.manager.info.get(_INSTRUMENTOR) if instrumenting_mapper: if Mapper._new_mappers: configure_mappers() if instrumenting_mapper._set_polymorphic_identity: instrumenting_mapper._set_polymorphic_identity(state) def _event_on_resurrect(state): # re-populate the primary key elements # of the dict based on the mapping. instrumenting_mapper = state.manager.info.get(_INSTRUMENTOR) if instrumenting_mapper: for col, val in zip(instrumenting_mapper.primary_key, state.key[1]): instrumenting_mapper._set_state_attr_by_column( state, state.dict, col, val) class _ColumnMapping(dict): """Error reporting helper for mapper._columntoproperty.""" def __init__(self, mapper): self.mapper = mapper def __missing__(self, column): prop = self.mapper._props.get(column) if prop: raise orm_exc.UnmappedColumnError( "Column '%s.%s' is not available, due to " "conflicting property '%s':%r" % ( column.table.name, column.name, column.key, prop)) raise orm_exc.UnmappedColumnError( "No column %s is configured on mapper %s..." % (column, self.mapper))
gpl-3.0
hojel/calibre
src/calibre/library/restore.py
14
11515
#!/usr/bin/env python2 # vim:fileencoding=UTF-8:ts=4:sw=4:sta:et:sts=4:ai __license__ = 'GPL v3' __copyright__ = '2010, Kovid Goyal <kovid@kovidgoyal.net>' __docformat__ = 'restructuredtext en' import re, os, traceback, shutil from threading import Thread from operator import itemgetter from calibre.ptempfile import TemporaryDirectory from calibre.ebooks.metadata.opf2 import OPF from calibre.library.database2 import LibraryDatabase2 from calibre.library.prefs import DBPrefs from calibre.constants import filesystem_encoding from calibre.utils.date import utcfromtimestamp from calibre import isbytestring NON_EBOOK_EXTENSIONS = frozenset([ 'jpg', 'jpeg', 'gif', 'png', 'bmp', 'opf', 'swp', 'swo' ]) class RestoreDatabase(LibraryDatabase2): PATH_LIMIT = 10 WINDOWS_LIBRARY_PATH_LIMIT = 180 def set_path(self, *args, **kwargs): pass def dirtied(self, *args, **kwargs): pass class Restore(Thread): def __init__(self, library_path, progress_callback=None): super(Restore, self).__init__() if isbytestring(library_path): library_path = library_path.decode(filesystem_encoding) self.src_library_path = os.path.abspath(library_path) self.progress_callback = progress_callback self.db_id_regexp = re.compile(r'^.* \((\d+)\)$') self.bad_ext_pat = re.compile(r'[^a-z0-9_]+') if not callable(self.progress_callback): self.progress_callback = lambda x, y: x self.dirs = [] self.ignored_dirs = [] self.failed_dirs = [] self.books = [] self.conflicting_custom_cols = {} self.failed_restores = [] self.mismatched_dirs = [] self.successes = 0 self.tb = None self.authors_links = {} @property def errors_occurred(self): return self.failed_dirs or self.mismatched_dirs or \ self.conflicting_custom_cols or self.failed_restores @property def report(self): ans = '' failures = list(self.failed_dirs) + [(x['dirpath'], tb) for x, tb in self.failed_restores] if failures: ans += 'Failed to restore the books in the following folders:\n' for dirpath, tb in failures: ans += '\t' + dirpath + ' with error:\n' ans += '\n'.join('\t\t'+x for x in tb.splitlines()) ans += '\n\n' if self.conflicting_custom_cols: ans += '\n\n' ans += 'The following custom columns have conflicting definitions ' \ 'and were not fully restored:\n' for x in self.conflicting_custom_cols: ans += '\t#'+x+'\n' ans += '\tused:\t%s, %s, %s, %s\n'%(self.custom_columns[x][1], self.custom_columns[x][2], self.custom_columns[x][3], self.custom_columns[x][5]) for coldef in self.conflicting_custom_cols[x]: ans += '\tother:\t%s, %s, %s, %s\n'%(coldef[1], coldef[2], coldef[3], coldef[5]) if self.mismatched_dirs: ans += '\n\n' ans += 'The following folders were ignored:\n' for x in self.mismatched_dirs: ans += '\t'+x+'\n' return ans def run(self): try: with TemporaryDirectory('_library_restore') as tdir: self.library_path = tdir self.scan_library() if not self.load_preferences(): # Something went wrong with preferences restore. Start over # with a new database and attempt to rebuild the structure # from the metadata in the opf dbpath = os.path.join(self.library_path, 'metadata.db') if os.path.exists(dbpath): os.remove(dbpath) self.create_cc_metadata() self.restore_books() if self.successes == 0 and len(self.dirs) > 0: raise Exception(('Something bad happened')) self.replace_db() except: self.tb = traceback.format_exc() def load_preferences(self): self.progress_callback(None, 1) self.progress_callback(_('Starting restoring preferences and column metadata'), 0) prefs_path = os.path.join(self.src_library_path, 'metadata_db_prefs_backup.json') if not os.path.exists(prefs_path): self.progress_callback(_('Cannot restore preferences. Backup file not found.'), 1) return False try: prefs = DBPrefs.read_serialized(self.src_library_path, recreate_prefs=False) db = RestoreDatabase(self.library_path, default_prefs=prefs, restore_all_prefs=True, progress_callback=self.progress_callback) db.commit() db.close() self.progress_callback(None, 1) if 'field_metadata' in prefs: self.progress_callback(_('Finished restoring preferences and column metadata'), 1) return True self.progress_callback(_('Finished restoring preferences'), 1) return False except: traceback.print_exc() self.progress_callback(None, 1) self.progress_callback(_('Restoring preferences and column metadata failed'), 0) return False def scan_library(self): for dirpath, dirnames, filenames in os.walk(self.src_library_path): leaf = os.path.basename(dirpath) m = self.db_id_regexp.search(leaf) if m is None or 'metadata.opf' not in filenames: self.ignored_dirs.append(dirpath) continue self.dirs.append((dirpath, filenames, m.group(1))) self.progress_callback(None, len(self.dirs)) for i, x in enumerate(self.dirs): dirpath, filenames, book_id = x try: self.process_dir(dirpath, filenames, book_id) except: self.failed_dirs.append((dirpath, traceback.format_exc())) self.progress_callback(_('Processed') + ' ' + dirpath, i+1) def is_ebook_file(self, filename): ext = os.path.splitext(filename)[1] if not ext: return False ext = ext[1:].lower() if ext in NON_EBOOK_EXTENSIONS or \ self.bad_ext_pat.search(ext) is not None: return False return True def process_dir(self, dirpath, filenames, book_id): book_id = int(book_id) formats = filter(self.is_ebook_file, filenames) fmts = [os.path.splitext(x)[1][1:].upper() for x in formats] sizes = [os.path.getsize(os.path.join(dirpath, x)) for x in formats] names = [os.path.splitext(x)[0] for x in formats] opf = os.path.join(dirpath, 'metadata.opf') mi = OPF(opf, basedir=dirpath).to_book_metadata() timestamp = os.path.getmtime(opf) path = os.path.relpath(dirpath, self.src_library_path).replace(os.sep, '/') if int(mi.application_id) == book_id: self.books.append({ 'mi': mi, 'timestamp': timestamp, 'formats': list(zip(fmts, sizes, names)), 'id': book_id, 'dirpath': dirpath, 'path': path, }) else: self.mismatched_dirs.append(dirpath) alm = mi.get('author_link_map', {}) for author, link in alm.iteritems(): existing_link, timestamp = self.authors_links.get(author, (None, None)) if existing_link is None or existing_link != link and timestamp < mi.timestamp: self.authors_links[author] = (link, mi.timestamp) def create_cc_metadata(self): self.books.sort(key=itemgetter('timestamp')) self.custom_columns = {} fields = ('label', 'name', 'datatype', 'is_multiple', 'is_editable', 'display') for b in self.books: for key in b['mi'].custom_field_keys(): cfm = b['mi'].metadata_for_field(key) args = [] for x in fields: if x in cfm: if x == 'is_multiple': args.append(bool(cfm[x])) else: args.append(cfm[x]) if len(args) == len(fields): # TODO: Do series type columns need special handling? label = cfm['label'] if label in self.custom_columns and args != self.custom_columns[label]: if label not in self.conflicting_custom_cols: self.conflicting_custom_cols[label] = [] if self.custom_columns[label] not in self.conflicting_custom_cols[label]: self.conflicting_custom_cols[label].append(self.custom_columns[label]) self.custom_columns[label] = args db = RestoreDatabase(self.library_path) self.progress_callback(None, len(self.custom_columns)) if len(self.custom_columns): for i,args in enumerate(self.custom_columns.values()): db.create_custom_column(*args) self.progress_callback(_('creating custom column ')+args[0], i+1) db.close() def restore_books(self): self.progress_callback(None, len(self.books)) self.books.sort(key=itemgetter('id')) db = RestoreDatabase(self.library_path) for i, book in enumerate(self.books): try: self.restore_book(book, db) except: self.failed_restores.append((book, traceback.format_exc())) self.progress_callback(book['mi'].title, i+1) for author in self.authors_links.iterkeys(): link, ign = self.authors_links[author] db.conn.execute('UPDATE authors SET link=? WHERE name=?', (link, author.replace(',', '|'))) db.conn.commit() db.close() def restore_book(self, book, db): db.create_book_entry(book['mi'], add_duplicates=True, force_id=book['id']) if book['mi'].uuid: db.set_uuid(book['id'], book['mi'].uuid, commit=False, notify=False) db.conn.execute('UPDATE books SET path=?,last_modified=? WHERE id=?', (book['path'], utcfromtimestamp(book['timestamp']), book['id'])) for fmt, size, name in book['formats']: db.conn.execute(''' INSERT INTO data (book,format,uncompressed_size,name) VALUES (?,?,?,?)''', (book['id'], fmt, size, name)) db.conn.commit() self.successes += 1 def replace_db(self): dbpath = os.path.join(self.src_library_path, 'metadata.db') ndbpath = os.path.join(self.library_path, 'metadata.db') save_path = self.olddb = os.path.splitext(dbpath)[0]+'_pre_restore.db' if os.path.exists(save_path): os.remove(save_path) os.rename(dbpath, save_path) shutil.copyfile(ndbpath, dbpath)
gpl-3.0
perlygatekeeper/glowing-robot
Little_Alchemy_2/Scraper_python/env/lib/python3.7/site-packages/pip/_vendor/html5lib/filters/sanitizer.py
56
26248
from __future__ import absolute_import, division, unicode_literals import re from xml.sax.saxutils import escape, unescape from pip._vendor.six.moves import urllib_parse as urlparse from . import base from ..constants import namespaces, prefixes __all__ = ["Filter"] allowed_elements = frozenset(( (namespaces['html'], 'a'), (namespaces['html'], 'abbr'), (namespaces['html'], 'acronym'), (namespaces['html'], 'address'), (namespaces['html'], 'area'), (namespaces['html'], 'article'), (namespaces['html'], 'aside'), (namespaces['html'], 'audio'), (namespaces['html'], 'b'), (namespaces['html'], 'big'), (namespaces['html'], 'blockquote'), (namespaces['html'], 'br'), (namespaces['html'], 'button'), (namespaces['html'], 'canvas'), (namespaces['html'], 'caption'), (namespaces['html'], 'center'), (namespaces['html'], 'cite'), (namespaces['html'], 'code'), (namespaces['html'], 'col'), (namespaces['html'], 'colgroup'), (namespaces['html'], 'command'), (namespaces['html'], 'datagrid'), (namespaces['html'], 'datalist'), (namespaces['html'], 'dd'), (namespaces['html'], 'del'), (namespaces['html'], 'details'), (namespaces['html'], 'dfn'), (namespaces['html'], 'dialog'), (namespaces['html'], 'dir'), (namespaces['html'], 'div'), (namespaces['html'], 'dl'), (namespaces['html'], 'dt'), (namespaces['html'], 'em'), (namespaces['html'], 'event-source'), (namespaces['html'], 'fieldset'), (namespaces['html'], 'figcaption'), (namespaces['html'], 'figure'), (namespaces['html'], 'footer'), (namespaces['html'], 'font'), (namespaces['html'], 'form'), (namespaces['html'], 'header'), (namespaces['html'], 'h1'), (namespaces['html'], 'h2'), (namespaces['html'], 'h3'), (namespaces['html'], 'h4'), (namespaces['html'], 'h5'), (namespaces['html'], 'h6'), (namespaces['html'], 'hr'), (namespaces['html'], 'i'), (namespaces['html'], 'img'), (namespaces['html'], 'input'), (namespaces['html'], 'ins'), (namespaces['html'], 'keygen'), (namespaces['html'], 'kbd'), (namespaces['html'], 'label'), (namespaces['html'], 'legend'), (namespaces['html'], 'li'), (namespaces['html'], 'm'), (namespaces['html'], 'map'), (namespaces['html'], 'menu'), (namespaces['html'], 'meter'), (namespaces['html'], 'multicol'), (namespaces['html'], 'nav'), (namespaces['html'], 'nextid'), (namespaces['html'], 'ol'), (namespaces['html'], 'output'), (namespaces['html'], 'optgroup'), (namespaces['html'], 'option'), (namespaces['html'], 'p'), (namespaces['html'], 'pre'), (namespaces['html'], 'progress'), (namespaces['html'], 'q'), (namespaces['html'], 's'), (namespaces['html'], 'samp'), (namespaces['html'], 'section'), (namespaces['html'], 'select'), (namespaces['html'], 'small'), (namespaces['html'], 'sound'), (namespaces['html'], 'source'), (namespaces['html'], 'spacer'), (namespaces['html'], 'span'), (namespaces['html'], 'strike'), (namespaces['html'], 'strong'), (namespaces['html'], 'sub'), (namespaces['html'], 'sup'), (namespaces['html'], 'table'), (namespaces['html'], 'tbody'), (namespaces['html'], 'td'), (namespaces['html'], 'textarea'), (namespaces['html'], 'time'), (namespaces['html'], 'tfoot'), (namespaces['html'], 'th'), (namespaces['html'], 'thead'), (namespaces['html'], 'tr'), (namespaces['html'], 'tt'), (namespaces['html'], 'u'), (namespaces['html'], 'ul'), (namespaces['html'], 'var'), (namespaces['html'], 'video'), (namespaces['mathml'], 'maction'), (namespaces['mathml'], 'math'), (namespaces['mathml'], 'merror'), (namespaces['mathml'], 'mfrac'), (namespaces['mathml'], 'mi'), (namespaces['mathml'], 'mmultiscripts'), (namespaces['mathml'], 'mn'), (namespaces['mathml'], 'mo'), (namespaces['mathml'], 'mover'), (namespaces['mathml'], 'mpadded'), (namespaces['mathml'], 'mphantom'), (namespaces['mathml'], 'mprescripts'), (namespaces['mathml'], 'mroot'), (namespaces['mathml'], 'mrow'), (namespaces['mathml'], 'mspace'), (namespaces['mathml'], 'msqrt'), (namespaces['mathml'], 'mstyle'), (namespaces['mathml'], 'msub'), (namespaces['mathml'], 'msubsup'), (namespaces['mathml'], 'msup'), (namespaces['mathml'], 'mtable'), (namespaces['mathml'], 'mtd'), (namespaces['mathml'], 'mtext'), (namespaces['mathml'], 'mtr'), (namespaces['mathml'], 'munder'), (namespaces['mathml'], 'munderover'), (namespaces['mathml'], 'none'), (namespaces['svg'], 'a'), (namespaces['svg'], 'animate'), (namespaces['svg'], 'animateColor'), (namespaces['svg'], 'animateMotion'), (namespaces['svg'], 'animateTransform'), (namespaces['svg'], 'clipPath'), (namespaces['svg'], 'circle'), (namespaces['svg'], 'defs'), (namespaces['svg'], 'desc'), (namespaces['svg'], 'ellipse'), (namespaces['svg'], 'font-face'), (namespaces['svg'], 'font-face-name'), (namespaces['svg'], 'font-face-src'), (namespaces['svg'], 'g'), (namespaces['svg'], 'glyph'), (namespaces['svg'], 'hkern'), (namespaces['svg'], 'linearGradient'), (namespaces['svg'], 'line'), (namespaces['svg'], 'marker'), (namespaces['svg'], 'metadata'), (namespaces['svg'], 'missing-glyph'), (namespaces['svg'], 'mpath'), (namespaces['svg'], 'path'), (namespaces['svg'], 'polygon'), (namespaces['svg'], 'polyline'), (namespaces['svg'], 'radialGradient'), (namespaces['svg'], 'rect'), (namespaces['svg'], 'set'), (namespaces['svg'], 'stop'), (namespaces['svg'], 'svg'), (namespaces['svg'], 'switch'), (namespaces['svg'], 'text'), (namespaces['svg'], 'title'), (namespaces['svg'], 'tspan'), (namespaces['svg'], 'use'), )) allowed_attributes = frozenset(( # HTML attributes (None, 'abbr'), (None, 'accept'), (None, 'accept-charset'), (None, 'accesskey'), (None, 'action'), (None, 'align'), (None, 'alt'), (None, 'autocomplete'), (None, 'autofocus'), (None, 'axis'), (None, 'background'), (None, 'balance'), (None, 'bgcolor'), (None, 'bgproperties'), (None, 'border'), (None, 'bordercolor'), (None, 'bordercolordark'), (None, 'bordercolorlight'), (None, 'bottompadding'), (None, 'cellpadding'), (None, 'cellspacing'), (None, 'ch'), (None, 'challenge'), (None, 'char'), (None, 'charoff'), (None, 'choff'), (None, 'charset'), (None, 'checked'), (None, 'cite'), (None, 'class'), (None, 'clear'), (None, 'color'), (None, 'cols'), (None, 'colspan'), (None, 'compact'), (None, 'contenteditable'), (None, 'controls'), (None, 'coords'), (None, 'data'), (None, 'datafld'), (None, 'datapagesize'), (None, 'datasrc'), (None, 'datetime'), (None, 'default'), (None, 'delay'), (None, 'dir'), (None, 'disabled'), (None, 'draggable'), (None, 'dynsrc'), (None, 'enctype'), (None, 'end'), (None, 'face'), (None, 'for'), (None, 'form'), (None, 'frame'), (None, 'galleryimg'), (None, 'gutter'), (None, 'headers'), (None, 'height'), (None, 'hidefocus'), (None, 'hidden'), (None, 'high'), (None, 'href'), (None, 'hreflang'), (None, 'hspace'), (None, 'icon'), (None, 'id'), (None, 'inputmode'), (None, 'ismap'), (None, 'keytype'), (None, 'label'), (None, 'leftspacing'), (None, 'lang'), (None, 'list'), (None, 'longdesc'), (None, 'loop'), (None, 'loopcount'), (None, 'loopend'), (None, 'loopstart'), (None, 'low'), (None, 'lowsrc'), (None, 'max'), (None, 'maxlength'), (None, 'media'), (None, 'method'), (None, 'min'), (None, 'multiple'), (None, 'name'), (None, 'nohref'), (None, 'noshade'), (None, 'nowrap'), (None, 'open'), (None, 'optimum'), (None, 'pattern'), (None, 'ping'), (None, 'point-size'), (None, 'poster'), (None, 'pqg'), (None, 'preload'), (None, 'prompt'), (None, 'radiogroup'), (None, 'readonly'), (None, 'rel'), (None, 'repeat-max'), (None, 'repeat-min'), (None, 'replace'), (None, 'required'), (None, 'rev'), (None, 'rightspacing'), (None, 'rows'), (None, 'rowspan'), (None, 'rules'), (None, 'scope'), (None, 'selected'), (None, 'shape'), (None, 'size'), (None, 'span'), (None, 'src'), (None, 'start'), (None, 'step'), (None, 'style'), (None, 'summary'), (None, 'suppress'), (None, 'tabindex'), (None, 'target'), (None, 'template'), (None, 'title'), (None, 'toppadding'), (None, 'type'), (None, 'unselectable'), (None, 'usemap'), (None, 'urn'), (None, 'valign'), (None, 'value'), (None, 'variable'), (None, 'volume'), (None, 'vspace'), (None, 'vrml'), (None, 'width'), (None, 'wrap'), (namespaces['xml'], 'lang'), # MathML attributes (None, 'actiontype'), (None, 'align'), (None, 'columnalign'), (None, 'columnalign'), (None, 'columnalign'), (None, 'columnlines'), (None, 'columnspacing'), (None, 'columnspan'), (None, 'depth'), (None, 'display'), (None, 'displaystyle'), (None, 'equalcolumns'), (None, 'equalrows'), (None, 'fence'), (None, 'fontstyle'), (None, 'fontweight'), (None, 'frame'), (None, 'height'), (None, 'linethickness'), (None, 'lspace'), (None, 'mathbackground'), (None, 'mathcolor'), (None, 'mathvariant'), (None, 'mathvariant'), (None, 'maxsize'), (None, 'minsize'), (None, 'other'), (None, 'rowalign'), (None, 'rowalign'), (None, 'rowalign'), (None, 'rowlines'), (None, 'rowspacing'), (None, 'rowspan'), (None, 'rspace'), (None, 'scriptlevel'), (None, 'selection'), (None, 'separator'), (None, 'stretchy'), (None, 'width'), (None, 'width'), (namespaces['xlink'], 'href'), (namespaces['xlink'], 'show'), (namespaces['xlink'], 'type'), # SVG attributes (None, 'accent-height'), (None, 'accumulate'), (None, 'additive'), (None, 'alphabetic'), (None, 'arabic-form'), (None, 'ascent'), (None, 'attributeName'), (None, 'attributeType'), (None, 'baseProfile'), (None, 'bbox'), (None, 'begin'), (None, 'by'), (None, 'calcMode'), (None, 'cap-height'), (None, 'class'), (None, 'clip-path'), (None, 'color'), (None, 'color-rendering'), (None, 'content'), (None, 'cx'), (None, 'cy'), (None, 'd'), (None, 'dx'), (None, 'dy'), (None, 'descent'), (None, 'display'), (None, 'dur'), (None, 'end'), (None, 'fill'), (None, 'fill-opacity'), (None, 'fill-rule'), (None, 'font-family'), (None, 'font-size'), (None, 'font-stretch'), (None, 'font-style'), (None, 'font-variant'), (None, 'font-weight'), (None, 'from'), (None, 'fx'), (None, 'fy'), (None, 'g1'), (None, 'g2'), (None, 'glyph-name'), (None, 'gradientUnits'), (None, 'hanging'), (None, 'height'), (None, 'horiz-adv-x'), (None, 'horiz-origin-x'), (None, 'id'), (None, 'ideographic'), (None, 'k'), (None, 'keyPoints'), (None, 'keySplines'), (None, 'keyTimes'), (None, 'lang'), (None, 'marker-end'), (None, 'marker-mid'), (None, 'marker-start'), (None, 'markerHeight'), (None, 'markerUnits'), (None, 'markerWidth'), (None, 'mathematical'), (None, 'max'), (None, 'min'), (None, 'name'), (None, 'offset'), (None, 'opacity'), (None, 'orient'), (None, 'origin'), (None, 'overline-position'), (None, 'overline-thickness'), (None, 'panose-1'), (None, 'path'), (None, 'pathLength'), (None, 'points'), (None, 'preserveAspectRatio'), (None, 'r'), (None, 'refX'), (None, 'refY'), (None, 'repeatCount'), (None, 'repeatDur'), (None, 'requiredExtensions'), (None, 'requiredFeatures'), (None, 'restart'), (None, 'rotate'), (None, 'rx'), (None, 'ry'), (None, 'slope'), (None, 'stemh'), (None, 'stemv'), (None, 'stop-color'), (None, 'stop-opacity'), (None, 'strikethrough-position'), (None, 'strikethrough-thickness'), (None, 'stroke'), (None, 'stroke-dasharray'), (None, 'stroke-dashoffset'), (None, 'stroke-linecap'), (None, 'stroke-linejoin'), (None, 'stroke-miterlimit'), (None, 'stroke-opacity'), (None, 'stroke-width'), (None, 'systemLanguage'), (None, 'target'), (None, 'text-anchor'), (None, 'to'), (None, 'transform'), (None, 'type'), (None, 'u1'), (None, 'u2'), (None, 'underline-position'), (None, 'underline-thickness'), (None, 'unicode'), (None, 'unicode-range'), (None, 'units-per-em'), (None, 'values'), (None, 'version'), (None, 'viewBox'), (None, 'visibility'), (None, 'width'), (None, 'widths'), (None, 'x'), (None, 'x-height'), (None, 'x1'), (None, 'x2'), (namespaces['xlink'], 'actuate'), (namespaces['xlink'], 'arcrole'), (namespaces['xlink'], 'href'), (namespaces['xlink'], 'role'), (namespaces['xlink'], 'show'), (namespaces['xlink'], 'title'), (namespaces['xlink'], 'type'), (namespaces['xml'], 'base'), (namespaces['xml'], 'lang'), (namespaces['xml'], 'space'), (None, 'y'), (None, 'y1'), (None, 'y2'), (None, 'zoomAndPan'), )) attr_val_is_uri = frozenset(( (None, 'href'), (None, 'src'), (None, 'cite'), (None, 'action'), (None, 'longdesc'), (None, 'poster'), (None, 'background'), (None, 'datasrc'), (None, 'dynsrc'), (None, 'lowsrc'), (None, 'ping'), (namespaces['xlink'], 'href'), (namespaces['xml'], 'base'), )) svg_attr_val_allows_ref = frozenset(( (None, 'clip-path'), (None, 'color-profile'), (None, 'cursor'), (None, 'fill'), (None, 'filter'), (None, 'marker'), (None, 'marker-start'), (None, 'marker-mid'), (None, 'marker-end'), (None, 'mask'), (None, 'stroke'), )) svg_allow_local_href = frozenset(( (None, 'altGlyph'), (None, 'animate'), (None, 'animateColor'), (None, 'animateMotion'), (None, 'animateTransform'), (None, 'cursor'), (None, 'feImage'), (None, 'filter'), (None, 'linearGradient'), (None, 'pattern'), (None, 'radialGradient'), (None, 'textpath'), (None, 'tref'), (None, 'set'), (None, 'use') )) allowed_css_properties = frozenset(( 'azimuth', 'background-color', 'border-bottom-color', 'border-collapse', 'border-color', 'border-left-color', 'border-right-color', 'border-top-color', 'clear', 'color', 'cursor', 'direction', 'display', 'elevation', 'float', 'font', 'font-family', 'font-size', 'font-style', 'font-variant', 'font-weight', 'height', 'letter-spacing', 'line-height', 'overflow', 'pause', 'pause-after', 'pause-before', 'pitch', 'pitch-range', 'richness', 'speak', 'speak-header', 'speak-numeral', 'speak-punctuation', 'speech-rate', 'stress', 'text-align', 'text-decoration', 'text-indent', 'unicode-bidi', 'vertical-align', 'voice-family', 'volume', 'white-space', 'width', )) allowed_css_keywords = frozenset(( 'auto', 'aqua', 'black', 'block', 'blue', 'bold', 'both', 'bottom', 'brown', 'center', 'collapse', 'dashed', 'dotted', 'fuchsia', 'gray', 'green', '!important', 'italic', 'left', 'lime', 'maroon', 'medium', 'none', 'navy', 'normal', 'nowrap', 'olive', 'pointer', 'purple', 'red', 'right', 'solid', 'silver', 'teal', 'top', 'transparent', 'underline', 'white', 'yellow', )) allowed_svg_properties = frozenset(( 'fill', 'fill-opacity', 'fill-rule', 'stroke', 'stroke-width', 'stroke-linecap', 'stroke-linejoin', 'stroke-opacity', )) allowed_protocols = frozenset(( 'ed2k', 'ftp', 'http', 'https', 'irc', 'mailto', 'news', 'gopher', 'nntp', 'telnet', 'webcal', 'xmpp', 'callto', 'feed', 'urn', 'aim', 'rsync', 'tag', 'ssh', 'sftp', 'rtsp', 'afs', 'data', )) allowed_content_types = frozenset(( 'image/png', 'image/jpeg', 'image/gif', 'image/webp', 'image/bmp', 'text/plain', )) data_content_type = re.compile(r''' ^ # Match a content type <application>/<type> (?P<content_type>[-a-zA-Z0-9.]+/[-a-zA-Z0-9.]+) # Match any character set and encoding (?:(?:;charset=(?:[-a-zA-Z0-9]+)(?:;(?:base64))?) |(?:;(?:base64))?(?:;charset=(?:[-a-zA-Z0-9]+))?) # Assume the rest is data ,.* $ ''', re.VERBOSE) class Filter(base.Filter): """Sanitizes token stream of XHTML+MathML+SVG and of inline style attributes""" def __init__(self, source, allowed_elements=allowed_elements, allowed_attributes=allowed_attributes, allowed_css_properties=allowed_css_properties, allowed_css_keywords=allowed_css_keywords, allowed_svg_properties=allowed_svg_properties, allowed_protocols=allowed_protocols, allowed_content_types=allowed_content_types, attr_val_is_uri=attr_val_is_uri, svg_attr_val_allows_ref=svg_attr_val_allows_ref, svg_allow_local_href=svg_allow_local_href): """Creates a Filter :arg allowed_elements: set of elements to allow--everything else will be escaped :arg allowed_attributes: set of attributes to allow in elements--everything else will be stripped :arg allowed_css_properties: set of CSS properties to allow--everything else will be stripped :arg allowed_css_keywords: set of CSS keywords to allow--everything else will be stripped :arg allowed_svg_properties: set of SVG properties to allow--everything else will be removed :arg allowed_protocols: set of allowed protocols for URIs :arg allowed_content_types: set of allowed content types for ``data`` URIs. :arg attr_val_is_uri: set of attributes that have URI values--values that have a scheme not listed in ``allowed_protocols`` are removed :arg svg_attr_val_allows_ref: set of SVG attributes that can have references :arg svg_allow_local_href: set of SVG elements that can have local hrefs--these are removed """ super(Filter, self).__init__(source) self.allowed_elements = allowed_elements self.allowed_attributes = allowed_attributes self.allowed_css_properties = allowed_css_properties self.allowed_css_keywords = allowed_css_keywords self.allowed_svg_properties = allowed_svg_properties self.allowed_protocols = allowed_protocols self.allowed_content_types = allowed_content_types self.attr_val_is_uri = attr_val_is_uri self.svg_attr_val_allows_ref = svg_attr_val_allows_ref self.svg_allow_local_href = svg_allow_local_href def __iter__(self): for token in base.Filter.__iter__(self): token = self.sanitize_token(token) if token: yield token # Sanitize the +html+, escaping all elements not in ALLOWED_ELEMENTS, and # stripping out all attributes not in ALLOWED_ATTRIBUTES. Style attributes # are parsed, and a restricted set, specified by ALLOWED_CSS_PROPERTIES and # ALLOWED_CSS_KEYWORDS, are allowed through. attributes in ATTR_VAL_IS_URI # are scanned, and only URI schemes specified in ALLOWED_PROTOCOLS are # allowed. # # sanitize_html('<script> do_nasty_stuff() </script>') # => &lt;script> do_nasty_stuff() &lt;/script> # sanitize_html('<a href="javascript: sucker();">Click here for $100</a>') # => <a>Click here for $100</a> def sanitize_token(self, token): # accommodate filters which use token_type differently token_type = token["type"] if token_type in ("StartTag", "EndTag", "EmptyTag"): name = token["name"] namespace = token["namespace"] if ((namespace, name) in self.allowed_elements or (namespace is None and (namespaces["html"], name) in self.allowed_elements)): return self.allowed_token(token) else: return self.disallowed_token(token) elif token_type == "Comment": pass else: return token def allowed_token(self, token): if "data" in token: attrs = token["data"] attr_names = set(attrs.keys()) # Remove forbidden attributes for to_remove in (attr_names - self.allowed_attributes): del token["data"][to_remove] attr_names.remove(to_remove) # Remove attributes with disallowed URL values for attr in (attr_names & self.attr_val_is_uri): assert attr in attrs # I don't have a clue where this regexp comes from or why it matches those # characters, nor why we call unescape. I just know it's always been here. # Should you be worried by this comment in a sanitizer? Yes. On the other hand, all # this will do is remove *more* than it otherwise would. val_unescaped = re.sub("[`\x00-\x20\x7f-\xa0\\s]+", '', unescape(attrs[attr])).lower() # remove replacement characters from unescaped characters val_unescaped = val_unescaped.replace("\ufffd", "") try: uri = urlparse.urlparse(val_unescaped) except ValueError: uri = None del attrs[attr] if uri and uri.scheme: if uri.scheme not in self.allowed_protocols: del attrs[attr] if uri.scheme == 'data': m = data_content_type.match(uri.path) if not m: del attrs[attr] elif m.group('content_type') not in self.allowed_content_types: del attrs[attr] for attr in self.svg_attr_val_allows_ref: if attr in attrs: attrs[attr] = re.sub(r'url\s*\(\s*[^#\s][^)]+?\)', ' ', unescape(attrs[attr])) if (token["name"] in self.svg_allow_local_href and (namespaces['xlink'], 'href') in attrs and re.search(r'^\s*[^#\s].*', attrs[(namespaces['xlink'], 'href')])): del attrs[(namespaces['xlink'], 'href')] if (None, 'style') in attrs: attrs[(None, 'style')] = self.sanitize_css(attrs[(None, 'style')]) token["data"] = attrs return token def disallowed_token(self, token): token_type = token["type"] if token_type == "EndTag": token["data"] = "</%s>" % token["name"] elif token["data"]: assert token_type in ("StartTag", "EmptyTag") attrs = [] for (ns, name), v in token["data"].items(): attrs.append(' %s="%s"' % (name if ns is None else "%s:%s" % (prefixes[ns], name), escape(v))) token["data"] = "<%s%s>" % (token["name"], ''.join(attrs)) else: token["data"] = "<%s>" % token["name"] if token.get("selfClosing"): token["data"] = token["data"][:-1] + "/>" token["type"] = "Characters" del token["name"] return token def sanitize_css(self, style): # disallow urls style = re.compile(r'url\s*\(\s*[^\s)]+?\s*\)\s*').sub(' ', style) # gauntlet if not re.match(r"""^([:,;#%.\sa-zA-Z0-9!]|\w-\w|'[\s\w]+'|"[\s\w]+"|\([\d,\s]+\))*$""", style): return '' if not re.match(r"^\s*([-\w]+\s*:[^:;]*(;\s*|$))*$", style): return '' clean = [] for prop, value in re.findall(r"([-\w]+)\s*:\s*([^:;]*)", style): if not value: continue if prop.lower() in self.allowed_css_properties: clean.append(prop + ': ' + value + ';') elif prop.split('-')[0].lower() in ['background', 'border', 'margin', 'padding']: for keyword in value.split(): if keyword not in self.allowed_css_keywords and \ not re.match(r"^(#[0-9a-fA-F]+|rgb\(\d+%?,\d*%?,?\d*%?\)?|\d{0,2}\.?\d{0,2}(cm|em|ex|in|mm|pc|pt|px|%|,|\))?)$", keyword): # noqa break else: clean.append(prop + ': ' + value + ';') elif prop.lower() in self.allowed_svg_properties: clean.append(prop + ': ' + value + ';') return ' '.join(clean)
artistic-2.0
zasdfgbnm/tensorflow
tensorflow/contrib/kernel_methods/python/losses_test.py
23
11003
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for third_party.tensorflow.contrib.kernel_methods.python.losses.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import numpy as np from tensorflow.contrib.kernel_methods.python import losses from tensorflow.python.framework import constant_op from tensorflow.python.framework import dtypes from tensorflow.python.framework import errors from tensorflow.python.ops import array_ops from tensorflow.python.platform import test class SparseMulticlassHingeLossTest(test.TestCase): def testInvalidLogitsShape(self): """An error is raised when logits have invalid shape.""" with self.test_session(): logits = constant_op.constant([-1.0, 2.1], shape=(2,)) labels = constant_op.constant([0, 1]) with self.assertRaises(ValueError): _ = losses.sparse_multiclass_hinge_loss(labels, logits) def testInvalidLabelsShape(self): """An error is raised when labels have invalid shape.""" with self.test_session(): logits = constant_op.constant([-1.0, 2.1], shape=(2, 1)) labels = constant_op.constant([1, 0], shape=(1, 1, 2)) with self.assertRaises(ValueError): _ = losses.sparse_multiclass_hinge_loss(labels, logits) def testInvalidWeightsShape(self): """An error is raised when weights have invalid shape.""" with self.test_session(): logits = constant_op.constant([-1.0, 2.1], shape=(2, 1)) labels = constant_op.constant([1, 0], shape=(2,)) weights = constant_op.constant([1.5, 0.2], shape=(2, 1, 1)) with self.assertRaises(ValueError): _ = losses.sparse_multiclass_hinge_loss(labels, logits, weights) def testInvalidLabelsDtype(self): """An error is raised when labels have invalid shape.""" with self.test_session(): logits = constant_op.constant([-1.0, 2.1], shape=(2, 1)) labels = constant_op.constant([1, 0], dtype=dtypes.float32) with self.assertRaises(ValueError): _ = losses.sparse_multiclass_hinge_loss(labels, logits) def testNoneWeightRaisesValueError(self): """An error is raised when weights are None.""" with self.test_session(): logits = constant_op.constant([-1.0, 2.1], shape=(2, 1)) labels = constant_op.constant([1, 0]) with self.assertRaises(ValueError): _ = losses.sparse_multiclass_hinge_loss(labels, logits, weights=None) def testInconsistentLabelsAndWeightsShapesSameRank(self): """Error raised when weights and labels have same ranks, different sizes.""" with self.test_session(): logits = constant_op.constant([-1.0, 2.1, 4.1], shape=(3, 1)) labels = constant_op.constant([1, 0, 2], shape=(3, 1)) weights = constant_op.constant([1.1, 2.0], shape=(2, 1)) with self.assertRaises(ValueError): _ = losses.sparse_multiclass_hinge_loss(labels, logits, weights) def testInconsistentLabelsAndWeightsShapesDifferentRank(self): """Error raised when weights and labels have different ranks and sizes.""" with self.test_session(): logits = constant_op.constant([-1.0, 2.1], shape=(2, 1)) labels = constant_op.constant([1, 0], shape=(2, 1)) weights = constant_op.constant([1.1, 2.0, 2.8], shape=(3,)) with self.assertRaises(ValueError): _ = losses.sparse_multiclass_hinge_loss(labels, logits, weights) def testOutOfRangeLabels(self): """An error is raised when labels are not in [0, num_classes).""" with self.test_session(): logits = constant_op.constant([[1.2, -1.4, -1.0], [1.4, 1.8, 4.0], [0.5, 1.8, -1.0]]) labels = constant_op.constant([1, 0, 4]) loss = losses.sparse_multiclass_hinge_loss(labels, logits) with self.assertRaises(errors.InvalidArgumentError): loss.eval() def testZeroLossInt32Labels(self): """Loss is 0 if true class logits sufficiently higher than other classes.""" with self.test_session(): logits = constant_op.constant([[1.2, -1.4, -1.0], [1.4, 1.8, 4.0], [0.5, 1.8, -1.0]]) labels = constant_op.constant([0, 2, 1], dtype=dtypes.int32) loss = losses.sparse_multiclass_hinge_loss(labels, logits) self.assertAlmostEqual(loss.eval(), 0.0, 3) def testZeroLossInt64Labels(self): """Loss is 0 if true class logits sufficiently higher than other classes.""" with self.test_session(): logits = constant_op.constant([[2.1, -0.4, -1.0], [1.4, 2.8, 4.0], [-0.5, 0.8, -1.0]]) labels = constant_op.constant([0, 2, 1], dtype=dtypes.int64) loss = losses.sparse_multiclass_hinge_loss(labels, logits) self.assertAlmostEqual(loss.eval(), 0.0, 3) def testUnknownShape(self): """Result keeps same with `testZeroLossInt32Labels`""" logits_np = np.array([[1.2, -1.4, -1.0], [1.4, 1.8, 4.0], [0.5, 1.8, -1.0]]) labels_np = np.array([0, 2, 1], dtype=np.int32) logits_shapes = [ [3, 3], # batch_size, num_classes [None, 3], [3, None], [None, None] ] for batch_size, num_classes in logits_shapes: with self.test_session(): logits = array_ops.placeholder( dtypes.float32, shape=(batch_size, num_classes)) labels = array_ops.placeholder(dtypes.int32, shape=(batch_size,)) loss = losses.sparse_multiclass_hinge_loss(labels, logits) result = loss.eval(feed_dict={logits: logits_np, labels: labels_np}) self.assertAlmostEqual(result, 0.0, 3) def testCorrectPredictionsSomeClassesInsideMargin(self): """Loss is > 0 even if true class logits are higher than other classes.""" with self.test_session(): logits = constant_op.constant([[1.2, -1.4, 0.8], [1.4, 1.8, 4.0], [1.5, 1.8, -1.0]]) labels = constant_op.constant([0, 2, 1]) loss = losses.sparse_multiclass_hinge_loss(labels, logits) # The first and third samples incur some loss (0.6 and 0.7 respectively). self.assertAlmostEqual(loss.eval(), 0.4333, 3) def testIncorrectPredictions(self): """Loss is >0 when an incorrect class has higher logits than true class.""" with self.test_session(): logits = constant_op.constant([[2.6, 0.4, 0.8], [1.4, 0.8, -1.0], [0.5, -1.8, 2.0]]) labels = constant_op.constant([1, 0, 2]) loss = losses.sparse_multiclass_hinge_loss(labels, logits) # The first examples incurs a high loss (3.2) since the logits of an # incorrect class (0) are higher than the logits of the ground truth. The # second example also incures a (smaller) loss (0.4). self.assertAlmostEqual(loss.eval(), 1.2, 3) def testIncorrectPredictionsColumnLabels(self): """Same as above but labels is a rank-2 tensor.""" with self.test_session(): logits = constant_op.constant([[1.6, -0.4, 0.8], [1.5, 0.8, -1.0], [0.2, -1.8, 4.0]]) labels = constant_op.constant([1, 0, 2], shape=(3, 1)) loss = losses.sparse_multiclass_hinge_loss(labels, logits) # The first examples incurs a high loss (3.0) since the logits of an # incorrect class (0) are higher than the logits of the ground truth. The # second example also incures a (smaller) loss (0.3). self.assertAlmostEqual(loss.eval(), 1.1, 3) def testIncorrectPredictionsZeroWeights(self): """Loss is 0 when all weights are missing even if predictions are wrong.""" with self.test_session(): logits = constant_op.constant([[1.6, -0.4, 0.8], [1.5, 0.8, -1.0], [0.2, -1.8, 4.0]]) labels = constant_op.constant([1, 0, 2], shape=(3, 1)) weights = constant_op.constant([0.0, 0.0, 0.0], shape=(3, 1)) loss = losses.sparse_multiclass_hinge_loss(labels, logits, weights) # No overall loss since all weights are 0. self.assertAlmostEqual(loss.eval(), 0.0, 3) def testNonZeroLossWithPythonScalarWeights(self): """Weighted loss is correctly computed when weights is a python scalar.""" with self.test_session(): logits = constant_op.constant([[1.6, -0.4, 0.8], [1.5, 0.8, -1.0], [0.2, -1.8, 4.0]]) labels = constant_op.constant([1, 0, 2], shape=(3, 1)) weights = 10.0 loss = losses.sparse_multiclass_hinge_loss(labels, logits, weights) self.assertAlmostEqual(loss.eval(), 11.0, 3) def testNonZeroLossWithScalarTensorWeights(self): """Weighted loss is correctly computed when weights is a rank-0 tensor.""" with self.test_session(): logits = constant_op.constant([[1.6, -0.4, 0.8], [1.5, 0.8, -1.0], [0.2, -1.8, 4.0]]) labels = constant_op.constant([1, 0, 2], shape=(3, 1)) weights = constant_op.constant(5.0) loss = losses.sparse_multiclass_hinge_loss(labels, logits, weights) self.assertAlmostEqual(loss.eval(), 5.5, 3) def testNonZeroLossWith1DTensorWeightsColumnLabels(self): """Weighted loss is correctly computed when weights is a rank-0 tensor.""" with self.test_session(): logits = constant_op.constant([[1.6, -0.4, 0.8], [1.5, 0.8, -1.0], [0.2, -1.8, 4.0]]) labels = constant_op.constant([1, 0, 2], shape=(3, 1)) weights = constant_op.constant([1.0, 0.5, 2.0], shape=(3,)) loss = losses.sparse_multiclass_hinge_loss(labels, logits, weights) # The overall loss is 1/3 *(3.0*1.0 + 0.5*0.3+ 2.0*0.0) = 1.05 self.assertAlmostEqual(loss.eval(), 1.05, 3) def testNonZeroLossWith2DTensorWeights1DLabelsSomeWeightsMissing(self): """Weighted loss is correctly computed when weights is a rank-0 tensor.""" with self.test_session(): logits = constant_op.constant([[1.6, -0.4, 0.8], [1.5, 0.8, -1.0], [0.2, -1.8, 4.0], [1.6, 1.8, -4.0]]) labels = constant_op.constant([1, 0, 2, 1]) weights = constant_op.constant([[1.0], [0.0], [2.0], [4.0]]) loss = losses.sparse_multiclass_hinge_loss(labels, logits, weights) # The overall loss is 1/3 *(3.0*1.0 + 0.0*0.3+ 2.0*0.0 + 4.0*0.8) = 6.2/3. self.assertAlmostEqual(loss.eval(), 2.06666, 3) if __name__ == '__main__': test.main()
apache-2.0
dhamaniasad/mythbox
resources/lib/twisted/twisted/test/test_logfile.py
59
9810
# Copyright (c) 2001-2007 Twisted Matrix Laboratories. # See LICENSE for details. from twisted.trial import unittest # system imports import os, time, stat # twisted imports from twisted.python import logfile, runtime class LogFileTestCase(unittest.TestCase): """ Test the rotating log file. """ def setUp(self): self.dir = self.mktemp() os.makedirs(self.dir) self.name = "test.log" self.path = os.path.join(self.dir, self.name) def tearDown(self): """ Restore back write rights on created paths: if tests modified the rights, that will allow the paths to be removed easily afterwards. """ os.chmod(self.dir, 0777) if os.path.exists(self.path): os.chmod(self.path, 0777) def testWriting(self): log = logfile.LogFile(self.name, self.dir) log.write("123") log.write("456") log.flush() log.write("7890") log.close() f = open(self.path, "r") self.assertEquals(f.read(), "1234567890") f.close() def testRotation(self): # this logfile should rotate every 10 bytes log = logfile.LogFile(self.name, self.dir, rotateLength=10) # test automatic rotation log.write("123") log.write("4567890") log.write("1" * 11) self.assert_(os.path.exists("%s.1" % self.path)) self.assert_(not os.path.exists("%s.2" % self.path)) log.write('') self.assert_(os.path.exists("%s.1" % self.path)) self.assert_(os.path.exists("%s.2" % self.path)) self.assert_(not os.path.exists("%s.3" % self.path)) log.write("3") self.assert_(not os.path.exists("%s.3" % self.path)) # test manual rotation log.rotate() self.assert_(os.path.exists("%s.3" % self.path)) self.assert_(not os.path.exists("%s.4" % self.path)) log.close() self.assertEquals(log.listLogs(), [1, 2, 3]) def testAppend(self): log = logfile.LogFile(self.name, self.dir) log.write("0123456789") log.close() log = logfile.LogFile(self.name, self.dir) self.assertEquals(log.size, 10) self.assertEquals(log._file.tell(), log.size) log.write("abc") self.assertEquals(log.size, 13) self.assertEquals(log._file.tell(), log.size) f = log._file f.seek(0, 0) self.assertEquals(f.read(), "0123456789abc") log.close() def testLogReader(self): log = logfile.LogFile(self.name, self.dir) log.write("abc\n") log.write("def\n") log.rotate() log.write("ghi\n") log.flush() # check reading logs self.assertEquals(log.listLogs(), [1]) reader = log.getCurrentLog() reader._file.seek(0) self.assertEquals(reader.readLines(), ["ghi\n"]) self.assertEquals(reader.readLines(), []) reader.close() reader = log.getLog(1) self.assertEquals(reader.readLines(), ["abc\n", "def\n"]) self.assertEquals(reader.readLines(), []) reader.close() # check getting illegal log readers self.assertRaises(ValueError, log.getLog, 2) self.assertRaises(TypeError, log.getLog, "1") # check that log numbers are higher for older logs log.rotate() self.assertEquals(log.listLogs(), [1, 2]) reader = log.getLog(1) reader._file.seek(0) self.assertEquals(reader.readLines(), ["ghi\n"]) self.assertEquals(reader.readLines(), []) reader.close() reader = log.getLog(2) self.assertEquals(reader.readLines(), ["abc\n", "def\n"]) self.assertEquals(reader.readLines(), []) reader.close() def testModePreservation(self): """ Check rotated files have same permissions as original. """ f = open(self.path, "w").close() os.chmod(self.path, 0707) mode = os.stat(self.path)[stat.ST_MODE] log = logfile.LogFile(self.name, self.dir) log.write("abc") log.rotate() self.assertEquals(mode, os.stat(self.path)[stat.ST_MODE]) def test_noPermission(self): """ Check it keeps working when permission on dir changes. """ log = logfile.LogFile(self.name, self.dir) log.write("abc") # change permissions so rotation would fail os.chmod(self.dir, 0555) # if this succeeds, chmod doesn't restrict us, so we can't # do the test try: f = open(os.path.join(self.dir,"xxx"), "w") except (OSError, IOError): pass else: f.close() return log.rotate() # this should not fail log.write("def") log.flush() f = log._file self.assertEquals(f.tell(), 6) f.seek(0, 0) self.assertEquals(f.read(), "abcdef") log.close() def test_maxNumberOfLog(self): """ Test it respect the limit on the number of files when maxRotatedFiles is not None. """ log = logfile.LogFile(self.name, self.dir, rotateLength=10, maxRotatedFiles=3) log.write("1" * 11) log.write("2" * 11) self.failUnless(os.path.exists("%s.1" % self.path)) log.write("3" * 11) self.failUnless(os.path.exists("%s.2" % self.path)) log.write("4" * 11) self.failUnless(os.path.exists("%s.3" % self.path)) self.assertEquals(file("%s.3" % self.path).read(), "1" * 11) log.write("5" * 11) self.assertEquals(file("%s.3" % self.path).read(), "2" * 11) self.failUnless(not os.path.exists("%s.4" % self.path)) def test_fromFullPath(self): """ Test the fromFullPath method. """ log1 = logfile.LogFile(self.name, self.dir, 10, defaultMode=0777) log2 = logfile.LogFile.fromFullPath(self.path, 10, defaultMode=0777) self.assertEquals(log1.name, log2.name) self.assertEquals(os.path.abspath(log1.path), log2.path) self.assertEquals(log1.rotateLength, log2.rotateLength) self.assertEquals(log1.defaultMode, log2.defaultMode) def test_defaultPermissions(self): """ Test the default permission of the log file: if the file exist, it should keep the permission. """ f = file(self.path, "w") os.chmod(self.path, 0707) currentMode = stat.S_IMODE(os.stat(self.path)[stat.ST_MODE]) f.close() log1 = logfile.LogFile(self.name, self.dir) self.assertEquals(stat.S_IMODE(os.stat(self.path)[stat.ST_MODE]), currentMode) def test_specifiedPermissions(self): """ Test specifying the permissions used on the log file. """ log1 = logfile.LogFile(self.name, self.dir, defaultMode=0066) mode = stat.S_IMODE(os.stat(self.path)[stat.ST_MODE]) if runtime.platform.isWindows(): # The only thing we can get here is global read-only self.assertEquals(mode, 0444) else: self.assertEquals(mode, 0066) def test_reopen(self): """ L{logfile.LogFile.reopen} allows to rename the currently used file and make L{logfile.LogFile} create a new file. """ log1 = logfile.LogFile(self.name, self.dir) log1.write("hello1") savePath = os.path.join(self.dir, "save.log") os.rename(self.path, savePath) log1.reopen() log1.write("hello2") log1.close() f = open(self.path, "r") self.assertEquals(f.read(), "hello2") f.close() f = open(savePath, "r") self.assertEquals(f.read(), "hello1") f.close() if runtime.platform.isWindows(): test_reopen.skip = "Can't test reopen on Windows" class RiggedDailyLogFile(logfile.DailyLogFile): _clock = 0.0 def _openFile(self): logfile.DailyLogFile._openFile(self) # rig the date to match _clock, not mtime self.lastDate = self.toDate() def toDate(self, *args): if args: return time.gmtime(*args)[:3] return time.gmtime(self._clock)[:3] class DailyLogFileTestCase(unittest.TestCase): """ Test rotating log file. """ def setUp(self): self.dir = self.mktemp() os.makedirs(self.dir) self.name = "testdaily.log" self.path = os.path.join(self.dir, self.name) def testWriting(self): log = RiggedDailyLogFile(self.name, self.dir) log.write("123") log.write("456") log.flush() log.write("7890") log.close() f = open(self.path, "r") self.assertEquals(f.read(), "1234567890") f.close() def testRotation(self): # this logfile should rotate every 10 bytes log = RiggedDailyLogFile(self.name, self.dir) days = [(self.path + '.' + log.suffix(day * 86400)) for day in range(3)] # test automatic rotation log._clock = 0.0 # 1970/01/01 00:00.00 log.write("123") log._clock = 43200 # 1970/01/01 12:00.00 log.write("4567890") log._clock = 86400 # 1970/01/02 00:00.00 log.write("1" * 11) self.assert_(os.path.exists(days[0])) self.assert_(not os.path.exists(days[1])) log._clock = 172800 # 1970/01/03 00:00.00 log.write('') self.assert_(os.path.exists(days[0])) self.assert_(os.path.exists(days[1])) self.assert_(not os.path.exists(days[2])) log._clock = 259199 # 1970/01/03 23:59.59 log.write("3") self.assert_(not os.path.exists(days[2]))
gpl-2.0
svn2github/django
django/core/management/commands/test.py
111
2965
import sys import os from optparse import make_option, OptionParser from django.conf import settings from django.core.management.base import BaseCommand from django.test.utils import get_runner class Command(BaseCommand): option_list = BaseCommand.option_list + ( make_option('--noinput', action='store_false', dest='interactive', default=True, help='Tells Django to NOT prompt the user for input of any kind.'), make_option('--failfast', action='store_true', dest='failfast', default=False, help='Tells Django to stop running the test suite after first ' 'failed test.'), make_option('--testrunner', action='store', dest='testrunner', help='Tells Django to use specified test runner class instead of ' 'the one specified by the TEST_RUNNER setting.'), make_option('--liveserver', action='store', dest='liveserver', default=None, help='Overrides the default address where the live server (used ' 'with LiveServerTestCase) is expected to run from. The ' 'default value is localhost:8081.'), ) help = ('Runs the test suite for the specified applications, or the ' 'entire site if no apps are specified.') args = '[appname ...]' requires_model_validation = False def __init__(self): self.test_runner = None super(Command, self).__init__() def run_from_argv(self, argv): """ Pre-parse the command line to extract the value of the --testrunner option. This allows a test runner to define additional command line arguments. """ option = '--testrunner=' for arg in argv[2:]: if arg.startswith(option): self.test_runner = arg[len(option):] break super(Command, self).run_from_argv(argv) def create_parser(self, prog_name, subcommand): test_runner_class = get_runner(settings, self.test_runner) options = self.option_list + getattr( test_runner_class, 'option_list', ()) return OptionParser(prog=prog_name, usage=self.usage(subcommand), version=self.get_version(), option_list=options) def handle(self, *test_labels, **options): from django.conf import settings from django.test.utils import get_runner TestRunner = get_runner(settings, options.get('testrunner')) options['verbosity'] = int(options.get('verbosity')) if options.get('liveserver') is not None: os.environ['DJANGO_LIVE_TEST_SERVER_ADDRESS'] = options['liveserver'] del options['liveserver'] test_runner = TestRunner(**options) failures = test_runner.run_tests(test_labels) if failures: sys.exit(bool(failures))
bsd-3-clause
MalloyPower/parsing-python
front-end/testsuite-python-lib/Python-2.2/Lib/plat-linux2/DLFCN.py
2
1632
# Generated by h2py from /usr/include/dlfcn.h _DLFCN_H = 1 # Included from features.h _FEATURES_H = 1 __USE_ANSI = 1 __FAVOR_BSD = 1 _ISOC99_SOURCE = 1 _POSIX_SOURCE = 1 _POSIX_C_SOURCE = 199506L _XOPEN_SOURCE = 600 _XOPEN_SOURCE_EXTENDED = 1 _LARGEFILE64_SOURCE = 1 _BSD_SOURCE = 1 _SVID_SOURCE = 1 _BSD_SOURCE = 1 _SVID_SOURCE = 1 __USE_ISOC99 = 1 _POSIX_SOURCE = 1 _POSIX_C_SOURCE = 2 _POSIX_C_SOURCE = 199506L __USE_POSIX = 1 __USE_POSIX2 = 1 __USE_POSIX199309 = 1 __USE_POSIX199506 = 1 __USE_XOPEN = 1 __USE_XOPEN_EXTENDED = 1 __USE_UNIX98 = 1 _LARGEFILE_SOURCE = 1 __USE_XOPEN2K = 1 __USE_ISOC99 = 1 __USE_XOPEN_EXTENDED = 1 __USE_LARGEFILE = 1 __USE_LARGEFILE64 = 1 __USE_FILE_OFFSET64 = 1 __USE_MISC = 1 __USE_BSD = 1 __USE_SVID = 1 __USE_GNU = 1 __USE_REENTRANT = 1 __STDC_IEC_559__ = 1 __STDC_IEC_559_COMPLEX__ = 1 __STDC_ISO_10646__ = 200009L __GNU_LIBRARY__ = 6 __GLIBC__ = 2 __GLIBC_MINOR__ = 2 # Included from sys/cdefs.h _SYS_CDEFS_H = 1 def __PMT(args): return args def __P(args): return args def __PMT(args): return args def __STRING(x): return #x __flexarr = [] __flexarr = [0] __flexarr = [] __flexarr = [1] def __ASMNAME(cname): return __ASMNAME2 (__USER_LABEL_PREFIX__, cname) def __attribute__(xyz): return def __attribute_format_arg__(x): return __attribute__ ((__format_arg__ (x))) def __attribute_format_arg__(x): return __USE_LARGEFILE = 1 __USE_LARGEFILE64 = 1 __USE_EXTERN_INLINES = 1 # Included from gnu/stubs.h # Included from bits/dlfcn.h RTLD_LAZY = 0x00001 RTLD_NOW = 0x00002 RTLD_BINDING_MASK = 0x3 RTLD_NOLOAD = 0x00004 RTLD_GLOBAL = 0x00100 RTLD_LOCAL = 0 RTLD_NODELETE = 0x01000
mit
hoevenvd/weewx_poller
bin/weedb/sqlite.py
1
4562
# # Copyright (c) 2012 Tom Keffer <tkeffer@gmail.com> # # See the file LICENSE.txt for your full rights. # # $Revision: 829 $ # $Author: tkeffer $ # $Date: 2013-01-19 08:05:49 -0800 (Sat, 19 Jan 2013) $ # """Driver for sqlite""" import os.path # Import sqlite3. If it does not support the 'with' statement, then # import pysqlite2, which might... import sqlite3 if not hasattr(sqlite3.Connection, "__exit__"): del sqlite3 from pysqlite2 import dbapi2 as sqlite3 #@Reimport @UnresolvedImport import weedb def connect(database='', root='', driver='', **argv): """Factory function, to keep things compatible with DBAPI. """ return Connection(database=database, root=root, **argv) def create(database='', root='', driver='', **argv): """Create the database specified by the db_dict. If it already exists, an exception of type DatabaseExists will be thrown.""" file_path = os.path.join(root, database) # Check whether the database file exists: if os.path.exists(file_path): raise weedb.DatabaseExists("Database %s already exists" % (file_path,)) else: # If it doesn't exist, create the parent directories fileDirectory = os.path.dirname(file_path) if not os.path.exists(fileDirectory): os.makedirs(fileDirectory) connection = sqlite3.connect(file_path, **argv) connection.close() def drop(database='', root='', driver='', **argv): file_path = os.path.join(root, database) try: os.remove(file_path) except OSError: raise weedb.NoDatabase("""Attempt to drop non-existent database %s""" % (file_path,)) class Connection(weedb.Connection): """A wrapper around a sqlite3 connection object.""" def __init__(self, database='', root='', **argv): """Initialize an instance of Connection. Parameters: file: Path to the sqlite file (required) fileroot: An optional path to be prefixed to parameter 'file'. If not given, nothing will be prefixed. If the operation fails, an exception of type weedb.OperationalError will be raised. """ self.file_path = os.path.join(root, database) if not os.path.exists(self.file_path): raise weedb.OperationalError("Attempt to open a non-existent database %s" % database) try: connection = sqlite3.connect(self.file_path, **argv) except sqlite3.OperationalError: # The Pysqlite driver does not include the database file path. # Include it in case it might be useful. raise weedb.OperationalError("Unable to open database '%s'" % (self.file_path,)) weedb.Connection.__init__(self, connection, database, 'sqlite') def cursor(self): """Return a cursor object.""" return Cursor(self.connection) def tables(self): """Returns a list of tables in the database.""" table_list = list() for row in self.connection.execute("""SELECT tbl_name FROM sqlite_master WHERE type='table';"""): # Extract the table name. Sqlite returns unicode, so always # convert to a regular string: table_list.append(str(row[0])) return table_list def columnsOf(self, table): """Return a list of columns in the specified table. If the table does not exist, None is returned.""" column_list = list() for row in self.connection.execute("""PRAGMA table_info(%s);""" % table): # Append this column to the list of columns. column_list.append(str(row[1])) # If there are no columns (which means the table did not exist) raise an exceptional if not column_list: raise weedb.OperationalError("No such table %s" % table) return column_list def begin(self): self.connection.execute("BEGIN TRANSACTION") class Cursor(sqlite3.Cursor): """A wrapper around the sqlite cursor object""" # The sqlite3 cursor object is very full featured. We need only turn # the sqlite exceptions into weedb exceptions. def __init__(self, *args, **kwargs): sqlite3.Cursor.__init__(self, *args, **kwargs) def execute(self, *args, **kwargs): try: return sqlite3.Cursor.execute(self, *args, **kwargs) except sqlite3.OperationalError, e: # Convert to a weedb exception raise weedb.OperationalError(e)
gpl-3.0
mozilla/make.mozilla.org
vendor-local/lib/python/celery/tests/test_backends/test_redis.py
14
3385
from __future__ import absolute_import from __future__ import with_statement import sys import socket from nose import SkipTest from celery.exceptions import ImproperlyConfigured from celery import states from celery.utils import uuid from celery.backends import redis from celery.backends.redis import RedisBackend from celery.tests.utils import Case, mask_modules _no_redis_msg = "* Redis %s. Will not execute related tests." _no_redis_msg_emitted = False try: from redis.exceptions import ConnectionError except ImportError: class ConnectionError(socket.error): # noqa pass class SomeClass(object): def __init__(self, data): self.data = data def get_redis_or_SkipTest(): def emit_no_redis_msg(reason): global _no_redis_msg_emitted if not _no_redis_msg_emitted: sys.stderr.write("\n" + _no_redis_msg % reason + "\n") _no_redis_msg_emitted = True if redis.redis is None: emit_no_redis_msg("not installed") raise SkipTest("redis library not installed") try: tb = RedisBackend(redis_db="celery_unittest") try: # Evaluate lazy connection tb.client.info() except ConnectionError, exc: emit_no_redis_msg("not running") raise SkipTest("can't connect to redis: %s" % (exc, )) return tb except ImproperlyConfigured, exc: if "need to install" in str(exc): return emit_no_redis_msg("not installed") return emit_no_redis_msg("not configured") class TestRedisBackend(Case): def test_mark_as_done(self): tb = get_redis_or_SkipTest() tid = uuid() self.assertEqual(tb.get_status(tid), states.PENDING) self.assertIsNone(tb.get_result(tid)) tb.mark_as_done(tid, 42) self.assertEqual(tb.get_status(tid), states.SUCCESS) self.assertEqual(tb.get_result(tid), 42) def test_is_pickled(self): tb = get_redis_or_SkipTest() tid2 = uuid() result = {"foo": "baz", "bar": SomeClass(12345)} tb.mark_as_done(tid2, result) # is serialized properly. rindb = tb.get_result(tid2) self.assertEqual(rindb.get("foo"), "baz") self.assertEqual(rindb.get("bar").data, 12345) def test_mark_as_failure(self): tb = get_redis_or_SkipTest() tid3 = uuid() try: raise KeyError("foo") except KeyError, exception: pass tb.mark_as_failure(tid3, exception) self.assertEqual(tb.get_status(tid3), states.FAILURE) self.assertIsInstance(tb.get_result(tid3), KeyError) class TestRedisBackendNoRedis(Case): def test_redis_None_if_redis_not_installed(self): prev = sys.modules.pop("celery.backends.redis") try: with mask_modules("redis"): from celery.backends.redis import redis self.assertIsNone(redis) finally: sys.modules["celery.backends.redis"] = prev def test_constructor_raises_if_redis_not_installed(self): from celery.backends import redis prev = redis.RedisBackend.redis redis.RedisBackend.redis = None try: with self.assertRaises(ImproperlyConfigured): redis.RedisBackend() finally: redis.RedisBackend.redis = prev
bsd-3-clause
hryamzik/ansible
lib/ansible/modules/windows/win_package.py
4
9677
#!/usr/bin/python # -*- coding: utf-8 -*- # Copyright: (c) 2014, Trond Hindenes <trond@hindenes.com>, and others # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) # this is a windows documentation stub. actual code lives in the .ps1 # file of the same name ANSIBLE_METADATA = {'metadata_version': '1.1', 'status': ['preview'], 'supported_by': 'core'} DOCUMENTATION = r''' --- module: win_package version_added: "1.7" short_description: Installs/uninstalls an installable package description: - Installs or uninstalls a package in either an MSI or EXE format. - These packages can be sources from the local file system, network file share or a url. - Please read the notes section around some caveats with this module. options: arguments: description: - Any arguments the installer needs to either install or uninstall the package. - If the package is an MSI do not supply the C(/qn), C(/log) or C(/norestart) arguments. - As of Ansible 2.5, this parameter can be a list of arguments and the module will escape the arguments as necessary, it is recommended to use a string when dealing with MSI packages due to the unique escaping issues with msiexec. creates_path: description: - Will check the existance of the path specified and use the result to determine whether the package is already installed. - You can use this in conjunction with C(product_id) and other C(creates_*). type: path version_added: '2.4' creates_service: description: - Will check the existing of the service specified and use the result to determine whether the package is already installed. - You can use this in conjunction with C(product_id) and other C(creates_*). version_added: '2.4' creates_version: description: - Will check the file version property of the file at C(creates_path) and use the result to determine whether the package is already installed. - C(creates_path) MUST be set and is a file. - You can use this in conjunction with C(product_id) and other C(creates_*). version_added: '2.4' expected_return_code: description: - One or more return codes from the package installation that indicates success. - Before Ansible 2.4 this was just 0 but since 2.4 this is both C(0) and C(3010). - A return code of C(3010) usually means that a reboot is required, the C(reboot_required) return value is set if the return code is C(3010). type: list default: [0, 3010] password: description: - The password for C(user_name), must be set when C(user_name) is. aliases: [ user_password ] path: description: - Location of the package to be installed or uninstalled. - This package can either be on the local file system, network share or a url. - If the path is on a network share and the current WinRM transport doesn't support credential delegation, then C(user_name) and C(user_password) must be set to access the file. - There are cases where this file will be copied locally to the server so it can access it, see the notes for more info. - If C(state=present) then this value MUST be set. - If C(state=absent) then this value does not need to be set if C(product_id) is. product_id: description: - The product id of the installed packaged. - This is used for checking whether the product is already installed and getting the uninstall information if C(state=absent). - You can find product ids for installed programs in the Windows registry editor either at C(HKLM:Software\Microsoft\Windows\CurrentVersion\Uninstall) or for 32 bit programs at C(HKLM:Software\Wow6432Node\Microsoft\Windows\CurrentVersion\Uninstall). - This SHOULD be set when the package is not an MSI, or the path is a url or a network share and credential delegation is not being used. The C(creates_*) options can be used instead but is not recommended. aliases: [ productid ] state: description: - Whether to install or uninstall the package. - The module uses C(product_id) and whether it exists at the registry path to see whether it needs to install or uninstall the package. default: present aliases: [ ensure ] username: description: - Username of an account with access to the package if it is located on a file share. - This is only needed if the WinRM transport is over an auth method that does not support credential delegation like Basic or NTLM. aliases: [ user_name ] validate_certs: description: - If C(no), SSL certificates will not be validated. This should only be used on personally controlled sites using self-signed certificates. - Before Ansible 2.4 this defaulted to C(no). type: bool default: 'yes' version_added: '2.4' notes: - For non Windows targets, use the M(package) module instead. - When C(state=absent) and the product is an exe, the path may be different from what was used to install the package originally. If path is not set then the path used will be what is set under C(UninstallString) in the registry for that product_id. - Not all product ids are in a GUID form, some programs incorrectly use a different structure but this module should support any format. - By default all msi installs and uninstalls will be run with the options C(/log, /qn, /norestart). - It is recommended you download the package first from the URL using the M(win_get_url) module as it opens up more flexibility with what must be set when calling C(win_package). - Packages will be temporarily downloaded or copied locally when path is a network location and credential delegation is not set, or path is a URL and the file is not an MSI. - All the installation checks under C(product_id) and C(creates_*) add together, if one fails then the program is considered to be absent. author: - Trond Hindenes (@trondhindenes) - Jordan Borean (@jborean93) ''' EXAMPLES = r''' - name: Install the Visual C thingy win_package: path: http://download.microsoft.com/download/1/6/B/16B06F60-3B20-4FF2-B699-5E9B7962F9AE/VSU_4/vcredist_x64.exe product_id: '{CF2BEA3C-26EA-32F8-AA9B-331F7E34BA97}' arguments: /install /passive /norestart - name: Install Visual C thingy with list of arguments instead of a string win_package: path: http://download.microsoft.com/download/1/6/B/16B06F60-3B20-4FF2-B699-5E9B7962F9AE/VSU_4/vcredist_x64.exe product_id: '{CF2BEA3C-26EA-32F8-AA9B-331F7E34BA97}' arguments: - /install - /passive - /norestart - name: Install Remote Desktop Connection Manager from msi win_package: path: https://download.microsoft.com/download/A/F/0/AF0071F3-B198-4A35-AA90-C68D103BDCCF/rdcman.msi product_id: '{0240359E-6A4C-4884-9E94-B397A02D893C}' state: present - name: Uninstall Remote Desktop Connection Manager win_package: product_id: '{0240359E-6A4C-4884-9E94-B397A02D893C}' state: absent - name: Install Remote Desktop Connection Manager locally omitting the product_id win_package: path: C:\temp\rdcman.msi state: present - name: Uninstall Remote Desktop Connection Manager from local MSI omitting the product_id win_package: path: C:\temp\rdcman.msi state: absent # 7-Zip exe doesn't use a guid for the Product ID - name: Install 7zip from a network share specifying the credentials win_package: path: \\domain\programs\7z.exe product_id: 7-Zip arguments: /S state: present user_name: DOMAIN\User user_password: Password - name: Install 7zip and use a file version for the installation check win_package: path: C:\temp\7z.exe creates_path: C:\Program Files\7-Zip\7z.exe creates_version: 16.04 state: present - name: Uninstall 7zip from the exe win_package: path: C:\Program Files\7-Zip\Uninstall.exe product_id: 7-Zip arguments: /S state: absent - name: Uninstall 7zip without specifying the path win_package: product_id: 7-Zip arguments: /S state: absent - name: Install application and override expected return codes win_package: path: https://download.microsoft.com/download/1/6/7/167F0D79-9317-48AE-AEDB-17120579F8E2/NDP451-KB2858728-x86-x64-AllOS-ENU.exe product_id: '{7DEBE4EB-6B40-3766-BB35-5CBBC385DA37}' arguments: '/q /norestart' state: present expected_return_code: [0, 666, 3010] ''' RETURN = r''' exit_code: description: See rc, this will be removed in favour of rc in Ansible 2.6. returned: change occured type: int sample: 0 log: description: The contents of the MSI log. returned: change occured and package is an MSI type: str sample: Installation completed successfully rc: description: The return code of the package process. returned: change occured type: int sample: 0 reboot_required: description: Whether a reboot is required to finalise package. This is set to true if the executable return code is 3010. returned: always type: bool sample: True restart_required: description: See reboot_required, this will be removed in favour of reboot_required in Ansible 2.6 returned: always type: bool sample: True stdout: description: The stdout stream of the package process. returned: failure during install or uninstall type: str sample: Installing program stderr: description: The stderr stream of the package process. returned: failure during install or uninstall type: str sample: Failed to install program '''
gpl-3.0
thaddeusdiamond/Social-Hierarchical-Learning
ext/protobuf/python/google/protobuf/internal/test_util.py
295
27103
# Protocol Buffers - Google's data interchange format # Copyright 2008 Google Inc. All rights reserved. # http://code.google.com/p/protobuf/ # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are # met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above # copyright notice, this list of conditions and the following disclaimer # in the documentation and/or other materials provided with the # distribution. # * Neither the name of Google Inc. nor the names of its # contributors may be used to endorse or promote products derived from # this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS # "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT # LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR # A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT # OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, # SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT # LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE # OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """Utilities for Python proto2 tests. This is intentionally modeled on C++ code in //google/protobuf/test_util.*. """ __author__ = 'robinson@google.com (Will Robinson)' import os.path from google.protobuf import unittest_import_pb2 from google.protobuf import unittest_pb2 def SetAllFields(message): """Sets every field in the message to a unique value. Args: message: A unittest_pb2.TestAllTypes instance. """ # # Optional fields. # message.optional_int32 = 101 message.optional_int64 = 102 message.optional_uint32 = 103 message.optional_uint64 = 104 message.optional_sint32 = 105 message.optional_sint64 = 106 message.optional_fixed32 = 107 message.optional_fixed64 = 108 message.optional_sfixed32 = 109 message.optional_sfixed64 = 110 message.optional_float = 111 message.optional_double = 112 message.optional_bool = True # TODO(robinson): Firmly spec out and test how # protos interact with unicode. One specific example: # what happens if we change the literal below to # u'115'? What *should* happen? Still some discussion # to finish with Kenton about bytes vs. strings # and forcing everything to be utf8. :-/ message.optional_string = '115' message.optional_bytes = '116' message.optionalgroup.a = 117 message.optional_nested_message.bb = 118 message.optional_foreign_message.c = 119 message.optional_import_message.d = 120 message.optional_nested_enum = unittest_pb2.TestAllTypes.BAZ message.optional_foreign_enum = unittest_pb2.FOREIGN_BAZ message.optional_import_enum = unittest_import_pb2.IMPORT_BAZ message.optional_string_piece = '124' message.optional_cord = '125' # # Repeated fields. # message.repeated_int32.append(201) message.repeated_int64.append(202) message.repeated_uint32.append(203) message.repeated_uint64.append(204) message.repeated_sint32.append(205) message.repeated_sint64.append(206) message.repeated_fixed32.append(207) message.repeated_fixed64.append(208) message.repeated_sfixed32.append(209) message.repeated_sfixed64.append(210) message.repeated_float.append(211) message.repeated_double.append(212) message.repeated_bool.append(True) message.repeated_string.append('215') message.repeated_bytes.append('216') message.repeatedgroup.add().a = 217 message.repeated_nested_message.add().bb = 218 message.repeated_foreign_message.add().c = 219 message.repeated_import_message.add().d = 220 message.repeated_nested_enum.append(unittest_pb2.TestAllTypes.BAR) message.repeated_foreign_enum.append(unittest_pb2.FOREIGN_BAR) message.repeated_import_enum.append(unittest_import_pb2.IMPORT_BAR) message.repeated_string_piece.append('224') message.repeated_cord.append('225') # Add a second one of each field. message.repeated_int32.append(301) message.repeated_int64.append(302) message.repeated_uint32.append(303) message.repeated_uint64.append(304) message.repeated_sint32.append(305) message.repeated_sint64.append(306) message.repeated_fixed32.append(307) message.repeated_fixed64.append(308) message.repeated_sfixed32.append(309) message.repeated_sfixed64.append(310) message.repeated_float.append(311) message.repeated_double.append(312) message.repeated_bool.append(False) message.repeated_string.append('315') message.repeated_bytes.append('316') message.repeatedgroup.add().a = 317 message.repeated_nested_message.add().bb = 318 message.repeated_foreign_message.add().c = 319 message.repeated_import_message.add().d = 320 message.repeated_nested_enum.append(unittest_pb2.TestAllTypes.BAZ) message.repeated_foreign_enum.append(unittest_pb2.FOREIGN_BAZ) message.repeated_import_enum.append(unittest_import_pb2.IMPORT_BAZ) message.repeated_string_piece.append('324') message.repeated_cord.append('325') # # Fields that have defaults. # message.default_int32 = 401 message.default_int64 = 402 message.default_uint32 = 403 message.default_uint64 = 404 message.default_sint32 = 405 message.default_sint64 = 406 message.default_fixed32 = 407 message.default_fixed64 = 408 message.default_sfixed32 = 409 message.default_sfixed64 = 410 message.default_float = 411 message.default_double = 412 message.default_bool = False message.default_string = '415' message.default_bytes = '416' message.default_nested_enum = unittest_pb2.TestAllTypes.FOO message.default_foreign_enum = unittest_pb2.FOREIGN_FOO message.default_import_enum = unittest_import_pb2.IMPORT_FOO message.default_string_piece = '424' message.default_cord = '425' def SetAllExtensions(message): """Sets every extension in the message to a unique value. Args: message: A unittest_pb2.TestAllExtensions instance. """ extensions = message.Extensions pb2 = unittest_pb2 import_pb2 = unittest_import_pb2 # # Optional fields. # extensions[pb2.optional_int32_extension] = 101 extensions[pb2.optional_int64_extension] = 102 extensions[pb2.optional_uint32_extension] = 103 extensions[pb2.optional_uint64_extension] = 104 extensions[pb2.optional_sint32_extension] = 105 extensions[pb2.optional_sint64_extension] = 106 extensions[pb2.optional_fixed32_extension] = 107 extensions[pb2.optional_fixed64_extension] = 108 extensions[pb2.optional_sfixed32_extension] = 109 extensions[pb2.optional_sfixed64_extension] = 110 extensions[pb2.optional_float_extension] = 111 extensions[pb2.optional_double_extension] = 112 extensions[pb2.optional_bool_extension] = True extensions[pb2.optional_string_extension] = '115' extensions[pb2.optional_bytes_extension] = '116' extensions[pb2.optionalgroup_extension].a = 117 extensions[pb2.optional_nested_message_extension].bb = 118 extensions[pb2.optional_foreign_message_extension].c = 119 extensions[pb2.optional_import_message_extension].d = 120 extensions[pb2.optional_nested_enum_extension] = pb2.TestAllTypes.BAZ extensions[pb2.optional_nested_enum_extension] = pb2.TestAllTypes.BAZ extensions[pb2.optional_foreign_enum_extension] = pb2.FOREIGN_BAZ extensions[pb2.optional_import_enum_extension] = import_pb2.IMPORT_BAZ extensions[pb2.optional_string_piece_extension] = '124' extensions[pb2.optional_cord_extension] = '125' # # Repeated fields. # extensions[pb2.repeated_int32_extension].append(201) extensions[pb2.repeated_int64_extension].append(202) extensions[pb2.repeated_uint32_extension].append(203) extensions[pb2.repeated_uint64_extension].append(204) extensions[pb2.repeated_sint32_extension].append(205) extensions[pb2.repeated_sint64_extension].append(206) extensions[pb2.repeated_fixed32_extension].append(207) extensions[pb2.repeated_fixed64_extension].append(208) extensions[pb2.repeated_sfixed32_extension].append(209) extensions[pb2.repeated_sfixed64_extension].append(210) extensions[pb2.repeated_float_extension].append(211) extensions[pb2.repeated_double_extension].append(212) extensions[pb2.repeated_bool_extension].append(True) extensions[pb2.repeated_string_extension].append('215') extensions[pb2.repeated_bytes_extension].append('216') extensions[pb2.repeatedgroup_extension].add().a = 217 extensions[pb2.repeated_nested_message_extension].add().bb = 218 extensions[pb2.repeated_foreign_message_extension].add().c = 219 extensions[pb2.repeated_import_message_extension].add().d = 220 extensions[pb2.repeated_nested_enum_extension].append(pb2.TestAllTypes.BAR) extensions[pb2.repeated_foreign_enum_extension].append(pb2.FOREIGN_BAR) extensions[pb2.repeated_import_enum_extension].append(import_pb2.IMPORT_BAR) extensions[pb2.repeated_string_piece_extension].append('224') extensions[pb2.repeated_cord_extension].append('225') # Append a second one of each field. extensions[pb2.repeated_int32_extension].append(301) extensions[pb2.repeated_int64_extension].append(302) extensions[pb2.repeated_uint32_extension].append(303) extensions[pb2.repeated_uint64_extension].append(304) extensions[pb2.repeated_sint32_extension].append(305) extensions[pb2.repeated_sint64_extension].append(306) extensions[pb2.repeated_fixed32_extension].append(307) extensions[pb2.repeated_fixed64_extension].append(308) extensions[pb2.repeated_sfixed32_extension].append(309) extensions[pb2.repeated_sfixed64_extension].append(310) extensions[pb2.repeated_float_extension].append(311) extensions[pb2.repeated_double_extension].append(312) extensions[pb2.repeated_bool_extension].append(False) extensions[pb2.repeated_string_extension].append('315') extensions[pb2.repeated_bytes_extension].append('316') extensions[pb2.repeatedgroup_extension].add().a = 317 extensions[pb2.repeated_nested_message_extension].add().bb = 318 extensions[pb2.repeated_foreign_message_extension].add().c = 319 extensions[pb2.repeated_import_message_extension].add().d = 320 extensions[pb2.repeated_nested_enum_extension].append(pb2.TestAllTypes.BAZ) extensions[pb2.repeated_foreign_enum_extension].append(pb2.FOREIGN_BAZ) extensions[pb2.repeated_import_enum_extension].append(import_pb2.IMPORT_BAZ) extensions[pb2.repeated_string_piece_extension].append('324') extensions[pb2.repeated_cord_extension].append('325') # # Fields with defaults. # extensions[pb2.default_int32_extension] = 401 extensions[pb2.default_int64_extension] = 402 extensions[pb2.default_uint32_extension] = 403 extensions[pb2.default_uint64_extension] = 404 extensions[pb2.default_sint32_extension] = 405 extensions[pb2.default_sint64_extension] = 406 extensions[pb2.default_fixed32_extension] = 407 extensions[pb2.default_fixed64_extension] = 408 extensions[pb2.default_sfixed32_extension] = 409 extensions[pb2.default_sfixed64_extension] = 410 extensions[pb2.default_float_extension] = 411 extensions[pb2.default_double_extension] = 412 extensions[pb2.default_bool_extension] = False extensions[pb2.default_string_extension] = '415' extensions[pb2.default_bytes_extension] = '416' extensions[pb2.default_nested_enum_extension] = pb2.TestAllTypes.FOO extensions[pb2.default_foreign_enum_extension] = pb2.FOREIGN_FOO extensions[pb2.default_import_enum_extension] = import_pb2.IMPORT_FOO extensions[pb2.default_string_piece_extension] = '424' extensions[pb2.default_cord_extension] = '425' def SetAllFieldsAndExtensions(message): """Sets every field and extension in the message to a unique value. Args: message: A unittest_pb2.TestAllExtensions message. """ message.my_int = 1 message.my_string = 'foo' message.my_float = 1.0 message.Extensions[unittest_pb2.my_extension_int] = 23 message.Extensions[unittest_pb2.my_extension_string] = 'bar' def ExpectAllFieldsAndExtensionsInOrder(serialized): """Ensures that serialized is the serialization we expect for a message filled with SetAllFieldsAndExtensions(). (Specifically, ensures that the serialization is in canonical, tag-number order). """ my_extension_int = unittest_pb2.my_extension_int my_extension_string = unittest_pb2.my_extension_string expected_strings = [] message = unittest_pb2.TestFieldOrderings() message.my_int = 1 # Field 1. expected_strings.append(message.SerializeToString()) message.Clear() message.Extensions[my_extension_int] = 23 # Field 5. expected_strings.append(message.SerializeToString()) message.Clear() message.my_string = 'foo' # Field 11. expected_strings.append(message.SerializeToString()) message.Clear() message.Extensions[my_extension_string] = 'bar' # Field 50. expected_strings.append(message.SerializeToString()) message.Clear() message.my_float = 1.0 expected_strings.append(message.SerializeToString()) message.Clear() expected = ''.join(expected_strings) if expected != serialized: raise ValueError('Expected %r, found %r' % (expected, serialized)) def ExpectAllFieldsSet(test_case, message): """Check all fields for correct values have after Set*Fields() is called.""" test_case.assertTrue(message.HasField('optional_int32')) test_case.assertTrue(message.HasField('optional_int64')) test_case.assertTrue(message.HasField('optional_uint32')) test_case.assertTrue(message.HasField('optional_uint64')) test_case.assertTrue(message.HasField('optional_sint32')) test_case.assertTrue(message.HasField('optional_sint64')) test_case.assertTrue(message.HasField('optional_fixed32')) test_case.assertTrue(message.HasField('optional_fixed64')) test_case.assertTrue(message.HasField('optional_sfixed32')) test_case.assertTrue(message.HasField('optional_sfixed64')) test_case.assertTrue(message.HasField('optional_float')) test_case.assertTrue(message.HasField('optional_double')) test_case.assertTrue(message.HasField('optional_bool')) test_case.assertTrue(message.HasField('optional_string')) test_case.assertTrue(message.HasField('optional_bytes')) test_case.assertTrue(message.HasField('optionalgroup')) test_case.assertTrue(message.HasField('optional_nested_message')) test_case.assertTrue(message.HasField('optional_foreign_message')) test_case.assertTrue(message.HasField('optional_import_message')) test_case.assertTrue(message.optionalgroup.HasField('a')) test_case.assertTrue(message.optional_nested_message.HasField('bb')) test_case.assertTrue(message.optional_foreign_message.HasField('c')) test_case.assertTrue(message.optional_import_message.HasField('d')) test_case.assertTrue(message.HasField('optional_nested_enum')) test_case.assertTrue(message.HasField('optional_foreign_enum')) test_case.assertTrue(message.HasField('optional_import_enum')) test_case.assertTrue(message.HasField('optional_string_piece')) test_case.assertTrue(message.HasField('optional_cord')) test_case.assertEqual(101, message.optional_int32) test_case.assertEqual(102, message.optional_int64) test_case.assertEqual(103, message.optional_uint32) test_case.assertEqual(104, message.optional_uint64) test_case.assertEqual(105, message.optional_sint32) test_case.assertEqual(106, message.optional_sint64) test_case.assertEqual(107, message.optional_fixed32) test_case.assertEqual(108, message.optional_fixed64) test_case.assertEqual(109, message.optional_sfixed32) test_case.assertEqual(110, message.optional_sfixed64) test_case.assertEqual(111, message.optional_float) test_case.assertEqual(112, message.optional_double) test_case.assertEqual(True, message.optional_bool) test_case.assertEqual('115', message.optional_string) test_case.assertEqual('116', message.optional_bytes) test_case.assertEqual(117, message.optionalgroup.a) test_case.assertEqual(118, message.optional_nested_message.bb) test_case.assertEqual(119, message.optional_foreign_message.c) test_case.assertEqual(120, message.optional_import_message.d) test_case.assertEqual(unittest_pb2.TestAllTypes.BAZ, message.optional_nested_enum) test_case.assertEqual(unittest_pb2.FOREIGN_BAZ, message.optional_foreign_enum) test_case.assertEqual(unittest_import_pb2.IMPORT_BAZ, message.optional_import_enum) # ----------------------------------------------------------------- test_case.assertEqual(2, len(message.repeated_int32)) test_case.assertEqual(2, len(message.repeated_int64)) test_case.assertEqual(2, len(message.repeated_uint32)) test_case.assertEqual(2, len(message.repeated_uint64)) test_case.assertEqual(2, len(message.repeated_sint32)) test_case.assertEqual(2, len(message.repeated_sint64)) test_case.assertEqual(2, len(message.repeated_fixed32)) test_case.assertEqual(2, len(message.repeated_fixed64)) test_case.assertEqual(2, len(message.repeated_sfixed32)) test_case.assertEqual(2, len(message.repeated_sfixed64)) test_case.assertEqual(2, len(message.repeated_float)) test_case.assertEqual(2, len(message.repeated_double)) test_case.assertEqual(2, len(message.repeated_bool)) test_case.assertEqual(2, len(message.repeated_string)) test_case.assertEqual(2, len(message.repeated_bytes)) test_case.assertEqual(2, len(message.repeatedgroup)) test_case.assertEqual(2, len(message.repeated_nested_message)) test_case.assertEqual(2, len(message.repeated_foreign_message)) test_case.assertEqual(2, len(message.repeated_import_message)) test_case.assertEqual(2, len(message.repeated_nested_enum)) test_case.assertEqual(2, len(message.repeated_foreign_enum)) test_case.assertEqual(2, len(message.repeated_import_enum)) test_case.assertEqual(2, len(message.repeated_string_piece)) test_case.assertEqual(2, len(message.repeated_cord)) test_case.assertEqual(201, message.repeated_int32[0]) test_case.assertEqual(202, message.repeated_int64[0]) test_case.assertEqual(203, message.repeated_uint32[0]) test_case.assertEqual(204, message.repeated_uint64[0]) test_case.assertEqual(205, message.repeated_sint32[0]) test_case.assertEqual(206, message.repeated_sint64[0]) test_case.assertEqual(207, message.repeated_fixed32[0]) test_case.assertEqual(208, message.repeated_fixed64[0]) test_case.assertEqual(209, message.repeated_sfixed32[0]) test_case.assertEqual(210, message.repeated_sfixed64[0]) test_case.assertEqual(211, message.repeated_float[0]) test_case.assertEqual(212, message.repeated_double[0]) test_case.assertEqual(True, message.repeated_bool[0]) test_case.assertEqual('215', message.repeated_string[0]) test_case.assertEqual('216', message.repeated_bytes[0]) test_case.assertEqual(217, message.repeatedgroup[0].a) test_case.assertEqual(218, message.repeated_nested_message[0].bb) test_case.assertEqual(219, message.repeated_foreign_message[0].c) test_case.assertEqual(220, message.repeated_import_message[0].d) test_case.assertEqual(unittest_pb2.TestAllTypes.BAR, message.repeated_nested_enum[0]) test_case.assertEqual(unittest_pb2.FOREIGN_BAR, message.repeated_foreign_enum[0]) test_case.assertEqual(unittest_import_pb2.IMPORT_BAR, message.repeated_import_enum[0]) test_case.assertEqual(301, message.repeated_int32[1]) test_case.assertEqual(302, message.repeated_int64[1]) test_case.assertEqual(303, message.repeated_uint32[1]) test_case.assertEqual(304, message.repeated_uint64[1]) test_case.assertEqual(305, message.repeated_sint32[1]) test_case.assertEqual(306, message.repeated_sint64[1]) test_case.assertEqual(307, message.repeated_fixed32[1]) test_case.assertEqual(308, message.repeated_fixed64[1]) test_case.assertEqual(309, message.repeated_sfixed32[1]) test_case.assertEqual(310, message.repeated_sfixed64[1]) test_case.assertEqual(311, message.repeated_float[1]) test_case.assertEqual(312, message.repeated_double[1]) test_case.assertEqual(False, message.repeated_bool[1]) test_case.assertEqual('315', message.repeated_string[1]) test_case.assertEqual('316', message.repeated_bytes[1]) test_case.assertEqual(317, message.repeatedgroup[1].a) test_case.assertEqual(318, message.repeated_nested_message[1].bb) test_case.assertEqual(319, message.repeated_foreign_message[1].c) test_case.assertEqual(320, message.repeated_import_message[1].d) test_case.assertEqual(unittest_pb2.TestAllTypes.BAZ, message.repeated_nested_enum[1]) test_case.assertEqual(unittest_pb2.FOREIGN_BAZ, message.repeated_foreign_enum[1]) test_case.assertEqual(unittest_import_pb2.IMPORT_BAZ, message.repeated_import_enum[1]) # ----------------------------------------------------------------- test_case.assertTrue(message.HasField('default_int32')) test_case.assertTrue(message.HasField('default_int64')) test_case.assertTrue(message.HasField('default_uint32')) test_case.assertTrue(message.HasField('default_uint64')) test_case.assertTrue(message.HasField('default_sint32')) test_case.assertTrue(message.HasField('default_sint64')) test_case.assertTrue(message.HasField('default_fixed32')) test_case.assertTrue(message.HasField('default_fixed64')) test_case.assertTrue(message.HasField('default_sfixed32')) test_case.assertTrue(message.HasField('default_sfixed64')) test_case.assertTrue(message.HasField('default_float')) test_case.assertTrue(message.HasField('default_double')) test_case.assertTrue(message.HasField('default_bool')) test_case.assertTrue(message.HasField('default_string')) test_case.assertTrue(message.HasField('default_bytes')) test_case.assertTrue(message.HasField('default_nested_enum')) test_case.assertTrue(message.HasField('default_foreign_enum')) test_case.assertTrue(message.HasField('default_import_enum')) test_case.assertEqual(401, message.default_int32) test_case.assertEqual(402, message.default_int64) test_case.assertEqual(403, message.default_uint32) test_case.assertEqual(404, message.default_uint64) test_case.assertEqual(405, message.default_sint32) test_case.assertEqual(406, message.default_sint64) test_case.assertEqual(407, message.default_fixed32) test_case.assertEqual(408, message.default_fixed64) test_case.assertEqual(409, message.default_sfixed32) test_case.assertEqual(410, message.default_sfixed64) test_case.assertEqual(411, message.default_float) test_case.assertEqual(412, message.default_double) test_case.assertEqual(False, message.default_bool) test_case.assertEqual('415', message.default_string) test_case.assertEqual('416', message.default_bytes) test_case.assertEqual(unittest_pb2.TestAllTypes.FOO, message.default_nested_enum) test_case.assertEqual(unittest_pb2.FOREIGN_FOO, message.default_foreign_enum) test_case.assertEqual(unittest_import_pb2.IMPORT_FOO, message.default_import_enum) def GoldenFile(filename): """Finds the given golden file and returns a file object representing it.""" # Search up the directory tree looking for the C++ protobuf source code. path = '.' while os.path.exists(path): if os.path.exists(os.path.join(path, 'src/google/protobuf')): # Found it. Load the golden file from the testdata directory. full_path = os.path.join(path, 'src/google/protobuf/testdata', filename) return open(full_path, 'rb') path = os.path.join(path, '..') raise RuntimeError( 'Could not find golden files. This test must be run from within the ' 'protobuf source package so that it can read test data files from the ' 'C++ source tree.') def SetAllPackedFields(message): """Sets every field in the message to a unique value. Args: message: A unittest_pb2.TestPackedTypes instance. """ message.packed_int32.extend([601, 701]) message.packed_int64.extend([602, 702]) message.packed_uint32.extend([603, 703]) message.packed_uint64.extend([604, 704]) message.packed_sint32.extend([605, 705]) message.packed_sint64.extend([606, 706]) message.packed_fixed32.extend([607, 707]) message.packed_fixed64.extend([608, 708]) message.packed_sfixed32.extend([609, 709]) message.packed_sfixed64.extend([610, 710]) message.packed_float.extend([611.0, 711.0]) message.packed_double.extend([612.0, 712.0]) message.packed_bool.extend([True, False]) message.packed_enum.extend([unittest_pb2.FOREIGN_BAR, unittest_pb2.FOREIGN_BAZ]) def SetAllPackedExtensions(message): """Sets every extension in the message to a unique value. Args: message: A unittest_pb2.TestPackedExtensions instance. """ extensions = message.Extensions pb2 = unittest_pb2 extensions[pb2.packed_int32_extension].extend([601, 701]) extensions[pb2.packed_int64_extension].extend([602, 702]) extensions[pb2.packed_uint32_extension].extend([603, 703]) extensions[pb2.packed_uint64_extension].extend([604, 704]) extensions[pb2.packed_sint32_extension].extend([605, 705]) extensions[pb2.packed_sint64_extension].extend([606, 706]) extensions[pb2.packed_fixed32_extension].extend([607, 707]) extensions[pb2.packed_fixed64_extension].extend([608, 708]) extensions[pb2.packed_sfixed32_extension].extend([609, 709]) extensions[pb2.packed_sfixed64_extension].extend([610, 710]) extensions[pb2.packed_float_extension].extend([611.0, 711.0]) extensions[pb2.packed_double_extension].extend([612.0, 712.0]) extensions[pb2.packed_bool_extension].extend([True, False]) extensions[pb2.packed_enum_extension].extend([unittest_pb2.FOREIGN_BAR, unittest_pb2.FOREIGN_BAZ]) def SetAllUnpackedFields(message): """Sets every field in the message to a unique value. Args: message: A unittest_pb2.TestUnpackedTypes instance. """ message.unpacked_int32.extend([601, 701]) message.unpacked_int64.extend([602, 702]) message.unpacked_uint32.extend([603, 703]) message.unpacked_uint64.extend([604, 704]) message.unpacked_sint32.extend([605, 705]) message.unpacked_sint64.extend([606, 706]) message.unpacked_fixed32.extend([607, 707]) message.unpacked_fixed64.extend([608, 708]) message.unpacked_sfixed32.extend([609, 709]) message.unpacked_sfixed64.extend([610, 710]) message.unpacked_float.extend([611.0, 711.0]) message.unpacked_double.extend([612.0, 712.0]) message.unpacked_bool.extend([True, False]) message.unpacked_enum.extend([unittest_pb2.FOREIGN_BAR, unittest_pb2.FOREIGN_BAZ])
mit
chatelak/RMG-Py
rmgpy/solver/__init__.py
11
1557
#!/usr/bin/env python # -*- coding: utf-8 -*- ################################################################################ # # RMG - Reaction Mechanism Generator # # Copyright (c) 2002-2010 Prof. William H. Green (whgreen@mit.edu) and the # RMG Team (rmg_dev@mit.edu) # # Permission is hereby granted, free of charge, to any person obtaining a # copy of this software and associated documentation files (the 'Software'), # to deal in the Software without restriction, including without limitation # the rights to use, copy, modify, merge, publish, distribute, sublicense, # and/or sell copies of the Software, and to permit persons to whom the # Software is furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED 'AS IS', WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER # DEALINGS IN THE SOFTWARE. # ################################################################################ from .base import ReactionSystem, TerminationTime, TerminationConversion from .simple import SimpleReactor
mit
danforthcenter/plantcv
plantcv/plantcv/photosynthesis/analyze_fvfm.py
2
5529
# Fluorescence Analysis import os import cv2 import numpy as np import pandas as pd from plotnine import ggplot, geom_label, aes, geom_line from plantcv.plantcv import print_image from plantcv.plantcv import plot_image from plantcv.plantcv import fatal_error from plantcv.plantcv import params from plantcv.plantcv import outputs def analyze_fvfm(fdark, fmin, fmax, mask, bins=256, label="default"): """Analyze PSII camera images. Inputs: fdark = grayscale fdark image fmin = grayscale fmin image fmax = grayscale fmax image mask = mask of plant (binary, single channel) bins = number of bins (1 to 256 for 8-bit; 1 to 65,536 for 16-bit; default is 256) label = optional label parameter, modifies the variable name of observations recorded Returns: analysis_images = list of images (fv image and fvfm histogram image) :param fdark: numpy.ndarray :param fmin: numpy.ndarray :param fmax: numpy.ndarray :param mask: numpy.ndarray :param bins: int :param label: str :return analysis_images: numpy.ndarray """ # Auto-increment the device counter params.device += 1 # Check that fdark, fmin, and fmax are grayscale (single channel) if not all(len(np.shape(i)) == 2 for i in [fdark, fmin, fmax]): fatal_error("The fdark, fmin, and fmax images must be grayscale images.") # QC Fdark Image fdark_mask = cv2.bitwise_and(fdark, fdark, mask=mask) if np.amax(fdark_mask) > 2000: qc_fdark = False else: qc_fdark = True # Mask Fmin and Fmax Image fmin_mask = cv2.bitwise_and(fmin, fmin, mask=mask) fmax_mask = cv2.bitwise_and(fmax, fmax, mask=mask) # Calculate Fvariable, where Fv = Fmax - Fmin (masked) fv = np.subtract(fmax_mask, fmin_mask) # When Fmin is greater than Fmax, a negative value is returned. # Because the data type is unsigned integers, negative values roll over, resulting in nonsensical values # Wherever Fmin is greater than Fmax, set Fv to zero fv[np.where(fmax_mask < fmin_mask)] = 0 analysis_images = [] # Calculate Fv/Fm (Fvariable / Fmax) where Fmax is greater than zero # By definition above, wherever Fmax is zero, Fvariable will also be zero # To calculate the divisions properly we need to change from unit16 to float64 data types fvfm = fv.astype(np.float64) analysis_images.append(fvfm) fmax_flt = fmax_mask.astype(np.float64) fvfm[np.where(fmax_mask > 0)] /= fmax_flt[np.where(fmax_mask > 0)] # Calculate the median Fv/Fm value for non-zero pixels fvfm_median = np.median(fvfm[np.where(fvfm > 0)]) # Calculate the histogram of Fv/Fm non-zero values fvfm_hist, fvfm_bins = np.histogram(fvfm[np.where(fvfm > 0)], bins, range=(0, 1)) # fvfm_bins is a bins + 1 length list of bin endpoints, so we need to calculate bin midpoints so that # the we have a one-to-one list of x (FvFm) and y (frequency) values. # To do this we add half the bin width to each lower bin edge x-value midpoints = fvfm_bins[:-1] + 0.5 * np.diff(fvfm_bins) # Calculate which non-zero bin has the maximum Fv/Fm value max_bin = midpoints[np.argmax(fvfm_hist)] # Create a dataframe dataset = pd.DataFrame({'Plant Pixels': fvfm_hist, 'Fv/Fm': midpoints}) # Make the histogram figure using plotnine fvfm_hist_fig = (ggplot(data=dataset, mapping=aes(x='Fv/Fm', y='Plant Pixels')) + geom_line(color='green', show_legend=True) + geom_label(label='Peak Bin Value: ' + str(max_bin), x=.15, y=205, size=8, color='green')) analysis_images.append(fvfm_hist_fig) if params.debug == 'print': print_image(fmin_mask, os.path.join(params.debug_outdir, str(params.device) + '_fmin_mask.png')) print_image(fmax_mask, os.path.join(params.debug_outdir, str(params.device) + '_fmax_mask.png')) print_image(fv, os.path.join(params.debug_outdir, str(params.device) + '_fv_convert.png')) fvfm_hist_fig.save(os.path.join(params.debug_outdir, str(params.device) + '_fv_hist.png'), verbose=False) elif params.debug == 'plot': plot_image(fmin_mask, cmap='gray') plot_image(fmax_mask, cmap='gray') plot_image(fv, cmap='gray') print(fvfm_hist_fig) outputs.add_observation(sample=label, variable='fvfm_hist', trait='Fv/Fm frequencies', method='plantcv.plantcv.fluor_fvfm', scale='none', datatype=list, value=fvfm_hist.tolist(), label=np.around(midpoints, decimals=len(str(bins))).tolist()) outputs.add_observation(sample=label, variable='fvfm_hist_peak', trait='peak Fv/Fm value', method='plantcv.plantcv.fluor_fvfm', scale='none', datatype=float, value=float(max_bin), label='none') outputs.add_observation(sample=label, variable='fvfm_median', trait='Fv/Fm median', method='plantcv.plantcv.fluor_fvfm', scale='none', datatype=float, value=float(np.around(fvfm_median, decimals=4)), label='none') outputs.add_observation(sample=label, variable='fdark_passed_qc', trait='Fdark passed QC', method='plantcv.plantcv.fluor_fvfm', scale='none', datatype=bool, value=qc_fdark, label='none') # Store images outputs.images.append(analysis_images) return analysis_images
mit
vicky2135/lucious
oscar/lib/python2.7/site-packages/urllib3/exceptions.py
223
6603
from __future__ import absolute_import from .packages.six.moves.http_client import ( IncompleteRead as httplib_IncompleteRead ) # Base Exceptions class HTTPError(Exception): "Base exception used by this module." pass class HTTPWarning(Warning): "Base warning used by this module." pass class PoolError(HTTPError): "Base exception for errors caused within a pool." def __init__(self, pool, message): self.pool = pool HTTPError.__init__(self, "%s: %s" % (pool, message)) def __reduce__(self): # For pickling purposes. return self.__class__, (None, None) class RequestError(PoolError): "Base exception for PoolErrors that have associated URLs." def __init__(self, pool, url, message): self.url = url PoolError.__init__(self, pool, message) def __reduce__(self): # For pickling purposes. return self.__class__, (None, self.url, None) class SSLError(HTTPError): "Raised when SSL certificate fails in an HTTPS connection." pass class ProxyError(HTTPError): "Raised when the connection to a proxy fails." pass class DecodeError(HTTPError): "Raised when automatic decoding based on Content-Type fails." pass class ProtocolError(HTTPError): "Raised when something unexpected happens mid-request/response." pass #: Renamed to ProtocolError but aliased for backwards compatibility. ConnectionError = ProtocolError # Leaf Exceptions class MaxRetryError(RequestError): """Raised when the maximum number of retries is exceeded. :param pool: The connection pool :type pool: :class:`~urllib3.connectionpool.HTTPConnectionPool` :param string url: The requested Url :param exceptions.Exception reason: The underlying error """ def __init__(self, pool, url, reason=None): self.reason = reason message = "Max retries exceeded with url: %s (Caused by %r)" % ( url, reason) RequestError.__init__(self, pool, url, message) class HostChangedError(RequestError): "Raised when an existing pool gets a request for a foreign host." def __init__(self, pool, url, retries=3): message = "Tried to open a foreign host with url: %s" % url RequestError.__init__(self, pool, url, message) self.retries = retries class TimeoutStateError(HTTPError): """ Raised when passing an invalid state to a timeout """ pass class TimeoutError(HTTPError): """ Raised when a socket timeout error occurs. Catching this error will catch both :exc:`ReadTimeoutErrors <ReadTimeoutError>` and :exc:`ConnectTimeoutErrors <ConnectTimeoutError>`. """ pass class ReadTimeoutError(TimeoutError, RequestError): "Raised when a socket timeout occurs while receiving data from a server" pass # This timeout error does not have a URL attached and needs to inherit from the # base HTTPError class ConnectTimeoutError(TimeoutError): "Raised when a socket timeout occurs while connecting to a server" pass class NewConnectionError(ConnectTimeoutError, PoolError): "Raised when we fail to establish a new connection. Usually ECONNREFUSED." pass class EmptyPoolError(PoolError): "Raised when a pool runs out of connections and no more are allowed." pass class ClosedPoolError(PoolError): "Raised when a request enters a pool after the pool has been closed." pass class LocationValueError(ValueError, HTTPError): "Raised when there is something wrong with a given URL input." pass class LocationParseError(LocationValueError): "Raised when get_host or similar fails to parse the URL input." def __init__(self, location): message = "Failed to parse: %s" % location HTTPError.__init__(self, message) self.location = location class ResponseError(HTTPError): "Used as a container for an error reason supplied in a MaxRetryError." GENERIC_ERROR = 'too many error responses' SPECIFIC_ERROR = 'too many {status_code} error responses' class SecurityWarning(HTTPWarning): "Warned when perfoming security reducing actions" pass class SubjectAltNameWarning(SecurityWarning): "Warned when connecting to a host with a certificate missing a SAN." pass class InsecureRequestWarning(SecurityWarning): "Warned when making an unverified HTTPS request." pass class SystemTimeWarning(SecurityWarning): "Warned when system time is suspected to be wrong" pass class InsecurePlatformWarning(SecurityWarning): "Warned when certain SSL configuration is not available on a platform." pass class SNIMissingWarning(HTTPWarning): "Warned when making a HTTPS request without SNI available." pass class DependencyWarning(HTTPWarning): """ Warned when an attempt is made to import a module with missing optional dependencies. """ pass class ResponseNotChunked(ProtocolError, ValueError): "Response needs to be chunked in order to read it as chunks." pass class BodyNotHttplibCompatible(HTTPError): """ Body should be httplib.HTTPResponse like (have an fp attribute which returns raw chunks) for read_chunked(). """ pass class IncompleteRead(HTTPError, httplib_IncompleteRead): """ Response length doesn't match expected Content-Length Subclass of http_client.IncompleteRead to allow int value for `partial` to avoid creating large objects on streamed reads. """ def __init__(self, partial, expected): super(IncompleteRead, self).__init__(partial, expected) def __repr__(self): return ('IncompleteRead(%i bytes read, ' '%i more expected)' % (self.partial, self.expected)) class InvalidHeader(HTTPError): "The header provided was somehow invalid." pass class ProxySchemeUnknown(AssertionError, ValueError): "ProxyManager does not support the supplied scheme" # TODO(t-8ch): Stop inheriting from AssertionError in v2.0. def __init__(self, scheme): message = "Not supported proxy scheme %s" % scheme super(ProxySchemeUnknown, self).__init__(message) class HeaderParsingError(HTTPError): "Raised by assert_header_parsing, but we convert it to a log.warning statement." def __init__(self, defects, unparsed_data): message = '%s, unparsed data: %r' % (defects or 'Unknown', unparsed_data) super(HeaderParsingError, self).__init__(message) class UnrewindableBodyError(HTTPError): "urllib3 encountered an error when trying to rewind a body" pass
bsd-3-clause
tutengfei/zerorpc-python
tests/zmqbug.py
94
5183
# -*- coding: utf-8 -*- # Open Source Initiative OSI - The MIT License (MIT):Licensing # # The MIT License (MIT) # Copyright (c) 2012 DotCloud Inc (opensource@dotcloud.com) # # Permission is hereby granted, free of charge, to any person obtaining a copy of # this software and associated documentation files (the "Software"), to deal in # the Software without restriction, including without limitation the rights to # use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies # of the Software, and to permit persons to whom the Software is furnished to do # so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in all # copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. # # Based on https://github.com/traviscline/gevent-zeromq/blob/master/gevent_zeromq/core.py import zmq import gevent.event import gevent.core STOP_EVERYTHING = False class ZMQSocket(zmq.Socket): def __init__(self, context, socket_type): super(ZMQSocket, self).__init__(context, socket_type) on_state_changed_fd = self.getsockopt(zmq.FD) self._readable = gevent.event.Event() self._writable = gevent.event.Event() try: # gevent>=1.0 self._state_event = gevent.hub.get_hub().loop.io( on_state_changed_fd, gevent.core.READ) self._state_event.start(self._on_state_changed) except AttributeError: # gevent<1.0 self._state_event = gevent.core.read_event(on_state_changed_fd, self._on_state_changed, persist=True) def _on_state_changed(self, event=None, _evtype=None): if self.closed: self._writable.set() self._readable.set() return events = self.getsockopt(zmq.EVENTS) if events & zmq.POLLOUT: self._writable.set() if events & zmq.POLLIN: self._readable.set() def close(self): if not self.closed and getattr(self, '_state_event', None): try: # gevent>=1.0 self._state_event.stop() except AttributeError: # gevent<1.0 self._state_event.cancel() super(ZMQSocket, self).close() def send(self, data, flags=0, copy=True, track=False): if flags & zmq.NOBLOCK: return super(ZMQSocket, self).send(data, flags, copy, track) flags |= zmq.NOBLOCK while True: try: return super(ZMQSocket, self).send(data, flags, copy, track) except zmq.ZMQError, e: if e.errno != zmq.EAGAIN: raise self._writable.clear() self._writable.wait() def recv(self, flags=0, copy=True, track=False): if flags & zmq.NOBLOCK: return super(ZMQSocket, self).recv(flags, copy, track) flags |= zmq.NOBLOCK while True: try: return super(ZMQSocket, self).recv(flags, copy, track) except zmq.ZMQError, e: if e.errno != zmq.EAGAIN: raise self._readable.clear() while not self._readable.wait(timeout=10): events = self.getsockopt(zmq.EVENTS) if bool(events & zmq.POLLIN): print "here we go, nobody told me about new messages!" global STOP_EVERYTHING STOP_EVERYTHING = True raise gevent.GreenletExit() zmq_context = zmq.Context() def server(): socket = ZMQSocket(zmq_context, zmq.REP) socket.bind('ipc://zmqbug') class Cnt: responded = 0 cnt = Cnt() def responder(): while not STOP_EVERYTHING: msg = socket.recv() socket.send(msg) cnt.responded += 1 gevent.spawn(responder) while not STOP_EVERYTHING: print "cnt.responded=", cnt.responded gevent.sleep(0.5) def client(): socket = ZMQSocket(zmq_context, zmq.DEALER) socket.connect('ipc://zmqbug') class Cnt: recv = 0 send = 0 cnt = Cnt() def recvmsg(): while not STOP_EVERYTHING: socket.recv() socket.recv() cnt.recv += 1 def sendmsg(): while not STOP_EVERYTHING: socket.send('', flags=zmq.SNDMORE) socket.send('hello') cnt.send += 1 gevent.sleep(0) gevent.spawn(recvmsg) gevent.spawn(sendmsg) while not STOP_EVERYTHING: print "cnt.recv=", cnt.recv, "cnt.send=", cnt.send gevent.sleep(0.5) gevent.spawn(server) client()
mit
chachan/nodeshot
nodeshot/community/notifications/migrations/0001_initial.py
5
4082
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations import django.utils.timezone class Migration(migrations.Migration): dependencies = [ ] operations = [ migrations.CreateModel( name='Notification', fields=[ ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)), ('added', models.DateTimeField(default=django.utils.timezone.now, verbose_name='created on')), ('updated', models.DateTimeField(default=django.utils.timezone.now, verbose_name='updated on')), ('type', models.CharField(max_length=64, verbose_name='type', choices=[(b'node_created', 'node_created'), (b'node_status_changed', 'node_status_changed'), (b'node_own_status_changed', 'node_own_status_changed'), (b'node_deleted', 'node_deleted'), (b'custom', 'custom')])), ('object_id', models.PositiveIntegerField(null=True, blank=True)), ('text', models.CharField(max_length=120, verbose_name='text', blank=True)), ('is_read', models.BooleanField(default=False, verbose_name='read?')), ], options={ 'ordering': ('-id',), }, ), migrations.CreateModel( name='UserEmailNotificationSettings', fields=[ ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)), ('node_created', models.IntegerField(default=30, help_text='-1 (less than 0): disabled; 0: enabled for all; 1 (less than 0): enabled for those in the specified distance range (km)', verbose_name='node_created')), ('node_status_changed', models.IntegerField(default=30, help_text='-1 (less than 0): disabled; 0: enabled for all; 1 (less than 0): enabled for those in the specified distance range (km)', verbose_name='node_status_changed')), ('node_own_status_changed', models.BooleanField(default=True, verbose_name='node_own_status_changed')), ('node_deleted', models.IntegerField(default=30, help_text='-1 (less than 0): disabled; 0: enabled for all; 1 (less than 0): enabled for those in the specified distance range (km)', verbose_name='node_deleted')), ], options={ 'db_table': 'notifications_user_email_settings', 'verbose_name': 'user email notification settings', 'verbose_name_plural': 'user email notification settings', }, ), migrations.CreateModel( name='UserWebNotificationSettings', fields=[ ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)), ('node_created', models.IntegerField(default=30, help_text='-1 (less than 0): disabled; 0: enabled for all; 1 (less than 0): enabled for those in the specified distance range (km)', verbose_name='node_created')), ('node_status_changed', models.IntegerField(default=30, help_text='-1 (less than 0): disabled; 0: enabled for all; 1 (less than 0): enabled for those in the specified distance range (km)', verbose_name='node_status_changed')), ('node_own_status_changed', models.BooleanField(default=True, verbose_name='node_own_status_changed')), ('node_deleted', models.IntegerField(default=30, help_text='-1 (less than 0): disabled; 0: enabled for all; 1 (less than 0): enabled for those in the specified distance range (km)', verbose_name='node_deleted')), ], options={ 'db_table': 'notifications_user_web_settings', 'verbose_name': 'user web notification settings', 'verbose_name_plural': 'user web notification settings', }, ), ]
gpl-3.0
waveform80/lars
lars/datatypes/url.py
1
10354
# vim: set et sw=4 sts=4 fileencoding=utf-8: # # Copyright (c) 2013-2017 Dave Jones <dave@waveform.org.uk> # Copyright (c) 2013 Mime Consulting Ltd. <info@mimeconsulting.co.uk> # All rights reserved. # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE # SOFTWARE. """ Defines the URL parsing specific parts of :mod:`lars.datatypes`. """ from __future__ import ( unicode_literals, absolute_import, print_function, division, ) from collections import namedtuple try: from urllib import parse except ImportError: import urlparse as parse from .ipaddress import hostname str = type('') # pylint: disable=redefined-builtin,invalid-name def path(s): """ Returns a :class:`Path` object for the given string. :param str s: The string containing the path to parse :returns: A :class:`Path` object representing the path """ i = s.rfind('/') + 1 dirname, basename = s[:i], s[i:] if dirname and dirname != '/' * len(dirname): dirname = dirname.rstrip('/') i = basename.rfind('.') if i > 0: ext = basename[i:] else: ext = '' return Path(dirname, basename, ext) def url(s): """ Returns a :class:`Url` object for the given string. :param str s: The string containing the URL to parse :returns: A :class:`Url` tuple representing the URL """ return Url(*parse.urlparse(s)) def request(s): """ Returns a :class:`Request` object for the given string. :param str s: The string containing the request line to parse :returns: A :class:`Request` tuple representing the request line """ try: method, s = s.split(' ', 1) except ValueError: raise ValueError('Request line is missing a space separated method') try: s, protocol = s.rsplit(' ', 1) except ValueError: raise ValueError('Request line is missing a space separated protocol') s = s.strip() if not s: raise ValueError('Request line URL cannot be blank') return Request(method, url(s) if s != '*' else None, protocol) class Path(namedtuple('Path', 'dirname basename ext')): """ Represents a path. This type is returned by the :func:`path` function and represents a path in POSIX format (forward slash separators and no drive portion). It is used to represent the path portion of URLs and provides attributes for extracting parts of the path there-in. The original path can be obtained as a string by asking for the string conversion of this class, like so:: p = datatypes.path('/foo/bar/baz.ext') assert p.dirname == '/foo/bar' assert p.basename == 'baz.ext' assert str(p) == '/foo/bar/baz.ext' .. attribute:: dirname A string containing all of the path except the basename at the end .. attribute:: basename A string containing the basename (filename and extension) at the end of the path .. attribute:: ext A string containing the filename's extension (including the leading dot) """ __slots__ = () @property def dirs(self): """ Returns a sequence of the directories making up :attr:`dirname` """ return [d for d in self.dirname.split('/') if d] @property def basename_no_ext(self): """ Returns a string containing basename with the extension removed (including the final dot separator). """ if self.ext: return self.basename[:-len(self.ext)] else: return self.basename @property def isabs(self): """ Returns True if the path is absolute (dirname begins with one or more forward slashes). """ return self.dirname.startswith('/') def join(self, *paths): """ Joins this path with the specified parts, returning a new :class:`Path` object. :param \\*paths: The parts to append to this path :returns: A new :class:`Path` object representing the extended path """ # pylint: disable=invalid-name result = str(self) for p in paths: if not isinstance(p, str): p = str(p) # Strip doubled slashes? Or leave this to normpath? if p.startswith('/'): result = p elif not result or result.endswith('/'): result += p else: result += '/' + p return path(result) def __str__(self): result = self.dirname if not result or result.endswith('/'): return result + self.basename else: return result + '/' + self.basename # This is rather hackish; in Python 2.x, urlparse.ResultMixin provides # functionality for extracting username, password, hostname and port from a # parsed URL. In Python 3 this changed to ResultBase, then to a whole bunch of # undocumented classes (split between strings and bytes) with ResultBase as an # alias try: _ResultMixin = parse.ResultBase # pylint: disable=invalid-name except AttributeError: _ResultMixin = parse.ResultMixin # pylint: disable=invalid-name class Url(namedtuple('Url', ('scheme', 'netloc', 'path_str', 'params', 'query_str', 'fragment')), _ResultMixin): """ Represents a URL. This type is returned by the :func:`url` function and represents the parts of the URL. You can obtain the original URL as a string by requesting the string conversion of this class, for example:: >>> u = datatypes.url('http://foo/bar/baz') >>> print u.scheme http >>> print u.hostname foo >>> print str(u) http://foo/bar/baz .. attribute:: scheme The scheme of the URL, before the first ``:`` .. attribute:: netloc The "network location" of the URL, comprising the hostname and port (separated by a colon), and historically the username and password (prefixed to the hostname and separated with an ampersand) .. attribute:: path_str The path of the URL from the first slash after the network location .. attribute:: path The path of the URL, parsed into a tuple which splits out the directory, filename, and extension:: >>> u = datatypes.url('foo/bar/baz.html') >>> u.path Path(dirname='foo/bar', basename='baz.html', ext='.html') >>> u.path.isabs False .. attribute:: params The parameters of the URL .. attribute:: query_str The query string of the URL from the first question-mark in the path .. attribute:: query The query string, parsed into a mapping of keys to lists of values. For example:: >>> u = datatypes.url('foo/bar?a=1&a=2&b=3&c=') >>> print u.query {'a': ['1', '2'], 'c': [''], 'b': ['3']} >>> print 'a' in u.query True .. attribute:: fragment The fragment of the URL from the last hash-mark to the end of the URL Additionally, the following attributes can be used to separate out the various parts of the :attr:`netloc` attribute: .. attribute:: username The username (historical, rare to see this used on the modern web) .. attribute:: password The password (historical, almost unheard of on the modern web as it's extremely insecure to include credentials in the URL) .. attribute:: hostname The hostname from the network location. This attribute returns a :class:`Hostname` object which can be used to resolve the hostname into an IP address if required. .. attribute:: port The optional network port """ __slots__ = () def geturl(self): """ Return the URL as a string string. """ return parse.urlunparse(self) def __str__(self): return self.geturl() @property def hostname(self): return hostname(super(Url, self).hostname) @property def query(self): # pylint: disable=missing-docstring return parse.parse_qs(self.query_str, keep_blank_values=True) @property def path(self): # pylint: disable=missing-docstring return path(self.path_str) class Request(namedtuple('Request', 'method url protocol')): """ Represents an HTTP request line. This type is returned by the :func:`request` function and represents the three parts of an HTTP request line: the method, the URL (optional, can be None in the case of methods like OPTIONS), and the protocol. The following attributes exist: .. attribute:: method The method of the request (typically GET, POST, or PUT but can technically be any valid HTTP token) .. attribute:: url The requested URL. May be an absolute URL, an absolute path, an authority token, or None in the case that the request line contained "*" for the URL. .. attribute:: protocol The HTTP protocol version requested. A string of the format 'HTTP/x.y' where x.y is the version number. At the time of writing only HTTP/1.0 and HTTP/1.1 are defined. """ def __str__(self): return '%s %s %s' % (self.method, self.url, self.protocol)
mit
adobdin/timmy
timmy/analyze_modules/rabbitmq.py
1
7526
#!/usr/bin/env python2 # -*- coding: utf-8 -*- # Copyright 2016 Mirantis, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from timmy.analyze_health import GREEN, UNKNOWN, YELLOW, RED from timmy.env import project_name import logging import re import yaml logger = logging.getLogger(project_name) def register(function_mapping): function_mapping['rabbitmqctl-list-queues'] = parse_list_queues function_mapping['rabbitmqctl-status'] = parse_status def parse_list_queues(stdout, script, node, stderr=None, exitcode=None): warning = 100 error = 1000 health = GREEN details = [] if exitcode: health = UNKNOWN return health, details data = [l.rstrip() for l in stdout.splitlines()] for line in data[1:]: elements = line.rstrip().split() if len(elements) < 2: logger.warning('no value in list_queues: "%s"' % line.rstrip()) else: count = int(elements[1]) if count < error and count >= warning: health = max(health, YELLOW) details.append(line) return health, details def prepare_status(stdout): bad_yaml = ''.join(stdout.splitlines()[1:]) # quoting string elements bad_yaml = re.sub(r'([,{])([a-z_A-Z]+)([,}])', r'\1"\2"\3', bad_yaml) # changing first element int a key - replacing , with : bad_yaml = re.sub(r'({[^,]+),', r'\1:', bad_yaml) bad_yaml_list = list(bad_yaml) good_yaml, _ = fix_dicts(bad_yaml_list, 0) status_list = yaml.load(''.join(good_yaml)) status_dict = squash_dicts(status_list) return status_dict def fix_dicts(json_str_list, pos): '''recursively puts all comma-separted values into square brackets to make data look like normal 'key: value' dicts ''' quoted_string = False value = True value_pos = 0 commas = False is_list = False in_list = 0 while pos < len(json_str_list): if not quoted_string: if json_str_list[pos] == '{': json_str_list, pos = fix_dicts(json_str_list, pos+1) elif json_str_list[pos] == '"': quoted_string = True elif json_str_list[pos] == ':': value = True value_pos = pos + 1 elif json_str_list[pos] == '[': if value and not commas: is_list = True in_list += 1 elif json_str_list[pos] == ']': in_list -= 1 elif json_str_list[pos] == ',': commas = True if not in_list: is_list = False elif json_str_list[pos] == '}': if not is_list and commas: json_str_list = (json_str_list[:value_pos] + ['['] + json_str_list[value_pos:pos] + [']'] + json_str_list[pos:]) pos += 2 return json_str_list, pos elif json_str_list[pos] == '"': quoted_string = False pos += 1 return json_str_list, pos def squash_dicts(input_data): # recursively converts [{a:1},{b:2},{c:3}...] into {a:1, b:2, c:3}''' if type(input_data) is list: for i in range(len(input_data)): input_data[i] = squash_dicts(input_data[i]) if all([type(i) is dict for i in input_data]): kv_list = [(k, v) for i in input_data for k, v in i.items()] input_data = dict(kv_list) elif type(input_data) is dict: for k, v in input_data.items(): input_data[k] = squash_dicts(v) return input_data def parse_status(stdout, script, node, stderr=None, exitcode=None): health = GREEN details = [] status = prepare_status(stdout) if not status: health = UNKNOWN details = ['Status unavailable'] if exitcode: if exitcode == 69: health = RED details = ['RabbitMQ is not running'] return health, details # disk free check try: dfree = int(status['disk_free']) dlimit = int(status['disk_free_limit']) dfree_ok = 10**9 # 1GB if dfree > dlimit and dfree < dfree_ok: health = max(health, YELLOW) details.append('disk_free: %s, disk_free_limit: %s' % (dfree, dlimit)) elif dfree <= dlimit: health = max(health, RED) details.append('disk_free: %s, disk_free_limit: %s' % (dfree, dlimit)) except ValueError: details.append('cannot convert disk_free* to int') health = max(health, UNKNOWN) except KeyError: details.append('disk_free* not present') health = max(health, UNKNOWN) # process limit check try: pused = float(status['processes']['used']) plimit = float(status['processes']['limit']) ok_ratio = 0.9 if pused < plimit and pused/plimit > ok_ratio: health = max(health, YELLOW) details.append('processes used: %s, processes limit: %s' % (pused, plimit)) elif pused >= plimit: health = max(health, RED) details.append('processes used: %s, processes limit: %s' % (pused, plimit)) except ValueError: details.append('cannot convert processes* to numbers') health = max(health, UNKNOWN) except KeyError: details.append('processes* not present') health = max(health, UNKNOWN) # fd check try: sused = float(status['file_descriptors']['sockets_used']) slimit = float(status['file_descriptors']['sockets_limit']) ok_ratio = 0.9 if sused < slimit and sused/slimit > ok_ratio: health = max(health, YELLOW) details.append('sockets used: %s, sockets limit: %s' % (sused, slimit)) elif sused >= slimit: health = max(health, RED) details.append('sockets used: %s, sockets limit: %s' % (sused, slimit)) fdused = float(status['file_descriptors']['total_used']) fdlimit = float(status['file_descriptors']['total_limit']) ok_ratio = 0.9 if fdused < fdlimit and fdused/fdlimit > ok_ratio: health = max(health, YELLOW) details.append('fd total used: %s, fd total limit: %s' % (fdused, fdlimit)) elif fdused >= fdlimit: health = max(health, RED) details.append('fd total used: %s, fd total limit: %s' % (fdused, fdlimit)) except ValueError: details.append('cannot convert file_descriptors* to numbers') health = max(health, UNKNOWN) except KeyError: details.append('file_descriptors* not present') health = max(health, UNKNOWN) return health, details
apache-2.0
smcantab/pele
playground/amber/driver_sanitycheck.py
7
1208
from pele.amber import amberSystem as amb # create a new amber system and load database to be pruned sys = amb.AMBERSystem('coords.prmtop', 'coords.inpcrd') dbcurr = sys.create_database(db="aladipep.db") print 'Collecting minima to delete .. ' listTODel = [] for minimum in dbcurr.minima(): testOutCome1 = sys.check_cistrans(minimum.coords) testOutCome2 = sys.check_CAchirality(minimum.coords) if testOutCome1 and testOutCome2: print 'PASS', minimum._id, minimum.energy else: listTODel.append(minimum) print 'FAIL', minimum._id, minimum.energy print '------------' print 'Number of minima to be deleted = ', len(listTODel) # now delete for minn in listTODel: dbcurr.removeMinimum(minn) #print 'Checking transition states .. ' #ct = 0 #print len(dbcurr.transition_states()) #for ts in dbcurr.transition_states() : # if sys.check_cistrans(ts.coords ): # print 'PASS', ts._id, ts.energy # ct = ct + 1 # # dbcurr.removeTS(ts) # not implemented yet # else: # print 'FAIL', ts._id, ts.energy # # print '------------' # #print 'Number of TS deleted = ', ct
gpl-3.0
sorenk/ansible
lib/ansible/utils/module_docs_fragments/ipa.py
27
2627
# Copyright (c) 2017-18, Ansible Project # Copyright (c) 2017-18, Abhijeet Kasurde (akasurde@redhat.com) # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) class ModuleDocFragment(object): # Parameters for FreeIPA/IPA modules DOCUMENTATION = ''' options: ipa_port: description: - Port of FreeIPA / IPA server. - If the value is not specified in the task, the value of environment variable C(IPA_PORT) will be used instead. - If both the environment variable C(IPA_PORT) and the value are not specified in the task, then default value is set. - 'Environment variable fallback mechanism is added in version 2.5.' default: 443 ipa_host: description: - IP or hostname of IPA server. - If the value is not specified in the task, the value of environment variable C(IPA_HOST) will be used instead. - If both the environment variable C(IPA_HOST) and the value are not specified in the task, then default value is set. - 'Environment variable fallback mechanism is added in version 2.5.' default: ipa.example.com ipa_user: description: - Administrative account used on IPA server. - If the value is not specified in the task, the value of environment variable C(IPA_USER) will be used instead. - If both the environment variable C(IPA_USER) and the value are not specified in the task, then default value is set. - 'Environment variable fallback mechanism is added in version 2.5.' default: admin ipa_pass: description: - Password of administrative user. - If the value is not specified in the task, the value of environment variable C(IPA_PASS) will be used instead. - If both the environment variable C(IPA_PASS) and the value are not specified in the task, then default value is set. - 'Environment variable fallback mechanism is added in version 2.5.' required: true ipa_prot: description: - Protocol used by IPA server. - If the value is not specified in the task, the value of environment variable C(IPA_PROT) will be used instead. - If both the environment variable C(IPA_PROT) and the value are not specified in the task, then default value is set. - 'Environment variable fallback mechanism is added in version 2.5.' default: https choices: ["http", "https"] validate_certs: description: - This only applies if C(ipa_prot) is I(https). - If set to C(no), the SSL certificates will not be validated. - This should only set to C(no) used on personally controlled sites using self-signed certificates. default: true '''
gpl-3.0
barykaed/Pelican-Test
activate/Lib/rlcompleter.py
100
5763
"""Word completion for GNU readline. The completer completes keywords, built-ins and globals in a selectable namespace (which defaults to __main__); when completing NAME.NAME..., it evaluates (!) the expression up to the last dot and completes its attributes. It's very cool to do "import sys" type "sys.", hit the completion key (twice), and see the list of names defined by the sys module! Tip: to use the tab key as the completion key, call readline.parse_and_bind("tab: complete") Notes: - Exceptions raised by the completer function are *ignored* (and generally cause the completion to fail). This is a feature -- since readline sets the tty device in raw (or cbreak) mode, printing a traceback wouldn't work well without some complicated hoopla to save, reset and restore the tty state. - The evaluation of the NAME.NAME... form may cause arbitrary application defined code to be executed if an object with a __getattr__ hook is found. Since it is the responsibility of the application (or the user) to enable this feature, I consider this an acceptable risk. More complicated expressions (e.g. function calls or indexing operations) are *not* evaluated. - When the original stdin is not a tty device, GNU readline is never used, and this module (and the readline module) are silently inactive. """ import atexit import builtins import __main__ __all__ = ["Completer"] class Completer: def __init__(self, namespace = None): """Create a new completer for the command line. Completer([namespace]) -> completer instance. If unspecified, the default namespace where completions are performed is __main__ (technically, __main__.__dict__). Namespaces should be given as dictionaries. Completer instances should be used as the completion mechanism of readline via the set_completer() call: readline.set_completer(Completer(my_namespace).complete) """ if namespace and not isinstance(namespace, dict): raise TypeError('namespace must be a dictionary') # Don't bind to namespace quite yet, but flag whether the user wants a # specific namespace or to use __main__.__dict__. This will allow us # to bind to __main__.__dict__ at completion time, not now. if namespace is None: self.use_main_ns = 1 else: self.use_main_ns = 0 self.namespace = namespace def complete(self, text, state): """Return the next possible completion for 'text'. This is called successively with state == 0, 1, 2, ... until it returns None. The completion should begin with 'text'. """ if self.use_main_ns: self.namespace = __main__.__dict__ if state == 0: if "." in text: self.matches = self.attr_matches(text) else: self.matches = self.global_matches(text) try: return self.matches[state] except IndexError: return None def _callable_postfix(self, val, word): if callable(val): word = word + "(" return word def global_matches(self, text): """Compute matches when text is a simple name. Return a list of all keywords, built-in functions and names currently defined in self.namespace that match. """ import keyword matches = [] n = len(text) for word in keyword.kwlist: if word[:n] == text: matches.append(word) for nspace in [builtins.__dict__, self.namespace]: for word, val in nspace.items(): if word[:n] == text and word != "__builtins__": matches.append(self._callable_postfix(val, word)) return matches def attr_matches(self, text): """Compute matches when text contains a dot. Assuming the text is of the form NAME.NAME....[NAME], and is evaluable in self.namespace, it will be evaluated and its attributes (as revealed by dir()) are used as possible completions. (For class instances, class members are also considered.) WARNING: this can still invoke arbitrary C code, if an object with a __getattr__ hook is evaluated. """ import re m = re.match(r"(\w+(\.\w+)*)\.(\w*)", text) if not m: return [] expr, attr = m.group(1, 3) try: thisobject = eval(expr, self.namespace) except Exception: return [] # get the content of the object, except __builtins__ words = dir(thisobject) if "__builtins__" in words: words.remove("__builtins__") if hasattr(thisobject, '__class__'): words.append('__class__') words.extend(get_class_members(thisobject.__class__)) matches = [] n = len(attr) for word in words: if word[:n] == attr and hasattr(thisobject, word): val = getattr(thisobject, word) word = self._callable_postfix(val, "%s.%s" % (expr, word)) matches.append(word) return matches def get_class_members(klass): ret = dir(klass) if hasattr(klass,'__bases__'): for base in klass.__bases__: ret = ret + get_class_members(base) return ret try: import readline except ImportError: pass else: readline.set_completer(Completer().complete) # Release references early at shutdown (the readline module's # contents are quasi-immortal, and the completer function holds a # reference to globals). atexit.register(lambda: readline.set_completer(None))
mit
feigames/Odoo
addons/survey/__init__.py
385
1037
# -*- encoding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution # Copyright (C) 2004-TODAY OpenERP S.A. <http://www.openerp.com> # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import survey import controllers import wizard
agpl-3.0
mfortner/MyPythonKoans
python2/koans/about_true_and_false.py
1
1502
#!/usr/bin/env python # -*- coding: utf-8 -*- from runner.koan import * class AboutTrueAndFalse(Koan): def truth_value(self, condition): if condition: return 'true stuff' else: return 'false stuff' def test_true_is_treated_as_true(self): self.assertEqual("true stuff", self.truth_value(True)) def test_false_is_treated_as_false(self): self.assertEqual("false stuff", self.truth_value(False)) def test_none_is_treated_as_false(self): self.assertEqual("false stuff", self.truth_value(None)) def test_zero_is_treated_as_false(self): self.assertEqual("false stuff", self.truth_value(0)) def test_empty_collections_are_treated_as_false(self): self.assertEqual("false stuff", self.truth_value([])) self.assertEqual("false stuff", self.truth_value(())) self.assertEqual("false stuff", self.truth_value({})) self.assertEqual("false stuff", self.truth_value(set())) def test_blank_strings_are_treated_as_false(self): self.assertEqual("false stuff", self.truth_value("")) def test_everything_else_is_treated_as_true(self): self.assertEqual("true stuff", self.truth_value(1)) self.assertEqual("true stuff", self.truth_value(1,)) self.assertEqual( "true stuff", self.truth_value("Python is named after Monty Python")) self.assertEqual("true stuff", self.truth_value(' '))
mit
llazzaro/analyzerstrategies
tests/test_sma_strategy.py
1
1444
import random import unittest from pyStock.models import ( Exchange, Stock, Account, Owner, ) from pyStock.models.money import Currency from analyzer.backtest.constant import ( SELL, BUY_TO_COVER, ) from analyzerstrategies.sma_strategy import SMAStrategy class TestSMAStrategy(unittest.TestCase): def setUp(self): pesos = Currency(name='Pesos', code='ARG') merval = Exchange(name='Merval', code='MERV', currency=pesos) owner = Owner(name='test user') self.account = Account(owner=owner) self.security = Stock(symbol='YPF', exchange=merval) self.tick = {'pattern': None, 'data': {'volume30d': '12165.08453826', 'timestamp': '1446070419', 'high': '305', 'ask': 302.7022, 'last': '302.632', 'bid': 301.0001, 'low': '294.51', 'volume': '437.07501250'}, 'type': 'message', 'security': self.security, 'channel': b'BTC'} def test_quotes_feeder(self): strategy = SMAStrategy(account=self.account, config=None, securities=[self.security], store=None) # not enough data to return action. self.assertTrue(strategy.update(self.tick) is None) tick = self.tick for i in range(0, 340): tick['data']['last'] = random.uniform(300, 350) action = strategy.update(tick) if action is not None: self.assertNotEquals(action, SELL) self.assertNotEquals(action, BUY_TO_COVER)
bsd-3-clause
takeshineshiro/keystone
keystone/common/extension.py
31
1652
# Copyright 2013 OpenStack Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. ADMIN_EXTENSIONS = {} PUBLIC_EXTENSIONS = {} def register_admin_extension(url_prefix, extension_data): """Register extension with collection of admin extensions. Extensions register the information here that will show up in the /extensions page as a way to indicate that the extension is active. url_prefix: unique key for the extension that will appear in the urls generated by the extension. extension_data is a dictionary. The expected fields are: 'name': short, human readable name of the extension 'namespace': xml namespace 'alias': identifier for the extension 'updated': date the extension was last updated 'description': text description of the extension 'links': hyperlinks to documents describing the extension """ ADMIN_EXTENSIONS[url_prefix] = extension_data def register_public_extension(url_prefix, extension_data): """Same as register_admin_extension but for public extensions.""" PUBLIC_EXTENSIONS[url_prefix] = extension_data
apache-2.0
bankonmeOS/cjdns
node_build/dependencies/libuv/build/gyp/pylib/gyp/generator/android.py
15
43658
# Copyright (c) 2012 Google Inc. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. # Notes: # # This generates makefiles suitable for inclusion into the Android build system # via an Android.mk file. It is based on make.py, the standard makefile # generator. # # The code below generates a separate .mk file for each target, but # all are sourced by the top-level GypAndroid.mk. This means that all # variables in .mk-files clobber one another, and furthermore that any # variables set potentially clash with other Android build system variables. # Try to avoid setting global variables where possible. import gyp import gyp.common import gyp.generator.make as make # Reuse global functions from make backend. import os import re import subprocess generator_default_variables = { 'OS': 'android', 'EXECUTABLE_PREFIX': '', 'EXECUTABLE_SUFFIX': '', 'STATIC_LIB_PREFIX': 'lib', 'SHARED_LIB_PREFIX': 'lib', 'STATIC_LIB_SUFFIX': '.a', 'SHARED_LIB_SUFFIX': '.so', 'INTERMEDIATE_DIR': '$(gyp_intermediate_dir)', 'SHARED_INTERMEDIATE_DIR': '$(gyp_shared_intermediate_dir)', 'PRODUCT_DIR': '$(gyp_shared_intermediate_dir)', 'SHARED_LIB_DIR': '$(builddir)/lib.$(TOOLSET)', 'LIB_DIR': '$(obj).$(TOOLSET)', 'RULE_INPUT_ROOT': '%(INPUT_ROOT)s', # This gets expanded by Python. 'RULE_INPUT_DIRNAME': '%(INPUT_DIRNAME)s', # This gets expanded by Python. 'RULE_INPUT_PATH': '$(RULE_SOURCES)', 'RULE_INPUT_EXT': '$(suffix $<)', 'RULE_INPUT_NAME': '$(notdir $<)', 'CONFIGURATION_NAME': '$(GYP_CONFIGURATION)', } # Make supports multiple toolsets generator_supports_multiple_toolsets = True # Generator-specific gyp specs. generator_additional_non_configuration_keys = [ # Boolean to declare that this target does not want its name mangled. 'android_unmangled_name', ] generator_additional_path_sections = [] generator_extra_sources_for_rules = [] SHARED_FOOTER = """\ # "gyp_all_modules" is a concatenation of the "gyp_all_modules" targets from # all the included sub-makefiles. This is just here to clarify. gyp_all_modules: """ header = """\ # This file is generated by gyp; do not edit. """ android_standard_include_paths = set([ # JNI_H_INCLUDE in build/core/binary.mk 'dalvik/libnativehelper/include/nativehelper', # from SRC_HEADERS in build/core/config.mk 'system/core/include', 'hardware/libhardware/include', 'hardware/libhardware_legacy/include', 'hardware/ril/include', 'dalvik/libnativehelper/include', 'frameworks/native/include', 'frameworks/native/opengl/include', 'frameworks/base/include', 'frameworks/base/opengl/include', 'frameworks/base/native/include', 'external/skia/include', # TARGET_C_INCLUDES in build/core/combo/TARGET_linux-arm.mk 'bionic/libc/arch-arm/include', 'bionic/libc/include', 'bionic/libstdc++/include', 'bionic/libc/kernel/common', 'bionic/libc/kernel/arch-arm', 'bionic/libm/include', 'bionic/libm/include/arm', 'bionic/libthread_db/include', ]) # Map gyp target types to Android module classes. MODULE_CLASSES = { 'static_library': 'STATIC_LIBRARIES', 'shared_library': 'SHARED_LIBRARIES', 'executable': 'EXECUTABLES', } def IsCPPExtension(ext): return make.COMPILABLE_EXTENSIONS.get(ext) == 'cxx' def Sourceify(path): """Convert a path to its source directory form. The Android backend does not support options.generator_output, so this function is a noop.""" return path # Map from qualified target to path to output. # For Android, the target of these maps is a tuple ('static', 'modulename'), # ('dynamic', 'modulename'), or ('path', 'some/path') instead of a string, # since we link by module. target_outputs = {} # Map from qualified target to any linkable output. A subset # of target_outputs. E.g. when mybinary depends on liba, we want to # include liba in the linker line; when otherbinary depends on # mybinary, we just want to build mybinary first. target_link_deps = {} class AndroidMkWriter(object): """AndroidMkWriter packages up the writing of one target-specific Android.mk. Its only real entry point is Write(), and is mostly used for namespacing. """ def __init__(self, android_top_dir): self.android_top_dir = android_top_dir def Write(self, qualified_target, relative_target, base_path, output_filename, spec, configs, part_of_all): """The main entry point: writes a .mk file for a single target. Arguments: qualified_target: target we're generating relative_target: qualified target name relative to the root base_path: path relative to source root we're building in, used to resolve target-relative paths output_filename: output .mk file name to write spec, configs: gyp info part_of_all: flag indicating this target is part of 'all' """ gyp.common.EnsureDirExists(output_filename) self.fp = open(output_filename, 'w') self.fp.write(header) self.qualified_target = qualified_target self.relative_target = relative_target self.path = base_path self.target = spec['target_name'] self.type = spec['type'] self.toolset = spec['toolset'] deps, link_deps = self.ComputeDeps(spec) # Some of the generation below can add extra output, sources, or # link dependencies. All of the out params of the functions that # follow use names like extra_foo. extra_outputs = [] extra_sources = [] self.android_class = MODULE_CLASSES.get(self.type, 'GYP') self.android_module = self.ComputeAndroidModule(spec) (self.android_stem, self.android_suffix) = self.ComputeOutputParts(spec) self.output = self.output_binary = self.ComputeOutput(spec) # Standard header. self.WriteLn('include $(CLEAR_VARS)\n') # Module class and name. self.WriteLn('LOCAL_MODULE_CLASS := ' + self.android_class) self.WriteLn('LOCAL_MODULE := ' + self.android_module) # Only emit LOCAL_MODULE_STEM if it's different to LOCAL_MODULE. # The library module classes fail if the stem is set. ComputeOutputParts # makes sure that stem == modulename in these cases. if self.android_stem != self.android_module: self.WriteLn('LOCAL_MODULE_STEM := ' + self.android_stem) self.WriteLn('LOCAL_MODULE_SUFFIX := ' + self.android_suffix) self.WriteLn('LOCAL_MODULE_TAGS := optional') if self.toolset == 'host': self.WriteLn('LOCAL_IS_HOST_MODULE := true') # Grab output directories; needed for Actions and Rules. self.WriteLn('gyp_intermediate_dir := $(call local-intermediates-dir)') self.WriteLn('gyp_shared_intermediate_dir := ' '$(call intermediates-dir-for,GYP,shared)') self.WriteLn() # List files this target depends on so that actions/rules/copies/sources # can depend on the list. # TODO: doesn't pull in things through transitive link deps; needed? target_dependencies = [x[1] for x in deps if x[0] == 'path'] self.WriteLn('# Make sure our deps are built first.') self.WriteList(target_dependencies, 'GYP_TARGET_DEPENDENCIES', local_pathify=True) # Actions must come first, since they can generate more OBJs for use below. if 'actions' in spec: self.WriteActions(spec['actions'], extra_sources, extra_outputs) # Rules must be early like actions. if 'rules' in spec: self.WriteRules(spec['rules'], extra_sources, extra_outputs) if 'copies' in spec: self.WriteCopies(spec['copies'], extra_outputs) # GYP generated outputs. self.WriteList(extra_outputs, 'GYP_GENERATED_OUTPUTS', local_pathify=True) # Set LOCAL_ADDITIONAL_DEPENDENCIES so that Android's build rules depend # on both our dependency targets and our generated files. self.WriteLn('# Make sure our deps and generated files are built first.') self.WriteLn('LOCAL_ADDITIONAL_DEPENDENCIES := $(GYP_TARGET_DEPENDENCIES) ' '$(GYP_GENERATED_OUTPUTS)') self.WriteLn() # Sources. if spec.get('sources', []) or extra_sources: self.WriteSources(spec, configs, extra_sources) self.WriteTarget(spec, configs, deps, link_deps, part_of_all) # Update global list of target outputs, used in dependency tracking. target_outputs[qualified_target] = ('path', self.output_binary) # Update global list of link dependencies. if self.type == 'static_library': target_link_deps[qualified_target] = ('static', self.android_module) elif self.type == 'shared_library': target_link_deps[qualified_target] = ('shared', self.android_module) self.fp.close() return self.android_module def WriteActions(self, actions, extra_sources, extra_outputs): """Write Makefile code for any 'actions' from the gyp input. extra_sources: a list that will be filled in with newly generated source files, if any extra_outputs: a list that will be filled in with any outputs of these actions (used to make other pieces dependent on these actions) """ for action in actions: name = make.StringToMakefileVariable('%s_%s' % (self.relative_target, action['action_name'])) self.WriteLn('### Rules for action "%s":' % action['action_name']) inputs = action['inputs'] outputs = action['outputs'] # Build up a list of outputs. # Collect the output dirs we'll need. dirs = set() for out in outputs: if not out.startswith('$'): print ('WARNING: Action for target "%s" writes output to local path ' '"%s".' % (self.target, out)) dir = os.path.split(out)[0] if dir: dirs.add(dir) if int(action.get('process_outputs_as_sources', False)): extra_sources += outputs # Prepare the actual command. command = gyp.common.EncodePOSIXShellList(action['action']) if 'message' in action: quiet_cmd = 'Gyp action: %s ($@)' % action['message'] else: quiet_cmd = 'Gyp action: %s ($@)' % name if len(dirs) > 0: command = 'mkdir -p %s' % ' '.join(dirs) + '; ' + command cd_action = 'cd $(gyp_local_path)/%s; ' % self.path command = cd_action + command # The makefile rules are all relative to the top dir, but the gyp actions # are defined relative to their containing dir. This replaces the gyp_* # variables for the action rule with an absolute version so that the # output goes in the right place. # Only write the gyp_* rules for the "primary" output (:1); # it's superfluous for the "extra outputs", and this avoids accidentally # writing duplicate dummy rules for those outputs. main_output = make.QuoteSpaces(self.LocalPathify(outputs[0])) self.WriteLn('%s: gyp_local_path := $(LOCAL_PATH)' % main_output) self.WriteLn('%s: gyp_intermediate_dir := ' '$(abspath $(gyp_intermediate_dir))' % main_output) self.WriteLn('%s: gyp_shared_intermediate_dir := ' '$(abspath $(gyp_shared_intermediate_dir))' % main_output) # Android's envsetup.sh adds a number of directories to the path including # the built host binary directory. This causes actions/rules invoked by # gyp to sometimes use these instead of system versions, e.g. bison. # The built host binaries may not be suitable, and can cause errors. # So, we remove them from the PATH using the ANDROID_BUILD_PATHS variable # set by envsetup. self.WriteLn('%s: export PATH := $(subst $(ANDROID_BUILD_PATHS),,$(PATH))' % main_output) for input in inputs: assert ' ' not in input, ( "Spaces in action input filenames not supported (%s)" % input) for output in outputs: assert ' ' not in output, ( "Spaces in action output filenames not supported (%s)" % output) self.WriteLn('%s: %s $(GYP_TARGET_DEPENDENCIES)' % (main_output, ' '.join(map(self.LocalPathify, inputs)))) self.WriteLn('\t@echo "%s"' % quiet_cmd) self.WriteLn('\t$(hide)%s\n' % command) for output in outputs[1:]: # Make each output depend on the main output, with an empty command # to force make to notice that the mtime has changed. self.WriteLn('%s: %s ;' % (self.LocalPathify(output), main_output)) extra_outputs += outputs self.WriteLn() self.WriteLn() def WriteRules(self, rules, extra_sources, extra_outputs): """Write Makefile code for any 'rules' from the gyp input. extra_sources: a list that will be filled in with newly generated source files, if any extra_outputs: a list that will be filled in with any outputs of these rules (used to make other pieces dependent on these rules) """ if len(rules) == 0: return rule_trigger = '%s_rule_trigger' % self.android_module did_write_rule = False for rule in rules: if len(rule.get('rule_sources', [])) == 0: continue did_write_rule = True name = make.StringToMakefileVariable('%s_%s' % (self.relative_target, rule['rule_name'])) self.WriteLn('\n### Generated for rule "%s":' % name) self.WriteLn('# "%s":' % rule) inputs = rule.get('inputs') for rule_source in rule.get('rule_sources', []): (rule_source_dirname, rule_source_basename) = os.path.split(rule_source) (rule_source_root, rule_source_ext) = \ os.path.splitext(rule_source_basename) outputs = [self.ExpandInputRoot(out, rule_source_root, rule_source_dirname) for out in rule['outputs']] dirs = set() for out in outputs: if not out.startswith('$'): print ('WARNING: Rule for target %s writes output to local path %s' % (self.target, out)) dir = os.path.dirname(out) if dir: dirs.add(dir) extra_outputs += outputs if int(rule.get('process_outputs_as_sources', False)): extra_sources.extend(outputs) components = [] for component in rule['action']: component = self.ExpandInputRoot(component, rule_source_root, rule_source_dirname) if '$(RULE_SOURCES)' in component: component = component.replace('$(RULE_SOURCES)', rule_source) components.append(component) command = gyp.common.EncodePOSIXShellList(components) cd_action = 'cd $(gyp_local_path)/%s; ' % self.path command = cd_action + command if dirs: command = 'mkdir -p %s' % ' '.join(dirs) + '; ' + command # We set up a rule to build the first output, and then set up # a rule for each additional output to depend on the first. outputs = map(self.LocalPathify, outputs) main_output = outputs[0] self.WriteLn('%s: gyp_local_path := $(LOCAL_PATH)' % main_output) self.WriteLn('%s: gyp_intermediate_dir := ' '$(abspath $(gyp_intermediate_dir))' % main_output) self.WriteLn('%s: gyp_shared_intermediate_dir := ' '$(abspath $(gyp_shared_intermediate_dir))' % main_output) # See explanation in WriteActions. self.WriteLn('%s: export PATH := ' '$(subst $(ANDROID_BUILD_PATHS),,$(PATH))' % main_output) main_output_deps = self.LocalPathify(rule_source) if inputs: main_output_deps += ' ' main_output_deps += ' '.join([self.LocalPathify(f) for f in inputs]) self.WriteLn('%s: %s $(GYP_TARGET_DEPENDENCIES)' % (main_output, main_output_deps)) self.WriteLn('\t%s\n' % command) for output in outputs[1:]: # Make each output depend on the main output, with an empty command # to force make to notice that the mtime has changed. self.WriteLn('%s: %s ;' % (output, main_output)) self.WriteLn('.PHONY: %s' % (rule_trigger)) self.WriteLn('%s: %s' % (rule_trigger, main_output)) self.WriteLn('') if did_write_rule: extra_sources.append(rule_trigger) # Force all rules to run. self.WriteLn('### Finished generating for all rules') self.WriteLn('') def WriteCopies(self, copies, extra_outputs): """Write Makefile code for any 'copies' from the gyp input. extra_outputs: a list that will be filled in with any outputs of this action (used to make other pieces dependent on this action) """ self.WriteLn('### Generated for copy rule.') variable = make.StringToMakefileVariable(self.relative_target + '_copies') outputs = [] for copy in copies: for path in copy['files']: # The Android build system does not allow generation of files into the # source tree. The destination should start with a variable, which will # typically be $(gyp_intermediate_dir) or # $(gyp_shared_intermediate_dir). Note that we can't use an assertion # because some of the gyp tests depend on this. if not copy['destination'].startswith('$'): print ('WARNING: Copy rule for target %s writes output to ' 'local path %s' % (self.target, copy['destination'])) # LocalPathify() calls normpath, stripping trailing slashes. path = Sourceify(self.LocalPathify(path)) filename = os.path.split(path)[1] output = Sourceify(self.LocalPathify(os.path.join(copy['destination'], filename))) self.WriteLn('%s: %s $(GYP_TARGET_DEPENDENCIES) | $(ACP)' % (output, path)) self.WriteLn('\t@echo Copying: $@') self.WriteLn('\t$(hide) mkdir -p $(dir $@)') self.WriteLn('\t$(hide) $(ACP) -rpf $< $@') self.WriteLn() outputs.append(output) self.WriteLn('%s = %s' % (variable, ' '.join(map(make.QuoteSpaces, outputs)))) extra_outputs.append('$(%s)' % variable) self.WriteLn() def WriteSourceFlags(self, spec, configs): """Write out the flags and include paths used to compile source files for the current target. Args: spec, configs: input from gyp. """ for configname, config in sorted(configs.iteritems()): extracted_includes = [] self.WriteLn('\n# Flags passed to both C and C++ files.') cflags, includes_from_cflags = self.ExtractIncludesFromCFlags( config.get('cflags', []) + config.get('cflags_c', [])) extracted_includes.extend(includes_from_cflags) self.WriteList(cflags, 'MY_CFLAGS_%s' % configname) self.WriteList(config.get('defines'), 'MY_DEFS_%s' % configname, prefix='-D', quoter=make.EscapeCppDefine) self.WriteLn('\n# Include paths placed before CFLAGS/CPPFLAGS') includes = list(config.get('include_dirs', [])) includes.extend(extracted_includes) includes = map(Sourceify, map(self.LocalPathify, includes)) includes = self.NormalizeIncludePaths(includes) self.WriteList(includes, 'LOCAL_C_INCLUDES_%s' % configname) self.WriteLn('\n# Flags passed to only C++ (and not C) files.') self.WriteList(config.get('cflags_cc'), 'LOCAL_CPPFLAGS_%s' % configname) self.WriteLn('\nLOCAL_CFLAGS := $(MY_CFLAGS_$(GYP_CONFIGURATION)) ' '$(MY_DEFS_$(GYP_CONFIGURATION))') # Undefine ANDROID for host modules # TODO: the source code should not use macro ANDROID to tell if it's host # or target module. if self.toolset == 'host': self.WriteLn('# Undefine ANDROID for host modules') self.WriteLn('LOCAL_CFLAGS += -UANDROID') self.WriteLn('LOCAL_C_INCLUDES := $(GYP_COPIED_SOURCE_ORIGIN_DIRS) ' '$(LOCAL_C_INCLUDES_$(GYP_CONFIGURATION))') self.WriteLn('LOCAL_CPPFLAGS := $(LOCAL_CPPFLAGS_$(GYP_CONFIGURATION))') # Android uses separate flags for assembly file invocations, but gyp expects # the same CFLAGS to be applied: self.WriteLn('LOCAL_ASFLAGS := $(LOCAL_CFLAGS)') def WriteSources(self, spec, configs, extra_sources): """Write Makefile code for any 'sources' from the gyp input. These are source files necessary to build the current target. We need to handle shared_intermediate directory source files as a special case by copying them to the intermediate directory and treating them as a genereated sources. Otherwise the Android build rules won't pick them up. Args: spec, configs: input from gyp. extra_sources: Sources generated from Actions or Rules. """ sources = filter(make.Compilable, spec.get('sources', [])) generated_not_sources = [x for x in extra_sources if not make.Compilable(x)] extra_sources = filter(make.Compilable, extra_sources) # Determine and output the C++ extension used by these sources. # We simply find the first C++ file and use that extension. all_sources = sources + extra_sources local_cpp_extension = '.cpp' for source in all_sources: (root, ext) = os.path.splitext(source) if IsCPPExtension(ext): local_cpp_extension = ext break if local_cpp_extension != '.cpp': self.WriteLn('LOCAL_CPP_EXTENSION := %s' % local_cpp_extension) # We need to move any non-generated sources that are coming from the # shared intermediate directory out of LOCAL_SRC_FILES and put them # into LOCAL_GENERATED_SOURCES. We also need to move over any C++ files # that don't match our local_cpp_extension, since Android will only # generate Makefile rules for a single LOCAL_CPP_EXTENSION. local_files = [] for source in sources: (root, ext) = os.path.splitext(source) if '$(gyp_shared_intermediate_dir)' in source: extra_sources.append(source) elif '$(gyp_intermediate_dir)' in source: extra_sources.append(source) elif IsCPPExtension(ext) and ext != local_cpp_extension: extra_sources.append(source) else: local_files.append(os.path.normpath(os.path.join(self.path, source))) # For any generated source, if it is coming from the shared intermediate # directory then we add a Make rule to copy them to the local intermediate # directory first. This is because the Android LOCAL_GENERATED_SOURCES # must be in the local module intermediate directory for the compile rules # to work properly. If the file has the wrong C++ extension, then we add # a rule to copy that to intermediates and use the new version. final_generated_sources = [] # If a source file gets copied, we still need to add the orginal source # directory as header search path, for GCC searches headers in the # directory that contains the source file by default. origin_src_dirs = [] for source in extra_sources: local_file = source if not '$(gyp_intermediate_dir)/' in local_file: basename = os.path.basename(local_file) local_file = '$(gyp_intermediate_dir)/' + basename (root, ext) = os.path.splitext(local_file) if IsCPPExtension(ext) and ext != local_cpp_extension: local_file = root + local_cpp_extension if local_file != source: self.WriteLn('%s: %s' % (local_file, self.LocalPathify(source))) self.WriteLn('\tmkdir -p $(@D); cp $< $@') origin_src_dirs.append(os.path.dirname(source)) final_generated_sources.append(local_file) # We add back in all of the non-compilable stuff to make sure that the # make rules have dependencies on them. final_generated_sources.extend(generated_not_sources) self.WriteList(final_generated_sources, 'LOCAL_GENERATED_SOURCES') origin_src_dirs = gyp.common.uniquer(origin_src_dirs) origin_src_dirs = map(Sourceify, map(self.LocalPathify, origin_src_dirs)) self.WriteList(origin_src_dirs, 'GYP_COPIED_SOURCE_ORIGIN_DIRS') self.WriteList(local_files, 'LOCAL_SRC_FILES') # Write out the flags used to compile the source; this must be done last # so that GYP_COPIED_SOURCE_ORIGIN_DIRS can be used as an include path. self.WriteSourceFlags(spec, configs) def ComputeAndroidModule(self, spec): """Return the Android module name used for a gyp spec. We use the complete qualified target name to avoid collisions between duplicate targets in different directories. We also add a suffix to distinguish gyp-generated module names. """ if int(spec.get('android_unmangled_name', 0)): assert self.type != 'shared_library' or self.target.startswith('lib') return self.target if self.type == 'shared_library': # For reasons of convention, the Android build system requires that all # shared library modules are named 'libfoo' when generating -l flags. prefix = 'lib_' else: prefix = '' if spec['toolset'] == 'host': suffix = '_host_gyp' else: suffix = '_gyp' if self.path: name = '%s%s_%s%s' % (prefix, self.path, self.target, suffix) else: name = '%s%s%s' % (prefix, self.target, suffix) return make.StringToMakefileVariable(name) def ComputeOutputParts(self, spec): """Return the 'output basename' of a gyp spec, split into filename + ext. Android libraries must be named the same thing as their module name, otherwise the linker can't find them, so product_name and so on must be ignored if we are building a library, and the "lib" prepending is not done for Android. """ assert self.type != 'loadable_module' # TODO: not supported? target = spec['target_name'] target_prefix = '' target_ext = '' if self.type == 'static_library': target = self.ComputeAndroidModule(spec) target_ext = '.a' elif self.type == 'shared_library': target = self.ComputeAndroidModule(spec) target_ext = '.so' elif self.type == 'none': target_ext = '.stamp' elif self.type != 'executable': print ("ERROR: What output file should be generated?", "type", self.type, "target", target) if self.type != 'static_library' and self.type != 'shared_library': target_prefix = spec.get('product_prefix', target_prefix) target = spec.get('product_name', target) product_ext = spec.get('product_extension') if product_ext: target_ext = '.' + product_ext target_stem = target_prefix + target return (target_stem, target_ext) def ComputeOutputBasename(self, spec): """Return the 'output basename' of a gyp spec. E.g., the loadable module 'foobar' in directory 'baz' will produce 'libfoobar.so' """ return ''.join(self.ComputeOutputParts(spec)) def ComputeOutput(self, spec): """Return the 'output' (full output path) of a gyp spec. E.g., the loadable module 'foobar' in directory 'baz' will produce '$(obj)/baz/libfoobar.so' """ if self.type == 'executable' and self.toolset == 'host': # We install host executables into shared_intermediate_dir so they can be # run by gyp rules that refer to PRODUCT_DIR. path = '$(gyp_shared_intermediate_dir)' elif self.type == 'shared_library': if self.toolset == 'host': path = '$(HOST_OUT_INTERMEDIATE_LIBRARIES)' else: path = '$(TARGET_OUT_INTERMEDIATE_LIBRARIES)' else: # Other targets just get built into their intermediate dir. if self.toolset == 'host': path = '$(call intermediates-dir-for,%s,%s,true)' % (self.android_class, self.android_module) else: path = '$(call intermediates-dir-for,%s,%s)' % (self.android_class, self.android_module) assert spec.get('product_dir') is None # TODO: not supported? return os.path.join(path, self.ComputeOutputBasename(spec)) def NormalizeIncludePaths(self, include_paths): """ Normalize include_paths. Convert absolute paths to relative to the Android top directory; filter out include paths that are already brought in by the Android build system. Args: include_paths: A list of unprocessed include paths. Returns: A list of normalized include paths. """ normalized = [] for path in include_paths: if path[0] == '/': path = gyp.common.RelativePath(path, self.android_top_dir) # Filter out the Android standard search path. if path not in android_standard_include_paths: normalized.append(path) return normalized def ExtractIncludesFromCFlags(self, cflags): """Extract includes "-I..." out from cflags Args: cflags: A list of compiler flags, which may be mixed with "-I.." Returns: A tuple of lists: (clean_clfags, include_paths). "-I.." is trimmed. """ clean_cflags = [] include_paths = [] for flag in cflags: if flag.startswith('-I'): include_paths.append(flag[2:]) else: clean_cflags.append(flag) return (clean_cflags, include_paths) def ComputeAndroidLibraryModuleNames(self, libraries): """Compute the Android module names from libraries, ie spec.get('libraries') Args: libraries: the value of spec.get('libraries') Returns: A tuple (static_lib_modules, dynamic_lib_modules) """ static_lib_modules = [] dynamic_lib_modules = [] for libs in libraries: # Libs can have multiple words. for lib in libs.split(): # Filter the system libraries, which are added by default by the Android # build system. if (lib == '-lc' or lib == '-lstdc++' or lib == '-lm' or lib.endswith('libgcc.a')): continue match = re.search(r'([^/]+)\.a$', lib) if match: static_lib_modules.append(match.group(1)) continue match = re.search(r'([^/]+)\.so$', lib) if match: dynamic_lib_modules.append(match.group(1)) continue # "-lstlport" -> libstlport if lib.startswith('-l'): if lib.endswith('_static'): static_lib_modules.append('lib' + lib[2:]) else: dynamic_lib_modules.append('lib' + lib[2:]) return (static_lib_modules, dynamic_lib_modules) def ComputeDeps(self, spec): """Compute the dependencies of a gyp spec. Returns a tuple (deps, link_deps), where each is a list of filenames that will need to be put in front of make for either building (deps) or linking (link_deps). """ deps = [] link_deps = [] if 'dependencies' in spec: deps.extend([target_outputs[dep] for dep in spec['dependencies'] if target_outputs[dep]]) for dep in spec['dependencies']: if dep in target_link_deps: link_deps.append(target_link_deps[dep]) deps.extend(link_deps) return (gyp.common.uniquer(deps), gyp.common.uniquer(link_deps)) def WriteTargetFlags(self, spec, configs, link_deps): """Write Makefile code to specify the link flags and library dependencies. spec, configs: input from gyp. link_deps: link dependency list; see ComputeDeps() """ for configname, config in sorted(configs.iteritems()): ldflags = list(config.get('ldflags', [])) self.WriteLn('') self.WriteList(ldflags, 'LOCAL_LDFLAGS_%s' % configname) self.WriteLn('\nLOCAL_LDFLAGS := $(LOCAL_LDFLAGS_$(GYP_CONFIGURATION))') # Libraries (i.e. -lfoo) libraries = gyp.common.uniquer(spec.get('libraries', [])) static_libs, dynamic_libs = self.ComputeAndroidLibraryModuleNames( libraries) # Link dependencies (i.e. libfoo.a, libfoo.so) static_link_deps = [x[1] for x in link_deps if x[0] == 'static'] shared_link_deps = [x[1] for x in link_deps if x[0] == 'shared'] self.WriteLn('') self.WriteList(static_libs + static_link_deps, 'LOCAL_STATIC_LIBRARIES') self.WriteLn('# Enable grouping to fix circular references') self.WriteLn('LOCAL_GROUP_STATIC_LIBRARIES := true') self.WriteLn('') self.WriteList(dynamic_libs + shared_link_deps, 'LOCAL_SHARED_LIBRARIES') def WriteTarget(self, spec, configs, deps, link_deps, part_of_all): """Write Makefile code to produce the final target of the gyp spec. spec, configs: input from gyp. deps, link_deps: dependency lists; see ComputeDeps() part_of_all: flag indicating this target is part of 'all' """ self.WriteLn('### Rules for final target.') if self.type != 'none': self.WriteTargetFlags(spec, configs, link_deps) # Add to the set of targets which represent the gyp 'all' target. We use the # name 'gyp_all_modules' as the Android build system doesn't allow the use # of the Make target 'all' and because 'all_modules' is the equivalent of # the Make target 'all' on Android. if part_of_all: self.WriteLn('# Add target alias to "gyp_all_modules" target.') self.WriteLn('.PHONY: gyp_all_modules') self.WriteLn('gyp_all_modules: %s' % self.android_module) self.WriteLn('') # Add an alias from the gyp target name to the Android module name. This # simplifies manual builds of the target, and is required by the test # framework. if self.target != self.android_module: self.WriteLn('# Alias gyp target name.') self.WriteLn('.PHONY: %s' % self.target) self.WriteLn('%s: %s' % (self.target, self.android_module)) self.WriteLn('') # Add the command to trigger build of the target type depending # on the toolset. Ex: BUILD_STATIC_LIBRARY vs. BUILD_HOST_STATIC_LIBRARY # NOTE: This has to come last! modifier = '' if self.toolset == 'host': modifier = 'HOST_' if self.type == 'static_library': self.WriteLn('include $(BUILD_%sSTATIC_LIBRARY)' % modifier) elif self.type == 'shared_library': self.WriteLn('LOCAL_PRELINK_MODULE := false') self.WriteLn('include $(BUILD_%sSHARED_LIBRARY)' % modifier) elif self.type == 'executable': if self.toolset == 'host': self.WriteLn('LOCAL_MODULE_PATH := $(gyp_shared_intermediate_dir)') else: # Don't install target executables for now, as it results in them being # included in ROM. This can be revisited if there's a reason to install # them later. self.WriteLn('LOCAL_UNINSTALLABLE_MODULE := true') self.WriteLn('include $(BUILD_%sEXECUTABLE)' % modifier) else: self.WriteLn('LOCAL_MODULE_PATH := $(PRODUCT_OUT)/gyp_stamp') self.WriteLn('LOCAL_UNINSTALLABLE_MODULE := true') self.WriteLn() self.WriteLn('include $(BUILD_SYSTEM)/base_rules.mk') self.WriteLn() self.WriteLn('$(LOCAL_BUILT_MODULE): $(LOCAL_ADDITIONAL_DEPENDENCIES)') self.WriteLn('\t$(hide) echo "Gyp timestamp: $@"') self.WriteLn('\t$(hide) mkdir -p $(dir $@)') self.WriteLn('\t$(hide) touch $@') def WriteList(self, value_list, variable=None, prefix='', quoter=make.QuoteIfNecessary, local_pathify=False): """Write a variable definition that is a list of values. E.g. WriteList(['a','b'], 'foo', prefix='blah') writes out foo = blaha blahb but in a pretty-printed style. """ values = '' if value_list: value_list = [quoter(prefix + l) for l in value_list] if local_pathify: value_list = [self.LocalPathify(l) for l in value_list] values = ' \\\n\t' + ' \\\n\t'.join(value_list) self.fp.write('%s :=%s\n\n' % (variable, values)) def WriteLn(self, text=''): self.fp.write(text + '\n') def LocalPathify(self, path): """Convert a subdirectory-relative path into a normalized path which starts with the make variable $(LOCAL_PATH) (i.e. the top of the project tree). Absolute paths, or paths that contain variables, are just normalized.""" if '$(' in path or os.path.isabs(path): # path is not a file in the project tree in this case, but calling # normpath is still important for trimming trailing slashes. return os.path.normpath(path) local_path = os.path.join('$(LOCAL_PATH)', self.path, path) local_path = os.path.normpath(local_path) # Check that normalizing the path didn't ../ itself out of $(LOCAL_PATH) # - i.e. that the resulting path is still inside the project tree. The # path may legitimately have ended up containing just $(LOCAL_PATH), though, # so we don't look for a slash. assert local_path.startswith('$(LOCAL_PATH)'), ( 'Path %s attempts to escape from gyp path %s !)' % (path, self.path)) return local_path def ExpandInputRoot(self, template, expansion, dirname): if '%(INPUT_ROOT)s' not in template and '%(INPUT_DIRNAME)s' not in template: return template path = template % { 'INPUT_ROOT': expansion, 'INPUT_DIRNAME': dirname, } return path def PerformBuild(data, configurations, params): # The android backend only supports the default configuration. options = params['options'] makefile = os.path.abspath(os.path.join(options.toplevel_dir, 'GypAndroid.mk')) env = dict(os.environ) env['ONE_SHOT_MAKEFILE'] = makefile arguments = ['make', '-C', os.environ['ANDROID_BUILD_TOP'], 'gyp_all_modules'] print 'Building: %s' % arguments subprocess.check_call(arguments, env=env) def GenerateOutput(target_list, target_dicts, data, params): options = params['options'] generator_flags = params.get('generator_flags', {}) builddir_name = generator_flags.get('output_dir', 'out') limit_to_target_all = generator_flags.get('limit_to_target_all', False) android_top_dir = os.environ.get('ANDROID_BUILD_TOP') assert android_top_dir, '$ANDROID_BUILD_TOP not set; you need to run lunch.' def CalculateMakefilePath(build_file, base_name): """Determine where to write a Makefile for a given gyp file.""" # Paths in gyp files are relative to the .gyp file, but we want # paths relative to the source root for the master makefile. Grab # the path of the .gyp file as the base to relativize against. # E.g. "foo/bar" when we're constructing targets for "foo/bar/baz.gyp". base_path = gyp.common.RelativePath(os.path.dirname(build_file), options.depth) # We write the file in the base_path directory. output_file = os.path.join(options.depth, base_path, base_name) assert not options.generator_output, ( 'The Android backend does not support options.generator_output.') base_path = gyp.common.RelativePath(os.path.dirname(build_file), options.toplevel_dir) return base_path, output_file # TODO: search for the first non-'Default' target. This can go # away when we add verification that all targets have the # necessary configurations. default_configuration = None toolsets = set([target_dicts[target]['toolset'] for target in target_list]) for target in target_list: spec = target_dicts[target] if spec['default_configuration'] != 'Default': default_configuration = spec['default_configuration'] break if not default_configuration: default_configuration = 'Default' srcdir = '.' makefile_name = 'GypAndroid' + options.suffix + '.mk' makefile_path = os.path.join(options.toplevel_dir, makefile_name) assert not options.generator_output, ( 'The Android backend does not support options.generator_output.') gyp.common.EnsureDirExists(makefile_path) root_makefile = open(makefile_path, 'w') root_makefile.write(header) # We set LOCAL_PATH just once, here, to the top of the project tree. This # allows all the other paths we use to be relative to the Android.mk file, # as the Android build system expects. root_makefile.write('\nLOCAL_PATH := $(call my-dir)\n') # Find the list of targets that derive from the gyp file(s) being built. needed_targets = set() for build_file in params['build_files']: for target in gyp.common.AllTargets(target_list, target_dicts, build_file): needed_targets.add(target) build_files = set() include_list = set() android_modules = {} for qualified_target in target_list: build_file, target, toolset = gyp.common.ParseQualifiedTarget( qualified_target) relative_build_file = gyp.common.RelativePath(build_file, options.toplevel_dir) build_files.add(relative_build_file) included_files = data[build_file]['included_files'] for included_file in included_files: # The included_files entries are relative to the dir of the build file # that included them, so we have to undo that and then make them relative # to the root dir. relative_include_file = gyp.common.RelativePath( gyp.common.UnrelativePath(included_file, build_file), options.toplevel_dir) abs_include_file = os.path.abspath(relative_include_file) # If the include file is from the ~/.gyp dir, we should use absolute path # so that relocating the src dir doesn't break the path. if (params['home_dot_gyp'] and abs_include_file.startswith(params['home_dot_gyp'])): build_files.add(abs_include_file) else: build_files.add(relative_include_file) base_path, output_file = CalculateMakefilePath(build_file, target + '.' + toolset + options.suffix + '.mk') spec = target_dicts[qualified_target] configs = spec['configurations'] part_of_all = (qualified_target in needed_targets and not int(spec.get('suppress_wildcard', False))) if limit_to_target_all and not part_of_all: continue relative_target = gyp.common.QualifiedTarget(relative_build_file, target, toolset) writer = AndroidMkWriter(android_top_dir) android_module = writer.Write(qualified_target, relative_target, base_path, output_file, spec, configs, part_of_all=part_of_all) if android_module in android_modules: print ('ERROR: Android module names must be unique. The following ' 'targets both generate Android module name %s.\n %s\n %s' % (android_module, android_modules[android_module], qualified_target)) return android_modules[android_module] = qualified_target # Our root_makefile lives at the source root. Compute the relative path # from there to the output_file for including. mkfile_rel_path = gyp.common.RelativePath(output_file, os.path.dirname(makefile_path)) include_list.add(mkfile_rel_path) root_makefile.write('GYP_CONFIGURATION ?= %s\n' % default_configuration) # Write out the sorted list of includes. root_makefile.write('\n') for include_file in sorted(include_list): root_makefile.write('include $(LOCAL_PATH)/' + include_file + '\n') root_makefile.write('\n') root_makefile.write(SHARED_FOOTER) root_makefile.close()
gpl-3.0
ichuang/sympy
sympy/polys/domains/gmpyrationalfield.py
3
3001
"""Implementaton of :class:`GMPYRationalField` class. """ from sympy.polys.domains.rationalfield import RationalField from sympy.polys.domains.groundtypes import ( GMPYRationalType, SymPyRationalType, gmpy_numer, gmpy_denom, gmpy_factorial, ) from sympy.polys.polyerrors import CoercionFailed class GMPYRationalField(RationalField): """Rational field based on GMPY mpq class. """ dtype = GMPYRationalType zero = dtype(0) one = dtype(1) alias = 'QQ_gmpy' def __init__(self): pass def to_sympy(self, a): """Convert `a` to a SymPy object. """ return SymPyRationalType(int(gmpy_numer(a)), int(gmpy_denom(a))) def from_sympy(self, a): """Convert SymPy's Integer to `dtype`. """ if a.is_Rational: return GMPYRationalType(a.p, a.q) elif a.is_Float: from sympy.polys.domains import RR return GMPYRationalType(*RR.as_integer_ratio(a)) else: raise CoercionFailed("expected `Rational` object, got %s" % a) def from_ZZ_python(K1, a, K0): """Convert a Python `int` object to `dtype`. """ return GMPYRationalType(a) def from_QQ_python(K1, a, K0): """Convert a Python `Fraction` object to `dtype`. """ return GMPYRationalType(a.numerator, a.denominator) def from_ZZ_sympy(K1, a, K0): """Convert a SymPy `Integer` object to `dtype`. """ return GMPYRationalType(a.p) def from_QQ_sympy(K1, a, K0): """Convert a SymPy `Rational` object to `dtype`. """ return GMPYRationalType(a.p, a.q) def from_ZZ_gmpy(K1, a, K0): """Convert a GMPY `mpz` object to `dtype`. """ return GMPYRationalType(a) def from_QQ_gmpy(K1, a, K0): """Convert a GMPY `mpq` object to `dtype`. """ return a def from_RR_sympy(K1, a, K0): """Convert a SymPy `Float` object to `dtype`. """ return GMPYRationalType(*K0.as_integer_ratio(a)) def from_RR_mpmath(K1, a, K0): """Convert a mpmath `mpf` object to `dtype`. """ return GMPYRationalType(*K0.as_integer_ratio(a)) def exquo(self, a, b): """Exact quotient of `a` and `b`, implies `__div__`. """ return GMPYRationalType(a.qdiv(b)) def quo(self, a, b): """Quotient of `a` and `b`, implies `__div__`. """ return GMPYRationalType(a.qdiv(b)) def rem(self, a, b): """Remainder of `a` and `b`, implies nothing. """ return self.zero def div(self, a, b): """Division of `a` and `b`, implies `__div__`. """ return GMPYRationalType(a.qdiv(b)), self.zero def numer(self, a): """Returns numerator of `a`. """ return gmpy_numer(a) def denom(self, a): """Returns denominator of `a`. """ return gmpy_denom(a) def factorial(self, a): """Returns factorial of `a`. """ return GMPYRationalType(gmpy_factorial(int(a)))
bsd-3-clause
cipriancraciun/extremely-simple-cluster-platform
components/py-tools/sources/escp/tools/enforcements.py
1
23625
import re import types import escp.tools.callables as callables import escp.tools.outcomes as outcomes _e_debug = False class Enforcer (object) : def __init__ (self) : return def __nonzero__ (self) : return _e_debug def enforce (self, _condition) : if _condition is not True : raise EnforcementError (self.enforce) return True def enforce_is (self, _value_a, _value_b) : if _value_a is not _value_b : raise EnforcementError (self.enforce_is, value_a = _value_a, value_b = _value_b) return True def enforce_is_any (self, _value, _values) : for _value_ in _values : if _value is _value_ : return True raise EnforcementError (self.enforce_is_any, value = _value, values = _values) def enforce_is_not (self, _value_a, _value_b) : if _value_a is _value_b : raise EnforcementError (self.enforce_is_not, value_a = _value_a, value_b = _value_b) return True def enforce_is_not_all (self, _value, _values) : for _value_ in _values : if _value is _value_ : raise EnforcementError (self.enforce_is_not_all, value = _value, values = _values) return True def enforce_is_not_any (self, _value, _values) : for _value_ in _values : if _value is not _value_ : return True raise EnforcementError (self.enforce_is_not_any, value = _value, values = _values) def enforce_is_none (self, _value) : if _value is not None : raise EnforcementError (self.enforce_is_none, value = _value) return True def enforce_is_none_all (self, _values) : for _value in _values : if _value is not None : raise EnforcementError (self.enforce_is_none_all, value = _value) return True def enforce_is_none_any (self, _values) : for _value in _values : if _value is None : return True raise EnforcementError (self.enforce_is_none_any, values = _values) def enforce_is_not_none (self, _value) : if _value is None : raise EnforcementError (self.enforce_is_not_none, value = _value) return True def enforce_is_not_none_all (self, _values) : for _value in _values : if _value is None : raise EnforcementError (self.enforce_is_not_none_all, value = _value) return True def enforce_is_not_none_any (self, _values) : for _value in _values : if _value is not None : return True raise EnforcementError (self.enforce_is_not_none_any, values = _values) def enforce_is_true (self, _value) : if _value is not True : raise EnforcementError (self.enforce_is_true, value = _value) return True def enforce_is_true_all (self, _values) : for _value in _values : if _value is not True : raise EnforcementError (self.enforce_is_true_all, values = _values) return True def enforce_is_true_any (self, _values) : for _value in _values : if _value is True : return raise EnforcementError (self.enforce_is_true_any, values = _values) def enforce_is_true_or_none (self, _value) : if _value is not None and _value is not True : raise EnforcementError (self.enforce_is_true_or_none, value = _value) return True def enforce_is_false (self, _value) : if _value is not False : raise EnforcementError (self.enforce_is_false, value = _value) return True def enforce_is_false_all (self, _values) : for _value in _values : if _value is not False : raise EnforcementError (self.enforce_is_false_all, values = _values) return True def enforce_is_false_any (self, _values) : for _value in _values : if _value is False : return True raise EnforcementError (self.enforce_is_false_any, values = _values) def enforce_is_false_or_none (self, _value) : if _value is not None and _value is not False : raise EnforcementError (self.enforce_is_false_or_none, value = _value) return True def enforce_instance (self, _value, _type) : if not isinstance (_value, _type) : raise EnforcementError (self.enforce_instance, value = _value, value_type = type (_value), expected_type = _type) return True def enforce_instance_or_none (self, _value, _type) : if _value is not None and not isinstance (_value, _type) : raise EnforcementError (self.enforce_instance_or_none, value = _value, value_type = type (_value), expected_type = _type) return True def enforce_int (self, _value) : if not isinstance (_value, int) : raise EnforcementError (self.enforce_int, value = _value, value_type = type (_value)) return True def enforce_positive_int (self, _value) : if not isinstance (_value, int) or _value < 0 : raise EnforcementError (self.enforce_positive_int, value = _value, value_type = type (_value)) return True def enforce_int_or_none (self, _value) : if _value is not None and not isinstance (_value, int) : raise EnforcementError (self.enforce_int_or_none, value = _value, value_type = type (_value)) return True def enforce_float (self, _value) : if not isinstance (_value, float) : raise EnforcementError (self.enforce_float, value = _value, value_type = type (_value)) return True def enforce_float_or_none (self, _value) : if _value is not None and not isinstance (_value, float) : raise EnforcementError (self.enforce_float_or_none, value = _value, value_type = type (_value)) return True def enforce_positive_float (self, _value) : if not isinstance (_value, float) or _value < 0.0 : raise EnforcementError (self.enforce_positive_float, value = _value, value_type = type (_value)) return True def enforce_positive_float_or_none (self, _value) : if _value is not None and (not isinstance (_value, float) or _value < 0.0) : raise EnforcementError (self.enforce_positive_float_or_none, value = _value, value_type = type (_value)) return True def enforce_bool (self, _value) : if not isinstance (_value, bool) : raise EnforcementError (self.enforce_bool, value = _value, value_type = type (_value)) return True def enforce_bool_all (self, _values) : for _value in _values : if not isinstance (_value, bool) : raise EnforcementError (self.enforce_bool_all, values = _values, value_type = type (_value)) return True def enforce_bool_or_none (self, _value) : if _value is not None and not isinstance (_value, bool) : raise EnforcementError (self.enforce_bool_or_none, value = _value, value_type = type (_value)) return True def enforce_string (self, _value) : if not isinstance (_value, basestring) : raise EnforcementError (self.enforce_string, value = _value, value_type = type (_value)) return True def enforce_string_all (self, _values) : for _value in _values : if not isinstance (_value, basestring) : raise EnforcementError (self.enforce_string_all, values = _values, value_type = type (_value)) return True def enforce_string_or_none (self, _value) : if _value is not None and not isinstance (_value, basestring) : raise EnforcementError (self.enforce_string_or_none, value = _value, value_type = type (_value)) return True def enforce_string_or_none_all (self, _values) : for _value in _values : if _value is not None and not isinstance (_value, basestring) : raise EnforcementError (self.enforce_string_or_none_all, values = _values, value_type = type (_value)) return True def enforce_str (self, _value) : if not isinstance (_value, str) : raise EnforcementError (self.enforce_str, value = _value, value_type = type (_value)) return True def enforce_str_or_none (self, _value) : if _value is not None and not isinstance (_value, str) : raise EnforcementError (self.enforce_str_or_none, value = _value, value_type = type (_value)) return True def enforce_unicode (self, _value) : if not isinstance (_value, unicode) : raise EnforcementError (self.enforce_unicode, value = _value, value_type = type (_value)) return True def enforce_unicode_or_none (self, _value) : if _value is not None and not isinstance (_value, unicode) : raise EnforcementError (self.enforce_unicode_or_none, value = _value, value_type = type (_value)) return True def enforce_tuple (self, _value, _length = None) : if not isinstance (_value, tuple) or (_length is not None and len (_value) != _length) : raise EnforcementError (self.enforce_tuple, value = _value, value_type = type (_value), length = _length) return True def enforce_tuple_or_none (self, _value, _length = None) : if _value is not None and (not isinstance (_value, tuple) or (_length is not None and len (_value) != _length)) : raise EnforcementError (self.enforce_tuple_or_none, value = _value, value_type = type (_value), length = _length) return True def enforce_list (self, _value, _length = None) : if not isinstance (_value, list) or (_length is not None and len (_value) != _length) : raise EnforcementError (self.enforce_list, value = _value, value_type = type (_value), length = _length) return True def enforce_not_empty_list (self, _value) : if not isinstance (_value, list) or len (_value) == 0 : raise EnforcementError (self.enforce_not_empty_list, value = _value, value_type = type (_value)) return True def enforce_list_or_none (self, _value, _length = None) : if _value is not None and (not isinstance (_value, list) or (_length is not None and len (_value) != _length)) : raise EnforcementError (self.enforce_list_or_none, value = _value, value_type = type (_value), length = _length) return True def enforce_not_empty_list_or_none (self, _value) : if _value is not None and (not isinstance (_value, list) or len (_value) == 0) : raise EnforcementError (self.enforce_not_empty_list_or_none, value = _value, value_type = type (_value)) return True def enforce_set (self, _value) : if not isinstance (_value, set) : raise EnforcementError (self.enforce_set, value = _value, value_type = type (_value)) return True def enforce_set_or_none (self, _value) : if _value is not None and not isinstance (_value, set) : raise EnforcementError (self.enforce_set_or_none, value = _value, value_type = type (_value)) return True def enforce_dict (self, _value) : if not isinstance (_value, dict) : raise EnforcementError (self.enforce_dict, value = _value, value_type = type (_value)) return True def enforce_dict_or_none (self, _value) : if _value is not None and not isinstance (_value, dict) : raise EnforcementError (self.enforce_dict_or_none, value = _value, value_type = type (_value)) return True def enforce_tuple_or_list (self, _value) : if not isinstance (_value, tuple) and not isinstance (_value, list) : raise EnforcementError (self.enforce_tuple_or_list, value = _value, value_type = type (_value)) return True def enforce_tuple_or_list_or_none (self, _value) : if _value is not None and not isinstance (_value, tuple) and not isinstance (_value, list) : raise EnforcementError (self.enforce_tuple_or_list_or_none, value = _value, value_type = type (_value)) return True def enforce_type (self, _value) : if type (_value) is not type : raise EnforcementError (self.enforce_type, value = _value, value_type = type (_value)) return True def enforce_type_or_none (self, _value) : if _value is not None and type (_value) is not type : raise EnforcementError (self.enforce_type_or_none, value = _value, value_type = type (_value)) return True def enforce_class (self, _value) : if type (_value) is not types.ClassType : raise EnforcementError (self.enforce_class, value = _value, value_type = type (_value)) return True def enforce_class_or_none (self, _value) : if _value is not None and type (_value) is not types.ClassType : raise EnforcementError (self.enforce_class_or_none, value = _value, value_type = type (_value)) return True def enforce_type_or_class (self, _value) : if type (_value) is not type and type (_value) is not types.ClassType : raise EnforcementError (self.enforce_type_or_class, value = _value, value_type = type (_value)) return True def enforce_type_or_class_or_none (self, _value) : if _value is not None and type (_value) is not type and type (_value) is not types.ClassType : raise EnforcementError (self.enforce_type_or_class_or_none, value = _value, value_type = type (_value)) return True def enforce_in (self, _collection, _value) : if _value not in _collection : raise EnforcementError (self.enforce_in, collection = _collection, value = _value) return True def enforce_in_all (self, _collection, _values) : for _value in _values : if _value not in _collection : raise EnforcementError (self.enforce_in_all, collection = _collection, values = _values) return True def enforce_in_any (self, _collection, _values) : for _value in _values : if _value in _collection : return True raise EnforcementError (self.enforce_in_any, collection = _collection, values = _values) def enforce_not_in (self, _collection, _value) : if _value in _collection : raise EnforcementError (self.enforce_not_in, collection = _collection, value = _value) return True def enforce_not_in_all (self, _collection, _values) : for _value in _values : if _value in _collection : raise EnforcementError (self.enforce_not_in_all, collection = _collection, values = _values) return True def enforce_not_in_any (self, _collection, _values) : for _value in _values : if _value not in _collection : return True raise EnforcementError (self.enforce_not_in_any, collection = _collection, values = _values) def enforce_empty (self, _value) : if len (_value) != 0 : raise EnforcementError (self.enforce_empty, value = _value, value_type = type (_value)) return True def enforce_not_empty (self, _value) : if len (_value) == 0 : raise EnforcementError (self.enforce_not_empty, value = _value, value_type = type (_value)) return True def enforce_eq (self, _value_a, _value_b) : if not (_value_a == _value_b) : raise EnforcementError (self.enforce_eq, value_a = _value_a, value_b = _value_b) return True def enforce_eq_or_none (self, _value_a, _value_b) : if _value_a is not None and _value_b is not None and not (_value_a == _value_b) : raise EnforcementError (self.enforce_eq_or_none, value_a = _value_a, value_b = _value_b) return True def enforce_ne (self, _value_a, _value_b) : if not (_value_a != _value_b) : raise EnforcementError (self.enforce_ne, value_a = _value_a, value_b = _value_b) return True def enforce_ne_or_none (self, _value_a, _value_b) : if _value_a is not None and _value_b is not None and not (_value_a != _value_b) : raise EnforcementError (self.enforce_ne_or_none, value_a = _value_a, value_b = _value_b) return True def enforce_lt (self, _value_a, _value_b) : if not (_value_a < _value_b) : raise EnforcementError (self.enforce_lt, value_a = _value_a, value_b = _value_b) return True def enforce_le (self, _value_a, _value_b) : if not (_value_a <= _value_b) : raise EnforcementError (self.enforce_le, value_a = _value_a, value_b = _value_b) return True def enforce_gt (self, _value_a, _value_b) : if not (_value_a > _value_b) : raise EnforcementError (self.enforce_gt, value_a = _value_a, value_b = _value_b) return True def enforce_ge (self, _value_a, _value_b) : if not (_value_a >= _value_b) : raise EnforcementError (self.enforce_ge, value_a = _value_a, value_b = _value_b) return True def enforce_zero (self, _value) : if _value != 0 : raise EnforcementError (self.enforce_zero, value = _value) return True def enforce_callable (self, _callable, _argument_count = None) : if not callable (_callable) or (_argument_count is not None and not callables.match_callable_spec_argument_count (_callable, _argument_count)) : raise EnforcementError (self.enforce_callable, callable = _callable, argument_count = _argument_count) return True def enforce_callable_all (self, _callables, _argument_count = None) : for _callable in _callables : if not callable (_callable) or (_argument_count is not None and not callables.match_callable_spec_argument_count (_callable, _argument_count)) : raise EnforcementError (self.enforce_callable_all, callables = _callables, argument_count = _argument_count) return True def enforce_callable_any (self, _callables, _argument_count = None) : for _callable in _callables : if callable (_callable) and (_argument_count is None or callables.match_callable_spec_argument_count (_callable, _argument_count)) : return True raise EnforcementError (self.enforce_callable_any, callables = _callables, argument_count = _argument_count) def enforce_callable_or_none (self, _callable, _argument_count = None) : if _callable is not None and (not callable (_callable) or (_argument_count is not None and not callables.match_callable_spec_argument_count (_callable, _argument_count))) : raise EnforcementError (self.enforce_callable_or_none, callable = _callable, argument_count = _argument_count) return True def enforce_callable_or_none_all (self, _callables, _argument_count = None) : for _callable in _callables : if _callable is not None and (not callable (_callable) or (_argument_count is not None and not callables.match_callable_spec_argument_count (_callable, _argument_count))) : raise EnforcementError (self.enforce_callable_or_none_all, callables = _callables, argument_count = _argument_count) return True def enforce_callable_or_none_any (self, _callables, _argument_count = None) : for _callable in _callables : if _callable is None or (callable (_callable) and (_argument_count is None or callables.match_callable_spec_argument_count (_callable, _argument_count))) : return True raise EnforcementError (self.enforce_callable_or_none_any, callables = _callables, argument_count = _argument_count) def enforce_callable_spec (self, _callable, _argument_count = None) : if (not isinstance (_callable, tuple) and not callable (_callable)) or (_argument_count is not None and not callables.match_callable_spec_argument_count (_callable, _argument_count)) : raise EnforcementError (self.enforce_callable_spec, callable = _callable, argument_count = _argument_count) return True def enforce_callable_spec_all (self, _callables, _argument_count = None) : for _callable in _callables : if (not isinstance (_callable, tuple) and not callable (_callable)) or (_argument_count is not None and not callables.match_callable_spec_argument_count (_callable, _argument_count)) : raise EnforcementError (self.enforce_callable_spec_all, callables = _callables, argument_count = _argument_count) return True def enforce_callable_spec_any (self, _callables, _argument_count = None) : for _callable in _callables : if (isinstance (_callable, tuple) or callable (_callable)) and (_argument_count is None or callables.match_callable_spec_argument_count (_callable, _argument_count)) : return True raise EnforcementError (self.enforce_callable_spec_any, callables = _callables, argument_count = _argument_count) def enforce_callable_spec_or_none (self, _callable, _argument_count = None) : if _callable is not None and ((not isinstance (_callable, tuple) and not callable (_callable)) or (_argument_count is not None and not callables.match_callable_spec_argument_count (_callable, _argument_count))) : raise EnforcementError (self.enforce_callable_spec_or_none, callable = _callable, argument_count = _argument_count) return True def enforce_callable_spec_or_none_all (self, _callables, _argument_count = None) : for _callable in _callables : if _callable is not None and ((not isinstance (_callable, tuple) and not callable (_callable)) or (_argument_count is not None and not callables.match_callable_spec_argument_count (_callable, _argument_count))) : raise EnforcementError (self.enforce_callable_spec_or_none_all, callables = _callables, argument_count = _argument_count) return True def enforce_callable_spec_or_none_any (self, _callables, _argument_count = None) : for _callable in _callables : if _callable is not None or ((isinstance (_callable, tuple) or not callable (_callable)) and (_argument_count is None or callables.match_callable_spec_argument_count (_callable, _argument_count))) : return True raise EnforcementError (self.enforce_callable_spec_or_none_any, callables = _callables, argument_count = _argument_count) def enforce_re_match (self, _value, _re) : if not isinstance (_value, basestring) or _re.match (_value) is None : raise EnforcementError (self.enforce_re_match, value = _value, re = _re) return True def enforce_re_match_all (self, _values, _re) : for _value in _values : if not isinstance (_value, basestring) or _re.match (_value) is None : raise EnforcementError (self.enforce_re_match_all, values = _values, re = _re) return True def enforce_re_match_or_none (self, _value, _re) : if _value is not None and _re.match (_value) is None : raise EnforcementError (self.enforce_re_match_or_none, value = _value, re = _re) return True def enforce_hasattr (self, _object, _attr) : if not hasattr (_object, _attr) : raise EnforcementError (self.enforce_hasattr, object = _object, attr = _attr) return True def enforce_hasattr_all (self, _object, _attrs) : for _attr in _attrs : if not hasattr (_object, _attr) : raise EnforcementError (self.enforce_hasattr_all, object = _object, attrs = _attrs) return True def enforce_not_hasattr (self, _object, _attr) : if hasattr (_object, _attr) : raise EnforcementError (self.enforce_not_hasattr, object = _object, attr = _attr) return True def enforce_not_hasattr_all (self, _object, _attrs) : for _attr in _attrs : if hasattr (_object, _attr) : raise EnforcementError (self.enforce_not_hasattr_all, object = _object, attrs = _attrs) return True def enforce_all (self, _enforcements) : for _enforcement in _enforcements : try : _enforcement[0] (* _enforcement[1:]) except Exception : raise return True def enforce_any (self, _enforcements) : for _enforcement in _enforcements : try : _enforcement[0] (* _enforcement[1:]) return True except Exception : pass raise EnforcementError (self.enforce_any, enforcements = _enforcements) def enforce_one (self, _enforcements) : _succeeded = 0 for _enforcement in _enforcements : try : _enforcement[0] (* _enforcement[1:]) _succeeded += 1 except Exception : pass if _succeeded == 1 : return True raise EnforcementError (self.enforce_one, enforcements = _enforcements) def enforcement_raise (self) : raise EnforcementError (self.enforcement_raise) class EnforcementError (outcomes.Failure) : def __init__ (self, _enforcement, ** _fields) : outcomes.Failure.__init__ (self, 'enforcement broken', None, _fields) def enforcer_ef (_owner, parent = None) : if _enforcer is not None : return _enforcer return Enforcer () _enforcer = None _enforcer = enforcer_ef (None)
gpl-3.0
linearregression/mpld3
mpld3/__init__.py
20
1109
""" Interactive D3 rendering of matplotlib images ============================================= Functions: General Use ---------------------- :func:`fig_to_html` convert a figure to an html string :func:`fig_to_dict` convert a figure to a dictionary representation :func:`show` launch a web server to view an d3/html figure representation :func:`save_html` save a figure to an html file :func:`save_json` save a JSON representation of a figure to file Functions: IPython Notebook --------------------------- :func:`display` display a figure in an IPython notebook :func:`enable_notebook` enable automatic D3 display of figures in the IPython notebook. :func:`disable_notebook` disable automatic D3 display of figures in the IPython """ __all__ = ["__version__", "fig_to_html", "fig_to_dict", "fig_to_d3", "display_d3", "display", "show_d3", "show", "save_html", "save_json", "enable_notebook", "disable_notebook", "plugins", "urls"] from .__about__ import __version__ from . import plugins from . import urls from ._display import *
bsd-3-clause
edgedb/edgedb
edb/edgeql/compiler/inference/multiplicity.py
1
21233
# # This source file is part of the EdgeDB open source project. # # Copyright 2020-present MagicStack Inc. and the EdgeDB authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # """EdgeQL multiplicity inference. A top-down multiplicity inferer that traverses the full AST populating multiplicity fields and performing multiplicity checks. """ from __future__ import annotations from typing import * import functools from edb import errors from edb.edgeql import qltypes from edb.schema import pointers as s_pointers from edb.ir import ast as irast from edb.ir import typeutils as irtyputils from . import cardinality from . import context as inference_context ZERO = qltypes.Multiplicity.ZERO ONE = qltypes.Multiplicity.ONE MANY = qltypes.Multiplicity.MANY def _max_multiplicity( args: Iterable[qltypes.Multiplicity] ) -> qltypes.Multiplicity: # Coincidentally, the lexical order of multiplicity is opposite of # order of multiplicity values. arg_list = list(args) if not arg_list: return ZERO else: return min(arg_list) def _common_multiplicity( args: Iterable[irast.Base], *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: return _max_multiplicity( infer_multiplicity(a, scope_tree=scope_tree, ctx=ctx) for a in args) @functools.singledispatch def _infer_multiplicity( ir: irast.Expr, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # return MANY raise ValueError(f'infer_multiplicity: cannot handle {ir!r}') @_infer_multiplicity.register def __infer_none( ir: None, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # Here for debugging purposes. raise ValueError('invalid infer_multiplicity(None, schema) call') @_infer_multiplicity.register def __infer_statement( ir: irast.Statement, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: return infer_multiplicity( ir.expr, scope_tree=scope_tree, ctx=ctx) @_infer_multiplicity.register def __infer_empty_set( ir: irast.EmptySet, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: return ZERO @_infer_multiplicity.register def __infer_type_introspection( ir: irast.TypeIntrospection, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # TODO: The result is always ONE, but we still want to actually # introspect the expression. Unfortunately, currently the # expression is not available at this stage. # # E.g. consider: # WITH X := Foo {bar := {Bar, Bar}} # SELECT INTROSPECT TYPEOF X.bar; return ONE def _infer_shape( ir: irast.Set, *, is_mutation: bool=False, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> None: for shape_set, _ in ir.shape: new_scope = cardinality._get_set_scope(shape_set, scope_tree, ctx=ctx) if shape_set.expr and shape_set.rptr: expr_mult = infer_multiplicity( shape_set.expr, scope_tree=new_scope, ctx=ctx) ptrref = shape_set.rptr.ptrref if expr_mult is MANY and irtyputils.is_object(ptrref.out_target): raise errors.QueryError( f'possibly not a strict set returned by an ' f'expression for a computable ' f'{ptrref.shortname.name}.', hint=( f'Use DISTINCT for the entire computable expression ' f'to resolve this.' ), context=shape_set.context ) _infer_shape( shape_set, is_mutation=is_mutation, scope_tree=scope_tree, ctx=ctx) def _infer_set( ir: irast.Set, *, is_mutation: bool=False, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: result = _infer_set_inner( ir, is_mutation=is_mutation, scope_tree=scope_tree, ctx=ctx) ctx.inferred_multiplicity[ir, scope_tree] = result # The shape doesn't affect multiplicity, but requires validation. _infer_shape(ir, is_mutation=is_mutation, scope_tree=scope_tree, ctx=ctx) return result def _infer_set_inner( ir: irast.Set, *, is_mutation: bool=False, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: rptr = ir.rptr new_scope = cardinality._get_set_scope(ir, scope_tree, ctx=ctx) if rptr is not None: # Validate the source infer_multiplicity(rptr.source, scope_tree=new_scope, ctx=ctx) if ir.expr: expr_mult = infer_multiplicity(ir.expr, scope_tree=new_scope, ctx=ctx) if rptr is not None: rptrref = rptr.ptrref if isinstance(rptr.ptrref, irast.TupleIndirectionPointerRef): # All bets are off for tuple elements. return MANY elif not irtyputils.is_object(ir.typeref): # This is not an expression and is some kind of scalar, so # multiplicity cannot be guaranteed to be ONE (most scalar # expressions don't have an implicit requirement to be sets) # unless we also have an exclusive constraint. if rptr is not None: schema = ctx.env.schema # We should only have some kind of path terminating in a # property here. assert isinstance(rptrref, irast.PointerRef) ptr = schema.get_by_id(rptrref.id, type=s_pointers.Pointer) if ptr.is_exclusive(schema): # Got an exclusive constraint return ONE return MANY else: # This is some kind of a link at the end of a path. # Therefore the target is a proper set. return ONE elif ir.expr is not None: return expr_mult else: # Evidently this is not a pointer, expression, or a scalar. # This is an object type and therefore a proper set. return ONE @_infer_multiplicity.register def __infer_func_call( ir: irast.FunctionCall, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # If the function returns a set (for any reason), all bets are off # and the maximum multiplicity cannot be inferred. card = cardinality.infer_cardinality( ir, scope_tree=scope_tree, ctx=ctx) # We still want to validate the multiplicity of the arguments, though. for arg in ir.args: infer_multiplicity(arg.expr, scope_tree=scope_tree, ctx=ctx) if card is not None and card.is_single(): return ONE elif str(ir.func_shortname) == 'std::enumerate': # Technically the output of enumerate is always of # multiplicity ONE because it's a set of tuples with first # elements being guaranteed to be distinct. return ONE else: return MANY @_infer_multiplicity.register def __infer_oper_call( ir: irast.OperatorCall, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: mult = [] cards = [] for arg in ir.args: cards.append( cardinality.infer_cardinality( arg.expr, scope_tree=scope_tree, ctx=ctx)) mult.append( infer_multiplicity( arg.expr, scope_tree=scope_tree, ctx=ctx)) op_name = str(ir.func_shortname) if op_name == 'std::UNION': # UNION will produce multiplicity MANY unless most or all of # the elements multiplicity is ZERO (from an empty set). result = ZERO for m in mult: if m is ONE and result is ZERO: result = m elif m is ONE and result is not ZERO: return MANY elif m is MANY: return MANY return result elif op_name == 'std::DISTINCT': if mult[0] is ZERO: return ZERO else: return ONE elif op_name == 'std::IF': # If the cardinality of the condition is more than ONE, then # the multiplicity cannot be inferred. if cards[1].is_single(): # Now it's just a matter of the multiplicity of the # possible results. return _max_multiplicity((mult[0], mult[2])) else: return MANY else: # The rest of the operators (other than UNION, DISTINCT, or # IF..ELSE). We can ignore the SET OF args because the results # are actually proportional to the element-wise args in our # operators. result = _max_multiplicity(mult) if result is MANY: return result # Even when arguments are of multiplicity ONE, we cannot # exclude the possibility of the result being of multiplicity # MANY. We need to check that at most one argument has # cardinality more than ONE. if len([card for card in cards if card.is_multi()]) > 1: return MANY else: return result @_infer_multiplicity.register def __infer_const( ir: irast.BaseConstant, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: return ONE @_infer_multiplicity.register def __infer_param( ir: irast.Parameter, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: return ONE @_infer_multiplicity.register def __infer_const_set( ir: irast.ConstantSet, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: if len(ir.elements) == len({el.value for el in ir.elements}): return ONE else: return MANY @_infer_multiplicity.register def __infer_typecheckop( ir: irast.TypeCheckOp, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # Unless this is a singleton, the multiplicity cannot be assumed to be ONE. card = cardinality.infer_cardinality( ir, scope_tree=scope_tree, ctx=ctx) if card is not None and card.is_single(): return ONE else: return MANY @_infer_multiplicity.register def __infer_typecast( ir: irast.TypeCast, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: return infer_multiplicity( ir.expr, scope_tree=scope_tree, ctx=ctx, ) def _infer_stmt_multiplicity( ir: irast.FilteredStmt, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: result = infer_multiplicity( ir.subject if isinstance(ir, irast.MutatingStmt) else ir.result, scope_tree=scope_tree, ctx=ctx, ) # WITH block bindings need to be validated, they don't have to # have multiplicity ONE, but their sub-expressions must be valid. # # Inferring how the FILTER clause affects multiplicity is in # general impossible, but we still want to ensure that the FILTER # expression has valid multiplicity. for part in ir.bindings + [ir.where]: if part: infer_multiplicity(part, scope_tree=scope_tree, ctx=ctx) return result def _infer_for_multiplicity( ir: irast.SelectStmt, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: assert ir.iterator_stmt is not None itexpr = ir.iterator_stmt.expr if isinstance(ir.result.expr, irast.SelectStmt): union = ir.result.expr if (isinstance(union.where, irast.Set) and isinstance(union.where.expr, irast.OperatorCall) and str(union.where.expr.func_shortname) == 'std::='): op = union.where.expr left, right = (a.expr for a in op.args) # The iterator set may be wrapped in an `enumerate`, this # requires different handling. has_enumerate = ( isinstance(itexpr, irast.SelectStmt) and isinstance(itfn := itexpr.result.expr, irast.FunctionCall) and str(itfn.func_shortname) == 'std::enumerate' ) # First make sure that the cardinality of the FILTER # expression is is no more than 1. Then make sure both # operands are paths. if union.where_card.is_single(): it = None if left.rptr is not None: it = right elif right.rptr is not None: it = left if it is not None: if has_enumerate: assert isinstance(itfn, irast.FunctionCall) enumerate_mult = infer_multiplicity( itfn.args[0].expr, scope_tree=scope_tree, ctx=ctx, ) if ( enumerate_mult is ONE and it.rptr is not None and isinstance( it.rptr, irast.TupleIndirectionPointer ) # Tuple comes from the iterator set and it.rptr.source.expr is itexpr # the indirection is accessing element 1 and str(it.rptr.ptrref.name) == '__tuple__::1' ): return ONE elif (it.is_binding and it.expr is itexpr): return ONE elif isinstance(ir.result.expr, irast.InsertStmt): # A union of inserts always has multiplicity ONE return ONE return MANY @_infer_multiplicity.register def __infer_select_stmt( ir: irast.SelectStmt, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: result = _infer_stmt_multiplicity(ir, scope_tree=scope_tree, ctx=ctx) itmult = None if ir.iterator_stmt: # If this is a FOR, then there's a common pattern which can be # detected and the multiplicity of it is ONE. Otherwise it # cannot be reliably inferred. # # The pattern is: FOR x IN {<set of multiplicity ONE>} UNION # (SELECT ... FILTER .prop = x) As long as the FILTER has just # a single .prop = x expression, this is going to be a bunch # of disjoint unions and the final multiplicity will be ONE. itmult = infer_multiplicity( ir.iterator_stmt, scope_tree=scope_tree, ctx=ctx, ) # OFFSET, LIMIT and ORDER BY have already been validated to be # singletons, but their sub-expressions (if any) still need to be # validated. for part in [ir.limit, ir.offset] + [sort.expr for sort in ir.orderby]: if part: new_scope = cardinality._get_set_scope(part, scope_tree, ctx=ctx) infer_multiplicity(part, scope_tree=new_scope, ctx=ctx) if itmult is not None: if itmult is ONE: return _infer_for_multiplicity( ir, scope_tree=scope_tree, ctx=ctx) return MANY else: return result @_infer_multiplicity.register def __infer_insert_stmt( ir: irast.InsertStmt, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # INSERT will always return a proper set, but we still want to # process the sub-expressions. infer_multiplicity( ir.subject, is_mutation=True, scope_tree=scope_tree, ctx=ctx ) new_scope = cardinality._get_set_scope(ir.result, scope_tree, ctx=ctx) infer_multiplicity( ir.result, is_mutation=True, scope_tree=new_scope, ctx=ctx ) if ir.on_conflict: for part in [ir.on_conflict.select_ir, ir.on_conflict.else_ir]: if part: infer_multiplicity(part, scope_tree=scope_tree, ctx=ctx) return ONE @_infer_multiplicity.register def __infer_update_stmt( ir: irast.UpdateStmt, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # Presumably UPDATE will always return a proper set, even if it's # fed something with higher multiplicity, but we still want to # process the expression being updated. infer_multiplicity( ir.result, is_mutation=True, scope_tree=scope_tree, ctx=ctx, ) result = _infer_stmt_multiplicity(ir, scope_tree=scope_tree, ctx=ctx) if result is ZERO: return ZERO else: return ONE @_infer_multiplicity.register def __infer_delete_stmt( ir: irast.DeleteStmt, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # Presumably DELETE will always return a proper set, even if it's # fed something with higher multiplicity, but we still want to # process the expression being deleted. infer_multiplicity( ir.result, is_mutation=True, scope_tree=scope_tree, ctx=ctx, ) result = _infer_stmt_multiplicity(ir, scope_tree=scope_tree, ctx=ctx) if result is ZERO: return ZERO else: return ONE @_infer_multiplicity.register def __infer_group_stmt( ir: irast.GroupStmt, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: raise NotImplementedError @_infer_multiplicity.register def __infer_slice( ir: irast.SliceIndirection, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # Slice indirection multiplicity is guaranteed to be ONE as long # as the cardinality of this expression is at most one, otherwise # the results of index indirection can contain values with # multiplicity > 1. card = cardinality.infer_cardinality( ir, scope_tree=scope_tree, ctx=ctx) if card is not None and card.is_single(): return ONE else: return MANY @_infer_multiplicity.register def __infer_index( ir: irast.IndexIndirection, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: # Index indirection multiplicity is guaranteed to be ONE as long # as the cardinality of this expression is at most one, otherwise # the results of index indirection can contain values with # multiplicity > 1. card = cardinality.infer_cardinality( ir, scope_tree=scope_tree, ctx=ctx) if card is not None and card.is_single(): return ONE else: return MANY @_infer_multiplicity.register def __infer_array( ir: irast.Array, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: return _common_multiplicity(ir.elements, scope_tree=scope_tree, ctx=ctx) @_infer_multiplicity.register def __infer_tuple( ir: irast.Tuple, *, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: return _common_multiplicity( [el.val for el in ir.elements], scope_tree=scope_tree, ctx=ctx ) def infer_multiplicity( ir: irast.Base, *, is_mutation: bool=False, scope_tree: irast.ScopeTreeNode, ctx: inference_context.InfCtx, ) -> qltypes.Multiplicity: result = ctx.inferred_multiplicity.get((ir, scope_tree)) if result is not None: return result # We can use cardinality as a helper in determining multiplicity, # since singletons have multiplicity one. card = cardinality.infer_cardinality( ir, is_mutation=is_mutation, scope_tree=scope_tree, ctx=ctx) if isinstance(ir, irast.Set): result = _infer_set( ir, is_mutation=is_mutation, scope_tree=scope_tree, ctx=ctx, ) else: result = _infer_multiplicity(ir, scope_tree=scope_tree, ctx=ctx) if card is not None and card.is_single(): # We've validated multiplicity, so now we can just override it # safely. result = ONE if result not in {ZERO, ONE, MANY}: raise errors.QueryError( 'could not determine the multiplicity of ' 'set produced by expression', context=ir.context) ctx.inferred_multiplicity[ir, scope_tree] = result return result
apache-2.0
luzheqi1987/nova-annotation
nova/vnc/__init__.py
38
1810
#!/usr/bin/env python # Copyright (c) 2010 OpenStack Foundation # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Module for VNC Proxying.""" from oslo.config import cfg vnc_opts = [ cfg.StrOpt('novncproxy_base_url', default='http://127.0.0.1:6080/vnc_auto.html', help='Location of VNC console proxy, in the form ' '"http://127.0.0.1:6080/vnc_auto.html"'), cfg.StrOpt('xvpvncproxy_base_url', default='http://127.0.0.1:6081/console', help='Location of nova xvp VNC console proxy, in the form ' '"http://127.0.0.1:6081/console"'), cfg.StrOpt('vncserver_listen', default='127.0.0.1', help='IP address on which instance vncservers should listen'), cfg.StrOpt('vncserver_proxyclient_address', default='127.0.0.1', help='The address to which proxy clients ' '(like nova-xvpvncproxy) should connect'), cfg.BoolOpt('vnc_enabled', default=True, help='Enable VNC related features'), cfg.StrOpt('vnc_keymap', default='en-us', help='Keymap for VNC'), ] CONF = cfg.CONF CONF.register_opts(vnc_opts)
apache-2.0