ext
stringclasses
9 values
sha
stringlengths
40
40
content
stringlengths
3
1.04M
py
1a2f2c37335270663f7395928bb8ea6e0ae5090e
from typing import Any from pyparsing import Word,nums,CaselessLiteral,ParseException from subprocess import Popen,PIPE,STDOUT,CREATE_NO_WINDOW from json import loads import os import errno def is_executable()->bool: """ Determine if the current script is packaged as an executable\n (EG: If packed into a .exe with PyInstaller)\n returns : True/False, if the script is an executable """ import sys return getattr(sys,'frozen',False) def script_dir()->str: """ Get the path to the current script's directory, whether running as an executable or in an interpreter.\n returns : A string containing the path to the script directory. """ from os import path import sys return path.dirname(sys.executable) if is_executable() else os.path.join(path.dirname(path.realpath(sys.argv[0])),'app') def local_path(dir_name:str='')->str: """ Get the absolute path to a local file/directory __MEIPASS or .), whether running as an executable or in an interpreter.\n returns : A string containing the path to the local file/directory """ from os import path import sys return path.join(sys._MEIPASS, dir_name) if is_executable() else path.join(script_dir(),dir_name) def convert_size_to_bytes(size_str:str): """ Converts a size string (eg: "12gb") to bytes. """ multipliers={"kb":1024,"mb":1024000,"gb":1024000000,"tb":1024000000000} #god help whoever converts a tb file expr=Word(nums+','+'.').setParseAction(lambda toks:float(toks[0])).setResultsName('size')+(CaselessLiteral('kb')^ CaselessLiteral('mb') ^ CaselessLiteral('gb') ^ CaselessLiteral('tb')).setParseAction(lambda toks:multipliers[toks[0]]).setResultsName('mult') result=None try: result=expr.parseString(size_str.replace(',','')) except ParseException: return None return result.size*result.mult def is_int(s:str): """ Return whether or not the str `s` is an int. """ try: int(s) return True except ValueError: return False def get_video_info(filename:str,ffprobe:str=os.path.join(local_path('resources'),'ffprobe.exe')): """ Get the video info from a video file.\n Returns a JSON object """ command = [ffprobe, "-loglevel", "quiet", "-print_format", "json", "-show_format", "-show_streams", filename ] pipe = Popen(command, stdout=PIPE, stderr=STDOUT,creationflags = CREATE_NO_WINDOW) out, err = pipe.communicate() return loads(out) def get_free_space_b(dirname): """Return folder/drive free space (in bytes).""" import ctypes import os import platform import sys if platform.system() == 'Windows': free_bytes = ctypes.c_ulonglong(0) ctypes.windll.kernel32.GetDiskFreeSpaceExW(ctypes.c_wchar_p(dirname), None, None, ctypes.pointer(free_bytes)) return free_bytes.value else: st = os.statvfs(dirname) return st.f_bavail * st.f_frsize def vid_time_to_ms(time_str:str): h,m,s=time_str.split(':') s,ms=s.split('.') ms=ms.rstrip('0') if not ms:ms='0' h,m,s,ms=map(float,(h,m,s,ms)) ms+=(s*1000) ms+=(m*60*1000) ms+=(h*60*60*1000) return ms class Custom_Zip(): """Same basic functionality as zip() builtin, but can be used with differently size iterators """ def __init__(self,*args,default:Any=None): """Same as zip() builtin, but returns default from any iters that are exhausted until all are exhausted Args: default (Any, optional): [description]. Defaults to None. """ self.args=[iter(c) for c in args] self.default=default def __iter__(self): return self def __next__(self): yields=[] for arg in self.args: yields.append(next(arg,self.default)) if all(c is None for c in yields): raise StopIteration return tuple(yields) def make_dirs(path:str): """Make the directory path specified. Args: path (str): The path to create. Should either be a file (eg: /foo/bar/baz.txt), or a directory ending in / (/foo/bar/) """ if not os.path.exists(os.path.dirname(path)): try: os.makedirs(os.path.dirname(path)) except OSError as exc: # Guard against race condition if exc.errno != errno.EEXIST: raise def normalize_extension(ext:str,dot=True): if dot: return '.'+ext.lstrip('.') else: return ext.lstrip('.') class file_str(str): """Empty subclass of str builtin, for use with type requiring """ pass def list_get (l:list, idx:int, default:Any=None): try: return l[idx] except IndexError: return default
py
1a2f2cb89e3306bc20b285de6f72ed3bd652f1c1
import functools import json import logging import math import os import time from functools import cached_property from typing import Callable, Dict, List, Tuple, Type # https://github.com/prius/python-leetcode import leetcode.api.default_api # type: ignore import leetcode.api_client # type: ignore import leetcode.auth # type: ignore import leetcode.configuration # type: ignore import leetcode.models.graphql_query # type: ignore import leetcode.models.graphql_query_get_question_detail_variables # type: ignore import leetcode.models.graphql_query_problemset_question_list_variables # type: ignore import leetcode.models.graphql_query_problemset_question_list_variables_filter_input # type: ignore import leetcode.models.graphql_question_detail # type: ignore import urllib3 # type: ignore from tqdm import tqdm # type: ignore CACHE_DIR = "cache" def _get_leetcode_api_client() -> leetcode.api.default_api.DefaultApi: """ Leetcode API instance constructor. This is a singleton, because we don't need to create a separate client each time """ configuration = leetcode.configuration.Configuration() session_id = os.environ["LEETCODE_SESSION_ID"] csrf_token = leetcode.auth.get_csrf_cookie(session_id) configuration.api_key["x-csrftoken"] = csrf_token configuration.api_key["csrftoken"] = csrf_token configuration.api_key["LEETCODE_SESSION"] = session_id configuration.api_key["Referer"] = "https://leetcode.com" configuration.debug = False api_instance = leetcode.api.default_api.DefaultApi( leetcode.api_client.ApiClient(configuration) ) return api_instance def retry(times: int, exceptions: Tuple[Type[Exception]], delay: float) -> Callable: """ Retry Decorator Retries the wrapped function/method `times` times if the exceptions listed in `exceptions` are thrown """ def decorator(func): @functools.wraps(func) def wrapper(*args, **kwargs): for attempt in range(times - 1): try: return func(*args, **kwargs) except exceptions: logging.exception( "Exception occured, try %s/%s", attempt + 1, times ) time.sleep(delay) logging.error("Last try") return func(*args, **kwargs) return wrapper return decorator class LeetcodeData: """ Retrieves and caches the data for problems, acquired from the leetcode API. This data can be later accessed using provided methods with corresponding names. """ def __init__(self, start: int, stop: int) -> None: """ Initialize leetcode API and disk cache for API responses """ if start < 0: raise ValueError(f"Start must be non-negative: {start}") if stop < 0: raise ValueError(f"Stop must be non-negative: {start}") if start > stop: raise ValueError(f"Start (){start}) must be not greater than stop ({stop})") self._start = start self._stop = stop @cached_property def _api_instance(self) -> leetcode.api.default_api.DefaultApi: return _get_leetcode_api_client() @cached_property def _cache( self, ) -> Dict[str, leetcode.models.graphql_question_detail.GraphqlQuestionDetail]: """ Cached method to return dict (problem_slug -> question details) """ problems = self._get_problems_data() return {problem.title_slug: problem for problem in problems} @retry(times=3, exceptions=(urllib3.exceptions.ProtocolError,), delay=5) def _get_problems_count(self) -> int: api_instance = self._api_instance graphql_request = leetcode.models.graphql_query.GraphqlQuery( query=""" query problemsetQuestionList($categorySlug: String, $limit: Int, $skip: Int, $filters: QuestionListFilterInput) { problemsetQuestionList: questionList( categorySlug: $categorySlug limit: $limit skip: $skip filters: $filters ) { totalNum } } """, variables=leetcode.models.graphql_query_problemset_question_list_variables.GraphqlQueryProblemsetQuestionListVariables( category_slug="", limit=1, skip=0, filters=leetcode.models.graphql_query_problemset_question_list_variables_filter_input.GraphqlQueryProblemsetQuestionListVariablesFilterInput( tags=[], # difficulty="MEDIUM", # status="NOT_STARTED", # list_id="7p5x763", # Top Amazon Questions # premium_only=False, ), ), operation_name="problemsetQuestionList", ) time.sleep(2) # Leetcode has a rate limiter data = api_instance.graphql_post(body=graphql_request).data return data.problemset_question_list.total_num or 0 @retry(times=3, exceptions=(urllib3.exceptions.ProtocolError,), delay=5) def _get_problems_data_page( self, offset: int, page_size: int, page: int ) -> List[leetcode.models.graphql_question_detail.GraphqlQuestionDetail]: api_instance = self._api_instance graphql_request = leetcode.models.graphql_query.GraphqlQuery( query=""" query problemsetQuestionList($categorySlug: String, $limit: Int, $skip: Int, $filters: QuestionListFilterInput) { problemsetQuestionList: questionList( categorySlug: $categorySlug limit: $limit skip: $skip filters: $filters ) { questions: data { questionFrontendId title titleSlug categoryTitle freqBar content isPaidOnly difficulty likes dislikes topicTags { name slug } stats hints } } } """, variables=leetcode.models.graphql_query_problemset_question_list_variables.GraphqlQueryProblemsetQuestionListVariables( category_slug="", limit=page_size, skip=offset + page * page_size, filters=leetcode.models.graphql_query_problemset_question_list_variables_filter_input.GraphqlQueryProblemsetQuestionListVariablesFilterInput(), ), operation_name="problemsetQuestionList", ) time.sleep(2) # Leetcode has a rate limiter data = api_instance.graphql_post( body=graphql_request ).data.problemset_question_list.questions return data def _get_problems_data( self, ) -> List[leetcode.models.graphql_question_detail.GraphqlQuestionDetail]: problem_count = self._get_problems_count() if self._start > problem_count: raise ValueError( "Start ({self._start}) is greater than problems count ({problem_count})" ) start = self._start stop = min(self._stop, problem_count) page_size = min(3000, stop - start + 1) problems: List[ leetcode.models.graphql_question_detail.GraphqlQuestionDetail ] = [] logging.info(f"Fetching {stop - start + 1} problems {page_size} per page") for page in tqdm( range(math.ceil((stop - start + 1) / page_size)), unit="problem", unit_scale=page_size, ): data = self._get_problems_data_page(start, page_size, page) problems.extend(data) return problems async def all_problems_handles(self) -> List[str]: """ Get all problem handles known. Example: ["two-sum", "three-sum"] """ return list(self._cache.keys()) def _get_problem_data( self, problem_slug: str ) -> leetcode.models.graphql_question_detail.GraphqlQuestionDetail: """ TODO: Legacy method. Needed in the old architecture. Can be replaced with direct cache calls later. """ cache = self._cache if problem_slug in cache: return cache[problem_slug] async def _get_description(self, problem_slug: str) -> str: """ Problem description """ data = self._get_problem_data(problem_slug) return data.content or "No content" async def _stats(self, problem_slug: str) -> Dict[str, str]: """ Various stats about problem. Such as number of accepted solutions, etc. """ data = self._get_problem_data(problem_slug) return json.loads(data.stats) async def submissions_total(self, problem_slug: str) -> int: """ Total number of submissions of the problem """ return int((await self._stats(problem_slug))["totalSubmissionRaw"]) async def submissions_accepted(self, problem_slug: str) -> int: """ Number of accepted submissions of the problem """ return int((await self._stats(problem_slug))["totalAcceptedRaw"]) async def description(self, problem_slug: str) -> str: """ Problem description """ return await self._get_description(problem_slug) async def difficulty(self, problem_slug: str) -> str: """ Problem difficulty. Returns colored HTML version, so it can be used directly in Anki """ data = self._get_problem_data(problem_slug) diff = data.difficulty if diff == "Easy": return "<font color='green'>Easy</font>" if diff == "Medium": return "<font color='orange'>Medium</font>" if diff == "Hard": return "<font color='red'>Hard</font>" raise ValueError(f"Incorrect difficulty: {diff}") async def paid(self, problem_slug: str) -> str: """ Problem's "available for paid subsribers" status """ data = self._get_problem_data(problem_slug) return data.is_paid_only async def problem_id(self, problem_slug: str) -> str: """ Numerical id of the problem """ data = self._get_problem_data(problem_slug) return data.question_frontend_id async def likes(self, problem_slug: str) -> int: """ Number of likes for the problem """ data = self._get_problem_data(problem_slug) likes = data.likes if not isinstance(likes, int): raise ValueError(f"Likes should be int: {likes}") return likes async def dislikes(self, problem_slug: str) -> int: """ Number of dislikes for the problem """ data = self._get_problem_data(problem_slug) dislikes = data.dislikes if not isinstance(dislikes, int): raise ValueError(f"Dislikes should be int: {dislikes}") return dislikes async def tags(self, problem_slug: str) -> List[str]: """ List of the tags for this problem (string slugs) """ data = self._get_problem_data(problem_slug) return list(map(lambda x: x.slug, data.topic_tags)) async def freq_bar(self, problem_slug: str) -> float: """ Returns percentage for frequency bar """ data = self._get_problem_data(problem_slug) return data.freq_bar or 0 async def title(self, problem_slug: str) -> float: """ Returns problem title """ data = self._get_problem_data(problem_slug) return data.title async def category(self, problem_slug: str) -> float: """ Returns problem category title """ data = self._get_problem_data(problem_slug) return data.category_title
py
1a2f2cfb3c68197b9fa17b3a971cc500b4db60d7
from unittest import TestCase from core_get.cli.parse.options_parsers.login_options_parser import LoginOptionsParser from test.cli.parse.options_parsers.options_parser_test_helper import do_test_parse class TestLoginOptionsParser(TestCase): def test_parses_access_token_correctly(self): login_options = do_test_parse(LoginOptionsParser, 'abc123') self.assertEqual('abc123', login_options.access_token) def test_fails_with_no_access_token(self): with self.assertRaises(SystemExit): do_test_parse(LoginOptionsParser)
py
1a2f2e10a41167a2443b155206a8a8b98bdeb8c0
import pytest import saltext.credstash.sdbs.credstash as credstash_sdb @pytest.fixture def configure_loader_modules(): module_globals = { "__salt__": {"this_does_not_exist.please_replace_it": lambda: True}, } return { credstash_sdb: module_globals, } def test_replace_this_this_with_something_meaningful(): assert "this_does_not_exist.please_replace_it" in credstash_sdb.__salt__ assert credstash_sdb.__salt__["this_does_not_exist.please_replace_it"]() is True
py
1a2f2e80c1717f84fc8c387bdf21036ed14f254e
import itertools import networkx as nx from spacy.tokens import Token def syntactic_depth(token): '''Finds token depth in the syntactic tree''' if token._._syntactic_depth is None: depth = 0 current_word = token while not current_word == current_word.head: depth += 1 current_word = current_word.head token._._syntactic_depth = depth return token._._syntactic_depth Token.set_extension('_syntactic_depth', default=None, force=True) Token.set_extension('syntactic_depth', getter=syntactic_depth, force=True) def filter_by_depth(depths, tokens): if isinstance(depths, int): depths = set([depths]) return [t for t in tokens if t._.syntactic_depth in depths] def shallowest_token(tokens): tokens = sort_by_depth(tokens) return tokens[0] def sort_by_depth(tokens): return sorted(tokens, key=lambda w: (w._.syntactic_depth, w.i)) def sort_by_idx(tokens): return sorted(tokens, key=lambda w: w.i) def siblings(token, side=None): try: siblings = token.head.children except: return [] if side == 'left': siblings = [s for s in siblings if s.i < token.i] elif side == 'left': siblings = [s for s in siblings if s.i > token.i] return siblings def doc_to_nx_graph(doc): edges = [] for token in doc: for child in token.children: edges.append(('{0}-{1}'.format(token.text, token.i), '{0}-{1}'.format(child.text, child.i))) graph = nx.Graph(edges) return graph def shortest_dependency_path(nx_graph, doc, source, target): source = '{0}-{1}'.format(source.text, source.i) target = '{0}-{1}'.format(target.text, target.i) try: path = nx.shortest_path(nx_graph, source=source, target=target) except nx.exception.NetworkXNoPath: path = [] dep_path = [] for node in path: idx = int(node.split('-')[-1]) token = doc[idx] dep_path.append(token) dep_path = sorted(dep_path, key=lambda t: t._.syntactic_depth) return dep_path def smallest_connected_subgraph(with_tokens, doc, nx_graph=None): # Find root nodes if not nx_graph: nx_graph = doc_to_nx_graph(doc) min_depth = min([t._.syntactic_depth for t in with_tokens]) roots = [t for t in with_tokens if t._.syntactic_depth == min_depth] non_roots = [t for t in with_tokens if t not in roots] tokens_touched = roots + non_roots # For each non-root token, trace paths to each root. This will touch every non-root token we're looking for for token in non_roots: for root in roots: path = shortest_dependency_path(nx_graph, doc, token, root) for t in path: if t not in tokens_touched: tokens_touched.append(t) tokens_touched = sorted(tokens_touched, key=lambda t: t.i) # Trace paths between roots for root_x, root_y in itertools.combinations(roots, 2): path = shortest_dependency_path(nx_graph, doc, root_x, root_y) for t in path: if t not in tokens_touched: tokens_touched.append(t) return tokens_touched def idxs_to_tokens(doc, idxs): return [doc[idx] for idx in idxs] def token_idxs(tokens): return [t.i for t in tokens] def de_duplicate_list(list_): unique_list = [] for item in list_: if item not in unique_list: unique_list.append(item) return unique_list def list_contains_duplicates(list_): unique_list = de_duplicate_list(list_) if len(list_) > len(unique_list): return True return False def features_are_in_pattern(features, pattern): for pattern_element in pattern: for feature in features: if feature not in pattern_element['PATTERN']: return False return True def flatten_list(list_): return list(itertools.chain(*list_))
py
1a2f2ee6eadc803b23db8246082ac439cb839c0c
import pytest from tests.mocks import MockOktaClient import okta.models as models from http import HTTPStatus from okta.errors.okta_api_error import OktaAPIError class TestTrustedOriginsResource: """ Integration Tests for the Trusted Origins Resource """ SDK_PREFIX = "python_sdk" @pytest.mark.vcr() @pytest.mark.asyncio async def test_create_get_origin(self, fs): # Instantiate Mock Client client = MockOktaClient(fs) # Create Trusted Origin TO_NAME = f"{TestTrustedOriginsResource.SDK_PREFIX}_test_TO" TO_ORIGIN = "http://example.com" trusted_origin_model = models.TrustedOrigin({ "name": TO_NAME, "origin": TO_ORIGIN, "scopes": [ models.Scope({ "type": models.ScopeType.CORS }), models.Scope({ "type": models.ScopeType.REDIRECT }), ] }) created_trusted_origin, _, err = await \ client.create_origin(trusted_origin_model) assert err is None assert isinstance(created_trusted_origin, models.TrustedOrigin) # Retrieve retrieved_origin, _, err = await \ client.get_origin(created_trusted_origin.id) assert err is None assert isinstance(retrieved_origin, models.TrustedOrigin) assert retrieved_origin.name == created_trusted_origin.name assert len(retrieved_origin.scopes) == 2 # Delete _, err = await client.delete_origin(created_trusted_origin.id) @pytest.mark.vcr() @pytest.mark.asyncio async def test_list_origins(self, fs): # Instantiate Mock Client client = MockOktaClient(fs) # Create Trusted Origin TO_NAME = f"{TestTrustedOriginsResource.SDK_PREFIX}_test_TO" TO_ORIGIN = "http://example.com" trusted_origin_model = models.TrustedOrigin({ "name": TO_NAME, "origin": TO_ORIGIN, "scopes": [ models.Scope({ "type": models.ScopeType.CORS }), models.Scope({ "type": models.ScopeType.REDIRECT }), ] }) created_trusted_origin, _, err = await \ client.create_origin(trusted_origin_model) assert err is None assert isinstance(created_trusted_origin, models.TrustedOrigin) # List trusted_origins, _, err = await client.list_origins() assert err is None assert isinstance(trusted_origins, list) assert len(trusted_origins) > 0 assert isinstance(trusted_origins[0], models.TrustedOrigin) assert next((to for to in trusted_origins if to.name == created_trusted_origin.name), None) \ is not None # Delete _, err = await client.delete_origin(created_trusted_origin.id) @pytest.mark.vcr() @pytest.mark.asyncio async def test_delete_origin(self, fs): # Instantiate Mock Client client = MockOktaClient(fs) # Create Trusted Origin TO_NAME = f"{TestTrustedOriginsResource.SDK_PREFIX}_test_TO" TO_ORIGIN = "http://example.com" trusted_origin_model = models.TrustedOrigin({ "name": TO_NAME, "origin": TO_ORIGIN, "scopes": [ models.Scope({ "type": models.ScopeType.CORS }), models.Scope({ "type": models.ScopeType.REDIRECT }), ] }) created_trusted_origin, _, err = await \ client.create_origin(trusted_origin_model) assert err is None assert isinstance(created_trusted_origin, models.TrustedOrigin) # Retrieve retrieved_origin, _, err = await \ client.get_origin(created_trusted_origin.id) assert err is None assert isinstance(retrieved_origin, models.TrustedOrigin) assert retrieved_origin.name == created_trusted_origin.name assert len(retrieved_origin.scopes) == 2 # Delete _, err = await client.delete_origin(created_trusted_origin.id) # Retrieve to validate retrieved_origin, resp, err = await \ client.get_origin(created_trusted_origin.id) assert err is not None assert isinstance(err, OktaAPIError) assert resp.get_status() == HTTPStatus.NOT_FOUND assert retrieved_origin is None @pytest.mark.vcr() @pytest.mark.asyncio async def test_update_origin(self, fs): # Instantiate Mock Client client = MockOktaClient(fs) # Create Trusted Origin TO_NAME = f"{TestTrustedOriginsResource.SDK_PREFIX}_test_TO" TO_ORIGIN = "http://example.com" trusted_origin_model = models.TrustedOrigin({ "name": TO_NAME, "origin": TO_ORIGIN, "scopes": [ models.Scope({ "type": models.ScopeType.CORS }), models.Scope({ "type": models.ScopeType.REDIRECT }), ] }) created_trusted_origin, _, err = await \ client.create_origin(trusted_origin_model) assert err is None assert isinstance(created_trusted_origin, models.TrustedOrigin) # Retrieve retrieved_origin, _, err = await \ client.get_origin(created_trusted_origin.id) assert err is None assert isinstance(retrieved_origin, models.TrustedOrigin) assert retrieved_origin.name == created_trusted_origin.name assert len(retrieved_origin.scopes) == 2 # Update updated_trusted_origin_model = models.TrustedOrigin({ "name": TO_NAME + "_updated", "origin": TO_ORIGIN, "scopes": [ models.Scope({ "type": models.ScopeType.CORS }), models.Scope({ "type": models.ScopeType.REDIRECT }), ] }) updated_origin, _, err = await \ client.update_origin(created_trusted_origin.id, updated_trusted_origin_model) assert err is None assert isinstance(updated_origin, models.TrustedOrigin) assert updated_origin.id == created_trusted_origin.id assert updated_origin.name == updated_trusted_origin_model.name # Retrieve to validate retrieved_origin, resp, err = await \ client.get_origin(created_trusted_origin.id) assert retrieved_origin.id == created_trusted_origin.id assert retrieved_origin.name == updated_origin.name # Delete _, err = await client.delete_origin(created_trusted_origin.id) @pytest.mark.vcr() @pytest.mark.asyncio async def test_activate_deactivate_origin(self, fs): # Instantiate Mock Client client = MockOktaClient(fs) # Create Trusted Origin TO_NAME = f"{TestTrustedOriginsResource.SDK_PREFIX}_test_TO" TO_ORIGIN = "http://example.com" trusted_origin_model = models.TrustedOrigin({ "name": TO_NAME, "origin": TO_ORIGIN, "scopes": [ models.Scope({ "type": models.ScopeType.CORS }), models.Scope({ "type": models.ScopeType.REDIRECT }), ] }) created_trusted_origin, _, err = await \ client.create_origin(trusted_origin_model) assert err is None assert isinstance(created_trusted_origin, models.TrustedOrigin) assert created_trusted_origin.status == "ACTIVE" # Deactivate deactivated_origin, _, err = await \ client.deactivate_origin(created_trusted_origin.id) assert err is None assert deactivated_origin.status == "INACTIVE" # Retrieve to validate retrieved_origin, resp, err = await \ client.get_origin(created_trusted_origin.id) assert retrieved_origin.id == created_trusted_origin.id assert retrieved_origin.status == "INACTIVE" # Reactivate reactivated_origin, _, err = await \ client.activate_origin(created_trusted_origin.id) assert err is None assert reactivated_origin.status == "ACTIVE" # Retrieve to validate retrieved_origin, resp, err = await \ client.get_origin(created_trusted_origin.id) assert retrieved_origin.id == created_trusted_origin.id assert retrieved_origin.status == "ACTIVE" # Delete _, err = await client.delete_origin(created_trusted_origin.id)
py
1a2f2ffdd7ac9f733b0540a9041999b6fa7df820
#unit test script to test sum module under my_sum import sys sys.path.append('/home/user/workarea/projects/learn-pyspark/jobs/samples/') import unittest import pytest from calc import basic def test_func(): assert basic.func(5)==6
py
1a2f30ad308da16cdbd261bb36d4ed8b0d26b32d
# Copyright 2017 Mycroft AI Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import hashlib import os import random import re import sys from abc import ABCMeta, abstractmethod from threading import Thread from time import time, sleep import os.path from os.path import dirname, exists, isdir, join import mycroft.util from mycroft.client.enclosure.api import EnclosureAPI from mycroft.configuration import Configuration from mycroft.messagebus.message import Message from mycroft.metrics import report_timing, Stopwatch from mycroft.util import ( play_wav, play_mp3, check_for_signal, create_signal, resolve_resource_file ) from mycroft.util.log import LOG from queue import Queue, Empty def send_playback_metric(stopwatch, ident): """ Send playback metrics in a background thread """ def do_send(stopwatch, ident): report_timing(ident, 'speech_playback', stopwatch) t = Thread(target=do_send, args=(stopwatch, ident)) t.daemon = True t.start() class PlaybackThread(Thread): """ Thread class for playing back tts audio and sending visime data to enclosure. """ def __init__(self, queue): super(PlaybackThread, self).__init__() self.queue = queue self._terminated = False self._processing_queue = False self._clear_visimes = False def init(self, tts): self.tts = tts def clear_queue(self): """ Remove all pending playbacks. """ while not self.queue.empty(): self.queue.get() try: self.p.terminate() except: pass def run(self): """ Thread main loop. get audio and visime data from queue and play. """ while not self._terminated: try: snd_type, data, visimes, ident = self.queue.get(timeout=2) self.blink(0.5) if not self._processing_queue: self._processing_queue = True self.tts.begin_audio() stopwatch = Stopwatch() with stopwatch: if snd_type == 'wav': self.p = play_wav(data) elif snd_type == 'mp3': self.p = play_mp3(data) if visimes: if self.show_visimes(visimes): self.clear_queue() else: self.p.communicate() self.p.wait() send_playback_metric(stopwatch, ident) if self.queue.empty(): self.tts.end_audio() self._processing_queue = False self.blink(0.2) except Empty: pass except Exception as e: LOG.exception(e) if self._processing_queue: self.tts.end_audio() self._processing_queue = False def show_visimes(self, pairs): """ Send visime data to enclosure Args: pairs(list): Visime and timing pair Returns: True if button has been pressed. """ start = time() for code, duration in pairs: if self._clear_visimes: self._clear_visimes = False return True if self.enclosure: # Include time stamp to assist with animation timing self.enclosure.mouth_viseme(code, start + duration) delta = time() - start if delta < duration: sleep(duration - delta) return False def clear_visimes(self): self._clear_visimes = True def blink(self, rate=1.0): """ Blink mycroft's eyes """ if self.enclosure and random.random() < rate: self.enclosure.eyes_blink("b") def stop(self): """ Stop thread """ self._terminated = True self.clear_queue() class TTS(object): """ TTS abstract class to be implemented by all TTS engines. It aggregates the minimum required parameters and exposes ``execute(sentence)`` and ``validate_ssml(sentence)`` functions. Args: lang (str): config (dict): Configuration for this specific tts engine validator (TTSValidator): Used to verify proper installation phonetic_spelling (bool): Whether to spell certain words phonetically ssml_tags (list): Supported ssml properties. Ex. ['speak', 'prosody'] """ __metaclass__ = ABCMeta def __init__(self, lang, config, validator, audio_ext='wav', phonetic_spelling=True, ssml_tags=None): super(TTS, self).__init__() self.lang = lang or 'en-us' self.config = config self.validator = validator self.phonetic_spelling = phonetic_spelling self.audio_ext = audio_ext self.ssml_tags = ssml_tags or [] self.voice = config.get("voice") self.filename = '/tmp/tts.wav' self.enclosure = None random.seed() self.queue = Queue() self.playback = PlaybackThread(self.queue) self.playback.start() self.clear_cache() self.spellings = self.load_spellings() def load_spellings(self): """Load phonetic spellings of words as dictionary""" path = join('text', self.lang, 'phonetic_spellings.txt') spellings_file = resolve_resource_file(path) if not spellings_file: return {} try: with open(spellings_file) as f: lines = filter(bool, f.read().split('\n')) lines = [i.split(':') for i in lines] return {key.strip(): value.strip() for key, value in lines} except ValueError: LOG.exception('Failed to load phonetic spellings.') return {} def begin_audio(self): """Helper function for child classes to call in execute()""" # Create signals informing start of speech self.ws.emit(Message("recognizer_loop:audio_output_start")) def end_audio(self): """ Helper function for child classes to call in execute(). Sends the recognizer_loop:audio_output_end message, indicating that speaking is done for the moment. It also checks if cache directory needs cleaning to free up disk space. """ self.ws.emit(Message("recognizer_loop:audio_output_end")) # Clean the cache as needed cache_dir = mycroft.util.get_cache_directory("tts") mycroft.util.curate_cache(cache_dir, min_free_percent=100) # This check will clear the "signal" check_for_signal("isSpeaking") def init(self, ws): self.ws = ws self.playback.init(self) self.enclosure = EnclosureAPI(self.ws) self.playback.enclosure = self.enclosure def get_tts(self, sentence, wav_file): """ Abstract method that a tts implementation needs to implement. Should get data from tts. Args: sentence(str): Sentence to synthesize wav_file(str): output file Returns: tuple: (wav_file, phoneme) """ pass def modify_tag(self, tag): """Override to modify each supported ssml tag""" return tag @staticmethod def remove_ssml(text): return re.sub('<[^>]*>', '', text).replace(' ', ' ') def validate_ssml(self, utterance): """ Check if engine supports ssml, if not remove all tags Remove unsupported / invalid tags Args: utterance(str): Sentence to validate Returns: validated_sentence (str) """ # if ssml is not supported by TTS engine remove all tags if not self.ssml_tags: return self.remove_ssml(utterance) # find ssml tags in string tags = re.findall('<[^>]*>', utterance) for tag in tags: if any(supported in tag for supported in self.ssml_tags): utterance = utterance.replace(tag, self.modify_tag(tag)) else: # remove unsupported tag utterance = utterance.replace(tag, "") # return text with supported ssml tags only return utterance.replace(" ", " ") def execute(self, sentence, ident=None): """ Convert sentence to speech, preprocessing out unsupported ssml The method caches results if possible using the hash of the sentence. Args: sentence: Sentence to be spoken ident: Id reference to current interaction """ sentence = self.validate_ssml(sentence) create_signal("isSpeaking") if self.phonetic_spelling: for word in re.findall(r"[\w']+", sentence): if word.lower() in self.spellings: sentence = sentence.replace(word, self.spellings[word.lower()]) key = str(hashlib.md5(sentence.encode('utf-8', 'ignore')).hexdigest()) wav_file = os.path.join(mycroft.util.get_cache_directory("tts"), key + '.' + self.audio_ext) if os.path.exists(wav_file): LOG.debug("TTS cache hit") phonemes = self.load_phonemes(key) else: wav_file, phonemes = self.get_tts(sentence, wav_file) if phonemes: self.save_phonemes(key, phonemes) vis = self.visime(phonemes) self.queue.put((self.audio_ext, wav_file, vis, ident)) def visime(self, phonemes): """ Create visimes from phonemes. Needs to be implemented for all tts backend Args: phonemes(str): String with phoneme data """ return None def clear_cache(self): """ Remove all cached files. """ if not os.path.exists(mycroft.util.get_cache_directory('tts')): return for f in os.listdir(mycroft.util.get_cache_directory("tts")): file_path = os.path.join(mycroft.util.get_cache_directory("tts"), f) if os.path.isfile(file_path): os.unlink(file_path) def save_phonemes(self, key, phonemes): """ Cache phonemes Args: key: Hash key for the sentence phonemes: phoneme string to save """ cache_dir = mycroft.util.get_cache_directory("tts") pho_file = os.path.join(cache_dir, key + ".pho") try: with open(pho_file, "w") as cachefile: cachefile.write(phonemes) except: LOG.debug("Failed to write .PHO to cache") pass def load_phonemes(self, key): """ Load phonemes from cache file. Args: Key: Key identifying phoneme cache """ pho_file = os.path.join(mycroft.util.get_cache_directory("tts"), key + ".pho") if os.path.exists(pho_file): try: with open(pho_file, "r") as cachefile: phonemes = cachefile.read().strip() return phonemes except: LOG.debug("Failed to read .PHO from cache") return None def __del__(self): self.playback.stop() self.playback.join() class TTSValidator(object): """ TTS Validator abstract class to be implemented by all TTS engines. It exposes and implements ``validate(tts)`` function as a template to validate the TTS engines. """ __metaclass__ = ABCMeta def __init__(self, tts): self.tts = tts def validate(self): self.validate_dependencies() self.validate_instance() self.validate_filename() self.validate_lang() self.validate_connection() def validate_dependencies(self): pass def validate_instance(self): clazz = self.get_tts_class() if not isinstance(self.tts, clazz): raise AttributeError('tts must be instance of ' + clazz.__name__) def validate_filename(self): filename = self.tts.filename if not (filename and filename.endswith('.wav')): raise AttributeError('file: %s must be in .wav format!' % filename) dir_path = dirname(filename) if not (exists(dir_path) and isdir(dir_path)): raise AttributeError('filename: %s is not valid!' % filename) @abstractmethod def validate_lang(self): pass @abstractmethod def validate_connection(self): pass @abstractmethod def get_tts_class(self): pass class TTSFactory(object): from mycroft.tts.espeak_tts import ESpeak from mycroft.tts.fa_tts import FATTS from mycroft.tts.google_tts import GoogleTTS from mycroft.tts.mary_tts import MaryTTS from mycroft.tts.mimic_tts import Mimic from mycroft.tts.spdsay_tts import SpdSay from mycroft.tts.ibm_tts import WatsonTTS from mycroft.tts.polly_tts import PollyTTS from mycroft.tts.bing_tts import BingTTS from mycroft.tts.beepspeak_tts import BeepSpeak from mycroft.tts.responsive_voice_tts import ResponsiveVoice CLASSES = { "mimic": Mimic, "google": GoogleTTS, "marytts": MaryTTS, "fatts": FATTS, "espeak": ESpeak, "spdsay": SpdSay, "polly": PollyTTS, "watson": WatsonTTS, "bing": BingTTS, "beep_speak": BeepSpeak, "responsive_voice": ResponsiveVoice } @staticmethod def create(): """ Factory method to create a TTS engine based on configuration. The configuration file ``mycroft.conf`` contains a ``tts`` section with the name of a TTS module to be read by this method. "tts": { "module": <engine_name> } """ config = Configuration.get() lang = config.get("lang", "en-us") tts_module = config.get('tts', {}).get('module', 'mimic') tts_config = config.get('tts', {}).get(tts_module, {}) tts_lang = tts_config.get('lang', lang) clazz = TTSFactory.CLASSES.get(tts_module) tts = clazz(tts_lang, tts_config) tts.validator.validate() return tts
py
1a2f30bb9423ca730b6a6febe87834f7af8a0fcc
import os # cur_directory = print(getcwd()) # Return the current working directory # # print(cur_directory) # print(os.chdir) # Change current working directory # os.system('mkdir today') # Run the command mkdir in the system shell. # There was today folder created in the root folder. # print(dir(os)) # returns a list of all module functions. # print(help(os)) # returns an extensive manual page crated from the module's docstring/ # import shutil # shutil.copyfile('data.db', 'archive.db') # shutil.move('/build/executables', 'installdir') # import glob # file_list = glob.glob('*.py') # print(file_list) # import sys # print(sys.argv) # # The argparse module provides more sophiscated mechanisms to process command line arguments # import argparse # parser = argparse.ArgumentParser(prog = 'top', description = 'Show top lines from each file') # parser.add_argument('filenames', nargs='+') # parser.add_argument('-l', '--lines', type=int, default=10) # args = parser.parse_args() # print(args) import re # f_words = re.findall(r'\bf[a-z]*', 'which foot or hand fell fastest') # print(f_words) # sub = re.sub(r'\(\b[a-z]+) \1', r'\1', 'cat in the hat') # print(sub) a = 'tea for too'.replace('too', 'two') print(a) # Mathematics import math cos = math.cos(math.pi / 4) print(cos) log = math.log(1024, 2) print(log) # Random module provides tools for making random selections. # import random # fruit = random.choice(['apple', 'pear', 'banana']) # print(fruit) # sample = random.sample(range(100), 10) # sampling without replacement # print(sample) # random_float = random.random() # print(random_float) # integer_number = random.randrange(6) # random integer chosen from range(6) # print(integer_number) # The statistics modue calculates basic statistical properties # (the mean, median, variance, etc) of numeric data. import statistics data = [2.75, 1.75, 1.25, 0.25, 0.5, 1.25, 3.5] print("mean", statistics.mean(data)) print("median :", statistics.median(data)) print("variance :", statistics.variance(data)) # # from urllib.request import urlopen # with urlopen('http://tycho.usno.navy.mil/cgi-bin/timer.pl') as response: # for line in response: # line = line.decode('utf-8') # Decoding the binary data to text. # if 'EST' in line or 'EDT' in line: # Look for Eastern Time # print(line) # # import smtplib # # server = smplib.SMTP('localhost') # server.sendmail('soothsayer@exmple.org', 'jcaesar@exmple.org', # """To : jcaesar@example.org # From : soothsayer@example.org # # Beware the Ideas of March. # """) # server.quit() # # import zlib # s = b'witch which has which witches wrist watch' # print(len(s)) # # t = zlib.compress(s) # print(len(t)) # # u = zlib.decompress(t) # print(u) # # print(zlib.crc32(s)) from timeit import Timer a = Timer('t=a; a=b; b=t', 'a=1;b=2').timeit() print(a) b = Timer('a,b = b,a', 'a=1; b=2').timeit() print(b) # # # def for_number(number): # for_list = [] # for i in range(0, 2000000): # for_list.append(i) # # def while_number(number): # while_list = [] # i = 0 # while i < 2000000: # while_list.append(i) # i = i + 1 # # d = Timer(while_number(2000000)).timeit() # # print(d) # # c = Timer(for_number(2000000)).timeit() # print(c) def average(values): """Computes the arithmetic mean of a list of numbers. >>> print(average([20, 30, 70])) 40.0 """ return sum(values) / len(values) import doctest z = doctest.testmod() print(z) import unittest class TestStatisticalFunctions(unittest.TestCase): def test_average(self): self.assertEqual(average([20,30,70]), 40.0) self.assertEqual(round(average([1, 5, 7]), 1), 4.3) with self.assertRaises((ZeroDivisionError)): average([]) with self.assertRaises(TypeError): average(20, 30, 70) unittest.main()
py
1a2f30dc447e062965e5d3168f863193937e133d
#!/usr/bin/env python # coding: utf-8 """ The Clear BSD License Copyright (c) – 2016, NetApp, Inc. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted (subject to the limitations in the disclaimer below) provided that the following conditions are met: * Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. * Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. * Neither the name of NetApp, Inc. nor the names of its contributors may be used to endorse or promote products derived from this software without specific prior written permission. NO EXPRESS OR IMPLIED LICENSES TO ANY PARTY'S PATENT RIGHTS ARE GRANTED BY THIS LICENSE. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """ import unittest from netapp.santricity.models.symbol.not_on_preferred_path import NotOnPreferredPath class NotOnPreferredPathTest(unittest.TestCase): """ NOTE: This class is auto generated by the swagger code generator program. Do not edit the class manually. """ # Try instantiating the model def test_not_on_preferred_path(self): not_on_preferred_path_obj = NotOnPreferredPath() self.assertNotEqual(not_on_preferred_path_obj, None)
py
1a2f3124887d8ffabdb5af92e1f6a3900185b604
#!/usr/local/bin/python3 from os import system, path, getcwd filePath = "assets/alexa-20180320.csv" print("read file: " + filePath) with open(filePath) as f: content = f.readlines() lines = [x.strip() for x in content] for i in range(0, 10): siteName = lines[i].split(",")[1] print("Dealing with " + str(i) + " " + siteName) script = "scripts/extractArchive.sh" siteName = siteName.replace(".", "_") command = "sh %s %s" % (script, siteName) print("run command %s" % command) system(command)
py
1a2f318380d27f90a9f3d30e3f8393cadd4da648
#!/usr/bin/python3 import sys def safe_print_integer_err(value): is_int = True try: print("{:d}".format(value)) except Exception as e: print("Exception:", e, file=sys.stderr) is_int = False return is_int
py
1a2f31eabf5896f9a474363a06a67a0704df22be
import hug import api @hug.cli() @hug.get('/health') @hug.local() def health_check(): return {'message': 'Status: Running'} # @hug.extend_api('/{BASE_URL_ROUTE}') @hug.extend_api() def apis(): return [api]
py
1a2f32f465d251d0b5f8bae0c4db8c8199036fb8
import pyspark.sql.functions as psf from pyspark.sql import SparkSession, DataFrame from .data_link import DataLink class JdbcDataLink(DataLink): def __init__( self, environment: str, session: SparkSession, url: str, username: str, password: str, driver: str, table: str, save_mode: str = "error", number_of_partitions: int = 1, partition_column: str = "", ): super().__init__(environment, session) self.number_of_partitions = number_of_partitions self.partition_column = partition_column self.save_mode = save_mode self.connection_properties = { "url": url, "user": username, "password": password, "driver": driver, "dbtable": table, } def read(self) -> DataFrame: # Partition parameters are only applicable for read operation # for now, order is important as some values can be overwritten reader = self.spark.read.format("jdbc").options(**self.connection_properties) if self.number_of_partitions == 1: return reader.load() else: # We need a partition column if self.partition_column == "": raise AssertionError("Partitioning column should not be empty.") col = psf.col(self.partition_column) # Retrieve lower and upper bound first to determine the degree of parallelism lower_bound, upper_bound = ( reader.load() .select( psf.min(col).alias("mmin"), psf.max(col).alias("mmax"), ) .collect()[0] ) return ( reader.option("partitionColumn", self.partition_column) .option("numPartitions", str(self.number_of_partitions)) .option("lowerBound", str(lower_bound)) .option("upperBound", str(upper_bound)) .load() ) def write(self, frame: DataFrame) -> None: ( frame.write.format("jdbc") .mode(self.save_mode) .options(**self.connection_properties) .save() )
py
1a2f334492020c13f496704486b147409467400f
# This file was automatically generated by SWIG (http://www.swig.org). # Version 3.0.12 # # Do not make changes to this file unless you know what you are doing--modify # the SWIG interface file instead. from sys import version_info as _swig_python_version_info if _swig_python_version_info >= (2, 7, 0): def swig_import_helper(): import importlib pkg = __name__.rpartition('.')[0] mname = '.'.join((pkg, '_coarse_sun_sensor')).lstrip('.') try: return importlib.import_module(mname) except ImportError: return importlib.import_module('_coarse_sun_sensor') _coarse_sun_sensor = swig_import_helper() del swig_import_helper elif _swig_python_version_info >= (2, 6, 0): def swig_import_helper(): from os.path import dirname import imp fp = None try: fp, pathname, description = imp.find_module('_coarse_sun_sensor', [dirname(__file__)]) except ImportError: import _coarse_sun_sensor return _coarse_sun_sensor try: _mod = imp.load_module('_coarse_sun_sensor', fp, pathname, description) finally: if fp is not None: fp.close() return _mod _coarse_sun_sensor = swig_import_helper() del swig_import_helper else: import _coarse_sun_sensor del _swig_python_version_info try: _swig_property = property except NameError: pass # Python < 2.2 doesn't have 'property'. try: import builtins as __builtin__ except ImportError: import __builtin__ def _swig_setattr_nondynamic(self, class_type, name, value, static=1): if (name == "thisown"): return self.this.own(value) if (name == "this"): if type(value).__name__ == 'SwigPyObject': self.__dict__[name] = value return method = class_type.__swig_setmethods__.get(name, None) if method: return method(self, value) if (not static): if _newclass: object.__setattr__(self, name, value) else: self.__dict__[name] = value else: raise AttributeError("You cannot add attributes to %s" % self) def _swig_setattr(self, class_type, name, value): return _swig_setattr_nondynamic(self, class_type, name, value, 0) def _swig_getattr(self, class_type, name): if (name == "thisown"): return self.this.own() method = class_type.__swig_getmethods__.get(name, None) if method: return method(self) raise AttributeError("'%s' object has no attribute '%s'" % (class_type.__name__, name)) def _swig_repr(self): try: strthis = "proxy of " + self.this.__repr__() except __builtin__.Exception: strthis = "" return "<%s.%s; %s >" % (self.__class__.__module__, self.__class__.__name__, strthis,) try: _object = object _newclass = 1 except __builtin__.Exception: class _object: pass _newclass = 0 class SwigPyIterator(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, SwigPyIterator, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, SwigPyIterator, name) def __init__(self, *args, **kwargs): raise AttributeError("No constructor defined - class is abstract") __repr__ = _swig_repr __swig_destroy__ = _coarse_sun_sensor.delete_SwigPyIterator __del__ = lambda self: None def value(self): return _coarse_sun_sensor.SwigPyIterator_value(self) def incr(self, n=1): return _coarse_sun_sensor.SwigPyIterator_incr(self, n) def decr(self, n=1): return _coarse_sun_sensor.SwigPyIterator_decr(self, n) def distance(self, x): return _coarse_sun_sensor.SwigPyIterator_distance(self, x) def equal(self, x): return _coarse_sun_sensor.SwigPyIterator_equal(self, x) def copy(self): return _coarse_sun_sensor.SwigPyIterator_copy(self) def next(self): return _coarse_sun_sensor.SwigPyIterator_next(self) def __next__(self): return _coarse_sun_sensor.SwigPyIterator___next__(self) def previous(self): return _coarse_sun_sensor.SwigPyIterator_previous(self) def advance(self, n): return _coarse_sun_sensor.SwigPyIterator_advance(self, n) def __eq__(self, x): return _coarse_sun_sensor.SwigPyIterator___eq__(self, x) def __ne__(self, x): return _coarse_sun_sensor.SwigPyIterator___ne__(self, x) def __iadd__(self, n): return _coarse_sun_sensor.SwigPyIterator___iadd__(self, n) def __isub__(self, n): return _coarse_sun_sensor.SwigPyIterator___isub__(self, n) def __add__(self, n): return _coarse_sun_sensor.SwigPyIterator___add__(self, n) def __sub__(self, *args): return _coarse_sun_sensor.SwigPyIterator___sub__(self, *args) def __iter__(self): return self SwigPyIterator_swigregister = _coarse_sun_sensor.SwigPyIterator_swigregister SwigPyIterator_swigregister(SwigPyIterator) def new_doubleArray(nelements): return _coarse_sun_sensor.new_doubleArray(nelements) new_doubleArray = _coarse_sun_sensor.new_doubleArray def delete_doubleArray(ary): return _coarse_sun_sensor.delete_doubleArray(ary) delete_doubleArray = _coarse_sun_sensor.delete_doubleArray def doubleArray_getitem(ary, index): return _coarse_sun_sensor.doubleArray_getitem(ary, index) doubleArray_getitem = _coarse_sun_sensor.doubleArray_getitem def doubleArray_setitem(ary, index, value): return _coarse_sun_sensor.doubleArray_setitem(ary, index, value) doubleArray_setitem = _coarse_sun_sensor.doubleArray_setitem def new_longArray(nelements): return _coarse_sun_sensor.new_longArray(nelements) new_longArray = _coarse_sun_sensor.new_longArray def delete_longArray(ary): return _coarse_sun_sensor.delete_longArray(ary) delete_longArray = _coarse_sun_sensor.delete_longArray def longArray_getitem(ary, index): return _coarse_sun_sensor.longArray_getitem(ary, index) longArray_getitem = _coarse_sun_sensor.longArray_getitem def longArray_setitem(ary, index, value): return _coarse_sun_sensor.longArray_setitem(ary, index, value) longArray_setitem = _coarse_sun_sensor.longArray_setitem def new_intArray(nelements): return _coarse_sun_sensor.new_intArray(nelements) new_intArray = _coarse_sun_sensor.new_intArray def delete_intArray(ary): return _coarse_sun_sensor.delete_intArray(ary) delete_intArray = _coarse_sun_sensor.delete_intArray def intArray_getitem(ary, index): return _coarse_sun_sensor.intArray_getitem(ary, index) intArray_getitem = _coarse_sun_sensor.intArray_getitem def intArray_setitem(ary, index, value): return _coarse_sun_sensor.intArray_setitem(ary, index, value) intArray_setitem = _coarse_sun_sensor.intArray_setitem def new_shortArray(nelements): return _coarse_sun_sensor.new_shortArray(nelements) new_shortArray = _coarse_sun_sensor.new_shortArray def delete_shortArray(ary): return _coarse_sun_sensor.delete_shortArray(ary) delete_shortArray = _coarse_sun_sensor.delete_shortArray def shortArray_getitem(ary, index): return _coarse_sun_sensor.shortArray_getitem(ary, index) shortArray_getitem = _coarse_sun_sensor.shortArray_getitem def shortArray_setitem(ary, index, value): return _coarse_sun_sensor.shortArray_setitem(ary, index, value) shortArray_setitem = _coarse_sun_sensor.shortArray_setitem def getStructSize(self): try: return eval('sizeof_' + repr(self).split(';')[0].split('.')[-1]) except (NameError) as e: typeString = 'sizeof_' + repr(self).split(';')[0].split('.')[-1] raise NameError(e.message + '\nYou tried to get this size macro: ' + typeString + '\n It appears to be undefined. \nYou need to run the SWIG GEN_SIZEOF' + ' SWIG macro against the class/struct in your SWIG file if you want to ' + ' make this call.\n') def protectSetAttr(self, name, value): if(hasattr(self, name) or name == 'this'): object.__setattr__(self, name, value) else: raise ValueError('You tried to add this variable: ' + name + '\n' + 'To this class: ' + str(self)) def protectAllClasses(moduleType): import inspect clsmembers = inspect.getmembers(sys.modules[__name__], inspect.isclass) for member in clsmembers: try: exec(str(member[0]) + '.__setattr__ = protectSetAttr') exec(str(member[0]) + '.getStructSize = getStructSize') except (AttributeError, TypeError) as e: pass def new_boolArray(nelements): return _coarse_sun_sensor.new_boolArray(nelements) new_boolArray = _coarse_sun_sensor.new_boolArray def delete_boolArray(ary): return _coarse_sun_sensor.delete_boolArray(ary) delete_boolArray = _coarse_sun_sensor.delete_boolArray def boolArray_getitem(ary, index): return _coarse_sun_sensor.boolArray_getitem(ary, index) boolArray_getitem = _coarse_sun_sensor.boolArray_getitem def boolArray_setitem(ary, index, value): return _coarse_sun_sensor.boolArray_setitem(ary, index, value) boolArray_setitem = _coarse_sun_sensor.boolArray_setitem class IntVector(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, IntVector, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, IntVector, name) __repr__ = _swig_repr def iterator(self): return _coarse_sun_sensor.IntVector_iterator(self) def __iter__(self): return self.iterator() def __nonzero__(self): return _coarse_sun_sensor.IntVector___nonzero__(self) def __bool__(self): return _coarse_sun_sensor.IntVector___bool__(self) def __len__(self): return _coarse_sun_sensor.IntVector___len__(self) def __getslice__(self, i, j): return _coarse_sun_sensor.IntVector___getslice__(self, i, j) def __setslice__(self, *args): return _coarse_sun_sensor.IntVector___setslice__(self, *args) def __delslice__(self, i, j): return _coarse_sun_sensor.IntVector___delslice__(self, i, j) def __delitem__(self, *args): return _coarse_sun_sensor.IntVector___delitem__(self, *args) def __getitem__(self, *args): return _coarse_sun_sensor.IntVector___getitem__(self, *args) def __setitem__(self, *args): return _coarse_sun_sensor.IntVector___setitem__(self, *args) def pop(self): return _coarse_sun_sensor.IntVector_pop(self) def append(self, x): return _coarse_sun_sensor.IntVector_append(self, x) def empty(self): return _coarse_sun_sensor.IntVector_empty(self) def size(self): return _coarse_sun_sensor.IntVector_size(self) def swap(self, v): return _coarse_sun_sensor.IntVector_swap(self, v) def begin(self): return _coarse_sun_sensor.IntVector_begin(self) def end(self): return _coarse_sun_sensor.IntVector_end(self) def rbegin(self): return _coarse_sun_sensor.IntVector_rbegin(self) def rend(self): return _coarse_sun_sensor.IntVector_rend(self) def clear(self): return _coarse_sun_sensor.IntVector_clear(self) def get_allocator(self): return _coarse_sun_sensor.IntVector_get_allocator(self) def pop_back(self): return _coarse_sun_sensor.IntVector_pop_back(self) def erase(self, *args): return _coarse_sun_sensor.IntVector_erase(self, *args) def __init__(self, *args): this = _coarse_sun_sensor.new_IntVector(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this def push_back(self, x): return _coarse_sun_sensor.IntVector_push_back(self, x) def front(self): return _coarse_sun_sensor.IntVector_front(self) def back(self): return _coarse_sun_sensor.IntVector_back(self) def assign(self, n, x): return _coarse_sun_sensor.IntVector_assign(self, n, x) def resize(self, *args): return _coarse_sun_sensor.IntVector_resize(self, *args) def insert(self, *args): return _coarse_sun_sensor.IntVector_insert(self, *args) def reserve(self, n): return _coarse_sun_sensor.IntVector_reserve(self, n) def capacity(self): return _coarse_sun_sensor.IntVector_capacity(self) __swig_destroy__ = _coarse_sun_sensor.delete_IntVector __del__ = lambda self: None IntVector_swigregister = _coarse_sun_sensor.IntVector_swigregister IntVector_swigregister(IntVector) class DoubleVector(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, DoubleVector, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, DoubleVector, name) __repr__ = _swig_repr def iterator(self): return _coarse_sun_sensor.DoubleVector_iterator(self) def __iter__(self): return self.iterator() def __nonzero__(self): return _coarse_sun_sensor.DoubleVector___nonzero__(self) def __bool__(self): return _coarse_sun_sensor.DoubleVector___bool__(self) def __len__(self): return _coarse_sun_sensor.DoubleVector___len__(self) def __getslice__(self, i, j): return _coarse_sun_sensor.DoubleVector___getslice__(self, i, j) def __setslice__(self, *args): return _coarse_sun_sensor.DoubleVector___setslice__(self, *args) def __delslice__(self, i, j): return _coarse_sun_sensor.DoubleVector___delslice__(self, i, j) def __delitem__(self, *args): return _coarse_sun_sensor.DoubleVector___delitem__(self, *args) def __getitem__(self, *args): return _coarse_sun_sensor.DoubleVector___getitem__(self, *args) def __setitem__(self, *args): return _coarse_sun_sensor.DoubleVector___setitem__(self, *args) def pop(self): return _coarse_sun_sensor.DoubleVector_pop(self) def append(self, x): return _coarse_sun_sensor.DoubleVector_append(self, x) def empty(self): return _coarse_sun_sensor.DoubleVector_empty(self) def size(self): return _coarse_sun_sensor.DoubleVector_size(self) def swap(self, v): return _coarse_sun_sensor.DoubleVector_swap(self, v) def begin(self): return _coarse_sun_sensor.DoubleVector_begin(self) def end(self): return _coarse_sun_sensor.DoubleVector_end(self) def rbegin(self): return _coarse_sun_sensor.DoubleVector_rbegin(self) def rend(self): return _coarse_sun_sensor.DoubleVector_rend(self) def clear(self): return _coarse_sun_sensor.DoubleVector_clear(self) def get_allocator(self): return _coarse_sun_sensor.DoubleVector_get_allocator(self) def pop_back(self): return _coarse_sun_sensor.DoubleVector_pop_back(self) def erase(self, *args): return _coarse_sun_sensor.DoubleVector_erase(self, *args) def __init__(self, *args): this = _coarse_sun_sensor.new_DoubleVector(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this def push_back(self, x): return _coarse_sun_sensor.DoubleVector_push_back(self, x) def front(self): return _coarse_sun_sensor.DoubleVector_front(self) def back(self): return _coarse_sun_sensor.DoubleVector_back(self) def assign(self, n, x): return _coarse_sun_sensor.DoubleVector_assign(self, n, x) def resize(self, *args): return _coarse_sun_sensor.DoubleVector_resize(self, *args) def insert(self, *args): return _coarse_sun_sensor.DoubleVector_insert(self, *args) def reserve(self, n): return _coarse_sun_sensor.DoubleVector_reserve(self, n) def capacity(self): return _coarse_sun_sensor.DoubleVector_capacity(self) __swig_destroy__ = _coarse_sun_sensor.delete_DoubleVector __del__ = lambda self: None DoubleVector_swigregister = _coarse_sun_sensor.DoubleVector_swigregister DoubleVector_swigregister(DoubleVector) class StringVector(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, StringVector, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, StringVector, name) __repr__ = _swig_repr def iterator(self): return _coarse_sun_sensor.StringVector_iterator(self) def __iter__(self): return self.iterator() def __nonzero__(self): return _coarse_sun_sensor.StringVector___nonzero__(self) def __bool__(self): return _coarse_sun_sensor.StringVector___bool__(self) def __len__(self): return _coarse_sun_sensor.StringVector___len__(self) def __getslice__(self, i, j): return _coarse_sun_sensor.StringVector___getslice__(self, i, j) def __setslice__(self, *args): return _coarse_sun_sensor.StringVector___setslice__(self, *args) def __delslice__(self, i, j): return _coarse_sun_sensor.StringVector___delslice__(self, i, j) def __delitem__(self, *args): return _coarse_sun_sensor.StringVector___delitem__(self, *args) def __getitem__(self, *args): return _coarse_sun_sensor.StringVector___getitem__(self, *args) def __setitem__(self, *args): return _coarse_sun_sensor.StringVector___setitem__(self, *args) def pop(self): return _coarse_sun_sensor.StringVector_pop(self) def append(self, x): return _coarse_sun_sensor.StringVector_append(self, x) def empty(self): return _coarse_sun_sensor.StringVector_empty(self) def size(self): return _coarse_sun_sensor.StringVector_size(self) def swap(self, v): return _coarse_sun_sensor.StringVector_swap(self, v) def begin(self): return _coarse_sun_sensor.StringVector_begin(self) def end(self): return _coarse_sun_sensor.StringVector_end(self) def rbegin(self): return _coarse_sun_sensor.StringVector_rbegin(self) def rend(self): return _coarse_sun_sensor.StringVector_rend(self) def clear(self): return _coarse_sun_sensor.StringVector_clear(self) def get_allocator(self): return _coarse_sun_sensor.StringVector_get_allocator(self) def pop_back(self): return _coarse_sun_sensor.StringVector_pop_back(self) def erase(self, *args): return _coarse_sun_sensor.StringVector_erase(self, *args) def __init__(self, *args): this = _coarse_sun_sensor.new_StringVector(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this def push_back(self, x): return _coarse_sun_sensor.StringVector_push_back(self, x) def front(self): return _coarse_sun_sensor.StringVector_front(self) def back(self): return _coarse_sun_sensor.StringVector_back(self) def assign(self, n, x): return _coarse_sun_sensor.StringVector_assign(self, n, x) def resize(self, *args): return _coarse_sun_sensor.StringVector_resize(self, *args) def insert(self, *args): return _coarse_sun_sensor.StringVector_insert(self, *args) def reserve(self, n): return _coarse_sun_sensor.StringVector_reserve(self, n) def capacity(self): return _coarse_sun_sensor.StringVector_capacity(self) __swig_destroy__ = _coarse_sun_sensor.delete_StringVector __del__ = lambda self: None StringVector_swigregister = _coarse_sun_sensor.StringVector_swigregister StringVector_swigregister(StringVector) class StringSet(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, StringSet, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, StringSet, name) __repr__ = _swig_repr def iterator(self): return _coarse_sun_sensor.StringSet_iterator(self) def __iter__(self): return self.iterator() def __nonzero__(self): return _coarse_sun_sensor.StringSet___nonzero__(self) def __bool__(self): return _coarse_sun_sensor.StringSet___bool__(self) def __len__(self): return _coarse_sun_sensor.StringSet___len__(self) def append(self, x): return _coarse_sun_sensor.StringSet_append(self, x) def __contains__(self, x): return _coarse_sun_sensor.StringSet___contains__(self, x) def __getitem__(self, i): return _coarse_sun_sensor.StringSet___getitem__(self, i) def add(self, x): return _coarse_sun_sensor.StringSet_add(self, x) def discard(self, x): return _coarse_sun_sensor.StringSet_discard(self, x) def __init__(self, *args): this = _coarse_sun_sensor.new_StringSet(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this def empty(self): return _coarse_sun_sensor.StringSet_empty(self) def size(self): return _coarse_sun_sensor.StringSet_size(self) def clear(self): return _coarse_sun_sensor.StringSet_clear(self) def swap(self, v): return _coarse_sun_sensor.StringSet_swap(self, v) def count(self, x): return _coarse_sun_sensor.StringSet_count(self, x) def begin(self): return _coarse_sun_sensor.StringSet_begin(self) def end(self): return _coarse_sun_sensor.StringSet_end(self) def rbegin(self): return _coarse_sun_sensor.StringSet_rbegin(self) def rend(self): return _coarse_sun_sensor.StringSet_rend(self) def erase(self, *args): return _coarse_sun_sensor.StringSet_erase(self, *args) def find(self, x): return _coarse_sun_sensor.StringSet_find(self, x) def lower_bound(self, x): return _coarse_sun_sensor.StringSet_lower_bound(self, x) def upper_bound(self, x): return _coarse_sun_sensor.StringSet_upper_bound(self, x) def equal_range(self, x): return _coarse_sun_sensor.StringSet_equal_range(self, x) def insert(self, __x): return _coarse_sun_sensor.StringSet_insert(self, __x) __swig_destroy__ = _coarse_sun_sensor.delete_StringSet __del__ = lambda self: None StringSet_swigregister = _coarse_sun_sensor.StringSet_swigregister StringSet_swigregister(StringSet) class intSet(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, intSet, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, intSet, name) __repr__ = _swig_repr def iterator(self): return _coarse_sun_sensor.intSet_iterator(self) def __iter__(self): return self.iterator() def __nonzero__(self): return _coarse_sun_sensor.intSet___nonzero__(self) def __bool__(self): return _coarse_sun_sensor.intSet___bool__(self) def __len__(self): return _coarse_sun_sensor.intSet___len__(self) def append(self, x): return _coarse_sun_sensor.intSet_append(self, x) def __contains__(self, x): return _coarse_sun_sensor.intSet___contains__(self, x) def __getitem__(self, i): return _coarse_sun_sensor.intSet___getitem__(self, i) def add(self, x): return _coarse_sun_sensor.intSet_add(self, x) def discard(self, x): return _coarse_sun_sensor.intSet_discard(self, x) def __init__(self, *args): this = _coarse_sun_sensor.new_intSet(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this def empty(self): return _coarse_sun_sensor.intSet_empty(self) def size(self): return _coarse_sun_sensor.intSet_size(self) def clear(self): return _coarse_sun_sensor.intSet_clear(self) def swap(self, v): return _coarse_sun_sensor.intSet_swap(self, v) def count(self, x): return _coarse_sun_sensor.intSet_count(self, x) def begin(self): return _coarse_sun_sensor.intSet_begin(self) def end(self): return _coarse_sun_sensor.intSet_end(self) def rbegin(self): return _coarse_sun_sensor.intSet_rbegin(self) def rend(self): return _coarse_sun_sensor.intSet_rend(self) def erase(self, *args): return _coarse_sun_sensor.intSet_erase(self, *args) def find(self, x): return _coarse_sun_sensor.intSet_find(self, x) def lower_bound(self, x): return _coarse_sun_sensor.intSet_lower_bound(self, x) def upper_bound(self, x): return _coarse_sun_sensor.intSet_upper_bound(self, x) def equal_range(self, x): return _coarse_sun_sensor.intSet_equal_range(self, x) def insert(self, __x): return _coarse_sun_sensor.intSet_insert(self, __x) __swig_destroy__ = _coarse_sun_sensor.delete_intSet __del__ = lambda self: None intSet_swigregister = _coarse_sun_sensor.intSet_swigregister intSet_swigregister(intSet) class ConstCharVector(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, ConstCharVector, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, ConstCharVector, name) __repr__ = _swig_repr def iterator(self): return _coarse_sun_sensor.ConstCharVector_iterator(self) def __iter__(self): return self.iterator() def __nonzero__(self): return _coarse_sun_sensor.ConstCharVector___nonzero__(self) def __bool__(self): return _coarse_sun_sensor.ConstCharVector___bool__(self) def __len__(self): return _coarse_sun_sensor.ConstCharVector___len__(self) def __getslice__(self, i, j): return _coarse_sun_sensor.ConstCharVector___getslice__(self, i, j) def __setslice__(self, *args): return _coarse_sun_sensor.ConstCharVector___setslice__(self, *args) def __delslice__(self, i, j): return _coarse_sun_sensor.ConstCharVector___delslice__(self, i, j) def __delitem__(self, *args): return _coarse_sun_sensor.ConstCharVector___delitem__(self, *args) def __getitem__(self, *args): return _coarse_sun_sensor.ConstCharVector___getitem__(self, *args) def __setitem__(self, *args): return _coarse_sun_sensor.ConstCharVector___setitem__(self, *args) def pop(self): return _coarse_sun_sensor.ConstCharVector_pop(self) def append(self, x): return _coarse_sun_sensor.ConstCharVector_append(self, x) def empty(self): return _coarse_sun_sensor.ConstCharVector_empty(self) def size(self): return _coarse_sun_sensor.ConstCharVector_size(self) def swap(self, v): return _coarse_sun_sensor.ConstCharVector_swap(self, v) def begin(self): return _coarse_sun_sensor.ConstCharVector_begin(self) def end(self): return _coarse_sun_sensor.ConstCharVector_end(self) def rbegin(self): return _coarse_sun_sensor.ConstCharVector_rbegin(self) def rend(self): return _coarse_sun_sensor.ConstCharVector_rend(self) def clear(self): return _coarse_sun_sensor.ConstCharVector_clear(self) def get_allocator(self): return _coarse_sun_sensor.ConstCharVector_get_allocator(self) def pop_back(self): return _coarse_sun_sensor.ConstCharVector_pop_back(self) def erase(self, *args): return _coarse_sun_sensor.ConstCharVector_erase(self, *args) def __init__(self, *args): this = _coarse_sun_sensor.new_ConstCharVector(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this def push_back(self, x): return _coarse_sun_sensor.ConstCharVector_push_back(self, x) def front(self): return _coarse_sun_sensor.ConstCharVector_front(self) def back(self): return _coarse_sun_sensor.ConstCharVector_back(self) def assign(self, n, x): return _coarse_sun_sensor.ConstCharVector_assign(self, n, x) def resize(self, *args): return _coarse_sun_sensor.ConstCharVector_resize(self, *args) def insert(self, *args): return _coarse_sun_sensor.ConstCharVector_insert(self, *args) def reserve(self, n): return _coarse_sun_sensor.ConstCharVector_reserve(self, n) def capacity(self): return _coarse_sun_sensor.ConstCharVector_capacity(self) __swig_destroy__ = _coarse_sun_sensor.delete_ConstCharVector __del__ = lambda self: None ConstCharVector_swigregister = _coarse_sun_sensor.ConstCharVector_swigregister ConstCharVector_swigregister(ConstCharVector) class MultiArray(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, MultiArray, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, MultiArray, name) __repr__ = _swig_repr def iterator(self): return _coarse_sun_sensor.MultiArray_iterator(self) def __iter__(self): return self.iterator() def __nonzero__(self): return _coarse_sun_sensor.MultiArray___nonzero__(self) def __bool__(self): return _coarse_sun_sensor.MultiArray___bool__(self) def __len__(self): return _coarse_sun_sensor.MultiArray___len__(self) def __getslice__(self, i, j): return _coarse_sun_sensor.MultiArray___getslice__(self, i, j) def __setslice__(self, *args): return _coarse_sun_sensor.MultiArray___setslice__(self, *args) def __delslice__(self, i, j): return _coarse_sun_sensor.MultiArray___delslice__(self, i, j) def __delitem__(self, *args): return _coarse_sun_sensor.MultiArray___delitem__(self, *args) def __getitem__(self, *args): return _coarse_sun_sensor.MultiArray___getitem__(self, *args) def __setitem__(self, *args): return _coarse_sun_sensor.MultiArray___setitem__(self, *args) def pop(self): return _coarse_sun_sensor.MultiArray_pop(self) def append(self, x): return _coarse_sun_sensor.MultiArray_append(self, x) def empty(self): return _coarse_sun_sensor.MultiArray_empty(self) def size(self): return _coarse_sun_sensor.MultiArray_size(self) def swap(self, v): return _coarse_sun_sensor.MultiArray_swap(self, v) def begin(self): return _coarse_sun_sensor.MultiArray_begin(self) def end(self): return _coarse_sun_sensor.MultiArray_end(self) def rbegin(self): return _coarse_sun_sensor.MultiArray_rbegin(self) def rend(self): return _coarse_sun_sensor.MultiArray_rend(self) def clear(self): return _coarse_sun_sensor.MultiArray_clear(self) def get_allocator(self): return _coarse_sun_sensor.MultiArray_get_allocator(self) def pop_back(self): return _coarse_sun_sensor.MultiArray_pop_back(self) def erase(self, *args): return _coarse_sun_sensor.MultiArray_erase(self, *args) def __init__(self, *args): this = _coarse_sun_sensor.new_MultiArray(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this def push_back(self, x): return _coarse_sun_sensor.MultiArray_push_back(self, x) def front(self): return _coarse_sun_sensor.MultiArray_front(self) def back(self): return _coarse_sun_sensor.MultiArray_back(self) def assign(self, n, x): return _coarse_sun_sensor.MultiArray_assign(self, n, x) def resize(self, *args): return _coarse_sun_sensor.MultiArray_resize(self, *args) def insert(self, *args): return _coarse_sun_sensor.MultiArray_insert(self, *args) def reserve(self, n): return _coarse_sun_sensor.MultiArray_reserve(self, n) def capacity(self): return _coarse_sun_sensor.MultiArray_capacity(self) __swig_destroy__ = _coarse_sun_sensor.delete_MultiArray __del__ = lambda self: None MultiArray_swigregister = _coarse_sun_sensor.MultiArray_swigregister MultiArray_swigregister(MultiArray) class CSSVector(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, CSSVector, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, CSSVector, name) __repr__ = _swig_repr def iterator(self): return _coarse_sun_sensor.CSSVector_iterator(self) def __iter__(self): return self.iterator() def __nonzero__(self): return _coarse_sun_sensor.CSSVector___nonzero__(self) def __bool__(self): return _coarse_sun_sensor.CSSVector___bool__(self) def __len__(self): return _coarse_sun_sensor.CSSVector___len__(self) def __getslice__(self, i, j): return _coarse_sun_sensor.CSSVector___getslice__(self, i, j) def __setslice__(self, *args): return _coarse_sun_sensor.CSSVector___setslice__(self, *args) def __delslice__(self, i, j): return _coarse_sun_sensor.CSSVector___delslice__(self, i, j) def __delitem__(self, *args): return _coarse_sun_sensor.CSSVector___delitem__(self, *args) def __getitem__(self, *args): return _coarse_sun_sensor.CSSVector___getitem__(self, *args) def __setitem__(self, *args): return _coarse_sun_sensor.CSSVector___setitem__(self, *args) def pop(self): return _coarse_sun_sensor.CSSVector_pop(self) def append(self, x): return _coarse_sun_sensor.CSSVector_append(self, x) def empty(self): return _coarse_sun_sensor.CSSVector_empty(self) def size(self): return _coarse_sun_sensor.CSSVector_size(self) def swap(self, v): return _coarse_sun_sensor.CSSVector_swap(self, v) def begin(self): return _coarse_sun_sensor.CSSVector_begin(self) def end(self): return _coarse_sun_sensor.CSSVector_end(self) def rbegin(self): return _coarse_sun_sensor.CSSVector_rbegin(self) def rend(self): return _coarse_sun_sensor.CSSVector_rend(self) def clear(self): return _coarse_sun_sensor.CSSVector_clear(self) def get_allocator(self): return _coarse_sun_sensor.CSSVector_get_allocator(self) def pop_back(self): return _coarse_sun_sensor.CSSVector_pop_back(self) def erase(self, *args): return _coarse_sun_sensor.CSSVector_erase(self, *args) def __init__(self, *args): this = _coarse_sun_sensor.new_CSSVector(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this def push_back(self, x): return _coarse_sun_sensor.CSSVector_push_back(self, x) def front(self): return _coarse_sun_sensor.CSSVector_front(self) def back(self): return _coarse_sun_sensor.CSSVector_back(self) def assign(self, n, x): return _coarse_sun_sensor.CSSVector_assign(self, n, x) def resize(self, *args): return _coarse_sun_sensor.CSSVector_resize(self, *args) def insert(self, *args): return _coarse_sun_sensor.CSSVector_insert(self, *args) def reserve(self, n): return _coarse_sun_sensor.CSSVector_reserve(self, n) def capacity(self): return _coarse_sun_sensor.CSSVector_capacity(self) __swig_destroy__ = _coarse_sun_sensor.delete_CSSVector __del__ = lambda self: None CSSVector_swigregister = _coarse_sun_sensor.CSSVector_swigregister CSSVector_swigregister(CSSVector) class SysModel(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, SysModel, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, SysModel, name) __repr__ = _swig_repr def __init__(self, *args): this = _coarse_sun_sensor.new_SysModel(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this __swig_destroy__ = _coarse_sun_sensor.delete_SysModel __del__ = lambda self: None def SelfInit(self): return _coarse_sun_sensor.SysModel_SelfInit(self) def CrossInit(self): return _coarse_sun_sensor.SysModel_CrossInit(self) def IntegratedInit(self): return _coarse_sun_sensor.SysModel_IntegratedInit(self) def UpdateState(self, CurrentSimNanos): return _coarse_sun_sensor.SysModel_UpdateState(self, CurrentSimNanos) def Reset(self, CurrentSimNanos): return _coarse_sun_sensor.SysModel_Reset(self, CurrentSimNanos) __swig_setmethods__["ModelTag"] = _coarse_sun_sensor.SysModel_ModelTag_set __swig_getmethods__["ModelTag"] = _coarse_sun_sensor.SysModel_ModelTag_get if _newclass: ModelTag = _swig_property(_coarse_sun_sensor.SysModel_ModelTag_get, _coarse_sun_sensor.SysModel_ModelTag_set) __swig_setmethods__["CallCounts"] = _coarse_sun_sensor.SysModel_CallCounts_set __swig_getmethods__["CallCounts"] = _coarse_sun_sensor.SysModel_CallCounts_get if _newclass: CallCounts = _swig_property(_coarse_sun_sensor.SysModel_CallCounts_get, _coarse_sun_sensor.SysModel_CallCounts_set) __swig_setmethods__["RNGSeed"] = _coarse_sun_sensor.SysModel_RNGSeed_set __swig_getmethods__["RNGSeed"] = _coarse_sun_sensor.SysModel_RNGSeed_get if _newclass: RNGSeed = _swig_property(_coarse_sun_sensor.SysModel_RNGSeed_get, _coarse_sun_sensor.SysModel_RNGSeed_set) __swig_setmethods__["moduleID"] = _coarse_sun_sensor.SysModel_moduleID_set __swig_getmethods__["moduleID"] = _coarse_sun_sensor.SysModel_moduleID_get if _newclass: moduleID = _swig_property(_coarse_sun_sensor.SysModel_moduleID_get, _coarse_sun_sensor.SysModel_moduleID_set) SysModel_swigregister = _coarse_sun_sensor.SysModel_swigregister SysModel_swigregister(SysModel) CSSFAULT_OFF = _coarse_sun_sensor.CSSFAULT_OFF CSSFAULT_STUCK_CURRENT = _coarse_sun_sensor.CSSFAULT_STUCK_CURRENT CSSFAULT_STUCK_MAX = _coarse_sun_sensor.CSSFAULT_STUCK_MAX CSSFAULT_STUCK_RAND = _coarse_sun_sensor.CSSFAULT_STUCK_RAND CSSFAULT_STUCK = _coarse_sun_sensor.CSSFAULT_STUCK CSSFAULT_RAND = _coarse_sun_sensor.CSSFAULT_RAND MAX_CSSFAULT = _coarse_sun_sensor.MAX_CSSFAULT class CoarseSunSensor(SysModel): __swig_setmethods__ = {} for _s in [SysModel]: __swig_setmethods__.update(getattr(_s, '__swig_setmethods__', {})) __setattr__ = lambda self, name, value: _swig_setattr(self, CoarseSunSensor, name, value) __swig_getmethods__ = {} for _s in [SysModel]: __swig_getmethods__.update(getattr(_s, '__swig_getmethods__', {})) __getattr__ = lambda self, name: _swig_getattr(self, CoarseSunSensor, name) __repr__ = _swig_repr __swig_destroy__ = _coarse_sun_sensor.delete_CoarseSunSensor __del__ = lambda self: None def CrossInit(self): return _coarse_sun_sensor.CoarseSunSensor_CrossInit(self) def SelfInit(self): return _coarse_sun_sensor.CoarseSunSensor_SelfInit(self) def UpdateState(self, CurrentSimNanos): return _coarse_sun_sensor.CoarseSunSensor_UpdateState(self, CurrentSimNanos) def setUnitDirectionVectorWithPerturbation(self, cssThetaPerturb, cssPhiPerturb): return _coarse_sun_sensor.CoarseSunSensor_setUnitDirectionVectorWithPerturbation(self, cssThetaPerturb, cssPhiPerturb) def setBodyToPlatformDCM(self, yaw, pitch, roll): return _coarse_sun_sensor.CoarseSunSensor_setBodyToPlatformDCM(self, yaw, pitch, roll) def readInputMessages(self): return _coarse_sun_sensor.CoarseSunSensor_readInputMessages(self) def computeSunData(self): return _coarse_sun_sensor.CoarseSunSensor_computeSunData(self) def computeTrueOutput(self): return _coarse_sun_sensor.CoarseSunSensor_computeTrueOutput(self) def applySensorErrors(self): return _coarse_sun_sensor.CoarseSunSensor_applySensorErrors(self) def scaleSensorValues(self): return _coarse_sun_sensor.CoarseSunSensor_scaleSensorValues(self) def applySaturation(self): return _coarse_sun_sensor.CoarseSunSensor_applySaturation(self) def writeOutputMessages(self, Clock): return _coarse_sun_sensor.CoarseSunSensor_writeOutputMessages(self, Clock) __swig_setmethods__["sunInMsgName"] = _coarse_sun_sensor.CoarseSunSensor_sunInMsgName_set __swig_getmethods__["sunInMsgName"] = _coarse_sun_sensor.CoarseSunSensor_sunInMsgName_get if _newclass: sunInMsgName = _swig_property(_coarse_sun_sensor.CoarseSunSensor_sunInMsgName_get, _coarse_sun_sensor.CoarseSunSensor_sunInMsgName_set) __swig_setmethods__["stateInMsgName"] = _coarse_sun_sensor.CoarseSunSensor_stateInMsgName_set __swig_getmethods__["stateInMsgName"] = _coarse_sun_sensor.CoarseSunSensor_stateInMsgName_get if _newclass: stateInMsgName = _swig_property(_coarse_sun_sensor.CoarseSunSensor_stateInMsgName_get, _coarse_sun_sensor.CoarseSunSensor_stateInMsgName_set) __swig_setmethods__["cssDataOutMsgName"] = _coarse_sun_sensor.CoarseSunSensor_cssDataOutMsgName_set __swig_getmethods__["cssDataOutMsgName"] = _coarse_sun_sensor.CoarseSunSensor_cssDataOutMsgName_get if _newclass: cssDataOutMsgName = _swig_property(_coarse_sun_sensor.CoarseSunSensor_cssDataOutMsgName_get, _coarse_sun_sensor.CoarseSunSensor_cssDataOutMsgName_set) __swig_setmethods__["sunEclipseInMsgName"] = _coarse_sun_sensor.CoarseSunSensor_sunEclipseInMsgName_set __swig_getmethods__["sunEclipseInMsgName"] = _coarse_sun_sensor.CoarseSunSensor_sunEclipseInMsgName_get if _newclass: sunEclipseInMsgName = _swig_property(_coarse_sun_sensor.CoarseSunSensor_sunEclipseInMsgName_get, _coarse_sun_sensor.CoarseSunSensor_sunEclipseInMsgName_set) __swig_setmethods__["faultState"] = _coarse_sun_sensor.CoarseSunSensor_faultState_set __swig_getmethods__["faultState"] = _coarse_sun_sensor.CoarseSunSensor_faultState_get if _newclass: faultState = _swig_property(_coarse_sun_sensor.CoarseSunSensor_faultState_get, _coarse_sun_sensor.CoarseSunSensor_faultState_set) __swig_setmethods__["theta"] = _coarse_sun_sensor.CoarseSunSensor_theta_set __swig_getmethods__["theta"] = _coarse_sun_sensor.CoarseSunSensor_theta_get if _newclass: theta = _swig_property(_coarse_sun_sensor.CoarseSunSensor_theta_get, _coarse_sun_sensor.CoarseSunSensor_theta_set) __swig_setmethods__["phi"] = _coarse_sun_sensor.CoarseSunSensor_phi_set __swig_getmethods__["phi"] = _coarse_sun_sensor.CoarseSunSensor_phi_get if _newclass: phi = _swig_property(_coarse_sun_sensor.CoarseSunSensor_phi_get, _coarse_sun_sensor.CoarseSunSensor_phi_set) __swig_setmethods__["B2P321Angles"] = _coarse_sun_sensor.CoarseSunSensor_B2P321Angles_set __swig_getmethods__["B2P321Angles"] = _coarse_sun_sensor.CoarseSunSensor_B2P321Angles_get if _newclass: B2P321Angles = _swig_property(_coarse_sun_sensor.CoarseSunSensor_B2P321Angles_get, _coarse_sun_sensor.CoarseSunSensor_B2P321Angles_set) __swig_setmethods__["dcm_PB"] = _coarse_sun_sensor.CoarseSunSensor_dcm_PB_set __swig_getmethods__["dcm_PB"] = _coarse_sun_sensor.CoarseSunSensor_dcm_PB_get if _newclass: dcm_PB = _swig_property(_coarse_sun_sensor.CoarseSunSensor_dcm_PB_get, _coarse_sun_sensor.CoarseSunSensor_dcm_PB_set) __swig_setmethods__["nHat_B"] = _coarse_sun_sensor.CoarseSunSensor_nHat_B_set __swig_getmethods__["nHat_B"] = _coarse_sun_sensor.CoarseSunSensor_nHat_B_get if _newclass: nHat_B = _swig_property(_coarse_sun_sensor.CoarseSunSensor_nHat_B_get, _coarse_sun_sensor.CoarseSunSensor_nHat_B_set) __swig_setmethods__["sHat_B"] = _coarse_sun_sensor.CoarseSunSensor_sHat_B_set __swig_getmethods__["sHat_B"] = _coarse_sun_sensor.CoarseSunSensor_sHat_B_get if _newclass: sHat_B = _swig_property(_coarse_sun_sensor.CoarseSunSensor_sHat_B_get, _coarse_sun_sensor.CoarseSunSensor_sHat_B_set) __swig_setmethods__["directValue"] = _coarse_sun_sensor.CoarseSunSensor_directValue_set __swig_getmethods__["directValue"] = _coarse_sun_sensor.CoarseSunSensor_directValue_get if _newclass: directValue = _swig_property(_coarse_sun_sensor.CoarseSunSensor_directValue_get, _coarse_sun_sensor.CoarseSunSensor_directValue_set) __swig_setmethods__["albedoValue"] = _coarse_sun_sensor.CoarseSunSensor_albedoValue_set __swig_getmethods__["albedoValue"] = _coarse_sun_sensor.CoarseSunSensor_albedoValue_get if _newclass: albedoValue = _swig_property(_coarse_sun_sensor.CoarseSunSensor_albedoValue_get, _coarse_sun_sensor.CoarseSunSensor_albedoValue_set) __swig_setmethods__["scaleFactor"] = _coarse_sun_sensor.CoarseSunSensor_scaleFactor_set __swig_getmethods__["scaleFactor"] = _coarse_sun_sensor.CoarseSunSensor_scaleFactor_get if _newclass: scaleFactor = _swig_property(_coarse_sun_sensor.CoarseSunSensor_scaleFactor_get, _coarse_sun_sensor.CoarseSunSensor_scaleFactor_set) __swig_setmethods__["sensedValue"] = _coarse_sun_sensor.CoarseSunSensor_sensedValue_set __swig_getmethods__["sensedValue"] = _coarse_sun_sensor.CoarseSunSensor_sensedValue_get if _newclass: sensedValue = _swig_property(_coarse_sun_sensor.CoarseSunSensor_sensedValue_get, _coarse_sun_sensor.CoarseSunSensor_sensedValue_set) __swig_setmethods__["trueValue"] = _coarse_sun_sensor.CoarseSunSensor_trueValue_set __swig_getmethods__["trueValue"] = _coarse_sun_sensor.CoarseSunSensor_trueValue_get if _newclass: trueValue = _swig_property(_coarse_sun_sensor.CoarseSunSensor_trueValue_get, _coarse_sun_sensor.CoarseSunSensor_trueValue_set) __swig_setmethods__["kellyFactor"] = _coarse_sun_sensor.CoarseSunSensor_kellyFactor_set __swig_getmethods__["kellyFactor"] = _coarse_sun_sensor.CoarseSunSensor_kellyFactor_get if _newclass: kellyFactor = _swig_property(_coarse_sun_sensor.CoarseSunSensor_kellyFactor_get, _coarse_sun_sensor.CoarseSunSensor_kellyFactor_set) __swig_setmethods__["fov"] = _coarse_sun_sensor.CoarseSunSensor_fov_set __swig_getmethods__["fov"] = _coarse_sun_sensor.CoarseSunSensor_fov_get if _newclass: fov = _swig_property(_coarse_sun_sensor.CoarseSunSensor_fov_get, _coarse_sun_sensor.CoarseSunSensor_fov_set) __swig_setmethods__["r_B"] = _coarse_sun_sensor.CoarseSunSensor_r_B_set __swig_getmethods__["r_B"] = _coarse_sun_sensor.CoarseSunSensor_r_B_get if _newclass: r_B = _swig_property(_coarse_sun_sensor.CoarseSunSensor_r_B_get, _coarse_sun_sensor.CoarseSunSensor_r_B_set) __swig_setmethods__["senBias"] = _coarse_sun_sensor.CoarseSunSensor_senBias_set __swig_getmethods__["senBias"] = _coarse_sun_sensor.CoarseSunSensor_senBias_get if _newclass: senBias = _swig_property(_coarse_sun_sensor.CoarseSunSensor_senBias_get, _coarse_sun_sensor.CoarseSunSensor_senBias_set) __swig_setmethods__["senNoiseStd"] = _coarse_sun_sensor.CoarseSunSensor_senNoiseStd_set __swig_getmethods__["senNoiseStd"] = _coarse_sun_sensor.CoarseSunSensor_senNoiseStd_get if _newclass: senNoiseStd = _swig_property(_coarse_sun_sensor.CoarseSunSensor_senNoiseStd_get, _coarse_sun_sensor.CoarseSunSensor_senNoiseStd_set) __swig_setmethods__["outputBufferCount"] = _coarse_sun_sensor.CoarseSunSensor_outputBufferCount_set __swig_getmethods__["outputBufferCount"] = _coarse_sun_sensor.CoarseSunSensor_outputBufferCount_get if _newclass: outputBufferCount = _swig_property(_coarse_sun_sensor.CoarseSunSensor_outputBufferCount_get, _coarse_sun_sensor.CoarseSunSensor_outputBufferCount_set) __swig_setmethods__["maxOutput"] = _coarse_sun_sensor.CoarseSunSensor_maxOutput_set __swig_getmethods__["maxOutput"] = _coarse_sun_sensor.CoarseSunSensor_maxOutput_get if _newclass: maxOutput = _swig_property(_coarse_sun_sensor.CoarseSunSensor_maxOutput_get, _coarse_sun_sensor.CoarseSunSensor_maxOutput_set) __swig_setmethods__["minOutput"] = _coarse_sun_sensor.CoarseSunSensor_minOutput_set __swig_getmethods__["minOutput"] = _coarse_sun_sensor.CoarseSunSensor_minOutput_get if _newclass: minOutput = _swig_property(_coarse_sun_sensor.CoarseSunSensor_minOutput_get, _coarse_sun_sensor.CoarseSunSensor_minOutput_set) __swig_setmethods__["walkBounds"] = _coarse_sun_sensor.CoarseSunSensor_walkBounds_set __swig_getmethods__["walkBounds"] = _coarse_sun_sensor.CoarseSunSensor_walkBounds_get if _newclass: walkBounds = _swig_property(_coarse_sun_sensor.CoarseSunSensor_walkBounds_get, _coarse_sun_sensor.CoarseSunSensor_walkBounds_set) def __init__(self, *args): this = _coarse_sun_sensor.new_CoarseSunSensor(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this CoarseSunSensor_swigregister = _coarse_sun_sensor.CoarseSunSensor_swigregister CoarseSunSensor_swigregister(CoarseSunSensor) class CSSConstellation(SysModel): __swig_setmethods__ = {} for _s in [SysModel]: __swig_setmethods__.update(getattr(_s, '__swig_setmethods__', {})) __setattr__ = lambda self, name, value: _swig_setattr(self, CSSConstellation, name, value) __swig_getmethods__ = {} for _s in [SysModel]: __swig_getmethods__.update(getattr(_s, '__swig_getmethods__', {})) __getattr__ = lambda self, name: _swig_getattr(self, CSSConstellation, name) __repr__ = _swig_repr __swig_destroy__ = _coarse_sun_sensor.delete_CSSConstellation __del__ = lambda self: None def CrossInit(self): return _coarse_sun_sensor.CSSConstellation_CrossInit(self) def SelfInit(self): return _coarse_sun_sensor.CSSConstellation_SelfInit(self) def UpdateState(self, CurrentSimNanos): return _coarse_sun_sensor.CSSConstellation_UpdateState(self, CurrentSimNanos) def appendCSS(self, newSensor): return _coarse_sun_sensor.CSSConstellation_appendCSS(self, newSensor) __swig_setmethods__["outputBufferCount"] = _coarse_sun_sensor.CSSConstellation_outputBufferCount_set __swig_getmethods__["outputBufferCount"] = _coarse_sun_sensor.CSSConstellation_outputBufferCount_get if _newclass: outputBufferCount = _swig_property(_coarse_sun_sensor.CSSConstellation_outputBufferCount_get, _coarse_sun_sensor.CSSConstellation_outputBufferCount_set) __swig_setmethods__["outputConstellationMessage"] = _coarse_sun_sensor.CSSConstellation_outputConstellationMessage_set __swig_getmethods__["outputConstellationMessage"] = _coarse_sun_sensor.CSSConstellation_outputConstellationMessage_get if _newclass: outputConstellationMessage = _swig_property(_coarse_sun_sensor.CSSConstellation_outputConstellationMessage_get, _coarse_sun_sensor.CSSConstellation_outputConstellationMessage_set) __swig_setmethods__["outputConstID"] = _coarse_sun_sensor.CSSConstellation_outputConstID_set __swig_getmethods__["outputConstID"] = _coarse_sun_sensor.CSSConstellation_outputConstID_get if _newclass: outputConstID = _swig_property(_coarse_sun_sensor.CSSConstellation_outputConstID_get, _coarse_sun_sensor.CSSConstellation_outputConstID_set) __swig_setmethods__["sensorList"] = _coarse_sun_sensor.CSSConstellation_sensorList_set __swig_getmethods__["sensorList"] = _coarse_sun_sensor.CSSConstellation_sensorList_get if _newclass: sensorList = _swig_property(_coarse_sun_sensor.CSSConstellation_sensorList_get, _coarse_sun_sensor.CSSConstellation_sensorList_set) def __init__(self, *args): this = _coarse_sun_sensor.new_CSSConstellation(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this CSSConstellation_swigregister = _coarse_sun_sensor.CSSConstellation_swigregister CSSConstellation_swigregister(CSSConstellation) class SCPlusStatesSimMsg(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, SCPlusStatesSimMsg, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, SCPlusStatesSimMsg, name) __repr__ = _swig_repr __swig_setmethods__["r_BN_N"] = _coarse_sun_sensor.SCPlusStatesSimMsg_r_BN_N_set __swig_getmethods__["r_BN_N"] = _coarse_sun_sensor.SCPlusStatesSimMsg_r_BN_N_get if _newclass: r_BN_N = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_r_BN_N_get, _coarse_sun_sensor.SCPlusStatesSimMsg_r_BN_N_set) __swig_setmethods__["v_BN_N"] = _coarse_sun_sensor.SCPlusStatesSimMsg_v_BN_N_set __swig_getmethods__["v_BN_N"] = _coarse_sun_sensor.SCPlusStatesSimMsg_v_BN_N_get if _newclass: v_BN_N = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_v_BN_N_get, _coarse_sun_sensor.SCPlusStatesSimMsg_v_BN_N_set) __swig_setmethods__["r_CN_N"] = _coarse_sun_sensor.SCPlusStatesSimMsg_r_CN_N_set __swig_getmethods__["r_CN_N"] = _coarse_sun_sensor.SCPlusStatesSimMsg_r_CN_N_get if _newclass: r_CN_N = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_r_CN_N_get, _coarse_sun_sensor.SCPlusStatesSimMsg_r_CN_N_set) __swig_setmethods__["v_CN_N"] = _coarse_sun_sensor.SCPlusStatesSimMsg_v_CN_N_set __swig_getmethods__["v_CN_N"] = _coarse_sun_sensor.SCPlusStatesSimMsg_v_CN_N_get if _newclass: v_CN_N = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_v_CN_N_get, _coarse_sun_sensor.SCPlusStatesSimMsg_v_CN_N_set) __swig_setmethods__["sigma_BN"] = _coarse_sun_sensor.SCPlusStatesSimMsg_sigma_BN_set __swig_getmethods__["sigma_BN"] = _coarse_sun_sensor.SCPlusStatesSimMsg_sigma_BN_get if _newclass: sigma_BN = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_sigma_BN_get, _coarse_sun_sensor.SCPlusStatesSimMsg_sigma_BN_set) __swig_setmethods__["omega_BN_B"] = _coarse_sun_sensor.SCPlusStatesSimMsg_omega_BN_B_set __swig_getmethods__["omega_BN_B"] = _coarse_sun_sensor.SCPlusStatesSimMsg_omega_BN_B_get if _newclass: omega_BN_B = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_omega_BN_B_get, _coarse_sun_sensor.SCPlusStatesSimMsg_omega_BN_B_set) __swig_setmethods__["omegaDot_BN_B"] = _coarse_sun_sensor.SCPlusStatesSimMsg_omegaDot_BN_B_set __swig_getmethods__["omegaDot_BN_B"] = _coarse_sun_sensor.SCPlusStatesSimMsg_omegaDot_BN_B_get if _newclass: omegaDot_BN_B = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_omegaDot_BN_B_get, _coarse_sun_sensor.SCPlusStatesSimMsg_omegaDot_BN_B_set) __swig_setmethods__["TotalAccumDVBdy"] = _coarse_sun_sensor.SCPlusStatesSimMsg_TotalAccumDVBdy_set __swig_getmethods__["TotalAccumDVBdy"] = _coarse_sun_sensor.SCPlusStatesSimMsg_TotalAccumDVBdy_get if _newclass: TotalAccumDVBdy = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_TotalAccumDVBdy_get, _coarse_sun_sensor.SCPlusStatesSimMsg_TotalAccumDVBdy_set) __swig_setmethods__["TotalAccumDV_BN_B"] = _coarse_sun_sensor.SCPlusStatesSimMsg_TotalAccumDV_BN_B_set __swig_getmethods__["TotalAccumDV_BN_B"] = _coarse_sun_sensor.SCPlusStatesSimMsg_TotalAccumDV_BN_B_get if _newclass: TotalAccumDV_BN_B = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_TotalAccumDV_BN_B_get, _coarse_sun_sensor.SCPlusStatesSimMsg_TotalAccumDV_BN_B_set) __swig_setmethods__["nonConservativeAccelpntB_B"] = _coarse_sun_sensor.SCPlusStatesSimMsg_nonConservativeAccelpntB_B_set __swig_getmethods__["nonConservativeAccelpntB_B"] = _coarse_sun_sensor.SCPlusStatesSimMsg_nonConservativeAccelpntB_B_get if _newclass: nonConservativeAccelpntB_B = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_nonConservativeAccelpntB_B_get, _coarse_sun_sensor.SCPlusStatesSimMsg_nonConservativeAccelpntB_B_set) __swig_setmethods__["MRPSwitchCount"] = _coarse_sun_sensor.SCPlusStatesSimMsg_MRPSwitchCount_set __swig_getmethods__["MRPSwitchCount"] = _coarse_sun_sensor.SCPlusStatesSimMsg_MRPSwitchCount_get if _newclass: MRPSwitchCount = _swig_property(_coarse_sun_sensor.SCPlusStatesSimMsg_MRPSwitchCount_get, _coarse_sun_sensor.SCPlusStatesSimMsg_MRPSwitchCount_set) def __init__(self, *args): this = _coarse_sun_sensor.new_SCPlusStatesSimMsg(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this __swig_destroy__ = _coarse_sun_sensor.delete_SCPlusStatesSimMsg __del__ = lambda self: None SCPlusStatesSimMsg_swigregister = _coarse_sun_sensor.SCPlusStatesSimMsg_swigregister SCPlusStatesSimMsg_swigregister(SCPlusStatesSimMsg) MAX_BODY_NAME_LENGTH = _coarse_sun_sensor.MAX_BODY_NAME_LENGTH class SpicePlanetStateSimMsg(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, SpicePlanetStateSimMsg, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, SpicePlanetStateSimMsg, name) __repr__ = _swig_repr __swig_setmethods__["J2000Current"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_J2000Current_set __swig_getmethods__["J2000Current"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_J2000Current_get if _newclass: J2000Current = _swig_property(_coarse_sun_sensor.SpicePlanetStateSimMsg_J2000Current_get, _coarse_sun_sensor.SpicePlanetStateSimMsg_J2000Current_set) __swig_setmethods__["PositionVector"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_PositionVector_set __swig_getmethods__["PositionVector"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_PositionVector_get if _newclass: PositionVector = _swig_property(_coarse_sun_sensor.SpicePlanetStateSimMsg_PositionVector_get, _coarse_sun_sensor.SpicePlanetStateSimMsg_PositionVector_set) __swig_setmethods__["VelocityVector"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_VelocityVector_set __swig_getmethods__["VelocityVector"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_VelocityVector_get if _newclass: VelocityVector = _swig_property(_coarse_sun_sensor.SpicePlanetStateSimMsg_VelocityVector_get, _coarse_sun_sensor.SpicePlanetStateSimMsg_VelocityVector_set) __swig_setmethods__["J20002Pfix"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_J20002Pfix_set __swig_getmethods__["J20002Pfix"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_J20002Pfix_get if _newclass: J20002Pfix = _swig_property(_coarse_sun_sensor.SpicePlanetStateSimMsg_J20002Pfix_get, _coarse_sun_sensor.SpicePlanetStateSimMsg_J20002Pfix_set) __swig_setmethods__["J20002Pfix_dot"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_J20002Pfix_dot_set __swig_getmethods__["J20002Pfix_dot"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_J20002Pfix_dot_get if _newclass: J20002Pfix_dot = _swig_property(_coarse_sun_sensor.SpicePlanetStateSimMsg_J20002Pfix_dot_get, _coarse_sun_sensor.SpicePlanetStateSimMsg_J20002Pfix_dot_set) __swig_setmethods__["computeOrient"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_computeOrient_set __swig_getmethods__["computeOrient"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_computeOrient_get if _newclass: computeOrient = _swig_property(_coarse_sun_sensor.SpicePlanetStateSimMsg_computeOrient_get, _coarse_sun_sensor.SpicePlanetStateSimMsg_computeOrient_set) __swig_setmethods__["PlanetName"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_PlanetName_set __swig_getmethods__["PlanetName"] = _coarse_sun_sensor.SpicePlanetStateSimMsg_PlanetName_get if _newclass: PlanetName = _swig_property(_coarse_sun_sensor.SpicePlanetStateSimMsg_PlanetName_get, _coarse_sun_sensor.SpicePlanetStateSimMsg_PlanetName_set) def __init__(self, *args): this = _coarse_sun_sensor.new_SpicePlanetStateSimMsg(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this __swig_destroy__ = _coarse_sun_sensor.delete_SpicePlanetStateSimMsg __del__ = lambda self: None SpicePlanetStateSimMsg_swigregister = _coarse_sun_sensor.SpicePlanetStateSimMsg_swigregister SpicePlanetStateSimMsg_swigregister(SpicePlanetStateSimMsg) class CSSRawDataSimMsg(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, CSSRawDataSimMsg, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, CSSRawDataSimMsg, name) __repr__ = _swig_repr __swig_setmethods__["OutputData"] = _coarse_sun_sensor.CSSRawDataSimMsg_OutputData_set __swig_getmethods__["OutputData"] = _coarse_sun_sensor.CSSRawDataSimMsg_OutputData_get if _newclass: OutputData = _swig_property(_coarse_sun_sensor.CSSRawDataSimMsg_OutputData_get, _coarse_sun_sensor.CSSRawDataSimMsg_OutputData_set) def __init__(self, *args): this = _coarse_sun_sensor.new_CSSRawDataSimMsg(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this __swig_destroy__ = _coarse_sun_sensor.delete_CSSRawDataSimMsg __del__ = lambda self: None CSSRawDataSimMsg_swigregister = _coarse_sun_sensor.CSSRawDataSimMsg_swigregister CSSRawDataSimMsg_swigregister(CSSRawDataSimMsg) class EclipseSimMsg(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, EclipseSimMsg, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, EclipseSimMsg, name) __repr__ = _swig_repr __swig_setmethods__["shadowFactor"] = _coarse_sun_sensor.EclipseSimMsg_shadowFactor_set __swig_getmethods__["shadowFactor"] = _coarse_sun_sensor.EclipseSimMsg_shadowFactor_get if _newclass: shadowFactor = _swig_property(_coarse_sun_sensor.EclipseSimMsg_shadowFactor_get, _coarse_sun_sensor.EclipseSimMsg_shadowFactor_set) def __init__(self, *args): this = _coarse_sun_sensor.new_EclipseSimMsg(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this __swig_destroy__ = _coarse_sun_sensor.delete_EclipseSimMsg __del__ = lambda self: None EclipseSimMsg_swigregister = _coarse_sun_sensor.EclipseSimMsg_swigregister EclipseSimMsg_swigregister(EclipseSimMsg) class CSSArraySensorIntMsg(_object): __swig_setmethods__ = {} __setattr__ = lambda self, name, value: _swig_setattr(self, CSSArraySensorIntMsg, name, value) __swig_getmethods__ = {} __getattr__ = lambda self, name: _swig_getattr(self, CSSArraySensorIntMsg, name) __repr__ = _swig_repr __swig_setmethods__["CosValue"] = _coarse_sun_sensor.CSSArraySensorIntMsg_CosValue_set __swig_getmethods__["CosValue"] = _coarse_sun_sensor.CSSArraySensorIntMsg_CosValue_get if _newclass: CosValue = _swig_property(_coarse_sun_sensor.CSSArraySensorIntMsg_CosValue_get, _coarse_sun_sensor.CSSArraySensorIntMsg_CosValue_set) def __init__(self, *args): this = _coarse_sun_sensor.new_CSSArraySensorIntMsg(*args) try: self.this.append(this) except __builtin__.Exception: self.this = this __swig_destroy__ = _coarse_sun_sensor.delete_CSSArraySensorIntMsg __del__ = lambda self: None CSSArraySensorIntMsg_swigregister = _coarse_sun_sensor.CSSArraySensorIntMsg_swigregister CSSArraySensorIntMsg_swigregister(CSSArraySensorIntMsg) sizeof_CSSRawDataSimMsg = _coarse_sun_sensor.sizeof_CSSRawDataSimMsg sizeof_EclipseSimMsg = _coarse_sun_sensor.sizeof_EclipseSimMsg sizeof_CSSArraySensorIntMsg = _coarse_sun_sensor.sizeof_CSSArraySensorIntMsg sizeof_SpicePlanetStateSimMsg = _coarse_sun_sensor.sizeof_SpicePlanetStateSimMsg sizeof_SCPlusStatesSimMsg = _coarse_sun_sensor.sizeof_SCPlusStatesSimMsg import sys protectAllClasses(sys.modules[__name__]) # This file is compatible with both classic and new-style classes.
py
1a2f343b727bcfb94c084c89c7d59ea6577e7319
""" Custom datasources for awx_manage information """ from insights.core.context import HostContext from insights.core.dr import SkipComponent from insights.core.plugins import datasource from insights.core.spec_factory import DatasourceProvider, simple_command from insights.core.filters import get_filters from insights.specs import Specs import json import collections class LocalSpecs(Specs): """ Local specs used only by awx_manage datasources """ awx_manage_check_license_data_raw = simple_command("/usr/bin/awx-manage check_license --data") """ Returns the output of command ``/usr/bin/awx-manage check_license --data`` """ @datasource(LocalSpecs.awx_manage_check_license_data_raw, HostContext) def awx_manage_check_license_data_datasource(broker): """ This datasource provides the not-sensitive information collected from ``/usr/bin/awx-manage check_license --data``. Typical content of ``/usr/bin/awx-manage check_license --data`` file is:: {"contact_email": "test@redhat.com", "company_name": "test Inc", "instance_count": 100, "license_date": 1655092799, "license_type": "enterprise", "subscription_name": "Red Hat Ansible Automation, Standard (100 Managed Nodes)", "sku": "MCT3691", "support_level": "Standard", "product_name": "Red Hat Ansible Automation Platform", "valid_key": true, "satellite": null, "pool_id": "2c92808179803e530179ea5989a157a4", "current_instances": 1, "available_instances": 100, "free_instances": 99, "time_remaining": 29885220, "trial": false, "grace_period_remaining": 32477220, "compliant": true, "date_warning": false, "date_expired": false} Returns: str: JSON string containing non-sensitive information. Raises: SkipComponent: When the filter/path does not exist or any exception occurs. """ try: filters = get_filters(Specs.awx_manage_check_license_data) content = broker[LocalSpecs.awx_manage_check_license_data_raw].content if content and filters: json_data = json.loads(content[0]) filter_result = {} for item in filters: filter_result[item] = json_data.get(item) if filter_result: return DatasourceProvider(content=json.dumps(collections.OrderedDict(sorted(filter_result.items()))), relative_path='insights_commands/awx-manage_check_license_--data') except Exception as e: raise SkipComponent("Unexpected exception:{e}".format(e=str(e))) raise SkipComponent
py
1a2f356a3b1a6c66f35b4a180bfe444be0803637
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # from collections import OrderedDict from typing import Dict, Type from .base import SecurityCenterTransport from .grpc import SecurityCenterGrpcTransport from .grpc_asyncio import SecurityCenterGrpcAsyncIOTransport # Compile a registry of transports. _transport_registry = OrderedDict() # type: Dict[str, Type[SecurityCenterTransport]] _transport_registry["grpc"] = SecurityCenterGrpcTransport _transport_registry["grpc_asyncio"] = SecurityCenterGrpcAsyncIOTransport __all__ = ( "SecurityCenterTransport", "SecurityCenterGrpcTransport", "SecurityCenterGrpcAsyncIOTransport", )
py
1a2f357640e44f8e721169a9ad18c8e63b8bd531
# Generated by Django 3.1.7 on 2021-09-04 16:59 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='BillCategory', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(max_length=150, unique=True)), ('balance', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ], options={ 'verbose_name_plural': '1. Bill Category', }, ), migrations.CreateModel( name='GenericExpenseCategory', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(max_length=150, unique=True)), ('balance', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ], options={ 'verbose_name_plural': '6. Expense Category', }, ), migrations.CreateModel( name='PaymentMethod', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(max_length=150, unique=True)), ], options={ 'verbose_name_plural': '0. Payment Method', }, ), migrations.CreateModel( name='PayrollCategory', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(max_length=150, unique=True)), ('balance', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ], options={ 'verbose_name_plural': '3. Payroll Category', }, ), migrations.CreateModel( name='Person', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(max_length=150, unique=True)), ('occupation', models.CharField(blank=True, max_length=100, null=True)), ('phone', models.CharField(blank=True, max_length=10, null=True)), ('balance', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ], options={ 'verbose_name_plural': '4. Persons', }, ), migrations.CreateModel( name='Payroll', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(blank=True, max_length=100, null=True)), ('date_expired', models.DateField()), ('final_value', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ('paid_value', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ('is_paid', models.BooleanField(default=False)), ('category', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='category_payroll', to='expenses.payrollcategory')), ('payment_method', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, to='expenses.paymentmethod')), ('person', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='person_payroll', to='expenses.person')), ], options={ 'verbose_name_plural': '5. Payroll', 'ordering': ['-date_expired'], }, ), migrations.CreateModel( name='GenericExpense', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(blank=True, max_length=100, null=True)), ('date_expired', models.DateField()), ('final_value', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ('paid_value', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ('is_paid', models.BooleanField(default=False)), ('category', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='category_expenses', to='expenses.genericexpensecategory')), ('payment_method', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, to='expenses.paymentmethod')), ], options={ 'verbose_name_plural': '7. Generic Expenses', 'ordering': ['-date_expired'], }, ), migrations.CreateModel( name='Bill', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('title', models.CharField(blank=True, max_length=100, null=True)), ('date_expired', models.DateField()), ('final_value', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ('paid_value', models.DecimalField(decimal_places=2, default=0, max_digits=20)), ('is_paid', models.BooleanField(default=False)), ('category', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, related_name='bills', to='expenses.billcategory')), ('payment_method', models.ForeignKey(null=True, on_delete=django.db.models.deletion.SET_NULL, to='expenses.paymentmethod')), ], options={ 'verbose_name_plural': '2. Bills', 'ordering': ['-date_expired'], }, ), ]
py
1a2f36f3b183465e4c10aefcedae56dc18b0d87f
# Python - 3.6.0 test.assert_equals(sumin(5), 55) test.assert_equals(sumax(8), 372) test.assert_equals(sumsum(8), 576) test.assert_equals(sumin(15), 1240)
py
1a2f371f7afdbbb5d3272fb08bed7e871560e86e
#!/usr/bin/env python # # A library that provides a Python interface to the Telegram Bot API # Copyright (C) 2015-2021 # Leandro Toledo de Souza <devs@python-telegram-bot.org> # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Lesser Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser Public License for more details. # # You should have received a copy of the GNU Lesser Public License # along with this program. If not, see [http://www.gnu.org/licenses/]. """This module contains an object that represents a Telegram Location.""" from typing import Any from telegram import TelegramObject class Location(TelegramObject): """This object represents a point on the map. Objects of this class are comparable in terms of equality. Two objects of this class are considered equal, if their :attr:`longitute` and :attr:`latitude` are equal. Args: longitude (:obj:`float`): Longitude as defined by sender. latitude (:obj:`float`): Latitude as defined by sender. horizontal_accuracy (:obj:`float`, optional): The radius of uncertainty for the location, measured in meters; 0-1500. live_period (:obj:`int`, optional): Time relative to the message sending date, during which the location can be updated, in seconds. For active live locations only. heading (:obj:`int`, optional): The direction in which user is moving, in degrees; 1-360. For active live locations only. proximity_alert_radius (:obj:`int`, optional): Maximum distance for proximity alerts about approaching another chat member, in meters. For sent live locations only. **kwargs (:obj:`dict`): Arbitrary keyword arguments. Attributes: longitude (:obj:`float`): Longitude as defined by sender. latitude (:obj:`float`): Latitude as defined by sender. horizontal_accuracy (:obj:`float`): Optional. The radius of uncertainty for the location, measured in meters. live_period (:obj:`int`): Optional. Time relative to the message sending date, during which the location can be updated, in seconds. For active live locations only. heading (:obj:`int`): Optional. The direction in which user is moving, in degrees. For active live locations only. proximity_alert_radius (:obj:`int`): Optional. Maximum distance for proximity alerts about approaching another chat member, in meters. For sent live locations only. """ def __init__( self, longitude: float, latitude: float, horizontal_accuracy: float = None, live_period: int = None, heading: int = None, proximity_alert_radius: int = None, **_kwargs: Any, ): # Required self.longitude = float(longitude) self.latitude = float(latitude) # Optionals self.horizontal_accuracy = float(horizontal_accuracy) if horizontal_accuracy else None self.live_period = int(live_period) if live_period else None self.heading = int(heading) if heading else None self.proximity_alert_radius = ( int(proximity_alert_radius) if proximity_alert_radius else None ) self._id_attrs = (self.longitude, self.latitude)
py
1a2f3777b3029c8ffb0ee5526be2beea1d046bea
#!/usr/bin/python3 # -*- coding:utf-8 -*- # Copyright 2020 Ozward Wang # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import struct HEADER_STRUCT_FORMAT = '!16s16s' N_STRUCT_FORMAT = '!H' def pack_number(number): n_struct = struct.Struct(N_STRUCT_FORMAT) return n_struct.pack(number) def unpack_number(data): n_struct = struct.Struct(N_STRUCT_FORMAT) return n_struct.unpack(data)[0] def pack_header(leads, identity): s_struct = struct.Struct(HEADER_STRUCT_FORMAT) if isinstance(identity, str): identity = identity.encode('utf-8')[:16] else: identity = identity[:16] return s_struct.pack(leads, identity) def unpack_header(header): s_struct = struct.Struct(HEADER_STRUCT_FORMAT) leads, identity = s_struct.unpack(header) return leads, identity def pack_iv(data, iv): return data[:2] + iv + data[2:] def unpack_iv(header): data = header[:2] + header[18:] iv = header[2:18] return iv, data def pack_body(data, tail): return data + tail def unpack_body(body): return body[:-4], body[-4:]
py
1a2f390c5688bbd1a52ab90c6fc0d7967e83b56d
#from app import app, db from app import db from sqlalchemy.orm import relationship import datetime class Post(db.Model): """ Table schema """ __tablename__ = "posts" id = db.Column(db.Integer, primary_key=True, autoincrement=True) post_id = db.Column(db.Text(), nullable=False) origin_blog = db.Column(db.Integer, db.ForeignKey('blogs.id')) date = db.Column(db.DateTime(), nullable=False) content = db.Column(db.Text(), nullable=False) title = db.Column(db.Text()) post_url = db.Column(db.Text()) blog = relationship("Blog", back_populates="posts") def __repr__(self): return "Post: {}.".format(self.title) @staticmethod def get_posts(): return Post.query.order_by(Post.date.desc()).limit(10).all() class Blog(db.Model): """ Blog model """ __tablename__ = "blogs" id = db.Column(db.Integer, primary_key=True, autoincrement=True) url = db.Column(db.Text()) feed_url = db.Column(db.Text()) last_update = db.Column(db.DateTime()) name = db.Column(db.Text()) posts = relationship("Post", order_by=Post.id, back_populates="blog")
py
1a2f3965932293b6381f41183d7f941abc26d964
# -*- coding: utf-8 -*- """setup.py""" import os import sys from setuptools import setup from setuptools.command.test import test as TestCommand class Tox(TestCommand): user_options = [('tox-args=', 'a', 'Arguments to pass to tox')] def initialize_options(self): TestCommand.initialize_options(self) self.tox_args = None def finalize_options(self): TestCommand.finalize_options(self) self.test_args = [] self.test_suite = True def run_tests(self): import tox import shlex if self.tox_args: errno = tox.cmdline(args=shlex.split(self.tox_args)) else: errno = tox.cmdline(self.test_args) sys.exit(errno) def read_content(filepath): with open(filepath) as fobj: return fobj.read() classifiers = [ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 3.5", "Programming Language :: Python :: 3.6", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: Implementation :: CPython", "Programming Language :: Python :: Implementation :: PyPy", ] long_description = ( read_content("README.rst") + read_content(os.path.join("docs/source", "CHANGELOG.rst"))) requires = [ 'pyyaml', 'setuptools', 'click', 'sphinx-click', 'texttable', 'coverage' ] extras_require = { 'reST': ['Sphinx'], } if os.environ.get('READTHEDOCS', None): extras_require['reST'].append('recommonmark') setup(name='pymerit', version='0.1.0', description='Standardized metadata', long_description=long_description, author='Venkata Pingali', author_email='pingali@scribbledata.io', url='https://github.com/pingali/pymerit', classifiers=classifiers, packages=['pymerit'], data_files = [("", ["LICENSE.txt"])], install_requires=requires, include_package_data=True, extras_require=extras_require, tests_require=['tox'], cmdclass={'test': Tox}, entry_points = { 'console_scripts': ['merit=pymerit.cli:main'], } )
py
1a2f39848bdc7abd1070f4ff458db1facf4a6ebd
# Copyright 2017 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Tests for tf.training.evaluation.""" import os import numpy as np from tensorflow.python.framework import constant_op from tensorflow.python.framework import dtypes from tensorflow.python.framework import ops from tensorflow.python.framework import random_seed from tensorflow.python.layers import layers from tensorflow.python.ops import array_ops from tensorflow.python.ops import math_ops from tensorflow.python.ops import metrics as metrics_module from tensorflow.python.ops import state_ops from tensorflow.python.ops import variable_scope from tensorflow.python.ops.losses import losses from tensorflow.python.platform import test from tensorflow.python.training import basic_session_run_hooks from tensorflow.python.training import evaluation from tensorflow.python.training import gradient_descent from tensorflow.python.training import monitored_session from tensorflow.python.training import saver from tensorflow.python.training import training _USE_GLOBAL_STEP = 0 def logistic_classifier(inputs): return layers.dense(inputs, 1, activation=math_ops.sigmoid) def local_variable(init_value, name): return variable_scope.get_variable( name, dtype=dtypes.float32, initializer=init_value, trainable=False, collections=[ops.GraphKeys.LOCAL_VARIABLES]) class EvaluateOnceTest(test.TestCase): def setUp(self): super(EvaluateOnceTest, self).setUp() # Create an easy training set: np.random.seed(0) self._inputs = np.zeros((16, 4)) self._labels = np.random.randint(0, 2, size=(16, 1)).astype(np.float32) for i in range(16): j = int(2 * self._labels[i] + np.random.randint(0, 2)) self._inputs[i, j] = 1 def _train_model(self, checkpoint_dir, num_steps): """Trains a simple classification model. Note that the data has been configured such that after around 300 steps, the model has memorized the dataset (e.g. we can expect %100 accuracy). Args: checkpoint_dir: The directory where the checkpoint is written to. num_steps: The number of steps to train for. """ with ops.Graph().as_default(): random_seed.set_random_seed(0) tf_inputs = constant_op.constant(self._inputs, dtype=dtypes.float32) tf_labels = constant_op.constant(self._labels, dtype=dtypes.float32) tf_predictions = logistic_classifier(tf_inputs) loss_op = losses.log_loss(labels=tf_labels, predictions=tf_predictions) optimizer = gradient_descent.GradientDescentOptimizer(learning_rate=1.0) train_op = optimizer.minimize(loss_op, training.get_or_create_global_step()) with monitored_session.MonitoredTrainingSession( checkpoint_dir=checkpoint_dir, hooks=[basic_session_run_hooks.StopAtStepHook(num_steps)]) as session: loss = None while not session.should_stop(): _, loss = session.run([train_op, loss_op]) if num_steps >= 300: assert loss < .015 def testEvaluatePerfectModel(self): checkpoint_dir = os.path.join(self.get_temp_dir(), 'evaluate_perfect_model_once') # Train a Model to completion: self._train_model(checkpoint_dir, num_steps=300) # Run inputs = constant_op.constant(self._inputs, dtype=dtypes.float32) labels = constant_op.constant(self._labels, dtype=dtypes.float32) logits = logistic_classifier(inputs) predictions = math_ops.round(logits) accuracy, update_op = metrics_module.accuracy(labels, predictions) checkpoint_path = saver.latest_checkpoint(checkpoint_dir) final_ops_values = evaluation._evaluate_once( checkpoint_path=checkpoint_path, eval_ops=update_op, final_ops={'accuracy': (accuracy, update_op)}, hooks=[ evaluation._StopAfterNEvalsHook(1), ]) self.assertGreater(final_ops_values['accuracy'], .99) def testEvaluateWithFiniteInputs(self): checkpoint_dir = os.path.join(self.get_temp_dir(), 'evaluate_with_finite_inputs') # Train a Model to completion: self._train_model(checkpoint_dir, num_steps=300) # Run evaluation. Inputs are fed through input producer for one epoch. all_inputs = constant_op.constant(self._inputs, dtype=dtypes.float32) all_labels = constant_op.constant(self._labels, dtype=dtypes.float32) single_input, single_label = training.slice_input_producer( [all_inputs, all_labels], num_epochs=1) inputs, labels = training.batch([single_input, single_label], batch_size=6, allow_smaller_final_batch=True) logits = logistic_classifier(inputs) predictions = math_ops.round(logits) accuracy, update_op = metrics_module.accuracy(labels, predictions) checkpoint_path = saver.latest_checkpoint(checkpoint_dir) final_ops_values = evaluation._evaluate_once( checkpoint_path=checkpoint_path, eval_ops=update_op, final_ops={ 'accuracy': (accuracy, update_op), 'eval_steps': evaluation._get_or_create_eval_step() }, hooks=[ evaluation._StopAfterNEvalsHook(None), ]) self.assertTrue(final_ops_values['accuracy'] > .99) # Runs evaluation for 4 iterations. First 2 evaluate full batch of 6 inputs # each; the 3rd iter evaluates the remaining 4 inputs, and the last one # triggers an error which stops evaluation. self.assertEqual(final_ops_values['eval_steps'], 4) def testEvalOpAndFinalOp(self): checkpoint_dir = os.path.join(self.get_temp_dir(), 'eval_ops_and_final_ops') # Train a model for a single step to get a checkpoint. self._train_model(checkpoint_dir, num_steps=1) checkpoint_path = saver.latest_checkpoint(checkpoint_dir) # Create the model so we have something to restore. inputs = constant_op.constant(self._inputs, dtype=dtypes.float32) logistic_classifier(inputs) num_evals = 5 final_increment = 9.0 my_var = local_variable(0.0, name='MyVar') eval_ops = state_ops.assign_add(my_var, 1.0) final_ops = array_ops.identity(my_var) + final_increment final_hooks = [evaluation._StopAfterNEvalsHook(num_evals),] initial_hooks = list(final_hooks) final_ops_values = evaluation._evaluate_once( checkpoint_path=checkpoint_path, eval_ops=eval_ops, final_ops={'value': final_ops}, hooks=final_hooks) self.assertEqual(final_ops_values['value'], num_evals + final_increment) self.assertEqual(initial_hooks, final_hooks) def testMultiEvalStepIncrements(self): checkpoint_dir = os.path.join(self.get_temp_dir(), 'eval_ops_and_final_ops') # Train a model for a single step to get a checkpoint. self._train_model(checkpoint_dir, num_steps=1) checkpoint_path = saver.latest_checkpoint(checkpoint_dir) # Create the model so we have something to restore. inputs = constant_op.constant(self._inputs, dtype=dtypes.float32) logistic_classifier(inputs) num_evals = 6 my_var = local_variable(0.0, name='MyVar') # In eval ops, we also increase the eval step one more time. eval_ops = [state_ops.assign_add(my_var, 1.0), state_ops.assign_add( evaluation._get_or_create_eval_step(), 1, use_locking=True)] expect_eval_update_counts = num_evals // 2 final_ops = array_ops.identity(my_var) final_ops_values = evaluation._evaluate_once( checkpoint_path=checkpoint_path, eval_ops=eval_ops, final_ops={'value': final_ops}, hooks=[evaluation._StopAfterNEvalsHook(num_evals),]) self.assertEqual(final_ops_values['value'], expect_eval_update_counts) def testOnlyFinalOp(self): checkpoint_dir = os.path.join(self.get_temp_dir(), 'only_final_ops') # Train a model for a single step to get a checkpoint. self._train_model(checkpoint_dir, num_steps=1) checkpoint_path = saver.latest_checkpoint(checkpoint_dir) # Create the model so we have something to restore. inputs = constant_op.constant(self._inputs, dtype=dtypes.float32) logistic_classifier(inputs) final_increment = 9.0 my_var = local_variable(0.0, name='MyVar') final_ops = array_ops.identity(my_var) + final_increment final_ops_values = evaluation._evaluate_once( checkpoint_path=checkpoint_path, final_ops={'value': final_ops}) self.assertEqual(final_ops_values['value'], final_increment) if __name__ == '__main__': test.main()
py
1a2f39b527b64a4f5dc39a9b4f4a86f5fc710033
class Solution: def decodeCiphertext(self, encodedText: str, rows: int) -> str: n = len(encodedText) cols = n // rows ans = [] matrix = [[' '] * cols for _ in range(rows)] for i in range(rows): for j in range(cols): matrix[i][j] = encodedText[i * cols + j] for col in range(cols): i = 0 j = col while i < rows and j < cols: ans.append(matrix[i][j]) i += 1 j += 1 return ''.join(ans).rstrip()
py
1a2f39cc0bde5b4bee29b348df8d7c8b3c876311
import sqlite3 conn = sqlite3.connect('emaildb.sqlite') cur = conn.cursor() cur.execute(''' DROP TABLE IF EXISTS Counts''') cur.execute(''' CREATE TABLE Counts (email TEXT, count INTEGER)''') fname = raw_input('Enter file name: ') if ( len(fname) < 1 ) : fname = 'mbox-short.txt' fh = open(fname) for line in fh: if not line.startswith('From: ') : continue pieces = line.split() email = pieces[1] email = email.split("@",1)[1] dir(email) #print email cur.execute('SELECT count FROM Counts WHERE email = ? ', (email, )) row = cur.fetchone() if row is None: cur.execute('''INSERT INTO Counts (email, count) VALUES ( ?, 1 )''', ( email, ) ) else : cur.execute('UPDATE Counts SET count=count+1 WHERE email = ?', (email, )) # This statement commits outstanding changes to disk each # time through the loop - the program can be made faster # by moving the commit so it runs only after the loop completes conn.commit() # https://www.sqlite.org/lang_select.html sqlstr = 'SELECT email, count FROM Counts ORDER BY count DESC' print "Counts:" for row in cur.execute(sqlstr) : print str(row[0]), row[1] cur.close()
py
1a2f3a32f18a9cbadd9eb5bd3299cb2ea83b186e
import os import re import numpy as np from PIL import Image from keras.models import load_model from keras.preprocessing.image import img_to_array from sklearn.preprocessing import LabelEncoder class ModelManager(): def __init__(self, db, s3, graph, backup_model, backup_label_encoder, collection_name='models', bucket_name='models'): self.conn = db[collection_name] self.s3_bucket = s3.Bucket(bucket_name) self.label_encoder = LabelEncoder() self.model = None self.backup_model = backup_model self.backup_label_encoder = backup_label_encoder self.graph = graph def predict(self, image_array): """Function that accepts a model and image_array and returns a prediction.""" if self.model: with self.graph.as_default(): out = self.model.predict_classes(image_array) return self.label_encoder.inverse_transform(out) with self.graph.as_default(): out = self.backup_model.predict_classes(image_array) return self.backup_label_encoder.inverse_transform(out) def preprocess(self, image_file): """Preprocess the given image file by resizing and turning it into an array.""" img = Image.open(image_file) image_resize = img.resize((40, 24)) image = img_to_array(image_resize) x = image.reshape(1, 40, 24, 1) return x / 255 def load(self, model_path, label_encoder_path): """Load a fitted model from the local filesystem into memory. "./models/classes.npy""" self.label_encoder.classes_ = np.load(label_encoder_path) self.model = load_model(model_path) def load_latest_model(self): """Load a fitted model from a remote filesystem into memory.""" latest_model = list(self.conn.find({'stage': 'PRODUCTION'}).sort([("version", -1)]).limit(1)) if len(latest_model) == 0: return latest_model = latest_model[0] if 'MODEL_NAME' in os.environ and latest_model['name'] == os.environ['MODEL_NAME']: if 'MODEL_VERSION' in os.environ and str(latest_model['version']) == os.environ['MODEL_VERSION']: return for f in os.listdir('./models'): if not f.endswith(".h5") and not f.endswith(".npy"): continue os.remove(os.path.join('./models', f)) # Download model h5 file model = re.sub('name', latest_model['name'], latest_model['naming_format']) model = re.sub('0', str(latest_model['version']), model) self.s3_bucket.download_file(model, f'./models/{model}') # Download model classes npy file classes = re.sub('name', latest_model['name'], latest_model['naming_format']) classes = re.sub('0', str(latest_model['version']), classes) classes = re.sub('h5', 'npy', classes) self.s3_bucket.download_file(classes, f'./models/{classes}') os.environ['MODEL_NAME'] = latest_model['name'] os.environ['MODEL_VERSION'] = str(latest_model['version']) def get_latest_online_model_version(self): initial_model = list(self.conn.find({"name": 'online-model'}).sort([("version", -1)]).limit(1))[0] return initial_model.version if initial_model else 0 def _insert(self, name, version, metrics_str, naming_format, stage): model_doc = { "name": name, "version": version, "naming_format": naming_format, "metrics": metrics_str, "stage": stage } self.conn.insert_one(model_doc) def publish_model(self, model_path, classes_path, name, metrics): latest_model = list(self.conn.find({'stage': 'PRODUCTION', 'name': name}).sort([("version", -1)]).limit(1)) # Save model file to AWS S3 bucket version = 1 if len(latest_model) > 0: version = latest_model[0].version + 1 self.s3_bucket.upload_file(Filename=model_path, Key=f'{name}-v{version}.h5') self.s3_bucket.upload_file(Filename=classes_path, Key=f'{name}-v{version}.npy') # Create and save a new model doc with related information self._insert(name, version, metrics, 'name-v0.h5', 'PRODUCTION')
py
1a2f3a96bacadd20f43b044dc038ace0d13d6a37
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.async_support.base.exchange import Exchange # ----------------------------------------------------------------------------- try: basestring # Python 3 except NameError: basestring = str # Python 2 import hashlib from ccxt.base.errors import ExchangeError from ccxt.base.errors import AuthenticationError from ccxt.base.errors import PermissionDenied from ccxt.base.errors import ArgumentsRequired from ccxt.base.errors import InsufficientFunds from ccxt.base.errors import InvalidAddress from ccxt.base.errors import InvalidOrder from ccxt.base.errors import OrderNotFound from ccxt.base.errors import NotSupported from ccxt.base.errors import RateLimitExceeded from ccxt.base.errors import OnMaintenance from ccxt.base.decimal_to_precision import TICK_SIZE from ccxt.base.precise import Precise class coinbasepro(Exchange): def describe(self): return self.deep_extend(super(coinbasepro, self).describe(), { 'id': 'coinbasepro', 'name': 'Coinbase Pro', 'countries': ['US'], 'rateLimit': 1000, 'userAgent': self.userAgents['chrome'], 'pro': True, 'has': { 'cancelAllOrders': True, 'cancelOrder': True, 'CORS': True, 'createDepositAddress': True, 'createOrder': True, 'deposit': True, 'fetchAccounts': True, 'fetchBalance': True, 'fetchCurrencies': True, 'fetchClosedOrders': True, 'fetchDepositAddress': False, # the exchange does not have self method, only createDepositAddress, see https://github.com/ccxt/ccxt/pull/7405 'fetchMarkets': True, 'fetchMyTrades': True, 'fetchOHLCV': True, 'fetchOpenOrders': True, 'fetchOrder': True, 'fetchOrderBook': True, 'fetchOrders': True, 'fetchOrderTrades': True, 'fetchTime': True, 'fetchTicker': True, 'fetchTrades': True, 'fetchTransactions': True, 'withdraw': True, 'fetchDeposits': True, 'fetchWithdrawals': True, }, 'timeframes': { '1m': 60, '5m': 300, '15m': 900, '1h': 3600, '6h': 21600, '1d': 86400, }, 'hostname': 'pro.coinbase.com', 'urls': { 'test': { 'public': 'https://api-public.sandbox.pro.coinbase.com', 'private': 'https://api-public.sandbox.pro.coinbase.com', }, 'logo': 'https://user-images.githubusercontent.com/1294454/41764625-63b7ffde-760a-11e8-996d-a6328fa9347a.jpg', 'api': { 'public': 'https://api.{hostname}', 'private': 'https://api.{hostname}', }, 'www': 'https://pro.coinbase.com/', 'doc': 'https://docs.pro.coinbase.com', 'fees': [ 'https://docs.pro.coinbase.com/#fees', 'https://support.pro.coinbase.com/customer/en/portal/articles/2945310-fees', ], }, 'requiredCredentials': { 'apiKey': True, 'secret': True, 'password': True, }, 'api': { 'public': { 'get': [ 'currencies', 'products', 'products/{id}', 'products/{id}/book', 'products/{id}/candles', 'products/{id}/stats', 'products/{id}/ticker', 'products/{id}/trades', 'time', ], }, 'private': { 'get': [ 'accounts', 'accounts/{id}', 'accounts/{id}/holds', 'accounts/{id}/ledger', 'accounts/{id}/transfers', 'coinbase-accounts', 'fills', 'funding', 'fees', 'margin/profile_information', 'margin/buying_power', 'margin/withdrawal_power', 'margin/withdrawal_power_all', 'margin/exit_plan', 'margin/liquidation_history', 'margin/position_refresh_amounts', 'margin/status', 'oracle', 'orders', 'orders/{id}', 'orders/client:{client_oid}', 'otc/orders', 'payment-methods', 'position', 'profiles', 'profiles/{id}', 'reports/{report_id}', 'transfers', 'transfers/{transfer_id}', 'users/self/exchange-limits', 'users/self/hold-balances', 'users/self/trailing-volume', 'withdrawals/fee-estimate', ], 'post': [ 'conversions', 'deposits/coinbase-account', 'deposits/payment-method', 'coinbase-accounts/{id}/addresses', 'funding/repay', 'orders', 'position/close', 'profiles/margin-transfer', 'profiles/transfer', 'reports', 'withdrawals/coinbase', 'withdrawals/coinbase-account', 'withdrawals/crypto', 'withdrawals/payment-method', ], 'delete': [ 'orders', 'orders/client:{client_oid}', 'orders/{id}', ], }, }, 'commonCurrencies': { 'CGLD': 'CELO', }, 'precisionMode': TICK_SIZE, 'fees': { 'trading': { 'tierBased': True, # complicated tier system per coin 'percentage': True, 'maker': 0.5 / 100, # highest fee of all tiers 'taker': 0.5 / 100, # highest fee of all tiers }, 'funding': { 'tierBased': False, 'percentage': False, 'withdraw': { 'BCH': 0, 'BTC': 0, 'LTC': 0, 'ETH': 0, 'EUR': 0.15, 'USD': 25, }, 'deposit': { 'BCH': 0, 'BTC': 0, 'LTC': 0, 'ETH': 0, 'EUR': 0.15, 'USD': 10, }, }, }, 'exceptions': { 'exact': { 'Insufficient funds': InsufficientFunds, 'NotFound': OrderNotFound, 'Invalid API Key': AuthenticationError, 'invalid signature': AuthenticationError, 'Invalid Passphrase': AuthenticationError, 'Invalid order id': InvalidOrder, 'Private rate limit exceeded': RateLimitExceeded, 'Trading pair not available': PermissionDenied, 'Product not found': InvalidOrder, }, 'broad': { 'Order already done': OrderNotFound, 'order not found': OrderNotFound, 'price too small': InvalidOrder, 'price too precise': InvalidOrder, 'under maintenance': OnMaintenance, 'size is too small': InvalidOrder, 'Cancel only mode': OnMaintenance, # https://github.com/ccxt/ccxt/issues/7690 }, }, }) async def fetch_currencies(self, params={}): response = await self.publicGetCurrencies(params) # # [ # { # id: 'XTZ', # name: 'Tezos', # min_size: '0.000001', # status: 'online', # message: '', # max_precision: '0.000001', # convertible_to: [], # details: { # type: 'crypto', # symbol: 'Τ', # network_confirmations: 60, # sort_order: 53, # crypto_address_link: 'https://tzstats.com/{{address}}', # crypto_transaction_link: 'https://tzstats.com/{{txId}}', # push_payment_methods: ['crypto'], # group_types: [], # display_name: '', # processing_time_seconds: 0, # min_withdrawal_amount: 1 # } # } # ] # result = {} for i in range(0, len(response)): currency = response[i] id = self.safe_string(currency, 'id') name = self.safe_string(currency, 'name') code = self.safe_currency_code(id) details = self.safe_value(currency, 'details', {}) precision = self.safe_number(currency, 'max_precision') status = self.safe_string(currency, 'status') active = (status == 'online') result[code] = { 'id': id, 'code': code, 'info': currency, 'type': self.safe_string(details, 'type'), 'name': name, 'active': active, 'fee': None, 'precision': precision, 'limits': { 'amount': { 'min': self.safe_number(details, 'min_size'), 'max': None, }, 'withdraw': { 'min': self.safe_number(details, 'min_withdrawal_amount'), 'max': None, }, }, } return result async def fetch_markets(self, params={}): response = await self.publicGetProducts(params) # # [ # { # "id":"ZEC-BTC", # "base_currency":"ZEC", # "quote_currency":"BTC", # "base_min_size":"0.01000000", # "base_max_size":"1500.00000000", # "quote_increment":"0.00000100", # "base_increment":"0.00010000", # "display_name":"ZEC/BTC", # "min_market_funds":"0.001", # "max_market_funds":"30", # "margin_enabled":false, # "post_only":false, # "limit_only":false, # "cancel_only":false, # "trading_disabled":false, # "status":"online", # "status_message":"" # } # ] # result = [] for i in range(0, len(response)): market = response[i] id = self.safe_string(market, 'id') baseId = self.safe_string(market, 'base_currency') quoteId = self.safe_string(market, 'quote_currency') base = self.safe_currency_code(baseId) quote = self.safe_currency_code(quoteId) symbol = base + '/' + quote priceLimits = { 'min': self.safe_number(market, 'quote_increment'), 'max': None, } precision = { 'amount': self.safe_number(market, 'base_increment'), 'price': self.safe_number(market, 'quote_increment'), } status = self.safe_string(market, 'status') active = (status == 'online') result.append(self.extend(self.fees['trading'], { 'id': id, 'symbol': symbol, 'baseId': baseId, 'quoteId': quoteId, 'base': base, 'quote': quote, 'precision': precision, 'limits': { 'amount': { 'min': self.safe_number(market, 'base_min_size'), 'max': self.safe_number(market, 'base_max_size'), }, 'price': priceLimits, 'cost': { 'min': self.safe_number(market, 'min_market_funds'), 'max': self.safe_number(market, 'max_market_funds'), }, }, 'active': active, 'info': market, })) return result async def fetch_accounts(self, params={}): await self.load_markets() response = await self.privateGetAccounts(params) # # [ # { # id: '4aac9c60-cbda-4396-9da4-4aa71e95fba0', # currency: 'BTC', # balance: '0.0000000000000000', # available: '0', # hold: '0.0000000000000000', # profile_id: 'b709263e-f42a-4c7d-949a-a95c83d065da' # }, # { # id: 'f75fa69a-1ad1-4a80-bd61-ee7faa6135a3', # currency: 'USDC', # balance: '0.0000000000000000', # available: '0', # hold: '0.0000000000000000', # profile_id: 'b709263e-f42a-4c7d-949a-a95c83d065da' # }, # ] # result = [] for i in range(0, len(response)): account = response[i] accountId = self.safe_string(account, 'id') currencyId = self.safe_string(account, 'currency') code = self.safe_currency_code(currencyId) result.append({ 'id': accountId, 'type': None, 'currency': code, 'info': account, }) return result async def fetch_balance(self, params={}): await self.load_markets() response = await self.privateGetAccounts(params) result = {'info': response} for i in range(0, len(response)): balance = response[i] currencyId = self.safe_string(balance, 'currency') code = self.safe_currency_code(currencyId) account = self.account() account['free'] = self.safe_string(balance, 'available') account['used'] = self.safe_string(balance, 'hold') account['total'] = self.safe_string(balance, 'balance') result[code] = account return self.parse_balance(result) async def fetch_order_book(self, symbol, limit=None, params={}): await self.load_markets() # level 1 - only the best bid and ask # level 2 - top 50 bids and asks(aggregated) # level 3 - full order book(non aggregated) request = { 'id': self.market_id(symbol), 'level': 2, # 1 best bidask, 2 aggregated, 3 full } response = await self.publicGetProductsIdBook(self.extend(request, params)) # # { # "sequence":1924393896, # "bids":[ # ["0.01825","24.34811287",2], # ["0.01824","72.5463",3], # ["0.01823","424.54298049",6], # ], # "asks":[ # ["0.01826","171.10414904",4], # ["0.01827","22.60427028",1], # ["0.01828","397.46018784",7], # ] # } # orderbook = self.parse_order_book(response, symbol) orderbook['nonce'] = self.safe_integer(response, 'sequence') return orderbook def parse_ticker(self, ticker, market=None): # # publicGetProductsIdTicker # # { # "trade_id":843439, # "price":"0.997999", # "size":"80.29769", # "time":"2020-01-28T02:13:33.012523Z", # "bid":"0.997094", # "ask":"0.998", # "volume":"1903188.03750000" # } # # publicGetProductsIdStats # # { # "open": "34.19000000", # "high": "95.70000000", # "low": "7.06000000", # "volume": "2.41000000" # } # timestamp = self.parse8601(self.safe_value(ticker, 'time')) bid = self.safe_number(ticker, 'bid') ask = self.safe_number(ticker, 'ask') last = self.safe_number(ticker, 'price') symbol = None if (market is None) else market['symbol'] return { 'symbol': symbol, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'high': self.safe_number(ticker, 'high'), 'low': self.safe_number(ticker, 'low'), 'bid': bid, 'bidVolume': None, 'ask': ask, 'askVolume': None, 'vwap': None, 'open': self.safe_number(ticker, 'open'), 'close': last, 'last': last, 'previousClose': None, 'change': None, 'percentage': None, 'average': None, 'baseVolume': self.safe_number(ticker, 'volume'), 'quoteVolume': None, 'info': ticker, } async def fetch_ticker(self, symbol, params={}): await self.load_markets() market = self.market(symbol) request = { 'id': market['id'], } # publicGetProductsIdTicker or publicGetProductsIdStats method = self.safe_string(self.options, 'fetchTickerMethod', 'publicGetProductsIdTicker') response = await getattr(self, method)(self.extend(request, params)) # # publicGetProductsIdTicker # # { # "trade_id":843439, # "price":"0.997999", # "size":"80.29769", # "time":"2020-01-28T02:13:33.012523Z", # "bid":"0.997094", # "ask":"0.998", # "volume":"1903188.03750000" # } # # publicGetProductsIdStats # # { # "open": "34.19000000", # "high": "95.70000000", # "low": "7.06000000", # "volume": "2.41000000" # } # return self.parse_ticker(response, market) def parse_trade(self, trade, market=None): # # { # type: 'match', # trade_id: 82047307, # maker_order_id: '0f358725-2134-435e-be11-753912a326e0', # taker_order_id: '252b7002-87a3-425c-ac73-f5b9e23f3caf', # order_id: 'd50ec984-77a8-460a-b958-66f114b0de9b', # side: 'sell', # size: '0.00513192', # price: '9314.78', # product_id: 'BTC-USD', # profile_id: '6244401d-c078-40d9-b305-7ad3551bc3b0', # sequence: 12038915443, # time: '2020-01-31T20:03:41.158814Z' # created_at: '2014-11-07T22:19:28.578544Z', # liquidity: 'T', # fee: '0.00025', # settled: True, # usd_volume: '0.0924556000000000', # user_id: '595eb864313c2b02ddf2937d' # } # timestamp = self.parse8601(self.safe_string_2(trade, 'time', 'created_at')) marketId = self.safe_string(trade, 'product_id') symbol = self.safe_symbol(marketId, market, '-') feeRate = None feeCurrency = None takerOrMaker = None cost = None if market is not None: feeCurrencyId = self.safe_string_lower(market, 'quoteId') costField = feeCurrencyId + '_value' cost = self.safe_number(trade, costField) feeCurrency = market['quote'] liquidity = self.safe_string(trade, 'liquidity') if liquidity is not None: takerOrMaker = 'taker' if (liquidity == 'T') else 'maker' feeRate = market[takerOrMaker] feeCost = self.safe_number_2(trade, 'fill_fees', 'fee') fee = { 'cost': feeCost, 'currency': feeCurrency, 'rate': feeRate, } type = None id = self.safe_string(trade, 'trade_id') side = 'sell' if (trade['side'] == 'buy') else 'buy' orderId = self.safe_string(trade, 'order_id') # Coinbase Pro returns inverted side to fetchMyTrades vs fetchTrades makerOrderId = self.safe_string(trade, 'maker_order_id') takerOrderId = self.safe_string(trade, 'taker_order_id') if (orderId is not None) or ((makerOrderId is not None) and (takerOrderId is not None)): side = 'buy' if (trade['side'] == 'buy') else 'sell' priceString = self.safe_string(trade, 'price') amountString = self.safe_string(trade, 'size') price = self.parse_number(priceString) amount = self.parse_number(amountString) if cost is None: cost = self.parse_number(Precise.string_mul(priceString, amountString)) return { 'id': id, 'order': orderId, 'info': trade, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'symbol': symbol, 'type': type, 'takerOrMaker': takerOrMaker, 'side': side, 'price': price, 'amount': amount, 'fee': fee, 'cost': cost, } async def fetch_my_trades(self, symbol=None, since=None, limit=None, params={}): # as of 2018-08-23 if symbol is None: raise ArgumentsRequired(self.id + ' fetchMyTrades() requires a symbol argument') await self.load_markets() market = self.market(symbol) request = { 'product_id': market['id'], } if limit is not None: request['limit'] = limit response = await self.privateGetFills(self.extend(request, params)) return self.parse_trades(response, market, since, limit) async def fetch_trades(self, symbol, since=None, limit=None, params={}): await self.load_markets() market = self.market(symbol) request = { 'id': market['id'], # fixes issue #2 } if limit is not None: request['limit'] = limit # default 100 response = await self.publicGetProductsIdTrades(self.extend(request, params)) return self.parse_trades(response, market, since, limit) def parse_ohlcv(self, ohlcv, market=None): # # [ # 1591514160, # 0.02507, # 0.02507, # 0.02507, # 0.02507, # 0.02816506 # ] # return [ self.safe_timestamp(ohlcv, 0), self.safe_number(ohlcv, 3), self.safe_number(ohlcv, 2), self.safe_number(ohlcv, 1), self.safe_number(ohlcv, 4), self.safe_number(ohlcv, 5), ] async def fetch_ohlcv(self, symbol, timeframe='1m', since=None, limit=None, params={}): await self.load_markets() market = self.market(symbol) granularity = self.timeframes[timeframe] request = { 'id': market['id'], 'granularity': granularity, } if since is not None: request['start'] = self.iso8601(since) if limit is None: # https://docs.pro.coinbase.com/#get-historic-rates limit = 300 # max = 300 else: limit = min(300, limit) request['end'] = self.iso8601(self.sum((limit - 1) * granularity * 1000, since)) response = await self.publicGetProductsIdCandles(self.extend(request, params)) # # [ # [1591514160,0.02507,0.02507,0.02507,0.02507,0.02816506], # [1591514100,0.02507,0.02507,0.02507,0.02507,1.63830323], # [1591514040,0.02505,0.02507,0.02505,0.02507,0.19918178] # ] # return self.parse_ohlcvs(response, market, timeframe, since, limit) async def fetch_time(self, params={}): response = await self.publicGetTime(params) # # { # "iso":"2020-05-12T08:00:51.504Z", # "epoch":1589270451.504 # } # return self.safe_timestamp(response, 'epoch') def parse_order_status(self, status): statuses = { 'pending': 'open', 'active': 'open', 'open': 'open', 'done': 'closed', 'canceled': 'canceled', 'canceling': 'open', } return self.safe_string(statuses, status, status) def parse_order(self, order, market=None): # # createOrder # # { # "id": "d0c5340b-6d6c-49d9-b567-48c4bfca13d2", # "price": "0.10000000", # "size": "0.01000000", # "product_id": "BTC-USD", # "side": "buy", # "stp": "dc", # "type": "limit", # "time_in_force": "GTC", # "post_only": False, # "created_at": "2016-12-08T20:02:28.53864Z", # "fill_fees": "0.0000000000000000", # "filled_size": "0.00000000", # "executed_value": "0.0000000000000000", # "status": "pending", # "settled": False # } # timestamp = self.parse8601(self.safe_string(order, 'created_at')) marketId = self.safe_string(order, 'product_id') market = self.safe_market(marketId, market, '-') status = self.parse_order_status(self.safe_string(order, 'status')) price = self.safe_number(order, 'price') filled = self.safe_number(order, 'filled_size') amount = self.safe_number(order, 'size', filled) cost = self.safe_number(order, 'executed_value') feeCost = self.safe_number(order, 'fill_fees') fee = None if feeCost is not None: feeCurrencyCode = None if market is not None: feeCurrencyCode = market['quote'] fee = { 'cost': feeCost, 'currency': feeCurrencyCode, 'rate': None, } id = self.safe_string(order, 'id') type = self.safe_string(order, 'type') side = self.safe_string(order, 'side') timeInForce = self.safe_string(order, 'time_in_force') postOnly = self.safe_value(order, 'post_only') stopPrice = self.safe_number(order, 'stop_price') clientOrderId = self.safe_string(order, 'client_oid') return self.safe_order({ 'id': id, 'clientOrderId': clientOrderId, 'info': order, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'lastTradeTimestamp': None, 'status': status, 'symbol': market['symbol'], 'type': type, 'timeInForce': timeInForce, 'postOnly': postOnly, 'side': side, 'price': price, 'stopPrice': stopPrice, 'cost': cost, 'amount': amount, 'filled': filled, 'remaining': None, 'fee': fee, 'average': None, 'trades': None, }) async def fetch_order(self, id, symbol=None, params={}): await self.load_markets() request = {} clientOrderId = self.safe_string_2(params, 'clientOrderId', 'client_oid') method = None if clientOrderId is None: method = 'privateGetOrdersId' request['id'] = id else: method = 'privateGetOrdersClientClientOid' request['client_oid'] = clientOrderId params = self.omit(params, ['clientOrderId', 'client_oid']) response = await getattr(self, method)(self.extend(request, params)) return self.parse_order(response) async def fetch_order_trades(self, id, symbol=None, since=None, limit=None, params={}): await self.load_markets() market = None if symbol is not None: market = self.market(symbol) request = { 'order_id': id, } response = await self.privateGetFills(self.extend(request, params)) return self.parse_trades(response, market, since, limit) async def fetch_orders(self, symbol=None, since=None, limit=None, params={}): request = { 'status': 'all', } return await self.fetch_open_orders(symbol, since, limit, self.extend(request, params)) async def fetch_open_orders(self, symbol=None, since=None, limit=None, params={}): await self.load_markets() request = {} market = None if symbol is not None: market = self.market(symbol) request['product_id'] = market['id'] if limit is not None: request['limit'] = limit # default 100 response = await self.privateGetOrders(self.extend(request, params)) return self.parse_orders(response, market, since, limit) async def fetch_closed_orders(self, symbol=None, since=None, limit=None, params={}): request = { 'status': 'done', } return await self.fetch_open_orders(symbol, since, limit, self.extend(request, params)) async def create_order(self, symbol, type, side, amount, price=None, params={}): await self.load_markets() market = self.market(symbol) request = { # common params -------------------------------------------------- # 'client_oid': clientOrderId, 'type': type, 'side': side, 'product_id': market['id'], # 'size': self.amount_to_precision(symbol, amount), # 'stp': 'dc', # self-trade prevention, dc = decrease and cancel, co = cancel oldest, cn = cancel newest, cb = cancel both # 'stop': 'loss', # "loss" = stop loss below price, "entry" = take profit above price # 'stop_price': self.price_to_precision(symbol, price), # limit order params --------------------------------------------- # 'price': self.price_to_precision(symbol, price), # 'size': self.amount_to_precision(symbol, amount), # 'time_in_force': 'GTC', # GTC, GTT, IOC, or FOK # 'cancel_after' [optional]* min, hour, day, requires time_in_force to be GTT # 'post_only': False, # invalid when time_in_force is IOC or FOK # market order params -------------------------------------------- # 'size': self.amount_to_precision(symbol, amount), # 'funds': self.cost_to_precision(symbol, amount), } clientOrderId = self.safe_string_2(params, 'clientOrderId', 'client_oid') if clientOrderId is not None: request['client_oid'] = clientOrderId params = self.omit(params, ['clientOrderId', 'client_oid']) stopPrice = self.safe_number_2(params, 'stopPrice', 'stop_price') if stopPrice is not None: request['stop_price'] = self.price_to_precision(symbol, stopPrice) params = self.omit(params, ['stopPrice', 'stop_price']) timeInForce = self.safe_string_2(params, 'timeInForce', 'time_in_force') if timeInForce is not None: request['time_in_force'] = timeInForce params = self.omit(params, ['timeInForce', 'time_in_force']) if type == 'limit': request['price'] = self.price_to_precision(symbol, price) request['size'] = self.amount_to_precision(symbol, amount) elif type == 'market': cost = self.safe_number_2(params, 'cost', 'funds') if cost is None: if price is not None: cost = amount * price else: params = self.omit(params, ['cost', 'funds']) if cost is not None: request['funds'] = self.cost_to_precision(symbol, cost) else: request['size'] = self.amount_to_precision(symbol, amount) response = await self.privatePostOrders(self.extend(request, params)) # # { # "id": "d0c5340b-6d6c-49d9-b567-48c4bfca13d2", # "price": "0.10000000", # "size": "0.01000000", # "product_id": "BTC-USD", # "side": "buy", # "stp": "dc", # "type": "limit", # "time_in_force": "GTC", # "post_only": False, # "created_at": "2016-12-08T20:02:28.53864Z", # "fill_fees": "0.0000000000000000", # "filled_size": "0.00000000", # "executed_value": "0.0000000000000000", # "status": "pending", # "settled": False # } # return self.parse_order(response, market) async def cancel_order(self, id, symbol=None, params={}): await self.load_markets() request = { # 'product_id': market['id'], # the request will be more performant if you include it } clientOrderId = self.safe_string_2(params, 'clientOrderId', 'client_oid') method = None if clientOrderId is None: method = 'privateDeleteOrdersId' request['id'] = id else: method = 'privateDeleteOrdersClientClientOid' request['client_oid'] = clientOrderId params = self.omit(params, ['clientOrderId', 'client_oid']) market = None if symbol is not None: market = self.market(symbol) request['product_id'] = market['symbol'] # the request will be more performant if you include it return await getattr(self, method)(self.extend(request, params)) async def cancel_all_orders(self, symbol=None, params={}): await self.load_markets() request = {} market = None if symbol is not None: market = self.market(symbol) request['product_id'] = market['symbol'] # the request will be more performant if you include it return await self.privateDeleteOrders(self.extend(request, params)) async def fetch_payment_methods(self, params={}): return await self.privateGetPaymentMethods(params) async def deposit(self, code, amount, address, params={}): await self.load_markets() currency = self.currency(code) request = { 'currency': currency['id'], 'amount': amount, } method = 'privatePostDeposits' if 'payment_method_id' in params: # deposit from a payment_method, like a bank account method += 'PaymentMethod' elif 'coinbase_account_id' in params: # deposit into Coinbase Pro account from a Coinbase account method += 'CoinbaseAccount' else: # deposit methodotherwise we did not receive a supported deposit location # relevant docs link for the Googlers # https://docs.pro.coinbase.com/#deposits raise NotSupported(self.id + ' deposit() requires one of `coinbase_account_id` or `payment_method_id` extra params') response = await getattr(self, method)(self.extend(request, params)) if not response: raise ExchangeError(self.id + ' deposit() error: ' + self.json(response)) return { 'info': response, 'id': response['id'], } async def withdraw(self, code, amount, address, tag=None, params={}): self.check_address(address) await self.load_markets() currency = self.currency(code) request = { 'currency': currency['id'], 'amount': amount, } method = 'privatePostWithdrawals' if 'payment_method_id' in params: method += 'PaymentMethod' elif 'coinbase_account_id' in params: method += 'CoinbaseAccount' else: method += 'Crypto' request['crypto_address'] = address if tag is not None: request['destination_tag'] = tag response = await getattr(self, method)(self.extend(request, params)) if not response: raise ExchangeError(self.id + ' withdraw() error: ' + self.json(response)) return { 'info': response, 'id': response['id'], } async def fetch_transactions(self, code=None, since=None, limit=None, params={}): await self.load_markets() await self.load_accounts() currency = None id = self.safe_string(params, 'id') # account id if id is None: if code is not None: currency = self.currency(code) accountsByCurrencyCode = self.index_by(self.accounts, 'currency') account = self.safe_value(accountsByCurrencyCode, code) if account is None: raise ExchangeError(self.id + ' fetchTransactions() could not find account id for ' + code) id = account['id'] request = {} if id is not None: request['id'] = id if limit is not None: request['limit'] = limit response = None if id is None: response = await self.privateGetTransfers(self.extend(request, params)) for i in range(0, len(response)): account_id = self.safe_string(response[i], 'account_id') account = self.safe_value(self.accountsById, account_id) code = self.safe_string(account, 'currency') response[i]['currency'] = code else: response = await self.privateGetAccountsIdTransfers(self.extend(request, params)) for i in range(0, len(response)): response[i]['currency'] = code return self.parse_transactions(response, currency, since, limit) async def fetch_deposits(self, code=None, since=None, limit=None, params={}): return self.fetch_transactions(code, since, limit, self.extend({'type': 'deposit'}, params)) async def fetch_withdrawals(self, code=None, since=None, limit=None, params={}): return self.fetch_transactions(code, since, limit, self.extend({'type': 'withdraw'}, params)) def parse_transaction_status(self, transaction): canceled = self.safe_value(transaction, 'canceled_at') if canceled: return 'canceled' processed = self.safe_value(transaction, 'processed_at') completed = self.safe_value(transaction, 'completed_at') if completed: return 'ok' elif processed and not completed: return 'failed' else: return 'pending' def parse_transaction(self, transaction, currency=None): details = self.safe_value(transaction, 'details', {}) id = self.safe_string(transaction, 'id') txid = self.safe_string(details, 'crypto_transaction_hash') timestamp = self.parse8601(self.safe_string(transaction, 'created_at')) updated = self.parse8601(self.safe_string(transaction, 'processed_at')) currencyId = self.safe_string(transaction, 'currency') code = self.safe_currency_code(currencyId, currency) status = self.parse_transaction_status(transaction) amount = self.safe_number(transaction, 'amount') type = self.safe_string(transaction, 'type') address = self.safe_string(details, 'crypto_address') tag = self.safe_string(details, 'destination_tag') address = self.safe_string(transaction, 'crypto_address', address) fee = None if type == 'withdraw': type = 'withdrawal' address = self.safe_string(details, 'sent_to_address', address) feeCost = self.safe_number(details, 'fee') if feeCost is not None: if amount is not None: amount -= feeCost fee = { 'cost': feeCost, 'currency': code, } return { 'info': transaction, 'id': id, 'txid': txid, 'timestamp': timestamp, 'datetime': self.iso8601(timestamp), 'address': address, 'tag': tag, 'type': type, 'amount': amount, 'currency': code, 'status': status, 'updated': updated, 'fee': fee, } async def create_deposit_address(self, code, params={}): await self.load_markets() currency = self.currency(code) accounts = self.safe_value(self.options, 'coinbaseAccounts') if accounts is None: accounts = await self.privateGetCoinbaseAccounts() self.options['coinbaseAccounts'] = accounts # cache it self.options['coinbaseAccountsByCurrencyId'] = self.index_by(accounts, 'currency') currencyId = currency['id'] account = self.safe_value(self.options['coinbaseAccountsByCurrencyId'], currencyId) if account is None: # eslint-disable-next-line quotes raise InvalidAddress(self.id + " fetchDepositAddress() could not find currency code " + code + " with id = " + currencyId + " in self.options['coinbaseAccountsByCurrencyId']") request = { 'id': account['id'], } response = await self.privatePostCoinbaseAccountsIdAddresses(self.extend(request, params)) address = self.safe_string(response, 'address') tag = self.safe_string(response, 'destination_tag') return { 'currency': code, 'address': self.check_address(address), 'tag': tag, 'info': response, } def sign(self, path, api='public', method='GET', params={}, headers=None, body=None): request = '/' + self.implode_params(path, params) query = self.omit(params, self.extract_params(path)) if method == 'GET': if query: request += '?' + self.urlencode(query) url = self.implode_hostname(self.urls['api'][api]) + request if api == 'private': self.check_required_credentials() nonce = str(self.nonce()) payload = '' if method != 'GET': if query: body = self.json(query) payload = body what = nonce + method + request + payload secret = self.base64_to_binary(self.secret) signature = self.hmac(self.encode(what), secret, hashlib.sha256, 'base64') headers = { 'CB-ACCESS-KEY': self.apiKey, 'CB-ACCESS-SIGN': signature, 'CB-ACCESS-TIMESTAMP': nonce, 'CB-ACCESS-PASSPHRASE': self.password, 'Content-Type': 'application/json', } return {'url': url, 'method': method, 'body': body, 'headers': headers} def handle_errors(self, code, reason, url, method, headers, body, response, requestHeaders, requestBody): if (code == 400) or (code == 404): if body[0] == '{': message = self.safe_string(response, 'message') feedback = self.id + ' ' + message self.throw_exactly_matched_exception(self.exceptions['exact'], message, feedback) self.throw_broadly_matched_exception(self.exceptions['broad'], message, feedback) raise ExchangeError(feedback) # unknown message raise ExchangeError(self.id + ' ' + body) async def request(self, path, api='public', method='GET', params={}, headers=None, body=None): response = await self.fetch2(path, api, method, params, headers, body) if not isinstance(response, basestring): if 'message' in response: raise ExchangeError(self.id + ' ' + self.json(response)) return response
py
1a2f3b88bb85969cdd684698c37db040612c9d5d
# -*- coding: utf-8 -*- # Generated by the protocol buffer compiler. DO NOT EDIT! # source: account.proto import sys _b=sys.version_info[0]<3 and (lambda x:x) or (lambda x:x.encode('latin1')) from google.protobuf import descriptor as _descriptor from google.protobuf import message as _message from google.protobuf import reflection as _reflection from google.protobuf import symbol_database as _symbol_database # @@protoc_insertion_point(imports) _sym_db = _symbol_database.Default() DESCRIPTOR = _descriptor.FileDescriptor( name='account.proto', package='types', syntax='proto3', serialized_options=None, serialized_pb=_b('\n\raccount.proto\x12\x05types\"\x1a\n\x07\x41\x63\x63ount\x12\x0f\n\x07\x61\x64\x64ress\x18\x01 \x01(\x0c\"/\n\x0b\x41\x63\x63ountList\x12 \n\x08\x61\x63\x63ounts\x18\x01 \x03(\x0b\x32\x0e.types.Accountb\x06proto3') ) _ACCOUNT = _descriptor.Descriptor( name='Account', full_name='types.Account', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='address', full_name='types.Account.address', index=0, number=1, type=12, cpp_type=9, label=1, has_default_value=False, default_value=_b(""), message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=24, serialized_end=50, ) _ACCOUNTLIST = _descriptor.Descriptor( name='AccountList', full_name='types.AccountList', filename=None, file=DESCRIPTOR, containing_type=None, fields=[ _descriptor.FieldDescriptor( name='accounts', full_name='types.AccountList.accounts', index=0, number=1, type=11, cpp_type=10, label=3, has_default_value=False, default_value=[], message_type=None, enum_type=None, containing_type=None, is_extension=False, extension_scope=None, serialized_options=None, file=DESCRIPTOR), ], extensions=[ ], nested_types=[], enum_types=[ ], serialized_options=None, is_extendable=False, syntax='proto3', extension_ranges=[], oneofs=[ ], serialized_start=52, serialized_end=99, ) _ACCOUNTLIST.fields_by_name['accounts'].message_type = _ACCOUNT DESCRIPTOR.message_types_by_name['Account'] = _ACCOUNT DESCRIPTOR.message_types_by_name['AccountList'] = _ACCOUNTLIST _sym_db.RegisterFileDescriptor(DESCRIPTOR) Account = _reflection.GeneratedProtocolMessageType('Account', (_message.Message,), { 'DESCRIPTOR' : _ACCOUNT, '__module__' : 'account_pb2' # @@protoc_insertion_point(class_scope:types.Account) }) _sym_db.RegisterMessage(Account) AccountList = _reflection.GeneratedProtocolMessageType('AccountList', (_message.Message,), { 'DESCRIPTOR' : _ACCOUNTLIST, '__module__' : 'account_pb2' # @@protoc_insertion_point(class_scope:types.AccountList) }) _sym_db.RegisterMessage(AccountList) # @@protoc_insertion_point(module_scope)
py
1a2f3b9c93e7776664f9189707036a7efce3ea75
from event_chain.app import controllers from event_chain.app import utils from event_chain.app.forms.admin import RegisterForm from flask import Blueprint from flask import g from flask import redirect from flask import render_template from flask import session from flask import url_for, flash admin = Blueprint( 'admin', __name__ ) @admin.route("/login", methods=['GET', 'POST']) def login(): form = RegisterForm() if form.validate_on_submit(): user = controllers.load_user( passphrase=form.passphrase.data, address=form.address.data, ) session['user'] = user if user: return redirect(url_for('events.all')) else: flash("Passphrase and address don't match!", 'error') return redirect(url_for('admin.login')) else: utils.flash_errors(form) return render_template('admin/login.html', form=form) @admin.route("/register", methods=['GET', 'POST']) def register(): form = RegisterForm() if form.validate_on_submit(): user = controllers.register_user( form.name.data, form.passphrase.data ) session['user'] = user g.logger.debug('New User registered!') else: utils.flash_errors(form) return render_template('admin/login.html', form=form) return redirect(url_for('events.all')) @admin.route("/logout", methods=['GET', 'POST']) def logout(): session['user'] = None return redirect(url_for('admin.login'))
py
1a2f3c714a578c0dace55275e4327436d1f9416e
import numpy as np import tensorflow as tf from mlagents.envs import UnityEnvironment initKernelAndBias={ 'kernel_initializer' : tf.random_normal_initializer(0., .1), 'bias_initializer' : tf.constant_initializer(0.1) } class Actor(object): def __init__(self, sess, observationDim, actionDim, learning_rate=0.001, update_frequency=10): self.sess=sess self.s=tf.placeholder(tf.float32, [1,observationDim],"state") self.a=tf.placeholder(tf.float32, [1,actionDim],"action") self.advantage=tf.placeholder(tf.float32,[1,1],"advantage") self.update_frequency=update_frequency with tf.variable_scope("ActorMain"): layer1 = tf.layers.dense( inputs=self.s, units=20, activation=tf.nn.relu, name='layer1', **initKernelAndBias ) layer2=tf.layers.dense( inputs=layer1, units=20, activation=tf.nn.relu, name='layer2', **initKernelAndBias ) self.mu = tf.layers.dense( inputs=layer2, units=actionDim, activation=None, name='mu', **initKernelAndBias ) self.norm_dist = tf.distributions.Normal(loc=self.mu,scale=[1.]*actionDim) self.var1=tf.get_variable_scope().global_variables() with tf.variable_scope("Actor2"): layer1 = tf.layers.dense( inputs=self.s, units=20, activation=tf.nn.relu, name='layer1', **initKernelAndBias, trainable=False ) layer2=tf.layers.dense( inputs=layer1, units=20, activation=tf.nn.relu, name='layer2', **initKernelAndBias, trainable=False ) self.mu = tf.layers.dense( inputs=layer2, units=actionDim, activation=None, name='mu', **initKernelAndBias, trainable=False ) self.norm_dist_behavior = tf.distributions.Normal(loc=self.mu,scale=[1.]*actionDim) self.sample_op = self.norm_dist_behavior.sample() self.var2=tf.get_variable_scope().global_variables() with tf.variable_scope('exp_v'): self.log_prob = self.norm_dist.log_prob(self.a) self.exp_v = tf.reduce_mean(self.log_prob*self.advantage) with tf.variable_scope('train'): self.train_op = tf.train.AdamOptimizer(learning_rate).minimize(-self.exp_v) with tf.variable_scope('assign'): self.assign_target_to_behavior=[tf.assign(r, v) for r, v in zip(self.var2, self.var1)] def choose_action(self, s): return self.sess.run(self.sample_op,feed_dict={ self.s:s }) def learn(self, s, a, advantage, step): if step % self.update_frequency == 0: self.sess.run([self.train_op, self.assign_target_to_behavior],feed_dict={ self.s:s, self.a:a, self.advantage:advantage }) else: self.sess.run(self.train_op,feed_dict={ self.s:s, self.a:a, self.advantage:advantage }) class Critic(object): def __init__(self, sess, observationDim, learning_rate=0.01, gamma=0.95): self.sess= sess self.s = tf.placeholder(tf.float32, [1,observationDim],"state") self.r = tf.placeholder(tf.float32, [1,1],"reward") self.v_ = tf.placeholder(tf.float32, [1,1], "value_of_next") with tf.variable_scope('Critic'): layer1 = tf.layers.dense( inputs=self.s, units=30, activation=tf.nn.relu, name='layer1', **initKernelAndBias ) layer2 = tf.layers.dense( inputs=layer1, units=10, activation=tf.nn.relu, name='layer2', **initKernelAndBias ) self.v = tf.layers.dense( inputs=layer2, units=1, activation=None, name='Value', **initKernelAndBias ) with tf.variable_scope('square_advantage'): self.advantage = tf.reduce_mean(self.r + gamma*self.v_-self.v) self.loss = tf.square(self.advantage) with tf.variable_scope('train'): self.train_op = tf.train.AdamOptimizer(learning_rate).minimize(self.loss) def learn(self, s, r, s_): v_ = self.sess.run(self.v, feed_dict={ self.s: s_ }) advantage, _ = self.sess.run([self.advantage, self.train_op], feed_dict={ self.s: s, self.v_: v_, self.r: r }) return advantage env = UnityEnvironment() brain_name = env.brain_names[0] brain = env.brains[brain_name] print(brain.vector_observation_space_size) print(brain.vector_action_space_size) sess = tf.Session() actor = Actor( sess=sess, observationDim=brain.vector_observation_space_size, actionDim=brain.vector_action_space_size[0], learning_rate=0.02, update_frequency=10 ) critic = Critic( sess=sess, observationDim=brain.vector_observation_space_size, learning_rate=0.01, gamma=0.95 ) sess.run(tf.global_variables_initializer()) time=0 gamma=0.9 for i_episode in range(5000): step=0 discounted_reward=0 observation = env.reset(train_mode=True)[brain_name] s=observation.vector_observations while True: time+=1 step+=1 action = np.squeeze(actor.choose_action(s), axis=0) # print(action) observation=env.step(action)[brain_name] reward=np.array(observation.rewards) discounted_reward*=gamma #有错 discounted_reward+=reward[0] advantage = critic.learn(s,reward[np.newaxis,:],observation.vector_observations) advantage=[[advantage]] # print(advantage) actor.learn(s,action[np.newaxis,:],advantage, time) s=observation.vector_observations if observation.local_done[0]: print("episode:", i_episode," steps:", step," rewards:", discounted_reward) break
py
1a2f3cd3370dbae7b4946f6dc729dbe64dad33be
# Resource object code (Python 3) # Created by: object code # Created by: The Resource Compiler for Qt version 6.2.0 # WARNING! All changes made in this file will be lost! from PySide6 import QtCore qt_resource_data = b"\ \x00\x00\x0b\x1b\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22 standalone=\x22\ no\x22?>\x0a<svg\x0a xm\ lns:dc=\x22http://p\ url.org/dc/eleme\ nts/1.1/\x22\x0a xml\ ns:cc=\x22http://cr\ eativecommons.or\ g/ns#\x22\x0a xmlns:\ rdf=\x22http://www.\ w3.org/1999/02/2\ 2-rdf-syntax-ns#\ \x22\x0a xmlns:svg=\x22\ http://www.w3.or\ g/2000/svg\x22\x0a x\ mlns=\x22http://www\ .w3.org/2000/svg\ \x22\x0a xmlns:sodip\ odi=\x22http://sodi\ podi.sourceforge\ .net/DTD/sodipod\ i-0.dtd\x22\x0a xmln\ s:inkscape=\x22http\ ://www.inkscape.\ org/namespaces/i\ nkscape\x22\x0a vers\ ion=\x221.1\x22\x0a wid\ th=\x2224\x22\x0a heigh\ t=\x2224\x22\x0a viewBo\ x=\x220 0 24 24\x22\x0a \ id=\x22svg845\x22\x0a \ sodipodi:docname\ =\x22content-paste-\ func_own.svg\x22\x0a \ inkscape:versio\ n=\x221.0.2-2 (e86c\ 870879, 2021-01-\ 15)\x22>\x0a <path\x0a \ id=\x22path1499\x22\ \x0a style=\x22str\ oke-width:1.0568\ 3\x22\x0a d=\x22M 5.2\ 636719 1.8613281\ A 2.2202235 2.0\ 122117 0 0 0 3.0\ 449219 3.8730469\ L 3.0449219 6.8\ 925781 L 5.26367\ 19 6.8925781 L 5\ .2636719 3.87304\ 69 L 18.585938 3\ .8730469 L 18.58\ 5938 9.8671875 L\ 20.806641 9.867\ 1875 L 20.806641\ 3.8730469 C 20.\ 806641 2.7562694\ 19.807059 1.861\ 3281 18.585938 1\ .8613281 L 5.263\ 6719 1.8613281 z\ M 9.7050781 5.8\ 867188 L 9.70507\ 81 8.9042969 L 0\ .82421875 8.9042\ 969 L 0.82421875\ 10.916016 L 9.7\ 050781 10.916016\ L 9.7050781 13.\ 935547 L 14.1445\ 31 9.9101562 L 9\ .7050781 5.88671\ 88 z M 3.0449219\ 12.927734 L 3.0\ 449219 15.947266\ A 2.2202235 2.0\ 122117 0 0 0 5.2\ 636719 17.958984\ L 9.1777344 17.\ 958984 L 9.17773\ 44 15.947266 L 5\ .2636719 15.9472\ 66 L 5.2636719 1\ 2.927734 L 3.044\ 9219 12.927734 z\ \x22 />\x0a <metadat\ a\x0a id=\x22metad\ ata851\x22>\x0a <rd\ f:RDF>\x0a <cc\ :Work\x0a r\ df:about=\x22\x22>\x0a \ <dc:format>\ image/svg+xml</d\ c:format>\x0a \ <dc:type\x0a \ rdf:resour\ ce=\x22http://purl.\ org/dc/dcmitype/\ StillImage\x22 />\x0a \ <dc:title\ ></dc:title>\x0a \ </cc:Work>\x0a \ </rdf:RDF>\x0a <\ /metadata>\x0a <de\ fs\x0a id=\x22defs\ 849\x22 />\x0a <sodip\ odi:namedview\x0a \ pagecolor=\x22#f\ fffff\x22\x0a bord\ ercolor=\x22#666666\ \x22\x0a borderopa\ city=\x221\x22\x0a ob\ jecttolerance=\x221\ 0\x22\x0a gridtole\ rance=\x2210\x22\x0a \ guidetolerance=\x22\ 10\x22\x0a inkscap\ e:pageopacity=\x220\ \x22\x0a inkscape:\ pageshadow=\x222\x22\x0a \ inkscape:win\ dow-width=\x221559\x22\ \x0a inkscape:w\ indow-height=\x2292\ 7\x22\x0a id=\x22name\ dview847\x22\x0a s\ howgrid=\x22false\x22\x0a\ inkscape:zo\ om=\x2223.208333\x22\x0a \ inkscape:cx=\ \x2212\x22\x0a inksca\ pe:cy=\x2212\x22\x0a \ inkscape:window-\ x=\x220\x22\x0a inksc\ ape:window-y=\x220\x22\ \x0a inkscape:w\ indow-maximized=\ \x220\x22\x0a inkscap\ e:current-layer=\ \x22svg845\x22\x0a in\ kscape:document-\ rotation=\x220\x22 />\x0a\ <text\x0a xml\ :space=\x22preserve\ \x22\x0a style=\x22fo\ nt-size:8.23937p\ x;line-height:1.\ 25;font-family:s\ ans-serif;stroke\ -width:1.02992\x22\x0a\ x=\x2212.07681\ 8\x22\x0a y=\x2220.66\ 8371\x22\x0a id=\x22t\ ext881\x22\x0a tra\ nsform=\x22scale(1.\ 0118915,0.988248\ 24)\x22><tspan\x0a \ sodipodi:role\ =\x22line\x22\x0a i\ d=\x22tspan879\x22\x0a \ x=\x2212.076818\ \x22\x0a y=\x2220.6\ 68371\x22\x0a st\ yle=\x22font-style:\ italic;font-vari\ ant:normal;font-\ weight:bold;font\ -stretch:normal;\ font-size:12.359\ 1px;font-family:\ Cambria;-inkscap\ e-font-specifica\ tion:'Cambria Bo\ ld Italic';strok\ e-width:1.02992\x22\ >fx</tspan></tex\ t>\x0a</svg>\x0a\ \x00\x00\x02\xf3\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M14.\ 73,20.83L17.58,1\ 8L14.73,15.17L16\ .15,13.76L19,16.\ 57L21.8,13.76L23\ .22,15.17L20.41,\ 18L23.22,20.83L2\ 1.8,22.24L19,19.\ 4L16.15,22.24L14\ .73,20.83M13,19.\ 88C13.04,20.18 1\ 2.94,20.5 12.71,\ 20.71C12.32,21.1\ 11.69,21.1 11.3\ ,20.71L7.29,16.7\ C7.06,16.47 6.96\ ,16.16 7,15.87V1\ 0.75L2.21,4.62C1\ .87,4.19 1.95,3.\ 56 2.38,3.22C2.5\ 7,3.08 2.78,3 3,\ 3V3H17V3C17.22,3\ 17.43,3.08 17.6\ 2,3.22C18.05,3.5\ 6 18.13,4.19 17.\ 79,4.62L13,10.75\ V19.88M5.04,5L9,\ 10.06V15.58L11,1\ 7.58V10.05L14.96\ ,5H5.04Z\x22 /></sv\ g>\ \x00\x00\x09\xfd\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22 standalone=\x22\ no\x22?>\x0d\x0a<svg\x0d\x0a \ xmlns:dc=\x22http:/\ /purl.org/dc/ele\ ments/1.1/\x22\x0d\x0a \ xmlns:cc=\x22http:/\ /creativecommons\ .org/ns#\x22\x0d\x0a xm\ lns:rdf=\x22http://\ www.w3.org/1999/\ 02/22-rdf-syntax\ -ns#\x22\x0d\x0a xmlns:\ svg=\x22http://www.\ w3.org/2000/svg\x22\ \x0d\x0a xmlns=\x22http\ ://www.w3.org/20\ 00/svg\x22\x0d\x0a xmln\ s:sodipodi=\x22http\ ://sodipodi.sour\ ceforge.net/DTD/\ sodipodi-0.dtd\x22\x0d\ \x0a xmlns:inksca\ pe=\x22http://www.i\ nkscape.org/name\ spaces/inkscape\x22\ \x0d\x0a version=\x221.\ 1\x22\x0d\x0a width=\x2224\ \x22\x0d\x0a height=\x2224\ \x22\x0d\x0a viewBox=\x220\ 0 24 24\x22\x0d\x0a id\ =\x22svg871\x22\x0d\x0a so\ dipodi:docname=\x22\ invert_own.svg\x22\x0d\ \x0a inkscape:ver\ sion=\x221.0.2-2 (e\ 86c870879, 2021-\ 01-15)\x22>\x0d\x0a <met\ adata\x0d\x0a id=\x22\ metadata877\x22>\x0d\x0a \ <rdf:RDF>\x0d\x0a \ <cc:Work\x0d\x0a \ rdf:about\ =\x22\x22>\x0d\x0a <d\ c:format>image/s\ vg+xml</dc:forma\ t>\x0d\x0a <dc:\ type\x0d\x0a \ rdf:resource=\x22h\ ttp://purl.org/d\ c/dcmitype/Still\ Image\x22 />\x0d\x0a \ <dc:title />\x0d\ \x0a </cc:Work\ >\x0d\x0a </rdf:RDF\ >\x0d\x0a </metadata>\ \x0d\x0a <defs\x0d\x0a \ id=\x22defs875\x22 />\x0d\ \x0a <sodipodi:nam\ edview\x0d\x0a pag\ ecolor=\x22#ffffff\x22\ \x0d\x0a bordercol\ or=\x22#666666\x22\x0d\x0a \ borderopacity\ =\x221\x22\x0d\x0a objec\ ttolerance=\x2210\x22\x0d\ \x0a gridtolera\ nce=\x2210\x22\x0d\x0a g\ uidetolerance=\x221\ 0\x22\x0d\x0a inkscap\ e:pageopacity=\x220\ \x22\x0d\x0a inkscape\ :pageshadow=\x222\x22\x0d\ \x0a inkscape:w\ indow-width=\x22135\ 8\x22\x0d\x0a inkscap\ e:window-height=\ \x22927\x22\x0d\x0a id=\x22\ namedview873\x22\x0d\x0a \ showgrid=\x22fa\ lse\x22\x0d\x0a inksc\ ape:zoom=\x2212.521\ 682\x22\x0d\x0a inksc\ ape:cx=\x2222.20088\ 3\x22\x0d\x0a inkscap\ e:cy=\x2212.69278\x22\x0d\ \x0a inkscape:w\ indow-x=\x22476\x22\x0d\x0a \ inkscape:win\ dow-y=\x2255\x22\x0d\x0a \ inkscape:window\ -maximized=\x220\x22\x0d\x0a\ inkscape:cu\ rrent-layer=\x22svg\ 871\x22\x0d\x0a inksc\ ape:document-rot\ ation=\x220\x22 />\x0d\x0a \ <path\x0d\x0a id=\x22\ path1451\x22\x0d\x0a \ style=\x22stroke-wi\ dth:0.812468\x22\x0d\x0a \ d=\x22m 5.90861\ 81,2.0811647 c -\ 2.2861249,0 -4.1\ 421232,1.7863062\ -4.1421232,3.98\ 58465 v 6.375866\ 8 c 0,2.199539 1\ .8559983,3.98398\ 7 4.1421232,3.98\ 3987 H 7.321002 \ v -4.621946 c 0,\ -2.1995404 1.854\ 135,-3.9858464 4\ .14026,-3.985846\ 4 h 5.215386 V 6\ .0670112 c 0,-2.\ 1995403 -1.85599\ 8,-3.9858465 -4.\ 142123,-3.985846\ 5 z\x22 />\x0d\x0a <path\ \x0d\x0a d=\x22m 12.5\ 45747,9.1785429 \ h 5.921856 c 2.0\ 43043,0 3.70116,\ 1.6002091 3.7011\ 6,3.5718951 v 5.\ 715032 c 0,1.971\ 686 -1.658117,3.\ 571897 -3.70116,\ 3.571897 h -5.92\ 1856 c -2.043034\ ,0 -3.70116,-1.6\ 00211 -3.70116,-\ 3.571897 v -5.71\ 5032 c 0,-1.9716\ 86 1.658126,-3.5\ 718951 3.70116,-\ 3.5718951 m 0,1.\ 4287581 c -1.228\ 78,0 -2.220696,0\ .957267 -2.22069\ 6,2.143137 v 5.7\ 15032 c 0,1.1858\ 71 0.991916,2.14\ 3138 2.220696,2.\ 143138 h 5.92185\ 6 c 1.22879,0 2.\ 220696,-0.957267\ 2.220696,-2.143\ 138 v -5.715032 \ c 0,-1.18587 -0.\ 991906,-2.143137\ -2.220696,-2.14\ 3137 z\x22\x0d\x0a id\ =\x22path1462\x22\x0d\x0a \ style=\x22stroke-\ width:0.727191\x22 \ />\x0d\x0a</svg>\x0d\x0a\ \x00\x00\x01\x94\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M19,\ 6.41L17.59,5L12,\ 10.59L6.41,5L5,6\ .41L10.59,12L5,1\ 7.59L6.41,19L12,\ 13.41L17.59,19L1\ 9,17.59L13.41,12\ L19,6.41Z\x22 /></s\ vg>\ \x00\x00\x05\xf8\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M12,\ 8A4,4 0 0,1 16,1\ 2A4,4 0 0,1 12,1\ 6A4,4 0 0,1 8,12\ A4,4 0 0,1 12,8M\ 12,10A2,2 0 0,0 \ 10,12A2,2 0 0,0 \ 12,14A2,2 0 0,0 \ 14,12A2,2 0 0,0 \ 12,10M10,22C9.75\ ,22 9.54,21.82 9\ .5,21.58L9.13,18\ .93C8.5,18.68 7.\ 96,18.34 7.44,17\ .94L4.95,18.95C4\ .73,19.03 4.46,1\ 8.95 4.34,18.73L\ 2.34,15.27C2.21,\ 15.05 2.27,14.78\ 2.46,14.63L4.57\ ,12.97L4.5,12L4.\ 57,11L2.46,9.37C\ 2.27,9.22 2.21,8\ .95 2.34,8.73L4.\ 34,5.27C4.46,5.0\ 5 4.73,4.96 4.95\ ,5.05L7.44,6.05C\ 7.96,5.66 8.5,5.\ 32 9.13,5.07L9.5\ ,2.42C9.54,2.18 \ 9.75,2 10,2H14C1\ 4.25,2 14.46,2.1\ 8 14.5,2.42L14.8\ 7,5.07C15.5,5.32\ 16.04,5.66 16.5\ 6,6.05L19.05,5.0\ 5C19.27,4.96 19.\ 54,5.05 19.66,5.\ 27L21.66,8.73C21\ .79,8.95 21.73,9\ .22 21.54,9.37L1\ 9.43,11L19.5,12L\ 19.43,13L21.54,1\ 4.63C21.73,14.78\ 21.79,15.05 21.\ 66,15.27L19.66,1\ 8.73C19.54,18.95\ 19.27,19.04 19.\ 05,18.95L16.56,1\ 7.95C16.04,18.34\ 15.5,18.68 14.8\ 7,18.93L14.5,21.\ 58C14.46,21.82 1\ 4.25,22 14,22H10\ M11.25,4L10.88,6\ .61C9.68,6.86 8.\ 62,7.5 7.85,8.39\ L5.44,7.35L4.69,\ 8.65L6.8,10.2C6.\ 4,11.37 6.4,12.6\ 4 6.8,13.8L4.68,\ 15.36L5.43,16.66\ L7.86,15.62C8.63\ ,16.5 9.68,17.14\ 10.87,17.38L11.\ 24,20H12.76L13.1\ 3,17.39C14.32,17\ .14 15.37,16.5 1\ 6.14,15.62L18.57\ ,16.66L19.32,15.\ 36L17.2,13.81C17\ .6,12.64 17.6,11\ .37 17.2,10.2L19\ .31,8.65L18.56,7\ .35L16.15,8.39C1\ 5.38,7.5 14.32,6\ .86 13.12,6.62L1\ 2.75,4H11.25Z\x22 /\ ></svg>\ \x00\x00\x09\x8b\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22 standalone=\x22\ no\x22?>\x0a<svg\x0a xm\ lns:dc=\x22http://p\ url.org/dc/eleme\ nts/1.1/\x22\x0a xml\ ns:cc=\x22http://cr\ eativecommons.or\ g/ns#\x22\x0a xmlns:\ rdf=\x22http://www.\ w3.org/1999/02/2\ 2-rdf-syntax-ns#\ \x22\x0a xmlns:svg=\x22\ http://www.w3.or\ g/2000/svg\x22\x0a x\ mlns=\x22http://www\ .w3.org/2000/svg\ \x22\x0a xmlns:sodip\ odi=\x22http://sodi\ podi.sourceforge\ .net/DTD/sodipod\ i-0.dtd\x22\x0a xmln\ s:inkscape=\x22http\ ://www.inkscape.\ org/namespaces/i\ nkscape\x22\x0a vers\ ion=\x221.1\x22\x0a wid\ th=\x2224\x22\x0a heigh\ t=\x2224\x22\x0a viewBo\ x=\x220 0 24 24\x22\x0a \ id=\x22svg845\x22\x0a \ sodipodi:docname\ =\x22content-import\ -fx_own.svg\x22\x0a \ inkscape:version\ =\x221.0.2-2 (e86c8\ 70879, 2021-01-1\ 5)\x22>\x0a <metadata\ \x0a id=\x22metada\ ta851\x22>\x0a <rdf\ :RDF>\x0a <cc:\ Work\x0a rd\ f:about=\x22\x22>\x0a \ <dc:format>i\ mage/svg+xml</dc\ :format>\x0a \ <dc:type\x0a \ rdf:resourc\ e=\x22http://purl.o\ rg/dc/dcmitype/S\ tillImage\x22 />\x0a \ <dc:title \ />\x0a </cc:Wo\ rk>\x0a </rdf:RD\ F>\x0a </metadata>\ \x0a <defs\x0a id\ =\x22defs849\x22 />\x0a \ <sodipodi:namedv\ iew\x0a pagecol\ or=\x22#ffffff\x22\x0a \ bordercolor=\x22#\ 666666\x22\x0a bor\ deropacity=\x221\x22\x0a \ objecttolera\ nce=\x2210\x22\x0a gr\ idtolerance=\x2210\x22\ \x0a guidetoler\ ance=\x2210\x22\x0a i\ nkscape:pageopac\ ity=\x220\x22\x0a ink\ scape:pageshadow\ =\x222\x22\x0a inksca\ pe:window-width=\ \x221559\x22\x0a inks\ cape:window-heig\ ht=\x22927\x22\x0a id\ =\x22namedview847\x22\x0a\ showgrid=\x22f\ alse\x22\x0a inksc\ ape:zoom=\x2223.208\ 333\x22\x0a inksca\ pe:cx=\x2212\x22\x0a \ inkscape:cy=\x2212\x22\ \x0a inkscape:w\ indow-x=\x220\x22\x0a \ inkscape:window\ -y=\x220\x22\x0a inks\ cape:window-maxi\ mized=\x220\x22\x0a i\ nkscape:current-\ layer=\x22svg845\x22\x0a \ inkscape:doc\ ument-rotation=\x22\ 0\x22 />\x0a <path\x0a \ id=\x22path843\x22\x0a\ d=\x22M 12 0 C\ 10.7 0 9.599687\ 5 0.84 9.1796875\ 2 L 5 2 A 2 2 0\ 0 0 3 4 L 3 20 \ A 2 2 0 0 0 5 22\ L 7.109375 22 L\ 7.109375 20 L 5\ 20 L 5 4 L 7 4 \ L 7 7 L 17 7 L 1\ 7 4 L 19 4 L 19 \ 10.642578 L 21 1\ 0.642578 L 21 4 \ A 2 2 0 0 0 19 2\ L 14.820312 2 C\ 14.400315 0.84 \ 13.3 0 12 0 z M \ 12 2 A 1 1 0 0 1\ 13 3 A 1 1 0 0 \ 1 12 4 A 1 1 0 0\ 1 11 3 A 1 1 0 \ 0 1 12 2 z \x22 />\x0a\ <text\x0a xml\ :space=\x22preserve\ \x22\x0a style=\x22fo\ nt-size:8.23937p\ x;line-height:1.\ 25;font-family:s\ ans-serif;stroke\ -width:1.02992\x22\x0a\ x=\x2211.35293\ 1\x22\x0a y=\x2220.84\ 2772\x22\x0a id=\x22t\ ext881\x22\x0a tra\ nsform=\x22scale(1.\ 0118915,0.988248\ 24)\x22><tspan\x0a \ sodipodi:role\ =\x22line\x22\x0a i\ d=\x22tspan879\x22\x0a \ x=\x2211.352931\ \x22\x0a y=\x2220.8\ 42772\x22\x0a st\ yle=\x22font-style:\ italic;font-vari\ ant:normal;font-\ weight:bold;font\ -stretch:normal;\ font-size:12.359\ 1px;font-family:\ Cambria;-inkscap\ e-font-specifica\ tion:'Cambria Bo\ ld Italic';strok\ e-width:1.02992\x22\ >fx</tspan></tex\ t>\x0a</svg>\x0a\ \x00\x00\x0b\xd2\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22 standalone=\x22\ no\x22?>\x0d\x0a<svg\x0d\x0a \ xmlns:dc=\x22http:/\ /purl.org/dc/ele\ ments/1.1/\x22\x0d\x0a \ xmlns:cc=\x22http:/\ /creativecommons\ .org/ns#\x22\x0d\x0a xm\ lns:rdf=\x22http://\ www.w3.org/1999/\ 02/22-rdf-syntax\ -ns#\x22\x0d\x0a xmlns:\ svg=\x22http://www.\ w3.org/2000/svg\x22\ \x0d\x0a xmlns=\x22http\ ://www.w3.org/20\ 00/svg\x22\x0d\x0a xmln\ s:sodipodi=\x22http\ ://sodipodi.sour\ ceforge.net/DTD/\ sodipodi-0.dtd\x22\x0d\ \x0a xmlns:inksca\ pe=\x22http://www.i\ nkscape.org/name\ spaces/inkscape\x22\ \x0d\x0a version=\x221.\ 1\x22\x0d\x0a width=\x2224\ \x22\x0d\x0a height=\x2224\ \x22\x0d\x0a viewBox=\x220\ 0 24 24\x22\x0d\x0a id\ =\x22svg1501\x22\x0d\x0a s\ odipodi:docname=\ \x22import-csv_own.\ svg\x22\x0d\x0a inkscap\ e:version=\x221.0.2\ -2 (e86c870879, \ 2021-01-15)\x22>\x0d\x0a \ <metadata\x0d\x0a \ id=\x22metadata150\ 7\x22>\x0d\x0a <rdf:RD\ F>\x0d\x0a <cc:Wo\ rk\x0d\x0a rdf\ :about=\x22\x22>\x0d\x0a \ <dc:format>i\ mage/svg+xml</dc\ :format>\x0d\x0a \ <dc:type\x0d\x0a \ rdf:resou\ rce=\x22http://purl\ .org/dc/dcmitype\ /StillImage\x22 />\x0d\ \x0a <dc:tit\ le></dc:title>\x0d\x0a\ </cc:Work>\ \x0d\x0a </rdf:RDF>\ \x0d\x0a </metadata>\x0d\ \x0a <defs\x0d\x0a i\ d=\x22defs1505\x22 />\x0d\ \x0a <sodipodi:nam\ edview\x0d\x0a pag\ ecolor=\x22#ffffff\x22\ \x0d\x0a bordercol\ or=\x22#666666\x22\x0d\x0a \ borderopacity\ =\x221\x22\x0d\x0a objec\ ttolerance=\x2210\x22\x0d\ \x0a gridtolera\ nce=\x2210\x22\x0d\x0a g\ uidetolerance=\x221\ 0\x22\x0d\x0a inkscap\ e:pageopacity=\x220\ \x22\x0d\x0a inkscape\ :pageshadow=\x222\x22\x0d\ \x0a inkscape:w\ indow-width=\x22172\ 0\x22\x0d\x0a inkscap\ e:window-height=\ \x22952\x22\x0d\x0a id=\x22\ namedview1503\x22\x0d\x0a\ showgrid=\x22f\ alse\x22\x0d\x0a inks\ cape:zoom=\x2230.70\ 8333\x22\x0d\x0a inks\ cape:cx=\x228.61032\ 64\x22\x0d\x0a inksca\ pe:cy=\x2212.194623\ \x22\x0d\x0a inkscape\ :window-x=\x220\x22\x0d\x0a \ inkscape:win\ dow-y=\x220\x22\x0d\x0a \ inkscape:window-\ maximized=\x220\x22\x0d\x0a \ inkscape:cur\ rent-layer=\x22svg1\ 501\x22 />\x0d\x0a <path\ \x0d\x0a d=\x22M 14.0\ 51328,8.6485752 \ 9.6108804,4.6241\ 52 V 7.6424694 H\ 0.72998644 V 9.\ 6546808 H 9.6108\ 804 V 12.672999 \ M 20.711998,14.6\ 85211 V 2.611940\ 3 c 0,-1.1167775\ -0.9991,-2.0122\ 1167 -2.220222,-\ 2.01221167 H 5.1\ 704335 A 2.22022\ 35,2.0122117 0 0\ 0 2.9502099,2.6\ 119403 V 5.63025\ 78 H 5.1704335 V\ 2.6119403 H 18.\ 491776 V 14.6852\ 11 H 5.1704335 V\ 11.666893 H 2.9\ 502099 v 3.01831\ 8 a 2.2202235,2.\ 0122117 0 0 0 2.\ 2202236,2.012212\ H 18.491776 a 2\ .2202235,2.01221\ 17 0 0 0 2.22022\ 2,-2.012212 z\x22\x0d\x0a\ id=\x22path149\ 9\x22\x0d\x0a style=\x22\ stroke-width:1.0\ 5683\x22 />\x0d\x0a <tex\ t\x0d\x0a xml:spac\ e=\x22preserve\x22\x0d\x0a \ style=\x22font-s\ tyle:normal;font\ -weight:normal;f\ ont-size:5.33333\ 333px;line-heigh\ t:1.25;font-fami\ ly:sans-serif;fi\ ll:#000000;fill-\ opacity:1;stroke\ :none;\x22\x0d\x0a x=\ \x229.1831751\x22\x0d\x0a \ y=\x2219.082767\x22\x0d\ \x0a id=\x22text15\ 11\x22><tspan\x0d\x0a \ sodipodi:role\ =\x22line\x22\x0d\x0a \ id=\x22tspan1509\x22\x0d\x0a\ x=\x229.1831\ 751\x22\x0d\x0a y=\x22\ 19.082767\x22></tsp\ an></text>\x0d\x0a <t\ ext\x0d\x0a xml:sp\ ace=\x22preserve\x22\x0d\x0a\ style=\x22font\ -size:9.19055px;\ line-height:0;fo\ nt-family:sans-s\ erif;stroke-widt\ h:3.44647\x22\x0d\x0a \ x=\x220.19669378\x22\x0d\ \x0a y=\x2227.3268\ 41\x22\x0d\x0a id=\x22te\ xt1519\x22\x0d\x0a tr\ ansform=\x22scale(1\ .1595229,0.86242\ 367)\x22><tspan\x0d\x0a \ sodipodi:ro\ le=\x22line\x22\x0d\x0a \ id=\x22tspan1517\x22\ \x0d\x0a x=\x220.19\ 669378\x22\x0d\x0a \ y=\x2227.326841\x22\x0d\x0a \ style=\x22fon\ t-style:normal;f\ ont-variant:norm\ al;font-weight:b\ old;font-stretch\ :normal;line-hei\ ght:1.25;font-fa\ mily:sans-serif;\ -inkscape-font-s\ pecification:'sa\ ns-serif Bold';s\ troke-width:3.44\ 647\x22>CSV</tspan>\ </text>\x0d\x0a</svg>\x0d\ \x0a\ \x00\x00\x01\x8a\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M6,1\ 9A2,2 0 0,0 8,21\ H16A2,2 0 0,0 18\ ,19V7H6V19M8,9H1\ 6V19H8V9M15.5,4L\ 14.5,3H9.5L8.5,4\ H5V6H19V4H15.5Z\x22\ /></svg>\ \x00\x00\x01\xb2\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M5,4\ H19A2,2 0 0,1 21\ ,6V18A2,2 0 0,1 \ 19,20H5A2,2 0 0,\ 1 3,18V6A2,2 0 0\ ,1 5,4M5,8V12H11\ V8H5M13,8V12H19V\ 8H13M5,14V18H11V\ 14H5M13,14V18H19\ V14H13Z\x22 /></svg\ >\ \x00\x00\x01\xe2\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M19,\ 3H5A2,2 0 0,0 3,\ 5V19A2,2 0 0,0 5\ ,21H19A2,2 0 0,0\ 21,19V5A2,2 0 0\ ,0 19,3M19,19H5V\ 5H19V19M17,8.4L1\ 3.4,12L17,15.6L1\ 5.6,17L12,13.4L8\ .4,17L7,15.6L10.\ 6,12L7,8.4L8.4,7\ L12,10.6L15.6,7L\ 17,8.4Z\x22 /></svg\ >\ \x00\x00\x02g\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M15,\ 19.88C15.04,20.1\ 8 14.94,20.5 14.\ 71,20.71C14.32,2\ 1.1 13.69,21.1 1\ 3.3,20.71L9.29,1\ 6.7C9.06,16.47 8\ .96,16.16 9,15.8\ 7V10.75L4.21,4.6\ 2C3.87,4.19 3.95\ ,3.56 4.38,3.22C\ 4.57,3.08 4.78,3\ 5,3V3H19V3C19.2\ 2,3 19.43,3.08 1\ 9.62,3.22C20.05,\ 3.56 20.13,4.19 \ 19.79,4.62L15,10\ .75V19.88M7.04,5\ L11,10.06V15.58L\ 13,17.58V10.05L1\ 6.96,5H7.04Z\x22 />\ </svg>\ \x00\x00\x01n\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M11,\ 4H13V16L18.5,10.\ 5L19.92,11.92L12\ ,19.84L4.08,11.9\ 2L5.5,10.5L11,16\ V4Z\x22 /></svg>\ \x00\x00\x92\xc2\ \x89\ PNG\x0d\x0a\x1a\x0a\x00\x00\x00\x0dIHDR\x00\ \x00\x02\x00\x00\x00\x02\x00\x08\x02\x00\x00\x00{\x1aC\xad\ \x00\x00\x00\x01sRGB\x00\xae\xce\x1c\xe9\x00\x00\x00\ \x04gAMA\x00\x00\xb1\x8f\x0b\xfca\x05\x00\x00\x00\ \x09pHYs\x00\x00\x0e\xc3\x00\x00\x0e\xc3\x01\xc7o\ \xa8d\x00\x00\x92WIDATx^\xed\x9d\x07`\ \x1c\xc5\xd5\xc7wv\xaf\xaa7\xdb*\x96+\xee\xb8\x00\ \xee\xc6\x06SM\x87PC \x10 \x01\x02\xa1\x84\x12\ \xc2G\x80@\x12j\x80$\x94@ \x94\x00!\xf4b\ \xc0\x06\xe3\x0666\x06\x83\x1b\xb87\xf5^O\xd7\xb6\ |\xef\xcd\x8e\x84e\xe9\xa4k\xaa\xfb~:\xddm\x99\ \x9d\x9d\x9d\xf2\xfe3\xb3\xb3\xb3\xcc0\x0c\x89 \x08\x82\ \xb0\x1e\xb2\xf8%\x08\x82 ,\x06\x09\x00A\x10\x84E\ !\x01 \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80\ \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\ \x8bB\x02@\x10\x04aQH\x00\x08\x82 ,\x0a\x09\ \x00A\x10\x84E!\x01 \x08\x82\xb0($\x00\x04A\ \x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\ \x12\x00\x82 \x08\x8bB\x02@\x10\x04aQH\x00\x08\ \x82 ,\x0a\x09\x00A\x10\x84E!\x01 \x08\x82\xb0\ ($\x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\ \x10\x04AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04\ aQH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E!\ \x01 \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80 \ \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\x8b\ B\x02@\x10\x04aQH\x00\x08\x82 ,\x0a\x09\x00\ A\x10\x84E!\x01 \x08\x82\xb0($\x00\x04A\x10\ \x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\ \x00\x82 \x08\x8bB\x02@\x10\x04aQH\x00\x08\x82\ ,\x0a\x09\x00A\x10\x84E!\x01 \x08\x82\xb0(\ $\x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\ \x04AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04a\ QH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E!\x01\ \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80 \x08\ \xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\x8bB\ \x02@\x10\x04aQH\x00\x08\x82 ,\x0a\x09\x00A\ \x10\x84E!\x01 \x08\x82\xb0($\x00\x04A\x10\x16\ \x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\ \x82 \x08\x8bB\x02@\x10\x04aQH\x00\x08\x82 \ ,\x0a\x09\x00A\x10\x84E!\x01 \x08\x82\xb0($\ \x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04\ AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04aQ\ H\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E!\x01 \ \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\ \xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\x8bB\x02\ @\x10\x04aQH\x00\x08\x82 ,\x0a\x09\x00A\x10\ \x84E!\x01 \x08\x82\xb0($\x00\x04A\x10\x16\x85\ \x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82\ \x08\x8bB\x02@\x10\x04aQH\x00\x08\x82 ,\ \x0a\x09\x00A\x10\x84E!\x01 \x08\x82\xb0($\x00\ \x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04A\ X\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04aQH\ \x00\x08\x82 ,\x0a\x09\x00A\x10\x84E!\x01 \x08\ \x82\xb0($\x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\ \x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\x8bB\x02@\ \x10\x04aQH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84\ E!\x01 \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\ \x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \ \x08\x8bB\x02@\x10\x04aQH\x00\x08\x82 ,\x0a\ \x09\x00A\x10\x84E!\x01 \x08\x82\xb0($\x00\x04\ A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\ \x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04aQH\x00\ \x08\x82 ,\x0a\x09\x00A\x10\x84E!\x01 \x08\x82\ \xb0($\x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\ \x00\x10\x04AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\ \x04aQH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E\ !\x01 \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80\ \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\ \x8bB\x02@\x10\x04aQH\x00\x08\x82 ,\x0a\x09\ \x00A\x10\x84E!\x01 \x08\x82\xb0($\x00\x04A\ \x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\ \x12\x00\x82 \x08\x8bB\x02@\x10\x04aQH\x00\x08\ \x82 ,\x0a\x09\x00A\x10\x84E!\x01 \x08\x82\xb0\ ($\x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\ \x10\x04AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04\ aQH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84Ea\ \x86a\x88E\xa2_\xc0\x13\xd4LS\x06\x7f\xb0\xca\xf8\ \x8a!1\xd9\x90\x0c\xa63P}fn#\x08\xc2\xd2\ \x90\x00\xf47L\x01\xd0\xb4 \xf3\x96\x1b\x9eB\x16\xf0\ HL\x97\xec)\xba;ON\xca\x91e\x05\xdc0\x12\ \x00\x82 H\x00\xfa\x1f\x90\xa0j\xfd~m\xcb3\xb6\ \xf2/\xf4\xfa\xbd\x92\x04U~\xa8\xf83=q\x04\xcb\ >B\x1fs\x9d+c\x08\x09\x00A\x10\x00\x09@?\ \x01\xd2\x11;x\x0c\xc9W\xfa\xad\xf2\xf9\xcf%\x7f\xfd\ \xc1V\xde\x90tp\x90\x98\xcf\x8e~N\xc9\x18\x03\xfb\ \x18\xa3;@\x04ai\xc8\x04\xf4\x1f\xc0\xa6\xab5;\ \xe45\xd7\xca\xbezq\x17\xe0\x00\x0c\xbc# \xb1\xc6\ B\xfd\xb3\x9f\x1a\xe5_\xe3:A\x10\xd6\x86\x04\xa0\xff\ \xa0k\xba\xfa\xed\x03Rc\x91\x0eu\xff6\xe6\x9da\ _\x10\xfc\x19\xccW\xed\xff\xfe\x05CW\xc5\x0e\x82 \ \xac\x0a\x09@\xffA-\xfd\xc2^\xfc\x09\xef\xdbi\xf7\ 6\xaf\xcc\x9b\x00\xb8\xdd^\xbc4\xd8Tan%\x08\ \xc2\xb2\x90\x00\xf4\x17\x0c]\xde\xf9*\xd6\xfd\xc3Ak\ *Z\xf7\x86X&\x08\xc2\xaa\x90\x00\xf4\x13\x0c]5\ \x9a\xca\x99\xde\xb6\xef\xa7\x0d\x86\xa1\x1bF\xbe\xb1Z\xac\ \x12\x04aUH\x00\xfa\x09\x86\xae\xe9\xbaf0\xc9h\ {\xff\xf7 x\xf7\x90\xe1k\x14\xab\x04AX\x15\x12\ \x80~\x02S\xec\xcc\xe6\xc0\x85\xce\xdb\x00\x06(\x80\x9a\ \x98-\xd6\x08\x82\xb0*$\x00\xfd\x04\xc6\x149s\x12\ ,t\xfaX\x07(\x84\xceXc\xd6\x99\xf8\xe8\x80\xae\ \xd3\x83 \x04aYH\x00\xfa\x0b\x8c\xa9\xc3/\x94l\ n\xec\x04\xea\x10|^Lv&\x0f>\x0c\x16\xb1-\ @\x10\x84U!\x01\xe8?\xb82F\xfaF\xdf(\x85\ \xf3|\xaf\xe6W\x96\x9e\xe7m\xa8\xe6jA-\x00\x82\ \xb0($\x00\xfd\x07\x83I\x09\x13.0\xf2\x8e\xc1%\ \xec\xdf\xe1\xe8\x12N\x01a.\xe2\xf0\x1fX\xc7\x87\x01\ XS![\xf5k_]\x09\x1e\xc8\xb7\x83\x03\xee\x0d\ A\x10V\x81\xe6\x02\xea'\x98\xe9\x88f>P\xeb_\ s\x97\xad\xe0\x03\xc9\xc0G\x7fq;\x13I\x0c?f\ \x8f\x8fxL\xcc\x90\xb4\xe4\xa1\xc6\xb1\xef8\x923`\ \x1d\x9f\x13\xa3\x1e!\x82\xb0\x12$\x00\xfd\x04L\xc7f\ \x03\xaf\xeb\xaaQ\xb8D\xdb\xf7\xb1V\xb5\xc5\xa6\xd6\xc1\ \x96&-\xc3\x99;N\x93\x14\xe7\xbe\xb7\x0dC3\xdd\ \x81\xbd\x87\xe3\xb4\xcc#\xa4#\x9ft\xa4d\xf3C\xf9\ v\xbeD\x10D\xbf\x87\x04\xa0\xdf\x822\x10\xa8g\x9a\ \x0au{U\xb7\xdb\x5cI\x90\xd4Mk\xefq\xec~\ \x917\x09t1\x1b\xa8!\xe9\xc9\xc3\xd4\xe3\xdeKH\ JE\x1d\x91\x98L\xb3\x84\x12\x845 \x01\xe8\xb7\xf0\ \x94E\x83n\xae\xc1\x87\x19,\xe0\xf7hk\xef\xb0\x15\ \xbe'k\x9a!\xe3.p\xc6\x98\xa1e\xcd\x90\xe6\xfc\ \xdd\x91\x9c\x83\xdd@\xd4\x04 \x08k@\x02\xd0oi\ NY\xb0\xe6\xad\x92XS\x03\xfeu\x7fr\xec|\x11\ S_\xbc.\x066\xcbz\xfa\xa1\xc6Q/9\x93\xd2\ \xcd\xae!\xd8KJ@\x10\xfd\x1bj\xec\xf7[\xd0~\ #-\x0b\x02\xc5f\xb7\x1f~[0\xffT\xa8\xf9\x83\ \x02\xa0\xa5\xc7n\x7fC\xa9\xda$\xad\xfce\xa0\xae\x00\ \x9b\x0dmd\x83 \x88\xfe\x07\x09\x80\x15\xb1;\xdd\xb6\ \xd9\x0f\x05\x86\x9d\xc7g\x0fm1\xf4\xa0\x01\xdf\xb0/\ \xae\xf66\xd5\xf2U\xaa\xfe\x13D?\x87\xba\x80\xa2\xa4\ 9\xdexe\x99\x7f\xfd\xb8\x8a@\xed\x9a\xff\xf6b|\ \x9ez\xf5\xab\xdb\x1c\x85\x1f\xe1\x93\x02\xba\xd1\xd2\xe5\xa3\ g\x1c.\xcd}\xca\x914\x08\xaf\x05\xef\x177\xef \ \x08\xa2\x7fA\x02\x10%\x9a\xa1\x9b\xad'MU\x8d`\ \x93\xea\xab\x81e\x9b+\x9d9\xdc\x8a\xe2\xe8\x13\x02\xa0\ \x1b\x86\x1a\xf0\x06\xbe\xbe\xcf\xbe\xfb%\x99? \x86]\ A\xa8mL\xcb<L\x9f\xfblBr\x06n$\x01\ \x88~\x0a\x09@d`\x9f9\x8f0#\xe8\xd1j\ wxv~\xe2*}[\xf1\x94\xf2h\xd4$\xc5!\ \xbb\x07\xfa\xf3Ow\x0e?MN\x19\xc9\xec\x09\xfc \ Ao3\xa3\xf8`\xb0$\xa9\xfe&\xfd\x8bk\xed%\ \xcb@\x0f`#o\xce\xc8\xf0\xa5e\x8cgG=\xef\ H\x86v\x80\x80d\x80 \xfa\x19$\x00\x91\x01\xd1e\ \x18\x9aZ\xbeA\xda\xf0\x90T\xb5^R\xbd\x07\x99E\ f0\x1d\xcc\xa7\xcd\xa1gM\xb7\x1dq\xbb\x92>\x1e\ m\xaa\xd9\xbf\xd2+\x0d(\x5cO\xc0\xef\xf1\xaf\xbb\xcf\ \xbd\xe7\x150\xfb\x07\xday|F\xec\xa8\x7f;\x13R\ a\x19tOf\x8a\xb9\x9d \x88\xfe\x01\x09@d\x80\ \xf1\xd7\xf6/\xd2\xd7\xdc*\x05\x1b\x98!\x83\xb5?\xc8\ \xacC|\xf2a\x94\x1a\x18LC\x96\x8d\x99\x0f\xd8G\ \x9cm\xc8\x8a\xdc[;R\xcc\x0c\x10\x0c4\x05V\xdd\ l/\xfaH\xd65I\xc6\xce-\xb3\xadc\xa4\x8d\x97\ \x8e\x86v\x00\xbe<@<8F\x10D\x7f\x81\x8at\ X\x80\x914\x0c]\xd7Uu\xfb\x8b\xc6\xea\x9b\xa4`\ #\x98C>\xaf~;6\xdd\xc0\x1b\xa7\x0a\x0e\xae7\ t\xf6\xcd\x9f\x02[_\xef\xfd\xb1lw$\xd8f>\ \x10\x18r\x9e\xc1lh\xf7\xe1\x0a\xf8#d\xacn\xab\ \xf1\xe5M\xbe\xa6Z3\x06@-P\x16\x08\x82\xe8\x17\ P\x0b ,x,\x19j\xd9\xd7l\xf9\xe5z\xb0\x0e\ b\x0d\x05 \x0c\xc0\xa1\xccd}\xfeK\xf6\xbc\xb9\xbd\ \xb3\x05\xd0\x02\x5cc\xc0\xd7\x10\x5c}\x8b\xa3\xe8c\x9c\ (\x02*\x07<\xbc`\xf7\xb5\xf4\xd1\xf2\xd1\xaf;\x92\ 2\xcc\xeb\xa6\xa6\x00A\xf4\x0f\xa8$\x87\x8b\x1e\xf0\xe9\ \xeb\xee4\x02h\xfd\xc3\x17M\xbc\xa7\xaa\xeb\xc6\xc6G\ \xb0\xd1\xd0\xcb1$\xbb+\xc9>\xfb\xe1\xc0\x88\x9fI\ \xcc\xc6\x1b\x00&L\xa9\xd9i\xac\xbe\xc6\xefm\x10\x1b\ \x08\x82\xe8\x17\x90\x00t\x02Z{n\xef\xf5=o\xcb\ 5?\xa0\x99\x84\x9a<\x9fE'<\x18N\xc1S\xbd\ \xb1a\xcfJ\xfe\xfeE>\xda\xb2W\x82\x93\xc01\xd9\ \xe9NvL\xbdC\xcd=\x1er\x86y\xe9X\xe3g\ \x92R\xfe\xa5\xbc\xe4,\xd5S\x81M!\xd1\x19Dm\ G\x82\xe8\xdb\x90\x00t\x82\xd9o\xa3\xfa\x1b\xb5\xddo\ \xf0\xbe\xf1\xc8\xbaq\xd0\x5c\x82\x1fz\xd0\xb9\xebi\xbd\ \x17[\xff\x03\xb1;\xddl\xd6\xc3\x9e\xa1\x17@\xd5\x1f\ ;\xbb\xf0\x22\x10V\xb3K_}}\xd0\x07M\x19>\ {P\xf3v\x82 \xfa(}[\x00x5\x14L\xd4\ \x8f\xdf|[\xf3vN\xcb\x8ap\x15!p\x18|3\ O\xa1\xd2T\xc2\xf0\x15+\x91\xc1\xd0VBu\x99\xb1\ \x9a\xad5\xe5\xfbqC_\xc0\xe5Nv\xcf\xbc\xdb\x9f\ \xbb\x00.\xd8\x8cS\xfcf\x86R\xb6\x9a-:9\xd0\ T\x8bC]\xf9\x0d$\x13q\x18A\x10}\x8a~\xd0\ \x020{#\xd0\xbe\xc3\x0a\xf6MhA]\x0d\xe8Z\ \x80\xbf\xf9\xc44N\xf8\xcdmV\xe4\xa6\xca\xf4\xb6\xb1\ \xd0\xf0V\xf3W\xa8G\xe8\x03\xde3\xc5\x09\xf6\x99\xe6\ K\x09n\x17\x1b{7\xd8\xc3\xc5d\xbb\xc3e\x9b\xf9\ \x80\x7f\xf8Ex\xcb\x17\xa3\x17/\x05\xfb\x83\x1a\xf7\xeb\ \xab~\x13\xf45`T\x98\x7f\x04A\xf4M\xfa\xf6(\ \xa0\xe6\xc0\xc3\x0f\xd3\xfd\x9e\xfa\xbd_\xf8~\xf8o\x96\ \xbcCV\xfd\x86\x22{\xa5\x9c\xa6\x01'fM<\x97\ %e*|\xe0\x0a\xb8\x8ex(\x0e\xb7q\xea\xae\xb7\ \xf5/o\x94\xb9\xb5\x8b\xd4\x03\x13C\x97\xb4\xa9w\xba\ &\x5c\xd17\x9a\x00\xcdq\xab\x06\xfc\x81/n\xb0\x15\ \x7f$\xeb\xfa\x8f\xcf\x07@c !O:e\x99\xcd\ \x99\x80\xc2\x8a\x91\xdaW.\x8b \x88\x1f\xe9\x93\x02`\ \x86\x19l\x0e,hj@\xda\xf7\x81\xba\xf7cV\xb6\ \x8ai\x1e\xd8\xdcl\xa2Q\x15\xd0\x89l\xd7R\xc6)\ \xf9\xc7\xb2\x11\xe7\xdaR\xf2p\x0f\xd6c\xf9_\x18\xf0\ s\x19\xc1\x9doI_\xde\x8c\x86\x0f\xcc_T\xc6\x0e\ \xbcQ\xa7\xdc\xee\x9a\xf8\xab.\x7f\xdfV\xe8\x14\x8d\xc8\ L\x9b\x17\x0e\x1f\xbf\xd7\xa3}\xfd'\xfb\xbe\xff\xc1&\ 8\xbeY\x04\x99\x9e3G\x9e\xfb\xb4\xe2L\xe2\xbdD\ \xe83\xc9\x00A\xf4-\xfap\x17\x10X\x9d@\xed^\ }\xe1\xf1\xda\xaa\x9b\xe4\xe2%\xb2\xee\xe5}\x17\xdc\x0a\ \xe1\xc7\x5c`\xccPmu\x9b\xd8\xa6G\xa5\x0f\x8f\xf5\ \xee_\xa5\xe3\xfc7:\xda\xe3p\xc1W\xab\xcb6'\ 7o\xe1\x1fu0\x10\x16\xd9\x91\x1a\xa6\xea\xc4B\x9b\ Fp\xa9\x07\x82\xf1\x87\xf1\xa98\xdd\xc9\xee\xd9\xf7\ J\xf9's\xbfa\x0b\xec\x82\xdd\x06+\xf9B\xfa\xe0\ \xa8\xa0\xb7\x01\xfb\xdd\x840\x10\x04\xd1\x97\xe8\xab\x02`\ \x18z\xb0`\xb5\xb4\xfcRV\xbfW\x06c\xd4\x99\xf5\ A\x13\x15\xf4\xdaV_\xa1myR\xd7\xb1\xd7B\xec\ \xe8\x1c|\xaeWsd\x19J\x02T\xfe\xa36rx\ `\xca!\xddf%\xe14\xbaa\xa8\x9a\x1e\xd0\x0c]\ \xd7`Y\xec\x88\x06CR\xec\xca\x8c?K\xf9'\xea\ \xa8\x9d\x22\xcf@\x9c3o\x85\xbe\xea*\xd5\x07\x1a\xc0\ \x1f\x7f&\x08\xa2O\xd1\xc7\xba\x80\xcc~\x09\xf8\x0f\x14\ \xacd+/\x97\xf5\x00n\xc5\x1ai'\xc0\x81\xe0\x08\ ,8\xd4\xe6\xf5\xd1\x978\xa6\xdd\xc9\x0f\xc2\x16\x83\xe9\ \x14xF&\x05\xebK\xb5\xcf.\xb2\xd5o\x85\xc3\ \xc39\xdd\x810]\xd2\xe0\xb4\x8e4\xed\xd4U\xce\xc4\ $\xacBw\x01\x10NP\x17\xee\xb5Q\xeb\xf5\xbf\xb8\ \xa5fQ\xb1Z\xd8\xa0z55\xd5a\x1f\x93\xe68\ \x7f\x98\xfd\x94\xd1\x03\x14\x19\x22\x0f\x5cA t\xa8\xdd\ \xf3C;\x83G:\xfe\x06\x9b\xd4/o\x91\xf6\xbdg\ \xb6\x03ph\x93\xf9\x95:B:\xe1C\xbb\xd3e\xf0\ \xbb\x04\x07\x88\x1c\x06\x07V\xba\xe8\x92\x09\x82\x88\x91\xbe\ '\x00\x10\x5c\xb5f\x9f\xb4\xe2R\xd6\xb0\x13\xfb\xe3#\ 1.\xe6\xc52\xd9\xa1O\xbf\xcf1\xf2ln\xff\xc3\ j\x03\xe9zP\x03\xdb\xb7\xebm0\xb1\x11\x193<\ #N\x10\xca<\xc3~\x91:\xefn\xb4\xbc]c\x0d\ 1f\x0ciG\xb5\xe7\xe9\x8dU\xef\x16\xf8\x1b\x828\ \x1eJF\x0b\x8d{\xcd\x16R^\x82\xf3\xf2q\x09\x97\ \x8cKKv9\xe1\x880\xaf\xbd\x05hu\xe9\xfez\ \xfd\xab\xdb\x8c\xbd\x1f\xc1*\xdaz~)x\xa6\xbc\xa3\ \xe5Y\x8f\xda\xdc\x19\xba\xe6W=\xe5R\xc0+\xbb\xd3\ \x14w&\xc408\x04\xd0\x1dA\x10\xbd\x8c>'\x00\ z0\xe07\x96\x9c+Wm\xe0\xb65B{\x8a\xc7\ \xa0\x0d\xd7\x13\xb2\xd5\x13?q'\xa5\x86yK\x16\x8e\ \x0bVm7>>^\xe6'\x15[\xc3\x01O(\x05\ \x9di\xb6\x05\xef\xdbS\x87\xc22\xeb\x1ak\x08Fx\ \xf1\xbe\xba\xdf~^V\xe53k\xe0:7\xd1\x10T\ \x1d/\x18\xbbh\xb0\xf5\xa3\xc8\xfa\x88d\xe7\xdb'\xe5\ \x0eHt\xc8\x11<\xcf\x8c\xa0\x00\x80\x17\xaaG]y\ \x9d^\xb4\x04_\x1a`F\x05WU\x96:<8\xf0\ (V\xf2\xa9\x1c\xf4\x82KI\xb6I\xae\x14u\xe4e\ \x8e\xd1\xe7\xda\xed.tF\x10D/C\xb9\xfbn\xa8\ \x96\xf6\x19\xd0\x80\xef{\x8fm\xfb\x0f\xaf\xbcGh\xfd\ \x01\xf3\x10\xf8\x04=R\xedNe\xd8)\xb2\x1c\xee\x1c\ \xf7\x8a;\xdd\x1f\xd4\x95\xaa\xaf\xd1\xbaq\x9f\xcc\xed\x1d\ \xc3\xbbIXp\xc4\x85\xae\x11\xa7\xca\xb2-\xe2\x00\x87\ \x01\xceR*\x19K\xf77^\xf7yy\xb5\x0f\x22\x08\ \xaf\xaf\xe5B\xc5\xafXA\x9b]\xe5S\xbf.\xf7.\ \x18\xe2v\xdb\xb1\x12\xcf\xc3\x18V\xa8\xc0\x19v\xe6\xc8\ v)g\x9eQ\xb7\x9b5\xec\xc1\x8dx>\xfe\xf1\xd7\ \xc8\xd5\xdf\xc9\xc1z\xa6y\x99\xeecZ\x93\xe4\xabV\ J\x97\xc8e\xeb\xb4\x94\x11\xcc=\x88\x9f'\xfe\x97O\ \x10D\xd4\xf4\xb5\xb694\x00\xf6|\x18\x97\x1b\x8e\xb6\ \x8a/\xd4\xca\xcdb%\x1c\x18\xb3\x8f\xbfL\xcb\x98h\ \x88\x9au\xe7\xf0\xf6\x86!g\x8cM8\xe2V\x065\ \xe2.\x82I\x15\x8d\xeau+\xca\xab\xfd:\xef\x7f\xef\ $r`\xf7\xd7\x15\xfe\xcb?\xc3\xb7\x98\xf1\xb9)\x22\ 6\xca\xcc\x95j\x9b\xf7\x844\xf4tX\xd41-\xda\ ?#\xec\xc1G\xa7\xcb\xbf4V\xfe\xda\xa8\xdc\x14\xf9\ y\x08\x82\xe8Z\xfa\x98\x00\xe8\xbe\x1a\xa5\xea\x1b\xb1\x12\ \x13\x86\xa1\xf9\xa5\xaa\x0db-<\xec\x09i\xca\x91\x8f\ \xcbY\x93\xd1\xb8\xa1\xdd3-<\xfe\xf3\x8fi\x09q\ \x03,\x98;\xa5\x01G\xe8G\xffGv\xb8M\x1f\xba\ \x82\x80\xaa\xdf\xb5\xba\xa8\xd2\xaf\xf1\xb3\xf3\x9a~\x87\xc8\ \x0c\x9f\x90\xfe\xb2\xcc\xff\xd6\xf6j\xbc%\x1c9p\x0e\ Yq\xc8\xd3\xef\xd1\x07\xcd\xe1\xb3F\x878#ow\ \xe8\xe0\xc0S\xaa}\xf3\x07\x03\x9f\xcd\x16\x08\x07\x04A\ \xf4(}L\x00\xb4\xf2oX\xa0N\xac\xc4\x06\x03\xd3\ T\xb1F\xact\x06v^\xf0A\xf1\xb6\x94!\xf2\xd1\ \xcf\x19\xc3O\x15\xf6\x1em\x19x\x04\x16\x10>\xa6\x08\ \xa0}\xe4\xc6X\xd6\x86\x9cc;\xeaY[\xe2\x00~\ lWU\x80w\xd5\x07>/\x0f\x808\xa2\xb9\x85\x8f\ \xd8\xdc\x012\xce\xf4\xcf\xa4w\xf745j\xa0\x05\x91\ aF\x05\xc8\x88\x0c\xed\x80c\x9e5\x06L\xe5\x17\xdd\ \x1ep\x12S\x1d\x98\xc1*6h\xdf?\xd5\x1cc\x04\ A\xf4\x0a\xfa\x98\x00x\xf6|\x16\x97\xf1\xe6h\xfft\ I\xaf\xd9!\xd6#\xc0\x90]\x99\xf69\x7f\x97\xe7=\ \xafe\xcf\xd7\x1c\x19h\x12\xd1\x04\x82a\x03S\x07\xf1\ \xa9\xe8\xce4=w\xbe>\xfb\xef\xae\xb9\x7fe\xeeL\ Y\x96\xbb\xb4\xda\xbb\xa9\xa4\xa1*\xc0\xef\xf1F\xc8\xda\ r\x7f\xa5GE\xa3\x1c%\xd0\x940\xa4\xea\xef:=\ ^\xc6>;\xc3(\xf8\xcc\x00\xbd\xe1\x82 v\x10\x04\ \xd1\xa3\xf4\x95Q@8\xa6\x05\x82Z\xf6\xda\xc9\x19\xc1\ \xcdh\xbec\xb3!x\xd5\x86\xe4\xb7g$_\x18Y\ /\xd0\x81\xd1\xa5\xab~\xa3\xa9$X_\x5c\xb5s\xb5\ \xb7j\x9f\xa1k\xc9\x83F\xa5\x8f\x9ekO\xcc2\x12\ \xf2\x14\xc5\x0en0\x98LDr\xdc\x1b\x01x\x0d\x86\ \xf4\x9bw\xbf\xf9_e2?\x91\xb99\x5c@\xb7^\ \x9a\xaa\x9c0\xe9\x10h\x9f\x88M\x91\x00\x17\xa5U\xef\ 6>\x9c\x8f\x01\xe9\xf0\xd2\xcc\xcb\xd7\xdd\xf9\xec\x847\ \x1c)\xb9\xb0\xdcu\xed!\x82 \xc2\xa7\xaf\xb4\x00\xcc\ J\xb4\x9e\xe9\xaa\xe3\x86+V\xd1\x02\xf3\x03&H\xd6\ \xa1\xfe\x1b\x19\xbc\x03D\xa0\xd8]J\xca0\xd7\xe0\xd9\ yG\xdf|\xc8\xd9\xff\x18u\xee\x93\xd9\xf3ntf\ \x1f!'\x0f\x03\xebo\xba1-\xa3X\x8e+hU\ \xf9\xdb\xc6\x0a\xed\x830F\x22\xf7^7\xa4B%\x0b\ \xef\xd5F\x05\x9c\xd0_\xb7\x0b\x03\xd19\x188#\xe8\ \xd5p\x0eQX\x89sT\x10\x04\x11\x1d}E\x00D\ \x0d\xba!\xe0\x06\xe3\xc1;[b\x05|\xd4\xe2y\xf9\ \xe0\x9f\x19\xaan\xb5n\x9an\xf85h\x1e\xc1\xd9\xc3\ 1\xc4\xad\x00\xcdhTA\x02\xa3\x0f\xb0\x92\x98\x05\xdf\ \xe1\x9eXv\xe0\x94J\xe8>\xe2\xa0\x12\x04\xd1\x15\xf4\ \x15\x01\xe0Uh&\x07\x13\x87\x8b\x0d1\xc2\xeb\xe36\ W\xaa\xb9\x165f\xb088D^,r\x84\x8b.\ \xc6\xa6\xb0TE\xc5;\x00\x917\x01\xe0\x80LY\x8d\ \xda\x1c\xc3a,i\x88\x14\xc6\x83\x14f\xc8\xe4\x84T\ [J6\xef\xcd#\x08\xa2W\xd0W\x04@\x90:|\ \x16\x0e)\x89\xd9\xba\x1a\xd8e\xa2\xcb)\x83\xc5z\x9f\ \x05\x94f|\x22\x98b\xb0\xc2\x11\xdbq\x1bc\xa3R\ \xf8Lz\xd1\xe2p\xa7J\x99\x87wz\xbc\x1929\ s\x9cbN\xaa\x1as\xf2\x11\x04\x11\x17\xfa\x98\x00\xb0\ AG\x1a\x8a3\xf6>\x04\x1c\x95\xc2\x14=k\x86X\ \xefv\xf8-\x8d\x1f\xd1t|a\xfcA\x08\xa7!h\ \xb1\xa3\xb3G\x0erD\xd6\x8f\x8f^\x83\xffy\x09\xf2\ \x88\x81\xe91\xf4\x00\x19\x12S\x8c\x11\xe7@L\xe2\xfd\ \x88\xb6\x01\xe6W\x01\xe0\x92+\xcd8\xfc\x1e\x1c?\x1a\ \x83\xde\x10\x04\x11_\xfa\x98\x00\xd8S\x87h)\xe3\xb9\ \x05\x8b\x09hD\x18\x92\xa2d\x1d&\xd6{\x02~\x0d\ h\x1f\xdf\xfe\xbe\xf4\xf2%\xc5W//\xd9Z\xed\xe3\ \xb7f\xc3\x05+\xd3\x8cM\xcct\x0eK\xb1\xe1\x90\xfb\ \xf0\xc0\xc6\x0f\x7f\x04\xf8\xa8\x5cW\xaa3\xdc\x990\xda\ \x03L9\x93\x87\x9d\xc1\x06\x1f\x8b]Pbc+\xb0\ \xb5\xc6\x98ns\xcbS\xfe\xcf\xeeL\xc1-\xa8[\xa4\ \x01\x04\xd1+\xe8c\x02 )\x0e[\xde\x5c\xb0\x22b\ 5Z\xc0\x04\xb2\xe4a\xf6\x9c\x1ek\x01\x00\xd8\x0a\x91\ \xa47v\xd6_\xb5\xban\xe1\xbe\xc6\xb7w\xd6_\xb0\ \xb8\xa8\xa8QC\xeb\x1c\x89\xc2\x0dp\xdb\xce\x1a\x91\xc2\ \xa4\xf0\x1f\xe9\x02\xcf\x99S\x96\xce\x1b\xea\xb4+1\xc6\ \xa4\xa18\xdc\xb6y\xff\xd2\x07\x1f\xdd\xae\x02\xe1;\xe5\ e7\x1b\x7f\xb52\xf2\x5c2\xfb\x04\xd1\xdb\xe8k\x02\ \x00fs\xe4\x85\x86+\x13\xfb\x1c\xf0\xe5$bk\xa4\ \x18\x8a;p\xf8\xbd\x8c\x0fJ\xe9\x11\xccG\x19|\x01\ \xf5\x9d\x1d8\xb0\x95\xdbFV\xecQO_X\xf0U\ \xa9\x1f\x87\xf5 \xd8-\xc4\x9dw\x04\xd4\xa9\xaf\x99\x98\ qr~\x22\xce\x04d\xbe\x9c\x0b\x11{\x0f\x00\xb6\x89\ z\xba]f\x7f\x99\x919}h\x16\x06$j\xf0P\ \x089\xc3\x89\xf9f\xfdC\x9f\xfe\x90\x94~\xa8\xc6\x1f\ z\xc0\x80`$\xdb\xb5\xdc\xe3\x94c\x9ewL\xbe\xa1\ \x8b\xe6@%\x08\x22\x16\xfa\xca\x83`\x02\x1dg\x17\x93\ \xea\xf7\x7f\x95\xb0\xea\x12\x16\xf4\x198\x0bC4&\xcc\ \x90\x1d\xec\xa8\x7f\xdbrgw\xe1\x1cm\x1d\xc2\xa3\xdd\ h\x0ah\x97}R\xb8\xb4\xcc\x0b\x01\xe2}#\xa8h\ \x03\x9d\xf2G\xa7\xe7\xe7\xa78`\x156a\xbfy\x87\ p\xa3\xcf\xca\x1b\xfd\x97|V\xf2]\x85\x1f\xfd\xe5\xdb\ M\x0f\x7f\x04\x05S\x05k\xad\xc8\xd2\x95c\x93\xff0\ 3\xd7\x06K\xf1\xc0\x9cS\x0e\xc7\xe6\x1aj\xd0\xeb\xf5\ \x97~\xa35T\xd8\xb3\x0eq\x0d\x9c\xa8\xd8\xb1\x8b\xa9\ \xd3K \x08\xa2G\xe8c\xd3A\x9b\xc6\xcd\x99\x9c\x1d\ \xd4e\xa9|5Z\xb8h\xec\xbf\xc4tU\xda\xbfX\ K\x19\xa1\xa4\x8f\x16\x9bz\x02\x9b,k\x86\xbe\xb8\xa0\ \xa9\xa5\xff\x04\xaa\xe8\x1eU\xffh_\xe3\xf4\x81\x09\xd9\ \x896.\x00\x9d\x5c!j\x093\x92\x1c\xb69\xd9.\ \x8f\xdf\xbf\xa3^S\xb9\x02\x1ct\xa0\xb9m\x90\xcbv\ \xf3\x94\xf4k&g9m\xf1\x9b\x9e\x08\xbcF\xbf\x0c\ \xc6\x14\xc5\xeet\xa6\x0fue\x8fw\xa4d\xa3\xd4\xe0\ \x05\xc1\xbe8\x9d\x88 \x88\xb8\xd2\xc7Z\x00&\x10f\ ]S\x1b\xbf\xba\xdf\xb5\xfb\xdfh\xca\xd1\x02\xe10\xfc\ v@\xdbd.\xe0u\xc2\x22\xefx\x17N\x0d\xbb[\ \x9a\xf7\xaa-\xe70\xb4\x85=TK\x85p\xbd\xba\xb9\ \xec\xe6\xaf\xeaU|S1b\x9a\xcb\x9c\x04\xdbgg\ \x0d\xcdp\x9a\xb3\xa9\xe1\xd0\x99\x8eCh\xa6#|/\ \xdbSs\xc7\xda\x9a\x22\x9f\x16\xd0\x99\x8e\xf3~\xa2o\ \x8a\xcc\xdc\x0a;,\xcb\xf9\xc8\xac\xf4!\x19\xc9d\x91\ \x09\x82\x00\xfa\xaa\x00\xf0\x1b\x00\x86\x7f\xfb[\xb6\x8d\x7f\ 1|5\xb0\x0d\xb6\xb7k\xd7\xc4\x05\xc2\x97\xcd\xce$\ \xc5\xd0|\xfc\xee+\xc3\xba)`O\x92\xa6\xfe\xc9~\ \xc8\x99\xe1\xbe 7\xee\xc0Uh\xfas\x1b\xca\xfe\xf8\ m\x83\xce;sL)\xd3\x0d#?\xd1\xf6\xd4\xfc\x9c\ \xe9\xd9\x09\xe6\xd5u\xfc\xf2\xb2\x03\xd31\xa0\xaa\xdb\xab\ }\x9bk\x82\xb5^5\xa8\x06\xdd\x0eGv\xb2}b\ \xba}h\x9a\xe9\x15\xca\x9dpJ\x10\x84\x85\xe9\xbb\x02\ \x00\x95b\xb4bj\xd0o\xac\xbe\xde(X,\x19\x9a\ 9\xcbqk\xc0)lbR\xc2 m\xce\xbf\xecF\ \x83\xb1\xec2I\xf3r\x8b\xda\xec\x85\xe2\xd4N|\xcf\ \x955\x017u?\xcd\x09\xf0\xdf\x8d%7|\xdd\x08\ M\x9b\xe6\x9a\xbe\xa1Klp\xa2m\xdd\xf9\xc3\xb1'\ \x08\x95\xa1\x936\x8a\xf0\xc9\x904\xa6+pexy\ \xfcH>\xae\x08#\x07\xef\xf9\xa3\x1b\x12\x00\x82 \x80\ >)\x00-\x98\x81\x87o\xadz;+[\xee+\xdf\ \x22Un\xb0\x1b\xb5\x92\xae\xa1\xa5\xb3\xa5\xf8\xdd#]\ yGH\x99\x87\xb1\xec#m6\x1bXBu\xff\x22\ \xe3\xcb[\xa4`#:\x80\xa3\xb9%d\x8aK\x9a\xfd\ \x88m\xd8\xa9\xb0\xb1\xa7\xfa\x82\x82\xaa\xf6\xaf\xcd\x95\x7f\ \xfc\xbaN\x97\xf0-[\x184n\xa6s\x12\x95\x17\x8f\ \xcd\x9d\x94\xe5B[\x8e\x8f\xed\xc6\xaf\xef\x9e \x08k\ \xd3\x1f\x04\x80/\x81q\x04\xbb\xc9\xff\xd0\xa8\xa3\xb5l\ \xfe6i1\x9aF\xb0r\x0b[r\x1e\xbe\x16\x18w\ \xa2\x1b<Rq\x1a'\xbc\xeb\xc8\x1a\xdfs7\x03\xf0\ \x1e\xc0k\x9bKo\x5c\xd3\xa0\x19\xaa$+f\xfb\x04\ \x14ip\x82\xf2\xc9\x19\xc3\xb2p\x1e<H0\x94\x00\ <\x80 \x08\x226\xfa\xdc(\xa0v\xe05b\xb4\x8b\ \xfcN0,\x9b\xf6\xb1\xe5\x1b?&\xe6\x16\xc5=\xc0\ H\x19i\x14\xaf\x90t\xbf\xf9:/\xdcahl\xcf\ \xdb\x9a;W\xce\x18c\x1e\xdb\xec\xbe;ac2\x13\ S\x14myi\xe0\x00\xfd2\xea\x83\xfa\xbb\xbbj\x8f\ \x18\xe8\xccM\xb2C\xb0L\xa7=\x13\xc0>\x82\xa8\x19\ \x98\x11t`\x15\x87\xc7\x19\xb6\xaf\xf8] q\x1f\xc8\ \xdc\xdf\xafcSD\x01\xfe\xe0\x95cq\xe9\xe2+6\ \x93\xa0%\x8f\xe2\x1a.\xf2\x13\xf3\xa5\xfe\x06\x5ca\xcb\ \xc5\xc2e\xe2\x85\xf2k\xee\xdd\xd7\xda\xb7[\x00Q\x83\ \xa3\x88\xaa6J\x9f\x9eo\xa8>3\xbd\x18\xce\xc1\xa0\ 0{\x82>\xf79\xc7\xe0\x99<\xeb\xf6@\xd2Ar\ \xa8\xba\xf6\xd2w\xc5\xbf\xff\xb6I7tY\xb2\x09#\ %IC\x13\xe5%g\x0eOq\x81#|\xc5\x16\xf5\ \x05u\x80\xf9lD;\xb1#\xcc\xd0A@d\xc2_\ \x7f\x8eL^\xcc\x0f\xbax\xcc>\x07\x99\xe98\xd2\xe2\ 3\xffm9;\x8fj\xde\x8e\xc5\xbd\xf8\xd5\xb7ie\ ?\xc5\xf5\xf0\xcb\x84?\xbcd\xbe\xde\x8b\xb3\x96U\x05\ \xc0\xd0\xc1B\x18\xfb\x16j\xabof8.\x08\xfb\x8f\ d\xd8\x0c\xfb\x14'\x9b\xf5\xa8m\xd8\xc9,\x8c\x89\x8e\ \xe3\x0e$\x07\x84\x0b\x02\xf7\xf8\xfa\xf2\xfb\xbe\xab\xd3\x0f\ \xc8=\x10\xb6\xe1I\xd2?\x8f\xce\x9b20\x81o\x83\ =\xbd7c\xf5,\xf8\x0c5\xb4\xefT\xcd\x13\x94\xca\ =\x81\x0ao\xb0\xc1\xaf\xfaU=\xa8\x1bv\x99%;\ \x95L\xb7=7\xc9\x95\xea\x92\x9cv;\x94Ts2\ oqp\xff\xa2\xa5\x80\x9b\x0b\xf0\xcf\xaf\x14'i2\ \xb7wi.\xe2\xd9\x19\xcd!\xff\xe6\xa7\xe5\xdb\xcd\x9f\ ~\x90\x81E\xac\xe27Fms\xf3\x0a\xbe\xf93\xaa\ X\x0dAL\xc7\xbd\x10\xab\x0a\x00\xff\x03\xd4\x8a\xcd\xc6\ \xd2\xf3\xe4\x80\xc7\x1c\x84\x8f\x19\x15\x12\xcb\x96 \xcd{\ \x1a'\x1d\xc2't\xbbU\x06Z\x92C\xd3\x8d\xffl\ ,\xb9\xed\x9b\x06\xec\xa2\xc2\xdb\x03|\x14\x8fd\xe4'\ \xd8\x16\x9e:8\x07\xfb\x82DN\xeb\xa9\x9b\x16=\x8e\ \x99\x84\x10cX\xce$\x19\xfea\xd9\xe3\xf3o\xab\xd3\ \xff\xb5\xb9bmY0\xa8K~\xdd\x00\xa3\xafAq\ DI\x10\xc8\x0c\xdaz\xcc&3\xa7,'\xda\xd8Y\ \xc3]?\x1d\xea\x1c20\xd5\x06i\x8d\xf7W\xc0a\ s\x89\xe5\x11,\x96#\xc1<\xd9\x8f\x05\x1f\xd6\xc3{\ 3(\xcf\x99f\x00\xb8Csn\x0f4\x9c\x9dL\xdb\ $r\x0e\x1ej\x1e.\xa9\xba^Y[\xbf\xa36\xb8\ \xa2\x5c\xfb\xba\xd4\xb3\xbfQ\xd3d\x1b\x1f\x1cfd9\ \xf4I\x99\xaey\xb9\x09\x93\x92\xd5\xdc\x81YN\xc5l\ h\x8a3\xe0\xe9\xc0U{\x83\xea\xdar\xf0y\xd1\xea\ a\xf3+\xa8\xaa\x15\xd5\xf5?x\xe4Ue\xbe\xf5\xa5\ \xf5{\x1b\x0c\x1d\xce\xc2I\x93\xd5\xf1\x19\xaeY\xf9)\ \xd3\x93\x82C\xb2R\x12\x9cN\xd4_H\xc0\x96 `\ !\x84\xcf\x8f*\xd5#\x98Y\x0b+a\xe6*\xff\xe5\ \xa9\x88a\xf3\x07\xb5\xa2\xea\x1a\x88\xd55\xd5lSE\ \xd3\xeez\x9fOrh<\xfd\x14&;\x8d\xc0\xb0\x14\ \xdb\xa4\xac\x84\x99\x99\x8e\x91\x89zvV\xba\xcb\x0e\x99\ \xce\xf4\xa9%\xae\xb9>\x08\xef{\x06\x8b\x0a@\x0b\xba\ \xae\xea\xfb\x16\xe9k\xccqA\x90\x14\xcd\xe9+;\xa4\ 9\x7f\xb7\x0f=\xb1\x07\xe7\x8a\x08\xe8\xc6\x93_\x97\xfe\ ec\x03\xd6&\xf8\xfd\x0d\xbe]\x1a\x93\xaa<3?\ wl\x96\x0bVa\x9bu\x05\x00\x93\x0a\xff*\x1a\xbc\ \x1b\xea\xa5\xcd\xa5u\xcbJ\x83\xeb\xcb\x9a\x82\xb0\x0d\xe2\ \x05\xfey\x94\xf1h\x0b\x01>N\xa2\x83\xed\xb1IF\ v\x82\xed\xd4a\x89G\x0cr\x1d\x96i\x1f\x9c\x9a\x00\ ;\xcd\x18\x8f\xce\x06\x99s\x96\xe0\xd1\xfc\x8b\x172\x0c\ j\xa7\xe5\xbd\xc6\x17\xd8X\x11\xe0\xb3\x5c\xa1;\x87\x22\ \xcd\x18\xe4\x96\xb9\x81\xec8 pF0V\xde\xa0\xbe\ \xa36\xf0]Y\xd3w\xd5\x81\xc5\x85\xdeZ\xaf\xaaq\ \x7fx\x94\x80\xdd1\xbd\xf8\xd1\xd0\xda\x99\xe4\x90\xd9\xcc\ \xbc\x84Y\x19\xf2\xc4A\xc9\x87\xa6;\xb2\x12\xa1n\x81\ \xc3\x851\xfa\xc2\xb8vn@\xcct\x90\xea\xbc\xfe\xad\ u\xfa\xb7%u_Ui_\x9445\x05\xf4 \xf8\ \x80\xd7\x8e\x0eM\xdfxT\x98@\x9b\x1bT\xc6\x18\x93\ \xe1\x9e?\x90\x8d\xcfN=,M\x19\x96\x91\xc8/V\ \x84\xb9\xe3g_\xba\x1a3\x83\xf1x\xc3\xc0\xf8\x02\xc1\ \xed\xb5\xc1-\xd5\x81\xed\x15\x8d_U\x1b[\xaa|\xaa\ \xa1A\xf4\xf2\x81\xe6\x10bp\x86\xb5~X\xe1\xd1\xc1\ 7`\x8c3\xc8]\xb0u\x5c\xa6{N\xba>&;\ mB\xbact\x9a\xd3\x09I\xcb\xa3$\x9cH\xee:\ ,/\x00\xd8Y`\xe8\x95\x9b\x8d\xcf.dj=$\ \x18~x\x8a\x18\xf6\x14}\xce\x13\xae!Gs\x87\xdd\ \x0d\xef\xc5\x96\x82\x9a\xfe\xef\xf5\x85wn\xf4B*\x99\ \x85\x01\xb3\xa4\xc4\xc6\xa6\xb2\x97N\x1c<,\xc5\x09!\ \xb5\xa6\x00`\xe14\x8c\xc2\xda\x86\xc7\xb6\xf8>\xdd\xd7\ P\x1b\xd0}\x9a\x0e\xe5\x0f\xe2\x02L\x8e\x99\x86X\xec\ Db\xb6\x0fw\xc7\x93\x1cm\x91\xa6K\x8aC\xd6\xd3\ \xec\xb69y\x09wNI\x1e\x9c\x9e\xc4mQ\x07\x1e\ \x84\x04\xc2\x06\xdf\x98\xb7\xf0Wl\x80\xa2\xae\x80\x8f\x1d\ z\xb8\xb6\xa8\xe1\xa7KJ\xa1\xd5b:\x82`\xfdb\ l\xc2\x1d3r\xa0R\xd9qH|\xc1\xe0\xff\xb6\xd5\ >\xfbCC\xb17X\x1f\x80*\xbc\x8c\xe3\xca\x18\xd7\ 7\x11\x86\x03\x81M\xe0\xb7Y\xbb\xc5\xa0\xc1_\x82\xc2\ \xd2\xed\xec\xc4\xe1\xa97\x8eO\x1c\x94\xeaf`\x89;\ \x8a\xbcf\xb0\x041_P\xfb\xf77{_\xdc\xaf\x97\ \xfa\x15o0\x881\x0a\x07\xf3\xf8\xe5^0\x1e\x0c\x8c\ \x02\x0c\x0b\xdf\xce\x8f\x85P\x80na\x18l\x8c\xa5\xd8\ \xd9\xa1\xe9\xae\x9b&'\xcd\x1c\x9c\xce3u'q\xd5\ \xd5\xf0\xc4\xc3\xa0V7y\xff\xb9\xa1\xf6\xad\xbdM\xf5\ \x01\xbd\x1e\xd2\xc6\xbc*\xf81m<\x82\xe1\x14Z\xc7\ \xc3\xcc\x0f\xc5x5}\xc0\x0f\xbf-\x00K\xc9\x0e6\ \xc0\xa9\x5cx\x88\xfb\x8aI\x03\x5cv|y8\x1e\xd5\ CX]\x00ZP\xf7,4V\xff\xd6P}\xbc\xf5\ \xcb\xad,$\x8c\xe2\xd6\xa6\xfd\xc91\xf2L&\xf7X\ :A1~hm\xf1c\xdf7\x04\xa1\x0cq;\x85\ \x16\xdf\x90\x86')\xcf\x1e\x9b3!\xcb\xc5Wyp\ {4'u\x03-\x05\xd2\x1b\xd4V\x157\xbe\xbe\xa5\ bq\x85\x04u^,Z\xbc\xa8\xc5\x0b(\xe2\x8a$\ \x9d>\xc4y\xf6\xd8\xac#\xb3]\x89PYCx,\ c\x01\xef\xe8\x5c\x10Hp\x00\xdfe\x8d\xfe\xeb?\xaf\ \xda]\xef\x17A3\xf4S\x06;\xee\x9e\x93o\x1a\x88\ P|YT\x7f\xce\xa2\x12h\xc44c8\x15\xf9\xcf\ \xd33/\x1a\x9f\x01\xcbhR\xf1~\x05/\xb6\xa6a\ \x91\xf4\xbdu\xc1O\xf6\xd6?\xf7}\xf5^>\xb0\x19\ \xb7\xc5\x92\x13\x0c=\xc9a\xbbp\x98\xfd\xbcqY\x13\ \xb2\xcc\xbbM@\x9b\x0b7\x13\xc30\xca=\xfe\x8f\xf7\ {\x9f\xd9\x5c\xbb\xb3\x1e\xab\xfb\xb1\x03\x9eLK\xd6.\ \x9b4p\xc1\x88T\x97\x1d\xda\xdf\x98\xb8\xe6%\xb7\x0a\ @\x17\x80\x92\x89\xf1\x87\x97\x064\xf9\x83+\x0a=\x1f\ \xeem\x5c\xb8\xdf\xefk\x9e\xa9%.\xf0\xc8\x932\x9d\ \xec\x82C\x92O\xccO\x9a\x9a\x93\x8833\xf2\xc4\xeb\ \xeak<\x08\x12\x00\x81\xae\xabj\xe9ji\xd9\xe58\ W\x04\xcf\x04<rt\xe6L\x93\xe6<a\x1f<\xb7\ \xa7*\xda\x10\x06\xa8R=\xbe\xae\xe4\xe1\x1f\xfc\xb0\xda\ \xfcL/f\xd2\x09i\x8e\xff\x9e\x94\x97\x9d\x08\x01C\ \x0b\xd5\xcdY\xa7\xfb1$\x0d\xae~K\x85\xe7\xf7k\ \xab7U\xf9= \x89\x98P<\xb1\xe2|\xe9hm\ \xe0\xc7mS\x0e\x1fh\x7frvzNZ\x92i\x1a\ x,wr288\xa8\x19\xb7\xae*~uG#\ \xe6#\xd3=\x93\xce\x1fj\xfb\xc7\xf1#\xc4j\x08\x0e\ \x12\x00\xfc5\xb4T\x87\xfd\xf9\xe3\xb3\xe7\xe4$A\xa2\ \xe3\xad\x80f\x01\xd0\x0c\xed\x99\xef\xca\x9f\xdb\xee-\xf4\ \x04t\xac\xd1\xc3F\xae\x851\xe4\x04\xee\x01\x9cV\xcf\ v\xdb\x7f?)\xf9\xfc\x09Y 8\xed(\x1f\x9e_\ \xdfW\xe3\xbdby\xc9\xe6\x1aM\x83\xa3\xa0\xb0\xc4p\ \xde\x16\xb0#\x05\xdfW\xc1&e:\xfe8=kj\ v2\xe6z<y\x1c<\xef\x18a\x0c1\xf1\x8d\xed\ e\x9e\xdf\x7fU\xf5M\xa5\xdf\x0b\xa6\xbfy\xb2\xdex\ \x817Ip\xdc\x89b0=\xcd&\x9d6\x22\xf9\x81\ 9\xd9\x0a6\xb9ZGr\xd7\xd3\x1f\x9e\x03\x88\x0bP\ \xbb\x90\x93\xf2\xa5\xe4\x91R\xe9J\xa6\xf9\xb1\xcb\x15\xb3\ \x9d,\xeb^i\xff\x87z\xda8\x962\x0cR\x07\x5c\ vs\x0a\x01vE\x99\x93\x9ff\x04|_U\x04$\ ]\xe6\xa6\x09\xc3R\xe9SW\x166\xce\xceM\xcat\ \x81\x00\xf0\xb0\xf1\xbf~\x03/\x90\x10\xddX,\xa1\xcc\ l\xab\x0e>\xfaM\xc9o\xd7T\x154\xaaA\xbd\xd9\ \xdc\x80\x1b,8q\x06\xfc\x06OU\xdd\xd8\xdf\xa8>\ \xf7C]Yc\xe0\x90t[\x9a\xd3\x81\xe74\xe5&\ D60M\xf3\xf3\xdf\x95\xfdcK\x03\x84\x0c\x9d\xc2\ \x05p\xdf&\xa6*'\x8d\x84\x8a|G\x146\xf8_\ \xdf\xd9x`m\x13\xce\xe3\xd5\x8dOv\xd7\xcf\xce\xb2\ \xe7\xa5\xday\xd7\x0d\xd4W\xf4o*\xbd\xd7//y\ i\xa7\xa7\x16\xe5\x02\xf3\xaa\xe8\xee\xc1xi?l\xe1\ \xd0|\xb0\xe1\x09J\x8b\x0a\x9b6W\xfbf\x0er%\ c\x1b\xa8\x95\x9f^U\xff\xd7\xa6\x9a\xcb\x96\x95\x16y\ \xb0\x0f\x074\x22^\xe5\x02\xcfo\x18\x9a\xc4J<\xea\ \x1b\xbb\x1a\x0d\xbfwxzB2\x7fu\x11\xe6\x06\xf3\ ;N\xe7jA\x98~\xce\x8eZ\xff\x1f\xbf\xac\xb8\xed\ \xab\xaa\xbd\x8d\xaa\x8a\xa6_FA\x8c\xeb\x19y7\x03\ D)\xfa\xeb3\xd8\xa6\xca\xc0?\xb7T\xe7'\xd8\x0e\ Is(\xfcD\xe6\x17w\xdb\xb5\x90\x00\x08 \xdba\ J\xa7\x8e\xd03\x0e\x97\x0a?\x92p\x92Q\x04\xf25\ \xe4FV\xb2BM\x1ciK?\x04]\xc6;\xf3u\ \x0c\x9c\xce\xe4\x88\xec\x04%\xd0\xb8\xaaJE\x8b\xc73\ \x07l\xac\xf0\xeb\xab\x8a<\xc7\xe4%\xa69m-\x19\ \x06\xb6\x8b\xa5>\x0eT\x06y\x99\xc7\xeb\xf9\xd7\x0f\xb5\ 7|^\xbe\xba\xc4\xcf\xab\xba\xc0\x8fW+\x16\xe2\x09\ \xf8\xc9O\x8b\x7f`\x5c\xe5\x8d\xd5\xea\xc2=\x8d\x8a\x1e\ \x9c\x9c\x99\xc0\xe7d\xc2\x141\x9d\xb6\x02\xcd\x88\xf1\xec\ \xf7u\x7fXW\x09&\xac\xc5\x89\xf9{h\x9a\xed\xa4\ \x91\xe9|1$\x07\x09\x00?\x10\xc3\xe13\xb4e\xc5\ \xfe\x9f\x8cHJ\xb4+\xaa\xae\x81\x10\xfe\xee\xcb\xda\x9d\ \x0d`\xfb\xcd\x9b\x03f\xe5\xb19\xd01\xc3}B\x1f\ w\xd5\x05?\xd9\xd7xL\xb6\x9c\x9e\xe0D\x1d\xc3\x12\ a\x04T\xe3\xaae\xc5/mm\xf4\xf2\xf6\x02\xbf\xcc\ 8\x9c\xb4\x053r!.5CZ]\x1e\x5cV\xe4\ \x99\x96i\x1b\x90\xe8l9I\xfb\x91\x1f5<\xd5\xe0\ \x13T\xf5\xe7\xbf\xaf\xbeiU\xc5\xda\x0a\x1f\xbf\x9f\xdb\ |\x9e\xf8\x9e\x0e#\x0b}\x84\x7f\xfc\xe0\xb9\xe1\xd4\xd2\ gEM_\x944BYNp\xf0\x9d\xf1>i\xbb\ \x90\x00\x1c\x04S\x92\x07K\x09C\x8c\xb2/\x98\x16\x10\ \xe5\x0aR\x08\xda\x04E\x8b\xb5\x94qJ\xea\xb0\x9e\xea\ \x0b\xb2\xc9lv~\xaa\xbf\xc9\xf3m\xb5\xae\xf1<k\ f\x91J\x9f\xbe\xba\xd8s\xec\x90\xa4\x14'\xd4\x92`\ {7e\x9dn\x01\xaeD\xde_\xe7\xbbfe\xc9s\ \x9b\xeb<\xaa\xc6\x14\xbc\xb7\xd9\xfd\x17\x08\xa2\xdb\x18\xd4\ \x97\x96\xf8w\x97WM\x1e\x94\x9c\xe2\xb4\xb5\x1b\x06(\ \xc9\xdf\x944\xfenUq\xa3\xc63\x09w\xd3\xe2\xee\ \xd04h\x01D&\x00?b\xd8\x1b\x82\xea;\xbb\x1a\ \xa6g\xbb\xef^W\xf9\xfc6\x8f\x17\xbb]\xb8\xff]\ \x1b\x1b\xac\xc6\xaf\xad*\xf6\xce\xc8\xb2\x0dHr\xc0\x09\ =A\xfd7\xcbK?\xdc\xe7U%\xdd\x86\xad\x9d\xae\ \x0b\x00x\x8c\x95o\xc8\xe1/\xef\xac\x1f\xe2\xd4\xc7\x0d\ pC\x85\xdc\xdca\xba\x88\x0b\x10\x8fx\x16O\xf0/\ _W>\xb6\xb1\xb6!\xa0as\xaaK#\xb55\x90\ \x8e\x86bh\x1a+\xf0\xe8\x9f\x17\xd5\xcf\x1c\x94\x98\xe5\ n?\x83\xc5\x9d^-\x00\x07\xb4\xcb\xcc\xba\x87\xd9?\ \x89Y\x0eW \x82`\x19r\x08\xdf\x19\xa7\xe8B_\ \xe4\xb4\xd1,c\x92^\xb8H\xd2\x03 \x01p>h\ \x97a;\xa0\xecs-y\xac-u8\x86\xa0[\x92\ \xa7\x15\xfc*\xa7f'i^\x0f\xd6/q\x9d\xf1Q\ \x8c\xac\xc2\xa7\xad)m\x9a?81\xc5\x81\xf3a\x1c\ \xa0\x0e}\x12LU\xbc\x06\xf85\x16\xee\xaa\xb9fE\ \xc9\xfa\xca\xa0y\xfd\x98\x13z\xe8\xc20\xba%\xe9\x87\ F\xb6\xa4\xa0qB\x8a\x94\x9f\xe2n\xde\xcc\xbb\x840\ O\xe8o\xee\xa8\xbb}uY\xa9\x97w\x9a7\x87S\ \xfc\xc4(\x00\xbc\x1f\xdc\xa3I\xef\xedi\xd8T\xe5\x87\ DG\xef\xcd\xbb\xdf-'\xe8\x0ax^\xaa\x0c\x18\xab\ K\xbc\x0b\xf2\xdc\xc9.\xfb\xcd\xab\xca\xdf\xdf\xd3\xc0\xef\ E\xf1\x8b\xee\xca\x00\xc05\xf2x\xc4\xf1L+K\x03\ M\x01\xed\xc8\x9c\x04\xd8\x80\xba\x13\x97\x13\xa3Gx\x85\ \xbb\xeb\xfc\x97/)\xfc\xb8\x10+\xfe\xb0\xda\xa5\x17\x15\ \x0a~\xadF\x99W_\xb4\xcf\x93\x9bh\x1f\x9df7\ 7viHz{\x0b\x00\x13\xc7\xd0\xab\xbc*\x0e\xbf\ \xad\xf6/\xd9[\xb3t_\xed\xf6\x1a_\xbd_\xab\xf4\ k\x89\x8a\xe1Tp\x80\x9cY\x1cbGd8&\xb3\ \x94!\x12\xce\x17\xb4\x1c\xea\xfe\xbc\x89\xcd\xbd\x87\xe5\xc2\ \x8f\xd4\xe41\xbc\x1d`\x8e\x0c\xe9>\xcc\xa09\x14\xf9\ \xc8\xa1)~O\xd3w5Z\x10\xc7\xfa\x99\x01f\xe5\ ^um\xa9\xf7\xa8\xbc\x844\xa7\x08\x18l4\x17\xfa\ \x1e\x5c\xd75]\xfa\xe7\xc6\xea\xdb\xd7TT\x07x\xfc\ \x9b\xd7\xdf\xa5\xa5!4\x18\x97xr(\xa1P#V\ _\xdf\xdd$\x05\x03S\xb2\x13\xf0\xbdj\x98Ia\xb7\ \x01U\xc8\xbb\xd6V|_'\xba\xad\xf8\x1f~Z\x88\ E\x00\xf0\xda1\x10R\x00M\x14.p\xcf\xcd\xc5\xae\ \x84\x9f\x06\xfe\xab\x03\xc6\xd7\x95\xfe-\x95\x9e\x97\xb6\x81\ \xf5\xc7\xf0\xe0\xc9\xf9w\x17\xc1/\x97\xc3O\x03\xf5\xf2\ 5e\xbeF\x7fpv\x8e\xdb\x8e\xddp\x18\xcdf\x9c\ D\x07\xaf_B6c\xdf\x97{.YR\xb4\xb5V\ 5+\x17]zQ\xed\xc3\xe3\x12\xfe0D\x12\xb65\ W\x157\x1d\x9a\xe1\x1c\x96j\xb6\x03\xba04|8\ Ao\x05\xc2\xd6\xe0\xf3\xffcs\xfd\x8a\xc2\xc6\x1fj\ \xd4\xa0\x86\x93\xe5CU\x08\x0a!\x18e\xbb\xa2O\xc9\ t\x1e\x95\x97\xfc\xebI\x19\x90#b\xc9\x0am1p\ t\xb3f\xec_\xaa\xaf\xf9\xad\xe4\xaf\xc7\x882\x13\xc8\ \x90dG\x9a4\xe7\xaf\xb6!'\x08\xa7\xdd\x8b\x99k\ \x1b}\xeac\xeb\x8a\xfe\xb1\xcd\xd7R\xd3\xc4\xcd\xcc\x98\ \x9c\xe1x\xe6\xd8\xbc\xe1)\x0e\xa8\xd7\xf4\xecC4\xb1\ \xa0\x1bz \xa0\xdd\xf1e\xf9\x7fw{\x02\xba\x81M\ \xb0.,\x02\x11``\xde\x83o\x88q\xc3\xa1(\xc7\ \xe6\xbb\x9e::\xd7m3\xe3\x99\xfdrY\xc9\xfb\xbb\ \xeb\x0d^3\xc6\xf0\xb6\x09\xf3\x05C\x1d\x7f?~\xb8\ X\x09A\x9ba\xa0\xbd\x07L\x06\xb86\xcd\x1c+\x09\ \xb1\xd0\x05\xf7\xde;\x00\x22\x1e2\xb9\xdb&];1\ \xe3\xe6\xc32e\xd0\x00\xd3jF\x8bY\x94\xd6\x975\ ]\xb1\xb4\xa4\xd0\xa3\x81G=\xd6\xbal\xc6\xcc\xe9\x18\ 0\xc62\x5c\xca\x13\xf3\x06\x1d\x9b\x9f\xd4\xa5\x81\xea\xa5\ -\x00\x88\x02\xb0\x02k\x8b\x1b.\xf8\xa4d\xf1\xfe\xa6\ 2\xaf\x86\xcfX\xf3\x98\xe0\x89\x8e\xe6X3X\xa1G\ \xff\xa2\xaci\xd9~\xcf\x8c\x81\xaet\x17\x96\xc3xE\ \x16\xf8\x83O\xed\xa4\x1db\xa4\x1fj\x14-e:\x98\ Z^\x00\x18\xd3u\xbfQ\xf8I0}\xa2\x9c8\x98\ w\x15vk\x9e\xe1\x01cN\xbb2oHZS}\ \xfdw\xb5 S\xa6\xb5\xc0\xedeM\xfa7e\xbe\x13\ \x86&&\xd9[\xac\x7f\x8f\xe7\xeap\xe1\x05\x12\x7f\x83\ A\xed\xee/K\x9e\xdf\x81\xd3``\xd0{M\xf0\xcd\ \xbc\x87\xf1\xc9\x18d\xc8\x9du\xea\xca\xa2\xc6\xa3\xb3]\ .\x87\xed\xbeu\x95\xff\xd9V\x83\xfd\x92\xb8\xb3\xfd0\ \xc7\xd6\x05\xd4\xe3\xe0%a\x19h\xc6\xdc\xdam\x98Q\ \x0f\xa5~uiS\xb6\xdbvh\xa6\x13\xda\xfe\xcdV\ !2ZrZ\xa5G\xfd\xd9\x92\xa2}\x8d\xfc\xaeR\ \xf7_R\x1b\xcc\x00\x98\x01\xf1\x06\x8d\xef\xca\x1b\xe6\xe6\ $d\x82\xe8uY)\xee\xa5\x02\xa0j\xfa\xf3\xdf\xd7\ \xdc\xbc\xba\xb2\xdc\x8bY.d\xda\xc0\x1e]*\xf3i\ \xef\xef\xae\xcbp\xd9&f\xb9\xe3\x1fMIy,\xe3\ P\xbd\xf03Y\x0b@0P\x9ba\xa3\x1e\x94K\x96\ j\x89\xa3l\xe9#\x9bS\xad\x07\x98\x9a\x93\xe8ol\ \x5cW\x03\xb9\x03\xf2\x87\xe8\x0e.i\x0a~Y\xea\x9d\ 9\x08\xf2\x8d\xd9\x17\x04Q\xd2c!\x8c\x10\xac\xeb7\ \xf9\xfd\xd7.+\xfb\xdf^/\xbf\xf3\x22v\xf4ZJ\ <\xfa\xca\xc2\xc6\x80$?\xbe\xa9&\xc8\xcdv\x07\xb1\ \xdd\xc7\x05\xa0\xf7\xc06\x94{\xa7\x0dr\xe5\xa58\xc5\ z\xc4\x19\x05\xadJq\x83\x1f\x1am\x9b\xaa\xf1\xf6R\ \xaf\xcbh(Q\xac6h|]\xee=}\x08\xd40\ \xba\xeaA\xd4\x96zb\xaf\x00k\xd9\xa0\xce\x86\xb1p\ O\xdd_\xd6W\xd7\x06 \x1a\xc0\xae\xf1\xcaw\xfb\xa0\ =\x86\x88\xa9\x0aHw}U\xbe\xae\xb8\x91G\x1c\x1e\ %\xf6\xc7\x8e\xac\xb0\xbc\xa3\x94y\x8f\x1b\xae\x0c\x19o\ |A\x13\x1fOj\xf8\xea\xe4/\xaf\xf6\xef\xf9\xcc\xc0\ \xb7\x8f\xf5\x04\x8c%9\x9dw\x1d=\xfc\x86167\ \x06H\xc1z\x0d\x8f\x8fo+|W.-,nR\ 1\x13\xf1l\x83q\x1a\xc78\x897\x18:\x0c\x9e\x0e\ \x0d\xbf\x87\xd6W\x7fP\xe4\xc1'ezq\x80[\x80\ X\xfd\xa1>x\xf7\xda\xf2&\x9c!\x80\xe7E\xa2\xeb\ )\x0b\x18W|VX\xd6\xe0\xc3\x95h2\x0a\xa6\xd3\ \xd3\x1b+\xd7\x95\xfb{gF\xe3!\xc2R\xb1\xb1*\ \xf8\xe0\x86z\x95\x97_,\xc2\xf1\x0ej\xef\x12\x00\xd3\ P\xed\xaa\xf5\xdd\xf5eE#\xf6\xferm\x87\xff\x10\ \xe5J\xec\xe3\xbb\xeb\xfc\xc6\xb5\x9fW\x144\xf81\xda\ \xe2\x17O\xb2\x84}\x8dr\xce\x5cy\xf6\xc3\x86-\x01\ \xef\xc0\xf1\x82\x8ecm4U^s\x83\xbao\x91\x8e\ \xe3\x95\xe3\x9d2\x9d\x01\xd7\x0c\xd7-\xcb\xf2u\xd3\xf3\ /\x19\xa1\xf0\xf7\x19\xc06\x0c\x06\xc4\xc9\x96Z\xf5\x92\ \xc5\x85{!R0\xe3tw\xd8\xa2\xc0\x0c\xe7S\x9b\ \xea\x9e\xdb\xda\xc8\x0c\xf3\x99H\xb1\xab7\x83\xf9\x0f\xe7\ \xe9\xc4\xb0B\x8e0\x17\x88\xae\x052\xba!\x95\xf9\xd9\ }\xeb\xab\x03\x1aT\x1a\xf8\xa6H\x80t\xfa\xa6\xb4\xf1\ \xd5]MPf\xc0\x02\x862/=\x08\xde\xd2\x040\ l\xc6\xbb{\x1a\xb6Wy\xb9m\x84\xcb\x8csY\xee\ ]\x02\x00\x17\x1c0\xf4\xbb\xd6V\x96\x04x\x89\x8a\xc0\ r1&\xcb{\x1a\x02\xcfl\xac\xc1>qH\xe1\xf8\ \xc2\x149\xefX\xf9\x98\xe7%W\x0a\x0f&3\xb3\x8e\ \x1e\xac\x93V]\xaf\x15\x7f-\xc5u\xaa\x90\x88Hr\ *\xf7\x1e=\xe2\x17\xc3\x15;\xce&\x82O\xb1\x9b\xdb\ 7T\xf9\xaf[QR\xe3\x17\x0d\x94\xdel\x9b\xf84\ \x8a\xd2\xdb\xbb\x1a\xff\xfcM\xb9\xef\x80\x19\xd4\x08\xa2]\ \xf8\xbdx\xf6\xfe\x9e\xc6\xf5\xa5\x0dQdk\x8f7p\ \xc3\xe7\x95uA>\x94\xbc7\xc3\x0bs\xa5W{\xf0\ \x9br\x8d\x0fP\xe5\x95\xbcx\xd2\xcbZ\x00\x12\xdbU\ \xe3__\xee\x15\xa6?\xfc\x8b\xe5V\x0f\xd4rqA\ ciC\xc0\xdc\x16O\xb0\x99\xc1\xe4A3\xe4\xd9\x8f\ J\x8ed\x1c\x97\xcc\xc7\xe0C\xfb\x00\xdf+\xb9\xe2\x97\ \xc1\xdd\xef\xeb\x5c\xa3\x85\xfbn\x05m\xfb\x1ds\x87]\ <\x14\x9b\x04\xa2.\x8d5\x05\xb6\xb6\xdc\x7f\xf6\x87\x05\ \xdbj B`\xd5\x0c[\xcbB/\x02.\xa0\xca\xe3\ {\xe0\x9b\x0aU\xe3s\x19\x88\xcd\x04\xd1>\xdc\x14\x1a\ \x0d\xaav\xcf\xba*\x8f\x1f\xdf\x9f\xca3}'\x19\x1b\ \x1f,\xe7%cia\xd3\xaez\xb3P\xc4\x07<{\ s\x00\xf8r\x9ch6\xf9\x9f\x15\x07?\xdfW\xd3\x15\ \xe5\xb6\xd7\x09\xc0\xba\x92\x86j?\x1f\xea\x8c-\xa0p\ M\x01:\xc7o\xb9\xc8\xab}WZ/\xb6\xc6\x0f3\ 8\xd0\xd8W\x06\x1f/\xcf\x7fAr\xa6\xe0\xc9\x0ch\ \x07`\xa2\xe8\x81\x1a\xb6\xf6fo\xd1\xba\x1ei\x07\xf0\ \x80\xc9I\x0e\xdb\xfd\xc7\x8c\xbc\xfc\xd0T|f\x0d\x22\ \x83Wn`\xc7\xe6\x1a\xffu\xcb\x8b\xbd*\xce.\x8d\ 9\x14\xc3\x1bn\xacv\x1b\xd0J9\xf1\xfd\xd2\xbd\x8d\ *\x04\x18S\x12\x13\x93 \xda\x073\x88\x99\xeb\x0dy\ ]\xa5\xfa\xce\xd6*\xcc\xd6\xe1X]\xde1\xa0K\xda\ \xc2=\x8d8\xa1\x0a\xefJ\x8a\x0b\xbc\xc8a\x80\xd0{\ ,_q\xcb\xc0f\x89\xf0\xeb\xfa\xcb\xbb|A\x15<\ \x8f\xb3\x85\xe9]\x02\x00i\xf2Ea\x839\xcfAt\ \x044\xe9\xa3\x02\xafX\xe9\x02 \x84\xca\x80\xc3\xe5\x99\ \x0fK67N\xbd\xc5+\xac8\x03\x83\xe6w,\xff\ y`\xdf\x22\xe1\xae'\x00\x19\xb8\xfd\xf0\x01?\x1f\x9f\ f\xae\xf1o\xcc\xf4\xdfV\xf9\xcf\x5c\xb8o;\xb6\x03\ \xf8\x86p\x8aJ\xf7\xf2\x9f-U\xfb\x9b\xf8\x0d=\x82\ \x08\x1b4\xb4\xccx~WS\x8dO\xfd1\xc3w\x00\ wR\xe3\x09./\xc5\xc1\x85\xb0\x1e\xc7\x92\xc0G\x88\ 0\x05?\x80\xa8n\xf1=\xf1a}ES\x89G\xeb\ \xe7]@\x10m\x1b\xea\x19\xf6\xab\xc4\xc0\xd6\xaa\xa6\xb8\ \xa6l+\xf0\xe6\x8c\xac(\xc3NR\x8e}\xc5\xb0'\ \xf1>:\xec\x0f\xc2\xcc\xa4y\xe4/o\x08\x96|\xa5\ \xe9Z\xf7\x1bY\xac\x101|\xdb\xed\xfd3\x07^6\ \x96\xb7\x03\xb0\xc1\x0b\x8a\x85c\x9c\xbf\xab\x0a\xfejY\ \xb1W\x03m\xc5\x80a\xde\x84\xb8\xeei\xf8\xd3v\x86\ \xc7\xef\x7flS\x03\x9fY \x8e\x99\x1b\xae\x9c\xe9L\ \xe5\x05\x86\xbf\xb6\x09/\x1a\xbf\xb1v\xc1\xa3\x85\xbb\xe2\ 1\x81\x91\xd2\xf3\xb1\xd1E\xe0\xb5\xf1\xeb\xe5\x91\xcd\xa3\ \x1c\xcb\x99X5c@8\xed\x1a\xf89\xf8\x82y.\ s\x99o\xc7\xa5X\xe0M\x81\x1fj\xb5o\xcb\xa1\xf6\ \xd0\xb9o\xa6UxqKE}\x80\xcf\xef\x8c\xf0\x1d\ \x11\xd3\x12w\x92\x8dI\x03\xdc\xca\xa5\xa3\x13\x97\x9c\x94\ \xb5\xe9\xcc\xccMg\x0f\xdcpV\xe6'\xa7\xe4\xfd\xf2\ \x10w~\x92\xe2\x80L\x8dN\xa3/l\x98R\x98C\ \xa5r\xbf\xbe\xa9\xa46\xeem\xe3\xde%\x00\xbe@@\ 5\x9f\xb5\x8c\x81\xba\xa6 \xb7&]\x0b\x1b0Y\x99\ q\x1f\xb3'\xf1\x14\xe1\xadR@\xf3J\x9f]\xa8\xef\ \xfd8\xae\xb6,2 S\xff\xdf\xf4\xccK\xc6\xa6\x9a\ =A-y|k\x8d\xff\xac\x85\x85\xbb\xeb\x9b'\xd5\ \xe9\xb1\x00\xfe\x08XgU7\xae\xfchw\xb9\x0f\xbb\ q\xa1\xa8\x88\x1dq\x811\xc5P\x18\x96=\xd9.I\ I69\xcb%\x0ft\xd9\xd3\x1d\xb2\x03J-X\x03\ \x14\x07\x88\xa4\x96\xbb\xe6\xfd\x16\xc8\x06\xd0p\xc5\xca\x0b\ F9\xf6\x9c\xf0|\x81\xcb\xb8\xb7\xabs\x828\x0f\xc4\ 2\x9cT\xe6*\x1c\xcf\x18\x87Z\xf1\xff\xb6U\x85\xe5\ %\x0e\x953>-\x0a\xc4xz\xc8Q\xe6x\xafi\ \x19\xec\xa9\xa3\xb3\xbf=\x7f\xd8\xfd\xf3\xf2&\xe4ff\ ffe\xa6\xa5\x0f\xcc\xc8\x9c\x92\x9d\xf8\xe7\xa3\x86\xac\ ;\x7f\xc4\x9b\x0b\xf2N\xc9\xc7\x07\xd6\xf8\xe5G\x0c\x1e\ \x83)\x85y\xd4&\xb1\xc2&s\x06\xb0x\xd2\xbb\xa6\ \x82\xf0\x05\x83s\xde\xdc\xbf\xbf1\x08\xd7,6E\x8a\ !\x0ds\xf8\xbf\xfa\xf9\xa4xG\xd4\xc1\xa0,K\x86\ V\xbd\xc9X|>S\x9bZz\xe6 \x9d4\xc5\x1e\ <\xe1\xe3\x84\x01\xa3\xc5\xa6\xee\x05\xb28\x97P\xe3\xda\ \x95eo\xec\xac\x97\xa4 \x93\xec<\x95\xb1\xf4M\x1f\ \xe0~\xe3\xe4|\x17\xbe\x99P\xb8\xefA \x0e?\xd9\ \xe7\xb9\xe8\xb3b>\xa7\x9d&\xf3\xd7\xda\xc4\x07\x88\x04\ IJs\xbb\xe6gK'\xa5\xf9O\x98<\xc2m\x07\ \xcf\xf1\xc1Qh\xb0\x19\x86\xbc\xb5\xa4\xf2?\xdf\x15,\ \xaaO.lD\xa7\xa8\x94\xe2\xc8.\xa4\x87\xa6\x82\x00\ \x09T\x86%)s\x12\x9b&'kN\x87}gu\ \xd3\xb7\xd2\xa0U\xa5\x1e\xa8'\x99\x86\xacK\x81\xdc\x98\ \x9dd\x9b\x97\xe4\x9d\x92\x1cLp\xda\xf64\xe8\xeb\x83\ \xe9k\xcb\x9a|Z\xf4\xa5\xbc5\x86K\x96\xd7\x9e\x9b\ \x9f\x9d\xec\xc4N\xd9\xd0@\xa4\x96\xd5{\x8f~o\x7f\ 5\x9fBEl\x8d\x1c(\xec %\xc7\xe7\xd8_>\ \xf9\x10\xfe\xd6\x10\xf0\x0a\xbe\xc5,\x8c\x12>\x9f\x8c\xb6\ \x81W\xaf!\xbf\xc9\xb7/\xdb\xfd\xdc\x9e\x00w\x16\x19\ \x98]%\x09\xd2\xee\xf2\xfc\xe0/f\x8c\xc5w\x05\xc4\ \xbb\xdc\xf6.\x01\x80\xa2y\xd2k[\xbem\xb4G\x9d\ +!\xca\xe6d\xd9\xde9cd\xdcc\xaa]\xc0\x84\ \xe9\x05\x8b\xb5/n\x90T/\xafM\xe2&\xc8\x01\x92\ \xe2\xd4\xa7\xdd\xef\x18u&\x7f`\xa0;B\xd2\x16O\ P\xfd\xc3\x9a\xca\xffl\xad\xe5\xf6\x0d\xa7\xa81C2\ y\x80\xf3\x89\xb99\x87\xa4\xdb\xb7\x95\xd5\xaf\xa9\xd6v\ TzJ\xab\xeb|A5\xd1\xe5\xc8\xcbL\x1f\x99\xe1\ \x9e\x9dn\x8c\x18\x98\xce/\xc4\xf4)\xce\xd9\xae%\xcb\ \xf95\xfd\xe7\x8b\x0b\x97\x97\xfa\xcc\x82\x13=\x18T\x08\ !$>x\xc4f\x0ft\x1e;4yv\x9629\ '\xc5\xa6t$*\x10\x92\x1f\xcaj?+\xd1\xde\xdb\ W\xbf\xb1\x92\x0f&\xc1\x0b\x85\xe8\xd2A-M7q\ \xa4\x9b\x05@\x97\xb4\xbcD\xe7E#\x13\xce\x18\x960\ j\x004U!\x07\xfc\x98\x8e\x8dM\xfe\x8f\x0a=\xaf\ \xefh\xf8\xbc\xd8\xa7\xf1\xf7\x8c\xc6-\xa3b\x02C\x0d\ CK\xb6\xdb~1&\xf1\x8c|\xc7\x84\x9c\x0c\x9e\x87\ DN\x04'\x0d\xfe\xe0'\xfb\xbdo\xee\xaa]^\x8c\ \x83\xf1\xa1\xe0hXl\xa2\x09\x82\xe9\xe1\xbd\xd3\xd3\xaf\ \x9c4\xb0\xb3\x8cj\xac)j8\x7f1\x9c2\xd6\x0c\ \x9d\x97 \xbf\xbd wxFB\xc7f\x1d;\x7f\x0c\ V\xee\x0d\xfc\xe2\xd3\xe2\xaf+\xf0\xe5\xa0\xad\xae\x10\x03\ \x0e\xfb\xb11f@%\x88K\x08\xdf\xc6\xc6\xa6\xd9g\ d\xcaS\xf3R\xa7\xa6\xb1\x91\x03SqS\x17dH\ \xa0w\x09\x00\xc4\xc3\x9d\xab\x8a\xff\xf9}c\xd4\xe9\x03\ \xd7s\xf90\xe5\xcf\xc7\x8d\x8a5\x85\xc3\x03#\xcf\xd0\ \xd5\xf2o\x8c\xa5?cA/NR\xc7k\x04\x90\xf0\ \xb2#Y\x9d\xf7\xb2+\xf7\xb0.J\xb9N\x81\x94\xf5\ \xab\xdamkk^\xd9Z\x09\x81\xe2\x93\xc3\xf1H1\ \xa4\xf1\xa9JF\x92c]\x997\x00UoM\xe7\x9d\ \x032\xe8\x97&\xe9vf$\xc8\xd2YC]\x0f\x1d\ 3\xd4\xf4\x06.\xa7\xb3r\x15)f\xa63\xb6\x97\xd5\ \x1f\xbd\xb0P\x95\xec\xb1\xfa\x8e\xdd\xfd\xba\xac\xb3\x04\xbb\ \xf4\xdc\xb1y\xb3r\x12\x5c\x8a\x19\xe8N\x0c\x0a\x86\x03\ _>exU\xe3\xcd\xed\xf5w}Y\xeae\x0a\xd8\ \x22\x8c\x8d\x8e\x8f\x8c\x8a\xee\x14\x00C\xd2'\xa4\xdb\xff\ }\x5c\xce\xb0d|\xaf/\xf8w\x90y5\xaf\x1dl\ \xffS\xdfU\xdd\xb7\xa1\x1a\x5c\xc4+\x91\xd1\x9a\x19\x0c\ j\x18\xaf\x1d\x9f\x93\x9b\xec\xc0\xdev<\xb1nV\x86\ \x9a\x0d\x0e\x1a\xbd\x80\xaa>\xffC\xdd]_U\xa1\x06\ D\x1b\xe3\xbc\x0c\x1a\xc7\xe7'<\x7f|\x9e\xd3\xd6a\ #\xd2\x90>\xdaS{\xc5\xf2R\xf3]\xa2Q\x03q\ \xfb\xc2\xdc\x81'\x8fN\xc7 w\xec\x93\xc8\xe9\xc6\xe3\ \x1b\xab\xefYW\xd1~QB7\xfc\xe2\x0d\xc3\xc5\xd8\ \xcfG\xbb.\x1b\x9b\x9c\x9d\x86o\xf7P\xf8\xbb\x87\xc0\ \x07\xd8\xd9\xce\x81\xf1\xa0glS( *fe'\ 8\x14\xbcf\xc8\x1fbkx`v6\x8c$;;\ c\xfc@\x9e\xe1\xbb\x07\x14p\xdb\xa0\xc3\xe5\xd9\x8f\x18\ \x8e$\xc6\xdfU\x02gG\x0b\x12h\xb4-?_\xdd\ \xf7\x89a\xde\x13\xc6\xd0\xe1\x8f8\xae[\x80\xf2\xf0\xe0\ \xcc\xcc\x8bG\xa7\xe3=/\x0c\x02\x88\x13\x06nK\x9d\ \xf6y\x91\xc7\xa7B\xfb\x052\x16>?\x86\xc1C\xc3\ '\xa9P;\xd3\xd9K{\xbc\x87\xffw\xd7{;j\ x\xde3\xc3\xce}\x8c\x07\xa6WAM\xbfnM-\ \xb7\xfe\x18\xaa\xa8\xe1\xb7pu(9\xbf\x99\x9c\xb6\xfd\ \xc2\x91\xf3\xf3\x13\x13\xec \xc4X2;\x8fm\xa8v\ \xe1\x83\xbcr\x92M\xb9dB\xfa\xe7g\xe5\x9f64\ A\x91\x148\x92\x0bysX\xfb\x18\x18\xe6c\xf2\xdc\ o,\x18jZ\x7fn\xa6\xc0&\x9b{\x05\xb8\xc6\x98\ \xcb\xc6n<\x22\xeb\xe1Y\x99\xa9\x0e\xb8V\xb0\xc3\xd1\ _/?\x96\x7f\x0cv\xc4@\xc7\x8b\xc7\x0d\xceG#\ \xc6S\x02\x03!4\x95\x7f\xe3\x02\x08\x92\xd3\xae\x5c5\ 1\xf3\x99\xa3\x06eB\xfaa\xfe\x8c\xee\xec\xf8L\xe6\ \xde\xfa`\xad_\xbc\xc5/$L\xaa\xf7\xab\xd8\xe5\x17\ \xf5U\xe2E\x1a\xf9I\x8e\xa3\x86$\xa0\xccu\xe6\x91\ p\xc0\xa4\xb9\xd9\xe6\x9d\x00\xbcF\xbcL\xbe\x03\xbe\xa0\ \xec\xa5;\x95\xd3\x87\xb8~3)\xed\xe3S\xf3w]\ 2\xfcOs\xf3G\x0cHK\xb0+\x8a\x22\x8c3O\ \xc1\xa8C\xdc\x09\xbdK\x00 N\xa7\x0cJ\x18\x9c\x04\ \x0a\xa0\xf2\xf8\x0d{\x9a\x1d\x8c\x1f\x8c\xa7q\xa9\xf2\xa8\ t7\xdeH\xee\x16 gc\x05\x87)\xca\xd0S\xd8\ \x09oJ\xf6D\x0c6\xe40|\xe53T\xae\xfdl\ \xd5\xf5\xfe\xe2U\xe0R\x04\xa8\x1b\x0d\x0a\x86\x8b1\xbb\ \xc2\xee\x9f=\xf0\xc2Q\xc9<z\xd0\xd6C\xdc@\x96\ \x12\x01\xe7\x85\x11\x7fp\xc1\xfc\x85%@.l\xd2\xae\ \xf9\xa2\xfc\xa1o* \x9b\x02z\xf8i\xd19h[\ \xd7\x96x\xbf\xad\xf4\xf3\x93E\x9b\x0912!\xa9\x95\ 4\xa7m\xfd\xb9\x83\xffoz\x96\xc3\xc9\xe7\xe7\xc7\x17\ \xa7\xf3+\xc3\xaf\x8e`8\xdf\x1c\x07\x9aG\x8c\x0d\xcd\ Lzz~\xce+Gg`g+f\xbf\xee\xcaF\ \xf1\x03b\x16\xd2k|\xaa\xfd\x1fs\xf3\xb2\x12lf\ \x04\x88\xd8\xe0\x0e~\xc4\xbcp\x5c\xd0/\x1e\x9fy\xed\ \xe4\x0c\xac\x87\x9a\xc6):\xf8c1\xf05<\xd9\xf6\ \xe6\x82\xfcCR\xed\xe2\x14\xe6IZ\xd2\x82\x87\xc6\x0c\ \x10\xd6\x93\x18;\xfd\x90\x94[\xa7da|G\x07\xf7\ \xb8\xcc\xa7\x955u\xee\x03\x9f:\x02\xealb5R\ \xcc\xc8\x19$5%\xba\xcc\xf7Sv\x92u\x9bK\x19\ \x1b\x9db\xb3c\xe5\x8a\x9f\x98A\xa3[J\xb4\xb1\xe3\ r\x1c\x9f\x9e\x92\xfb\xf5\xf9C\xfey\x5c\xfe\x1dS\x07\ L\xcdIt\xd8\x1d\xfc\x10~\xd4\x8f\xd1\x07\x9f\xae2\ \xd4\xbdL\x00$c\x90\xdb~\xfa\xb0D\x1e0,\xdb\ bs\xa7\xf0\x88\x95\x99~\xfe\x98\x8c4\xb7]\xee\xfa\ Q@\x07\x01\xf9\xc2\x96>\x96\xcd\xf8\x8b\x84I\x08\x0a\ \x00\x81\xc7\x9bA\xba\xda$/\xbf,\xb8\xfbC\x03$\ \x0d]E\x9b\xf5b\x004\xe0\xa1\xd9\xd9?\x1f\x9d\xc6\ \x87\xc4\x84_\xbe\xf1=\xa5On\xae}{{\x1df\ \xe4xf\x15\x05\xaaa\xef\xefm\x14kQ\x03e\xc3\ P\x86&\xc9o\x9e\x9c; \xd9\x8d\x83[b\x03\x13\ Q\x91\x8f\x19\x91\xf9\xe4\x91Y\x89\x0e\x10\xcb\xe8-E\ O\x01\xb5\x0f0\xc4\xbf\x99\x9c\x01\xd6\xbfy\x93\xf8\x0d\ \x01\xec\xc6\xab\xbctL\xca\xb84l:\x85\x9f?\xda\ \xa0\x83\x8c\x1a\x86|\xc3\xe4\xb4\x04>\xfe1L\xaf\xc0\ \xc0]:>u\xc1`h\x05`\x07\x11 v\x84\x09\ \xd6\xe8\x99_\x93J\xea\x9b\xc4\x96\x10\x98^\xa3\xb4G\ z\x8a\x16 \xa6\x0cIe\x0az\x80~\x84\xe5\x0f\xe4\ #Yf\xa9\x0e%\xdd!\x1f\x91\xa2]>&\xf3_\ \xb3R6\x9e?\xf4\xd5\x93\x87M\xc8NJt:\xa0\ \xda\x82\xaf7\xe0!4\x0f\xe96\xba\xdbPv\x0c\xe4\ \x00\x88\xa9\xdf\x1e6`\xfa '\x84\x8d\xe7\xa0p\x1f\ \xae\x05\xc7\x0e&\x8fO\x87\x03\xc3<\x22\x9e`)\x93\ ee\xc4\x99\xf2\x09\xef\x1a\xb6D\xcc\x1e(\x02<w\ \xeb\x01\xe9\xab\xdb\xbc\x05kMW\xdd\x8cyJ\xbb\x8d\ \x9d>\x22\x09\xe7X\x0b;\xd7\x02p\x09MA\xe3\xce\ \xaf*\x9b|\x81\xb8\xc6\xaaQ\xd3\xa4.\xd9\xdf \xd6\ \x22\x07\xeb\xaa\x18\x1e\x96\xee`\x1f\x9d\x91;)\xcb\x0d\ \xcbq2\xd6\xe0\x0f;mT\xfaG'\x0fI\xc6\xc1\ S\x10a|\x8e\x0f\xb1\xb7\xf7\x02a\xc4`26+\ '\xe1\xb4\xe1\x89\xbc\xbf!L\xb0\x8a\x99\xear<>\ ;\xdd)\xf3,\x1b%`\xfd\xa5,\xa7r\xfa\xc8T\ \x8cF\xfc\x84\x0bd\xcc{ge\xa79\xc4jd\xf0\ \xfb5\x1d\x1a\x01\xb0\xd6AZA\xeb\xc4\x0eg\x02\ \x05\x886\xaf\x98\xad\x94=~[e#\xde\xd1\x0d\x1b\ C\xb1;\x1e\x9a\xc4\x96\xfed\xc8\x1bg\x8c\xfa\xcb\x9c\ AgM\xc8Nv\xc1\xd5b,\xe1\xad\xf98\xe5\xdd\ (\xe8u\x02\x00\xb8\xed\xb6Gg\x0d\x9a\x94a\x83u\ \x14O\xdc\xd3I\x01\x84T\x87\xb4\xf1\xe9\xc6\x0d\xab+\ \xbe\xc3)^ya\xe8F0\xdc\xf8%\xcb\x19c\x95\ \xd9\x7fe\x8a\x0bg\x0a\xc2N\x04\x9e\xb4\xc1:\xc7\x17\ ?\xc7v\x80\x1e\xe4\xce\xbb\x0f\x1e\x0b\x86_3^\xfa\ \xbe\x1a\x15\x15\xb3pX\xb9\x8d_\x0e|I\x15~\xed\ \xeae\xc5\xd0t\x16;\xe2\xc1\xb6j\x8fx\x183*\ \xe0\x02\x0cINV\xa4\xc7\xe7\x0e\x18\xe8r\xc15\x99\ \xd1/vG\x0b^0\x07\x96\xc6d\xda\x1f\x98\x91\x85\ \xf60\xfc(\xebQx\xb0\xf1\x8e\xcaUc\x13\xec6\ \xbc\x8df^\x0bn\xed\x00\xbcV\x1eqL\x9a\x98\x97\ 6\x05\x9f\x22\x8fV\xe9\xd1+\xe3\x0fG\xa4$B\x93\ \x93\xc7!~\xc2\x81\xbb\x1a\xe8V\x0e\xcb4\xfbU\xa2\ \x00'\x8f\xd9[\xe5\xc1*ch\x1f @\x89v\xbc\ \x17\x1eC\xceC\xea\xfd\xc6\x9f\xbe\xae\x0ch\xfc\xfd\x84\ \xe1a\x93\xa5\x05\x93G\xe5%9\xc0\xee\x83\x06\x99\xfe\ \x98\x08\x17\x9c\x83V\xbb\x81\xde%\x00&`\xbbGe\ \xb9_<.{l\xaa\x03\x22\xc4\xbcu\xd91-}\ d[\xab\x03?\xfb\xb4hCE\x8f\xcd+\x00\x19P\ \x1e\xba\x80\x9d\xf8\xa6l\x13\xcf\x09\x9b\x12\xc0\xd4\xa0\xb4\ \xf6\xd6`\xc1r\xee\xaa;\x81\xd3\xb3\x92z\xff\xe2\x02\ /\xde\xf1\xc4L&v\x84\x0d[[\xa9\xfdP\x1b7\ \xe9\x82HyxCug7\xec:\x02\xa5U2~\ >6y\xfe\xd0\x14\x1c?\x17\xf7R\x03\xd5~&\xff\ dt\xc6\x83\xb3\xb2\xa0\xb0r\xff\xa37\x1a\xdd\xc9\x00\ \xb7<n@\x12\xb6\xdd\x22\x07r\xeb\xdc\x9c\x84\xe8\xbb\ \x9b\x0d\xc3.\x1b\xa7\x8c\xc8\x10\xab\xe1\x83u\x0b\xbc\x1d\ =9\xcb\x15M4c^\xc0j}am\xa7\x9d\x8a\ r\x9a\xcbn\x8f\xa4q\xd4.p\xc2\xd7v7\xbd\xb2\ -\xdc&,\x9au<i\xdc\xb3i\x1c\xe8\x8d\x02\xc0\ uQ\x1e\x9c\x9a\xb8\xf0\xb4\xfcK\xc7$g\xb9\xb0:\ \x83\x95z\x94\x06\xac\xd8\x8b\x054\xad,\xcb\xc9\x5c8\ \x17\x0f\x88\x84\x99w\x8c\x0a\x9fv\xcd\xf2\xe2-\xd5~\ \xf3iw\xbe\xb1\xfb\xe0w\x13\x15[\xd6d\xe5\xc8\xbf\ I\xf6T\xcc\x99\xa6\x80A&P\x1b\xd8\xaa_\x07\xf7\ .\x12\xe3\x82\xba\x09\x8c\xb77w\xd4\xf81\xba\xb0Z\ (6\x87\x8fa\xd4\x05\xb4\xad\x15\x9dt\xb0v\x0aO\ 1\x1d\x92\xa4\xae\xc9\xf7M\xb9\x8a\xe1\x8a\x160U\x87\ \xa6*w\xce\x1a\xa4\xe0\x13mJ\xdc\x8b\x16\xe4?\xa8\ \xc4\x82\xe7?\x9f\x90y\xd2P>\xd6;*\x93\xda\x9d\ `\x11\x91X\xae\xd3H\xb1\xf1\x08\x8a\x10\xb3R:<\ \x8d\x8f\x1a\x8a\x0e&\x1d\x92lK\xc4\xd4\x88\xcc\x0b\xec\ \xf6\xc5\xfb\x9e\xb6\xc1\x09\xfc\xe6g\xa4\x87\x83s~H\ \xb5\xc7\x87W\x11:3@Q\xccscwh,m\ \x00\x00\xca4\xe4\xde\xbb\xd6V<\xbf\xa1\xb2\xc2\xe3\xe7\ \x99\x9a\xdb\x1a\xd38\xb5\x07\x0fc\xc4\x97\xd6\x0d\xf4F\ \x01h!\xd5\xa5\xfcyv\xf6\xc2\xd3\x86\xfeq\xb2{\ t*\xb4\xf4\xb1}gF!\xb4\xa2&e:\x1f\x9e\ \xe2\xfc\xf0\xb4!\xb7L\xc9\x90A\x03\xf0\x9e\x01\xc64\ \xa4\xce\xf6\xba\xe0E\x9f\x16m\xab\x16\xd3\x9f\xf5\x08l\ \xf0\xb1\xd21/J\x8e$\x19o%b\xa8\xb1Tj\ \x01i\xed\xcd\xfe\xfd\xcb\xbb-#\xc0\x894]zi\ GC\xd4y\x1e\x8b\x8b!}\xbb\xb7H\xac\xc7\x04\x96\ \x80\xcfvV\xa2\x1a\xc5\x10\x05N\x99\xdd:u@l\ \xa58,\x98\xa4\xdf<)\xd5\x8e5\x0d\xb1\xa5\xf7\x82\ \x0a`\xa4\xba\x9cN|\xe69\xf2\xe0\xf2\xc8L\xe4\xfd\ \x13\xd1\x01\xc5s@\x82CQ\xa0\xba\x16U\xc20#\ '\xd9\x8e\xd7\x10\xe5\xe1\x92\xaa\xe2\x8b\xb3:8\x1cv\ \x0cMs\x0f\x8c\xeeN\xc3\x01\xe0hY\x86\x1d\xce\xb7\ }]u\xe6G\x05\xefl\xabR5\xbc\xfd\xc0u\xa5\ \xf7g\x94V\xf4j\x01\x80\xa4\x84\x1a\xde\x884\xc7\x95\ \xd3\x86|q\xce\x90\xef\xcf\xcd\xfbx\x9e\xf3\xadi\xd2\ \xe2\xa3\x5c\xdb.\x18\xf2\xd9O\x86^|\xc4\x90\xa1i\ \xce\xab&\xa6_?1-\xc1\xce\xab\x0f\xbc\xa6\x06\xbf\ E\x8d\x81+\x96\x95l\xac\xf4u\x9c'\xba\x10&\xdb\ \x07\x1e\xa1\xccz\xd4pA\xa3\x98\xb7P\xd0\xe6\x19\xcc\ _'\xaf\xba\xd2\xbb\xeb}]\x0bbn7K^\x97\ \x01'\x0d\xa8\xaa7h\x8b\xda\x86\x81\xea\x1a\xcc(P\ :y\x99mx`Zl\xf0\xe0\xab\x5c\xa3K\x13<\ \xc8\x90F&+sr\xa0b\xde\xf5\xb9Wg\x87\x0e\ J\xbe~r2,\xc2\xa9yz\xf5R \xcfCt\ \xd8\x14|\xac\x81\x9b\xa1\x08\x93\x1bk\xb5,\x06\xfb\xcf\ 3\xbc\xa8\xc1G\xe9\x87\x1dG\xc2@\x04G|8\x86\ ]\xf4$!!\x13\x89A\xfc(G\xe7\xba`1\x96\ \xa2\xc7\x8b\x126\x03T\x1d\xea\x9a\xeaU\xab\xaa\x8fz\ c\xe7S\xdf\x14\xad-\xf1\xd4y\xc5\x0dH\xee\xbb\xf8\ 17\x88\x83{\x19\xbdZ\x00 -\xb1G\x05r\x16\ &\xab\x9c\x95\x962u\xf4\xf0\xb9\x93\xc7\x1c>jX\ Z\x12/\xff\x98\xe9\x99\xdd\xa6\xdcrx\xe6\xcd\x93\xd3\ \xf1\x8e\x01>Rn\x16V\xb6\xbd\xc6\x7f\xf5\xf2\xe2\xad\ 5]8;t\x07\xf0\xbe &\xe7\x1f\xcb\x8e\xfc;\ S\xdc\x98\x11P\x9d0\xc2\x0d\xcdo\xff\xe6\xff\x82\x05\ +x8M\xe7]\x08d\xd3\x00>\xf1%V#\x05\ \xeb5\x86T\xc7\x87X\xc4\x06\x16\x9b\xa0\xa6\xef\xaen\ \x844\x05\xc4\xe6\xf0\xc1'~\xb0$\x9d=,!\xc9\ \xe9\xe0\xaa\xda\xc5`\x18\x8d\xeb\xa7d\x0dI\xb2C\x92\ \xf2^\xb4^Z\x929\xa2\x8b\x0f\xbe\x22\x8f]4k\ \xbc\xfa\x14=\xb1E\x0d?7\x06\x9d\xafE\x08?H\ \x04?\x94\x07\x18<C\xbahl:\xef\x042\xd7\xa2\ A\x84\xd1|\x8a\x84/\xeeh\x94\xee\xfe\xb6\xf1\xa2O\ \x8a\xcf\xfd\xb8\xe0\xd6\xcfK\xb6VxP\x8f\xc4\xa0k\ 1\x8e\x91\xab@t'\xecBz\xb5\x00\x84\x0f4<\ \xaf\x9e\x94\xfe\xcbCS\xa1\x09\x8a1\xcd\xb32\xa4\xcc\ \x8eZ\xf5\xb2%%?`\x07x\x0fE\xbd\xac(\xb9\ s\xd9\xdc'dG\x0af\x06\xec/\xc4Nk\xcdW\ +\x7f~\xa5\xba\x7f\x09\x9f4\xaa+1\xf0)`\x19\ kE\xd1\xc7\x00\xc4db\xc7\x0f\xd9\x87\x03\xaf\x995\ iF\xa1/Z\xc3\xcd\x1bx\x896\xe9\xca\xc9\x19\xe0\ \x19\xe3C\xa7\xbb\x14\xecs\x94\x98\xc3\xa1\xfcbL\x12\ \xc3'\xa5qd/\xd1GA\x93\xc0\xa4q\x99\x09\x13\ \xb9\x04\x98\xa9\x1b\x1f\xb8\x12\xd4\xaa\xd2\xb7\xd5\xea\x0b?\ \xd4\xcf{\xbf`\xca\xab\xbb\x9f\xfc\xb6\xfc\xabbO\xb9\ '\xc0\xdb&\xbd\x94~\x22\x00\x10\xfb\x8a,\xffa\xda\ \x80\xdf\x1d\x96\x8e\x8fU\xa0\xa91\xbf\xa4\x9d\xf5\x81+\ >+\xf9\xbe\xda\x8b&\xb0\x1b\xea\x8cm\xe0\xed\x80c\ \xd8Q\xcfJ\xceT3\x0bb\xb8\x98\x8cCB\xbf\xbc\ 1\xb0\xfb\x03s\xa6\x08\xe1:\xee0\xc9icN\xbc\ G\x12%\x182&\x0f2b~\xcf\x1a/o\x9e\x80\ \x1a\xfd\x00P\xa6\x1a\xba|\xca\xb0D\x87-\xea\xf1\x82\ \x91\x01'\x81\x8f,\xc9\xf3\xf3\x13\x5c\x0a\xb68c\xac\ #\x13=\x0c&';kd\x12>\xd8\x18\xcf\xa4\xc4\ \xcc\x8d7\xf9\xf0\x07sM\xb1W\xbb\xe7\xdb\xda\x9f,\ *>\xeb\xa3\xc2K\x17\x17\xbc\xb1\xb9\xa4\xd1\xcb\xf3\xbd\ \x01r n\x19C\xe3\xc0,^=H?\x11\x00\x1e\ \xed\xa8\x01\xbf<4\xe3\x97\xe3\xd2\xec\x0a\xdaWL\x09\ \xa8'\x1alG\xad\xef\xf2%%\xf0\xcd\xd3\xbf[\xe1\ \xd9\x01\x83\xa6d\xcf\x94g<`\xe0;\xe51\xcd1\ \x1c\xcc\xd0\xfdul\xed\xad\xc1=Kt\xad\xcb\x9e\x0f\ 0$\x9b\xacLM\x8b\xc1ta\x1c\xea\x87fG>\ \xbc\xaf=<A\xa3\xd6\x1f\xa5\x0c\x1b\x9a\x0c\xcd\x90\xf9\ \x03\xedf\x9c\x8a\xad]\x0a\x96e<\xd1\xd8Lw^\ \x82\x1d7\xf4p\x81%\xa2\x07\x13\x133\x0e;.?\ %\xc5\xce+bh\x85\xe3\x00x\xdb\xf2\x83\x99\xd3\xcc\ 4\x12\x0b\xe8\xc6\xf6z\xed\xe3b\xef\xb5k\xea'\xbf\ Qp\xdb\xca\xc25\xfb+\xcb\x1a\xfd\x1a\xbf'\xc8\xab\ \xa4\xe8\xae\x07\xe9/\x02\xd0\x8c\xd3&\xdf>5\xeb\xe6\ \xc9\xe9\xf8,\x0c\x7fX\x94\xc70\xdb\xd5\x10\xbcjY\ \xe9\xae:?w\xd5#\xe8l\xc8\x09\xf2\x9c\xbf\x1b\xb6\ \x04\x9e9 d\xfc\xfe\x86\xe6S\xd6\xde\x10\xd8\xfb)\ \xdf\xd6E\x187N\x1b\xc0\xf3{t0\xb7]\x9e\x98\ \x83\xd2\x15;\xbb\x8a\xcb\xf0\xfdFQ\x01\xa5\xcan\xd3\ \xb3R\x13\xc5z\xb7\x22_16\x01~\xa8\x05\xd0\x0f\ \x18\x9e\xea\xbc\xe5\xf0L\xbc\xdf\xdd\xe5\xa9\x89uPn\ \xeaY\xa3\xdfx~\xbb\xe7\xac%\x15g~Xp\xd5\ \xd2\xa2\x0fv\xd6`WuO\xd3\xdf\x04\x00b\xdc\xae\ \xc8\xd7L\xce\xbcxT\x92\x18T\x00\x12\x00\xf1lH\ \x9b\xaa\x02\x97~R\xb4\x15\xfb\x82\x10\xe1\xbc\xbb`8\ \xda\xd9&\xe7\xcd\xb7\xcd}\x1c\x9f\x0f\x80L\xc1@\xa0\ \xd0\x18\x1a\xc1\x06e\xcdo\xbd{>\xd35\x15\xb7\xc4\ \x1b\xb8\xfaq\x99I\x133\x1d\xcd-\xcf\xf0\xae\x1d\xdc\ 1\xcd\x8c\xa7\xe1\xa9\xca\xc4\xcc\xe6\xb9eb\xc2\xd8]\ \x8e3\x0fG\x07\xc4XN\x82}H2\xd6\xc4\xbb\x19\ \x88\xc3s\x87\xf3&\x00)@\xdf\x87\xc9\xd2/&\xa4\ \x1d9\xc8\x01\xb6\x01Kb\x17Z\x03\xf4\x19\xef\xc11\ |!\x07d#U\x97w7\xe8\xef\xed\xf5^\xb1\xa2\ r\xea\x7fw\xbc\xb4\xa5j_\x9d_\xe5S\xd4\x99\x98\ \x87u\x1b\xfdM\x00\xb0\x81\x87S`\xca\x7f\x9c1\xe8\ \xea\x09i\xcef\x09\xe0\xe8\xdb\xeb\xd5\xebV\x94\xedk\ \xe8\xee\xf9\x188f\xd0\x98\x9c{\x14\x9b)\xde%i\ 6\x19q\xa7\xdad[s\x9d\xbe\xef\xd3\xae\xe8a\x80\ \x08p\xd9\xe5K\xc7$\xf3sE`\xbf\x98.C\xdc\ A\x1c\xde;-\xcbe\x8f\xd9\xec\xf23\x97:r \ \x12\xf8z4\xa4(j\xba\xd2\xc5\xb7\xcd\xdb\x03\x92%\ ))!\xd9\x01\xcd\xca\xee.\xa2D\xdc\x81r\x079\ \xfbw\x87ed%\x98\x0f\x99F\x9f!;\xc6,\xf2\ b\xd1\xfc\xf0\x1f\xc8EP\x19+\xf0J\xb7\xae\xa9<\ \xf3\xc3}\xf7\xac-\xad\xf5\xf8P\x8a\xb8\xd3\xee\xa4\xff\ \xb5\x00\x04\x0e\x1b\xfb\xdd\xd4\xac\x8b\xc6\xa6(\x0c\xe7\x0a\ \xe0\xa9\x0c*\xac\xaf\xaf\xf2\x9e\xf5Q\xc1\x9e\xfa\xee\x18\ \x83\xdf.\x86l\x93\x87\x9d\xca\x8e\xfa\x97aO\xc1Q\ \xa1h\xf2q\x8c&\x0b6\xe8\xab\xaf\xf5\x15\xaf2\x9f\ \x13\x8ec\xd8 \xc3)L:oL\xc6\xb4\x01n\xcc\ \xea\xe0q\x875\x0d3#\xe27N\xc7\xca\xae\x18\x9b\ 4'\x0f\xe4*\xe6B\xc2\xfd-\xf4aeGl\x89\ \x10\x08\x83\xd3\xa6\xb8b\x1f\x8f\x149\x90J\xa0\xdd\x87\ f\xc5\xfc\xee\x1a\xa2\x97\xc0\xd8\xb4\xbc\x94\x07g\x0dL\ \xc4\x01A\xe2\x9el\xfc\xca\x5cg\x98:\xc0\xc7\x04\x16\ 6\xe9Oni\x98\xf8\xbf}\xcfl\xa8n\xf0\x05y\ H\xf0\x8b\x17\xc28\x9a\x81\xf6\xe9\xb7\x02\x00q\x8c\xed\ \x80\xe9\x03o\x98\x98\x0e\xf56~\x0f\x14o\x09@I\ .lP\xaf^^\xbc\xbf>\xc0\xe3\xb8\xbb1k\x00\ xOx\xce\xc3\x92#\x096`\x05\xc4\xcc}z\xc0\ \xf6\xf9\xd5\xfe\x1d\xef\xf0\x91\x02\xa6\xf3x\x81/\x06x\ \xf5\xf8\x81G\x0e\xb4s\xc9\xe98_a\xc7%\x0f\x13\ ;eh\xe25\x87\x0d\x8aS>\xc1\xf9\xe2\xab}j\ \xd4Z\x02qb\xb3)\x0e{\x5c:\xa3\x22\x04s\x8f\ 1\xc2\x0d\xf1\xd0\x9f%\x80g\x0d\x0b \xb2 ;i\ x\xf2\x1f\xa7g\xbal8\x22\x08\xeb\xe4|gw \ \x1a\x1d\x10\x0e,e\xb0\x1c0\xd8]_W\x9e\xf3Q\ \xc1\x07;\xaau|r\xc0\xd0\xf1\x11\xfe.W\x80\xfe\ )\x00f\xcb\x0bp\xd8\x94\x1b\x0f\xcf<wd2\xa4\ 0\xaff\xc3\x07.\xd9X_\xee?\xed\x83\x82\x82\xc6\ Xf$\x8b\x12\xae\xfd\xf0\xaf\xc8\xf9'\xb2\xb9O\xea\ \xf6$\x9c\xc2\x82\xc9P\x17\xc0@\xfa\xeb\x94\xaf\x7f\xef\ +X\xc7\x07\xb1\xc6\x013\x1e\x00XLMp=3\ o\xd0\xa1\x19\x89\xcdo\x9a\x0b\x85\xc6\xc3'\x9d72\ \xe9\x99\xf9\xb9Yn\xa8#\x99>\xc4\x04^\xb4$y\ \xa1\xe9\x15\xb5\x99a\xb2\x0d\xfeq(Pw\x03!\x97\ %#Y\xf3\x9a\x05\x97\xe8\xd3\xf0\x22\x88\x09\x099\xf1\ \xa2q\x19w\x1d\x9e\xe2\x82\x22\xd1\x8dC|yq4\ K\x84Y\xae\xc0\xe4\x83\xfe\xb0ok\xd4+WT>\ \xb0\xaa\xd0\x8f\xc6\xdf\xdc\xd5\xb5\xf4O\x018\x10\x87M\ \xbe\x7f\xd6\xc0\x1b'\xa5\xbbd^\xa9m~\x14\xa0\xac\ I\xbffYqQ\x83\x0a\x1a\xdbS\x0fj\xc89s\ \x959\x7f\xc5wI\xc2\x0a\xafcb\x96T}\xf6/\ .\x09\xee\xfd\x98w\x99\xc49dY)\xeeE\xa7\xe7\ \xfdij\xc6@'\x9f\xb3\xc5l\x04a\xb4\x88E\x80\ I\xca\xe1\x99\xce\xc7fgB\x03\x99\xc5\xf0\x00\xc1\xc1\ p\xef\xf1Iy\xbe\x16\x05\xfc\x0dq\x10\x9e\xa8=\x88\ \x1e~J\x99\xd9\xb1\xf6\xc07\x10}\x1f\xc8Jx\x7f\ V\xbal\xe2\xa0g\xe6\x0f\x1a\x9c\xe0\x80\xa4\xe5\xed\x00\ \xf1\x17\xa7:X8\x98\xb6\x1e\x9b\xdd*3\x1e\xdb\xe6\ \xbd\xe4\x93\xe2\xe2\x06s*3\x08\x06\xc2\x97\xe3O?\ \x17\x00^s\x95\xdd\x0e\xdbM\x87g\xfdtT\x0a\x93\ [\xaa\xfc\xba&ik\xca}\xa7|\xb0{O\x9d\x9f\ +C\xb7\x82\xc1\x82\x8f\xac(\x83\x17(\xc7\xbc\xac\xdb\ @\x030\x0c\xb2\xa4\xf0\x5c\xe0e\xab\x7f\x13(\xfb\xd6\ \xd0\xe2\x160\x1e\x15xJ\x87\x22_>)s\xc3\x85\ \xc3\x9f\x9e\x931.M\xc9t\xe2\xbb\xd4\xa1\x96\x9f\xe4\ \x90\xb3\x9c\xca\xe4L\xdb\xfb'\xe5-<=\xff\xbc\xb1\ \x03\x12]\xfc\x8d\xdeq\xad\x88\xd8y\xb1\x13+\x11b\ \xde0\xe9\x11\xcc\xf3\xf2\xe4\x88gl\x10=\x88\xd9\x0e\ \xc0\x82(\xcb\x0b\x86%/?=o\xca\x00\x87\x0dV\ \xd1\xdc\x9a&\xb7\x9b\xb2\x1b\x86\xc2\x0c\x0c|\xf3\x02\xb2\ \xb4\xa8\xe9\xb8\xf7\xf7\xed\xaa\x0d`P 0\xc2a\xfc\ \xe9\xff-\x00\x13E\x91\xef\x9d\x99u\xed\xf8\xac\xe6w\ \x1e\xc1\x0f\xbe\x1e\xab\xd8\xa3\xff\xfa\xb3\x82\x92\xeaZ\xe1\ \xae\xfb\x81\x04\x1f0\xc5v$\xb4\x03\x92\xb9y\xd31\ \x13@5DW\xd9\x92\x0b\xd5=\xef\x9a\xae\xe2\x0aZ\ 1E\x96\xcf\x1c\x97\xb5\xf4\x9c\x91\xdf^0\xe2\x8b3\ \xf3?;5g\xf5YC\xbe\xfd\xe9\xb0Eg\x8c\x98\ \x91\x9b\xc8gW\x8ds\xae\xe3\x0dl\x86\xfdI\xe6z\ ThPC\x0b\xf7\x1dq\xf1\x84\x87\xd9\xa8\x97\xdd\xdd\ f\x14\x88n\x04-lJ\xa2\xf3\xf5\x13\xf3\xff43\ s\xa0\x8b\xe7\xd4\x9eMg&U\xf9\xb5K\x96\x14o\ \xa84\x1f]\xea\xaa\xd0XE\x00\xc0\xaa:m\xb6\xdb\ \xa6e\xe1\xabq\xf1\x85\xed\x86\x84\xdfh\x95\xd6W\xeb\ '|\x5c^Z]\xc7\xbb\x82\xba\xdf\xb8\x80\xee\xcb\xf2\ \x90\x13\x95\xe3\xdf2\xec\xc9\x10N\xbc\x03,c3@\ \xd2\x9a\xd8\xda\xdb\x9a\x8a7Cm$\x8e7\x85\xf9\xd0\ #\xb3\xae!\xcb\x92\xec\xb6+\x83\xd3\xdc#\xb3\x92\xb2\ \x93\x9d8\x99/7u\xbc2\x12\xef\xbc\xc1\x1b\xb2)\ \xf6\xe8K\x16\x1c\xaf\x06\x83\xfe@O\xcd\xf2\xcd\xf6\xf5\ \xcc\xbc\x82\xdd\x07\x17i+\xc2\x8b\x03Ks+\x97M\ \xc8\x5cv\xd6\xd0\x13\x06\xe3KF\xb1,\xf2\x82\xc8s\ .v\xcawe]\xbc\x15pf\xd9P\xb6\xd7\x06.\ \xf8\xb8\xa0\xdc\x83\xe3&x\x10\xe0?\xce\x01\xb0\x8c\x00\ p\xec\x8a\xfc\x7f\xd33\xae\x1e\x9f\x8aS\xa6\xf3\xf7u\ \x99\xa9[\xe63\xae\x5cYQZ\xe7\xe9\xfe\x12\xc0-\ 1\xffJ\x1f-\xcfyX\xb7\xe3\xebGp\xf8%\x86\ \x04\x0c\x9e\xd7\xbe\xec\xac\xc0\xee\x85-\xb7.\xe2\x00\x9c\ L\x9c\x13\xcf\x8e\xeb|\xd5D,rGqG\x96\xd9\ \xa0$G\xd4y\x18\xc2\xe4Su\xaf\xda\x03-\x00\x00\ \xf2\xc9f\x9c]\x5c\xac\x12\xfd\x86\xe6L\x8f\xb6\x1e\xc8\ Jt\xbdxb\xfe{'f]<2)\xc1n\x88\ W\xce\xa1\xb9\xd0\xbb\xe5\xbe,G\x14AV\x1d\xd0\xaf\ ]Q\x5c\xeb\xe7\xb7*Q\x90\xc4\xfexa\x15\x01h\ \x01*\xbcw\xcc\x18\xf0\xd3\x91P\xd7\x86\xb84c\x14\ l-[]\xa1\x1e\xb7\xb0\xb4\xb4\xf3\xb7\xcau\x15\x10\ \x08y\xc8\x02\xf9\xf87d\x9c+\x02\xc7\xe0\xe3VX\ \xd2|\xf2\x97\xd7\xabEk\xb8\xab>\x0c\x160&\xa5\ 7\x14\xc1\xb5\x8aM\x91\xc2\xa4j\xddV\xa1\xf6\xc0(\ \x08s\xa3\xc7[\xef\xd7\xba\xcb\x00\x10=\x09\xb4\x83\ g\x0dI\xff\xeb\xd1y\x1b/\x18\xf1\xeb\x09\xa9\x0e\x10\ \x7fn\x8e\xcdB\xd9m\xa0\x11`\xf2\xcab\xef\xb3\x1b\ \xcb\xf1\xec\xf8\x893\x16\x13\x000B\xb2l\x93m\x7f\ \x9c9\xe8\xf2qiv\xd9\x90\xb91\xc2\xe6\x15\x93*\ \xbc\xfa5+\xcb\x8bk\x1a\x0c\xa3;_\xd9(\x90\xf1\ \xb6\x04S2\xc6K\xb3\xffj\xd8\x13\xb1\xce\x01[1\ \xe3\x81\xf4\x07\xa5\xcf/\x0d\xec\xfeP\xef\xd6wI\xc6\ \x1b\x8ccix\x1a\x7f\x18-*\xe0\xda\xcb=Za\ }\x8fL\xe8\xc4\xde\xdd\xdb\x18\x88\x7f\x0d\x8c\xe8E`\ \x1d\x05\xab\x19\xf0\x0d\x86\x11\xedm\xb2\xd3v\xd7\xcc\x9c\ \x8d?\x1b\xf9\xd0\xac\x01\x17\x8cJ\x1c\x8c#\xa8\xb1?\ \x96w\x1c\xe0\x17,\xeb]f.\x18\xb48\xb0_\x9a\ \xfd\xf3{\xcf\xea\xfd\xd5\xd8\x04\x11{\xe2\x86\xb5\x04\x00\ \x12\x16S\x95\xb1\x04\x87\xfc\xc7Y\x03/\xc1\xd9\x110\ \xa5\xcd\x84\xd7\x99\xf4yy\xe0\xcc\x8fK\xcbj=\xa6\ \xfb\xee\x04\x03\x81\x1a/+CNb'\xbe#\xd9\xf1\ \x9e0\xe6E\xcc\x04\xcc\x08\xfa\xd9\x9a\x9b\xfc%_\x0b\ \xd7}\x96\xe1\xb9\x99\xa6\xb4E\x05\x0b\x18zI]\xac\ o'\x8e\x8e\x7fn\xf5`6\xc1\x9cB\xf4[x\x19\ \x84o\xd3(\xe0\x97\xccX\xba\xdbv\xc9\xb8\xf4G\xe6\ \xe6.?{\xe8\xf2\xb3\xf2\xcf\xcd1\x12\xa1\x8d\xc0\x8b\ &\x96M\x09\x1f4\x15\xc7\xc7\x17\xccoH}P\x7f\ hC\x9d\xc7\x1f\xff\xe7\x96,\xd7\x05\xd4\x0c\xb3\xcb\xec\ \xff\xa6\x0d\xfc\xc5\x98$gs+\x00?\x8c\xedi\xd4\ \xafZY^P\xdb\x88\xab=Q\xe1\xc3\x17\x9d\xa4\x8d\ \x91\xe7<\x02\x1a\x00\xe1\x81\x0c\x06\x1f\xf85\xd4&\xdb\ \xb2\x8b\xfc;\xdf\xc5v\x00\x0f\xae8\xa0\x8f\xc0\x03l\ \xe4\xa6\xa7:\x95hs\x1d/m\xff\xdd\xe3\xe3>\xc1\ z\x97\xc7\x00\xaf\xe1a>\xd8\x5c\xee)lT\xb1=\ \xd6\xe5\xe7$z\x1d\xa6\x12@\xb6Mv\xd8\xc6d\xb8\ \x9f8e\xcc\xf6\x8b\x0fy\xe3\xb8\x81\xb7LN?!\ \xcf\x9e\x97\xa0@\xb64\xb3$\xe6\x0e\xd3\x924\xafp\ \x0fb\x05\x84\xe0\xbb*\xff\xb7eX\xf5\xc1\x0c\x89Y\ 2>XT\x00P\xbe%\x96\xe8\xb0\xdd=3\xfb\xe7\ \xa3\xcd\xe9\x85a\x1b\x97\x02f\xac\xaeP\x7f\xfeYY\ E}c\xfc\xe29\x5cx 0\xb7\xc9\xf9'\xc8\xc7\ >\xcfl\x89x\xc7\x93\xe7(l\x0cj~\xe5\xeb;\ |\x85k\x9a\xb3W\x9f\x83%\xd8X\x963\xeaN|\ \xbc\x15\xbe\xb1\xc2_\xdf\xe4\x8b\xeb\xc0\xa8\xd0\xf0\x1e8\ I3\xd6\x94z\xfd\xf8\x10@\x9f\x8ct\x22F\xb8\xb9\ \xe0\xdf\xbc\x0d\x08?\x0eE9jh\xfa-G\x0cx\ \xfa\xb8\xfcE\xa7\x0d\xfe\xcfQ\xe9\xa7\x0eM\xb0\x9b\xc3\ \x0b\xf1\xb1^X\xc2\x85\xb8e\x17&yT\xf9\xbf\xdb\ b~)S\x1b\xac\xda\x02\xc0\xa4D\x9c6\xe5\xf7\xd3\ \x06\xfd\xec\x10|\xa5<\xe8*\xa63gs\x9dz\xe5\ \xca\xca\xd2\xfan\xefm\xc0\x0c\x86a\xc3|6`\x1a\ \x9b\xf3\xb0ls\xc1\x16\x05\x9b\x02|R\xd9`\xbd}\ \xe5%\xfe]\x1fv\xf9\xbb$\xbb\x04\xc3m\x93\x07\xe1\ ;\xe1\xa3\x81\x19\xf8\x10\x87\xdf0\xfe\xb0\xa6\x8c\x17\xc5\ \xae7\xc7X\x8cY\xc0\xd0\xfe\xfdC\x8df\x0e\x11\xee\ \xa2\xc6>\xd1\x9b1\x8b\xa4\xf8G\xcc-`/\x12\x1d\ JN\xb2\xe3\xf8C\xb2\xfeul\xde\x96\xf3\x86\xbdp\ L\xf6\xcfF\xa7\x1e\x9e\xe9p\x98m\xc58\xe5P^\ \xd7\xd1?-\x09\x14\xd6z\xe2\x9b\xe9-\xdb\x05$\x80\ DLr\xda\x1f:2\xf7\xd21\xc92/\xe1`c\ Q\xb8uiUi\xe0\xa7\x9f\x14\x15U\xd5\xf1!\xc0\ \xa6\xf3n\x05\x1f\x15\xce?I9\xf1m\xc9\x91\xc4C\ \x80\x0d\x01\xc8x\x86\xee\x97\xd7\xdd\xea\xdf\xf7\x19(V\ \x9f\x13\x01P\xda\xfc\x147\xbf\xcd.\xb6D\x82y\x0c\ [Q\x12\xd8[\xe3\xc5R\xd8\xe5`\xa3\xf0\xbeu\x15\ \xbb\x1ax\x95\xae\xdfc\x85k\x8c\x1f\xbc2\x00\xff\xa8\ \x08PEKKr-\x18\x9e\xf2\xd7\xb99o\x9c\x92\ \xff\xd1\xa9\xf9s\x068\xa1\x08c\x94B\xe9\x8d\xb1\xd7\ \x86\x9f\xa8>\xa0}R\x08\xd9\x1e\xd5\xc7\xdc\x1c;$\ \x008\xd9\x81\x1d\xda\x01Gd\x9e<$Y\x91\xf9\xe3\ a\xb8\x1dSmK\x9dt\xd5\xe7\x95\xc55M\xa8\x01\ \xdd\x0bV3\x00Y\x963&(\xd3\xef\xc3w\xca\x83\ \xe9\xe79\x0e\x094(\xab\xae\x0e\xec\xfd\x88u\xdd\xbb\ $\xbb\x00(\x04PNF\xa4\xe1\x94n\xd8\x9d\x1f)\ \x22RX\x99O]VP\xdf\x1d\xc6\x8a\xe9[*\x1a\ \xfe\xf9}\x03\x9c\xcb\xcc*q+y\xbd\x92\xfe}u\ q\x877\x03Lx\xde\xc0\x1f(\xb1r\x8aC\x99<\ 0\xe1\xad\xd3\x87\xbdv\xec\xc0#\x068%\x99[\x0f\ \x14\x81(\xf3\xaci\x93\xa0\xf6\xf7E\x897\xa0\xe2-\ \xc0xi\xb5\xd5\x05\xa0\x85D\x97\xfd\xa9\xf9\x83.\x1b\ \x95h\xa6$\xd6\xbfa\xab\xa1\xaf\xad\x0c^\xb2\xb4\xb4\ \xb2\xfb\xfb\x82\x0e@\x1ev\xaa<\xff?\x92=I\xe6\ 9\x01u\x00~\xf5\xa0\xb4\xf6w\xfe}\xcbM7}\ \x044\xa4\x93\x13T\xf8\x8e\xa5B\xadJ\xf2}\xebk\ +=]>X\x0b\xa2\xfa\xb1\xcd\x8d\xfc\x0d\xae\x04\x11\ \x11:\x94\xd2\xa3\x87\xa6\xfe\xe7\xc4\xbc\xab\xc7\xa5\xd8d\ n\xb4\xa3\x87\xd7<\x0ci_}\xb0.\x10\xcf\xdcH\ \x02 \x001w\xdb\xed\xbf\x9b>\xe0\xfcaI\xb6\x03\ n\xe0\x18\x06\xdbP\x13\xbcby\xf9\xbe\xeaFsL\ \x08\xdf\xdc\x8d\xa0\x1e)l\xe0a\xca\x9cG\x0c{\x12\ \xe4#\x1d\xefV@f`,P'\x7fy\x8dw\xfb\ ;\xba\x86\xef6h\x81\x8f\x5c19`\x111k!\ b\xa3\xf0\xbf\xbb\x80\x18\x86\xfa\x11\x93\xa5\xf9\xa3\x06\xb9\ m\x0a\x18W\xb1#\x0a\x0c\xa3V5\xae^YU\x8f\ /\xd0\xe8\x82D\xc1\xe8\xd15]{vC\xc9\x07{\ \x1a0\x13`\xf9\xeb\xff\xfcx\x13\x8ch\x03f\x8af\ \xc4\xa6P`4\xf2\x1a\xa4,g\xbal\xb7N\x1ft\ \xe9\x984\xb3\x1a\x1f=\xbc\xb9Q\x174j\x03\xf1\xec\ \xf7%\x01\x10`\xa5\x9aI\xc9N\xc7\xfds\x07\x9e?\ <Q\x86\xe4\x16;@o\xd9\x9a\xca\xe0\xafW\x96\xd7\ 5\xf9\xccm\xdd\x09\x06\x0b5\x80\xc9y\xc7\xb0yO\ 0\xfe\x8c\x18~p\x5c\x10c\xaa\xd7\xb6\xfeN\xb5`\ \x05\xe4O\xee\x1c\xbfa\xaf\xaek\xde\xf2\xede_\xbd\ X\xf6\xe9=\xf5_\xfc\xb1l\xcdsM%\x9b4\x1c\ \xc7\x82\x97\xd3rq\xdd\x0a\x06\x1ac\xd9e\x97\xe7\xa6\ \xf8b*\x0e\xd8~`\xabK\x9b\xfe\xbb\xad\xaaK\xae\ \x04{n\x8d\x0fv5\xdc\xf5u\xbd\x0e\xd5\x7f\x86\xaf\ l \x880\x11%\x16\xe0Y>\xc1\xa6\xdczD\xd6\ \xb84\x1b\x96\xe5\xd8P\x0d\x06\x1f\xf4'NZ\xdd_\ \x04\x00EY\xd4o\x11\xb4p\xe6\x0f\x87\xff\x9a\xdb\x84\ \xfb\x90\x18\x09\x0e\xfb\x1fg\x0d<}h\x12\xce\xfd\x0e\ \x11\x8d\xf5m<\xee\xabJ\xdf\xc5KJK\x1b\xf8;\ \xe5\x85\xb5\xedV\x0c\xd9.\xe7\xceeG\xfe]r\xa4\ \xf2\x0c\xc0\xe7\x06\x81l\x10\xa8\x95\xbe\xb8*X\xb0\xdc\ 0T-\xd0\x14,Y\xeb\xfd\xfc\xe6\xe0\x1b\x13m\x8b\ \x8fO\xdfvgf\xe9\xb3\xee=\xcfe\xec\xb8\xd7\xfe\ \xd9i\xfa\xff\xc6\x06V\x5c\x17(Zi\x04\xea\xf8\xfd\ c~!\xdd~)\xa0[7\xcc\x1c\xca\xe7\x85\x8e\x12\ \xb8z\xa8_\x055\xe9\xce\xaf\xeb>\xdb[\xaf\xeb\xd0\ \x5c\xe3\xc9\x12\x1b\xa6'\x1c\xe9\xbbr\xdf\x1d_U\xe0\ \xb4sP\xdc\xe2\xff\x00&\xd1\xcb\xc1\xa2\xc13\x02\xff\ \xe6?\xb0\xd5\x1bT\xcb\x1a\x03Q\xe4\xdc4\x97r\xcb\ \x94L%\xe6\x8a\x84\x8e\xcf!c\xd9\x8f\xa1\xf4\xb4\xa2\ \x9f\x08\x00&\x8f\x18\xaa\x83_\xc1@\xb0\xb6\xbe\xd1\xef\ \xf7\xf3UL=s\x07,\xa1\x93\xd0\xe0\x1c\xd1\x12K\ q9\xfevT\xf6%\xa3\x92p\x05o\xbdp\xff\x0d\ \x05\x0c\xc2/\x97\x96T54\xf2\xf1\xe8\xdd\x0d\x04\x83\ I\x8a\x92w\xac|\xd4\xd3\x92#\x05r\x01\xdf\xc2[\ \x02z\x90\xad\xfa\x8d\xb6\xe9I\xf5\xb3\x9fJ\xcb/\xb6\ \xed{C\x0ex\xb8\x95\xc4l\x82\x17\xc1u\x8ci^\ \xb6\xff}e\xd9\xa5\xdag?\xf3\x14o\xe0\x17e\xde\ P\xea^\x981,\xd59$)\xa6\xbc\x07\x81\x86\xc4\ \xd2\x0d\xe3W\xcbJ\xde\xdb\x0dz\x06)\x12\xf3\x85p\ \x0f Z>\xda[{\xfe\xa2\xa22/\xa62\xc4/\ o\xcdcL\x12\xd6\x00\xb2\x80i4\xf8\xaf$iZ\ p}Q\xcd\xcd\x9f\x17\xff\xe4\xe3\x82\xdb\xbf,\xd7\xb4\ H;a0\xff\x1c\x9a\xe5\xcap\xc4\x9a\x8b <\xf1\ \xad\x8d\xf4\x97\x16\x00\x93\x1aUu}\x99\xe7O\xab\x8a\ \xe6\xbc\xb1{\xd8\xcb{\xc6\xbfQ8\xf2\x95\xbd\xd3\xff\ \xb7\xe7\xde\xd5Ek\x8a\x1b\xeb}P\x99\x83\xd8\xeb\xec\ zyI\x87V\x9b\xdbn\xbb}Z\xd6i\xf9n\x85\ w\xe5\xc1V\xc6\xa0\x9e)\xaf\xad\xd4.\xf8\xb4\xac\xb0\ \xae'\xfa\x828\x12\x93\xe5\xec\xd9l\xce\xdf$w\x86\ \xd8\x0e\x86\x1e.-\xd8`lx\xd8V\xf1\x9d\xa4b\ \xef\x0a\x17\x07\x1er<\xa4\xe5\x1b=\x91\x0c\xd5\xa8\xda\ \xecZz\xae\xba\xe5)MU!Z\xb87\xdd\x09K\ s\xd9\x8e\xcb7\x9f\xbf\x8b\x0e\xf3z\xb0\x946j\xf2\ 5+\xca^\xd8TU\xef\x8b\xf5Ay\xf0\xd1\xe3\x0f\ \xbe\xb7\xb3\xf6\xfa\x15\x1558\xef\x0f\xafS`\x94\xe1\ .\xc2\x22\xa0\xe5\x97XP\xd3\xf7\xd6\xf9>/\xa8\xbf\ wu\xf1\x94\xff\xed>yQ\xf9K\xdb\x1b\xd6\x97\x07\ \xbe\xad\x0e\xd4G:%\x03\xcf=n;K\xb1\xc7\x9a\ \x8f\xcc\xb2\xca3\x7f|\xe8\x12\x01@\x01m\x86\xafa\ A\xe2R\xca\x7fx#F8\x8d\x0d\xf3\x14\xc0\x9e\xaa\ \xa6\x1b\x96\x17\x9f\xbb\xa8\xf8\x1f?xv\xd4\x07\x83\x9a\ \xa4\xea,\xa8\x1b{\x1b\x82\xb0\xe5gKJ\xafZZ\ \xf4C\x154\x08t\x1d>\xe1\x9d>\xd9\xe5\xfc\xeb\xdc\ A\x97\x1c\x92\xa0\x80\xcd\xc4?\x88v\x15\xac\xe7\xc6\x1a\ \xf5\x86/\xca\xc0R\xe8\xf1\xbb\x90\x88\x80\xc4W\xf2\x8e\ V@\x03ln\xac\x9f\x8al\x81_\xe1t\x0c\x9a\x93\ \xdb\x1a\x92jlxL\xfd\xfanX\x82#\xbb\xf9Z\ lL:i\xb0K\xc1\x1b\x12\x86\x16\xf5y\xf1b\xf1\ _5\xa4\xff[Ww\xe3\x8a\xe2jO\x93\x88\x07\xfe\ \xcfs_\xfb\xbeC\x9e\xc1o\xb3\xd9h\xe6!\xc3\xa8\ \x0f\xe87\x7f^v\xd5\xe7\xa5uA=\xa6QJD\ \xef\x06\xf3\xbc\xf8\x98\x80M\xf8\x91\x92:\xcf?7\x94\ \xff\xf2\xb3\xe2\x0b\x16\x15\xfe\xf4\xd3\xd2\xc7\x7fh(k\ \x02\x07\x98\xd3 OT7\xa9%\xfce\x8d<\x07\x85\ \x05\xda\x0f<\x15\x962\xb1)&\x9asv<\xe8\xb2\ \x16\x80\x19\xbb\xba\xae\xf9\x1a5O\xb5\xea\xaf\x87\xca&\ \xc6.\x169\x1e!q\x02N\xf3\xd9\xfe\xfac\x16\x16\ \xbd\xbf\xcf\xdb\x80/\x1c\xc7\xba\xe1\x8f\xff\xfc\xa71h\ ,)\x0e\x9e\xfc\xe1\xfe\x15\xbb\xab!\x14\xd0\xa4\xeb4\ \xf5\xccCS\x13\x9c\xff7#\xfb\xb4!\x89\xf8Zt\ \x88*\x1d\xfee\xc3`+K\x82\x0f~]%rD\ \xf7b\x06\x8c\xc969w\x9er\xd4\xbf\x0cWj\xf3\ &\xc4\x5c\x16NC\xf0\xa3+\xcd\xc7v\xbc\xacmx\ X2\x1f&\xe8\xaek\x81$\x83\x10\xcc\x18\x9c6\x02\ \xe7\xd9\xc5x\x8d\x16~\x1dx\xb9L\xd5\xb5\x0f\x0a\xfd\ S\xfe\xb7\xff\xd9MU\xc5\xf5>(j\xfc\x05\xc2\x98\ \x0b\xd1\x92\xb7\xbe4\xcc\x86\xe67\xfe\xa1\xf2\xd5\xf8\x82\ on\xad\x9c\xf4\xda\xae\xb7\xf66\xe9\x86\xd2\x1cI8\ $\xd8<\x84\xe8_\x80!6?\x98\x0d #\xd4\xfa\ \x82;\xaa}/m\xae8\xfe\x9d=\xd3\xde,\xbek\ ]\xcdG\xfb\xbd{\x1at\xb0\x16\xcd9\x01\xeb\x1a\xf0\ \x09h\xd2\xc6\x1a\x8d\xdfB\x0bw\xf8\x99\xe9\xaa!\xa0\ \xd5\x05c\x9d\xc9\x1cLg|\xf3d\x17\x09\x00X\xfe\ \x80\xff\x87\x97\xd4\x15W\x06?\xbb8\xf8\xc9\x05\xea\xa7\ \x17i\xb0\xfc\xfd\xbfUU\x8b\xe6\x09\xa0\xd0\xac-\xf5\ \x5c\xffyY\x13\x9au\x88\x9e\x90>C\x9cy\x82\xd2\ U_T-\xde]\xc3\xd7\xc2%\xc9a{t\xde\xc0\ KG'\xc88\x18\x04\xf2\x82y\xfbE\xdfZ\xd9\xc0\ \xfd\x89\xe7\xb5D\x0a\xcb9\xb2n\xf85b%J\x98\ \xb1\xed%\xadvgw^\x07\x8f}\x9c\xd8\xe2\x99\xa3\ s\x14|F&\xf6amL\xe7\x0f\x96y5\xe3\xff\ \xd6V\x9d\xf1\xfe\xbe\x1bW\x16o\xac\xf4\x06\xb1\x83\x8b\ \xd77Z'8\x14\xa1\x96RT\x5c\xeb\xf9\xf3\xba\xca\ s?.\xbcnuuS\x80'.FEO&+\ \xd1\xf5\xf0Z\x01\xd8\xfd\xc6\x86\xffm\xab\xfe\xf5\xb2\xa2\ s?*<\xe1\xfd\xfd\xb7\xac\xad\xdaP\x1d\x08\x88\xd6\ 0|\xb7\xce7\x00\xc3\x97\xb6\x7f\xb4\xaf\x91w&\xb7\ \xd9\xdb\x01\x86\xf4\xd1\x8e\x9a\x9a@\xac\xbd\x94\x0ac6\ \xcc\xd2p\xeaH\xce\x1e\x9ax\x0a\x00\xd7R\xdd\xd0\xb4\ `\xdd>}\xd1\xe9\xca7wJ\x85\x8b\x95\xaa\xf5\x8a\ g\x9b\x5c\xfb\xad\x5c\xbcD\xfa\xe6\x1e\xe3\xfdYF\xd5\ f\xacv\xc52\x896\xa4\x0f\x9c\xc80\xf6\xd7\xfb.\ \xfa\xac\xac\xdc\x8b\xf5=s{h\xc0@\x18\xd5\x01\xfd\ \x8eo\xea\x1a\x22\xea\xc2c,\xd1\xe9\xf8\xbf\x19\xb9\x0b\ \x86%\xa1q\xc0S\xa0\x08\x8f\x1f\x90\xc8W\xe2\x93\x0c\ Q\xa2\xf9S\xca\xdf\x13\xcb\x91\x03u\x1a|\xb2,P\ \xaf\x7f~%\xber\xa0\xdb\xe0\xd9\x97I\xf2\xa8t\xe7\ q\xd9n\xc8\xd5b{\xb4\xa0_\x06\xe6d\xfc\x95\xa4\ }>\xe9\xd5\x1d\x8d'\xbcWt\xc4k{\xee\xfd\xaa\ r\xc9\xe6\xbd{\xeb\xbc\xc5\x9e`y\x93Z\xd1\xa4\x96\ {\x82E\x8d\xc1\x8d\xc55/\xaf\xdd1\xff\xad\xfd\xd3\ \xdf*\xfa\xdb\x86\xea\x8d\xd5\x01\xcc\x8e\xdc,\xf0O\x8f\ &k\x0f\xd1\xff/\x1b\xbb$to0X\xea\x09,\ \xf9\xa1\xf8\x82E\x05\x87\xbd^r\xc3\x17Uo\xedn\ \xdaX\x13\xf0\xa8P\x9e\xb1s\x82G\x82\x99E!J\ Z\x03E\xdf\x90>-\xf0\xac.\xf6p\xf3\xc5-^\ \xe8\x01\x08\xe6^p\xb8\xad\xc6\xf7\xd8\xf7M\x90\xc3\xc4\ \x8eHA\x8b\x87\xdf6\x19>\x10.\xb4~bWl\ \xc4\xb9\x05\x00\x81\xf2Wl4\x96]$Um\x850\ c=\x8b\xff\xf1\xf6\x13\x8fTO\xb9\xbe\xfc\x0a\xad\x84\ \xbf\xdf*\xdak\xe0\x87\xa1\x87\xcfo\xa9n\xf0r\xcb\ \xc5\xebt\x00.\xb7\x0b\x06\x01\xcf_\xdc\x18\xb8sM\ \xb9\xaaa\xeb/\x1cLo\x93\x1c\xf2\xdf\xe6e\xffz\ |R\xa2]\xb1+\xec\xa7\xc3\x9c\xbf\x99\x92\x19uj\ \xc6\x0b\xa3|-\xab\xdf-V\xa2\x00\xe3\x03\xfe\x0c\xa9\ a\x7f`\xd7\xfbbc\xd7\x83)\x81H\x8a\xcc~6\ &\x19\xb2\x89\xd8\x11\x03\xe8\x05\xbf\x9c\x16 \x93\x94\xfa\ \xf4\x7fl\xaa\xbah\x8dw\xf6\xeb\x05G\xbe\xb9\xeb\xb8\ w\xf7\x9e\xf0\xee\xdec\xde\xd9;\xfb\x8d\xdd'|T\ v\xf3&ms\x8d\xaf\xf9\xb1J<\xd8\xcc\xa8f\xd8\ \x88\xfe\x07d\xb4W\xb7T\xfc|q\xe1q\xef\x16\x5c\ \xbc\xbaqi\x91\xd7\xc3{\xe51\xd5\xcdt\xe7\xdf\x22\ \x0b\xf0\x05q\xa4\x80oa\x0cZ\xac\xbfZZ\xb2\x1e\ *\x9e\xb0\x0ds/\x8e\x14\x0f\x01\xf6\x18|S\xd6p\ \xe5\xd2\x12\x8f\xae\xc5\x94\xd3\xf9\xb1i\x0e%\xdd\xd9\x8e\ 0EM\x9c\x05@\xf5\xd4\xc9k\xafc\x0d\xfb\xf9\xa5\ \xb6w\xb9\xbad4\x95\xea+\xaeP\x1bJc\x88\x0e\ \xd0?C\xd5\x8c\xf7\xf6b[\x8c\xfb\x13v\x940\xf6\ q\x81wOC\x10\xe5'\x92xLv\xc8\x7f\x98\x99\ \xbb\xed\xa2\x91\xbb/:\xe4\xd1\xf9C\xd3\xdd\xf68&\ C\x14@\xe8\xfd\xfb\x96\x1b\xc18\xccQ\xc14\x95\xed\ \xff\xc0h~\x9cXl\xedz \xf3\x1d7\x22}A\ ~\x02_\x83\xf3\xc6\xed\xd4\xe8\x11\xfc\xf3G8t\x83\ A\xb3\x1d\x12\xbc\xa4I-\xf2\xa8\xe5^\xcd\xa3\x99\xdd\ \xb726\x80\xf8P\xdf\xe8s\x22\xd1\x870\x8c\x12\x9f\ \xb4\xb2\xd4W\xee\x0dB\xf2c{\x8f\xdb\xf9\xc81*\ \xfc\xfaU\xcbKwV5\xf951z\x02\xbe\xf9G\ ,\xf2\x7f\xa3\xde\x1f|i}\xd1\xd9\x8bK\xbf\xafU\ 1\xbf\xf3\xdd\xd1\x02\xed\x07}d\xb2\x9cj\xc7`\xb7\ k\x5c\xa3 \x9e\x02\x80\xcf\xe3\xec|Q\xaa\xdb\xc7\xd7\ \xc0,\xb7w\xb9\x0a6\xb1\x98\xda\xa8m\x7f%\xa6\xd1\ \xf4\x8c}SRo\xbe !\xfc\xa80\x95\xa21\xa0\ o.\xf7\x80\x12\x99)\x17\x1ep(c\xb2\xec\xb0\xc9\ N\x9b,\xf3{BQf\x9ex\xa1\x06\xed\x85\x0bc\ \xcaQ\x1cL\x0f\xb8\x9e\xba\x1d\x86\xafBl\xea6\x18\ \x0e\x07zl\xce\x80\x5c'\x0e\xc3\x88\xadx\xb4\xc2L\ \x17\xb4\xee\x90; \x99\xb8\xcf\xb8\x91\xa7\x1b\xef-\xe2\ 9\x9f\x97\xa3\x9eL\xc4^F\xfcR\xa0w\xc2&d\ :mXjE\xbe03F\xa4@\xbe\x82\xef}\x0d\ \xea)\x0bKn\xfd\xa2\xe4\xad\x1f\xca\xab<~s\x90\ \x0bD!d\xe5\x80\xa6\xad/\xa9\xff\xc7w\x95\x97,\ .\xfc\xbf\x0d\x8d\xbe \x1aC\xfcD\x9b\xd5\x9a\x83)\ \x1f38Q\xb1\xd9\xc4Z<\x88\xa7\x00H\xfeZ\xa3\ \xe8\x13\x1eZn\x1b\xdb^.n\xe0[\xc1I\xc9\x0a\ \xa8x\xf2\xadQ\x80\x9e|\xb4\xb5\x04\xaa\xf1\xcdg\x0a\ +j\xcd\x14Puc7T\x08#)\xf8\xfc\x1cx\ \xb0\xf8\x15\xeb|\xa1\x87\x084\x96\xe9\xfe*\x0cC\xec\ @\xce\xf5\x95K\x81\x06\x8c!\xb1\xa9\xfb\xc8Hr^\ =e\x10\xaf\xad\xc7\xef\xe4<\xa9\xf0\xd7\x8c H\xa8\ \x96M-\xb4^#\xac\x10\x1f\xd3\x07\xb92\x5c\xe6s\ }\xed\xe4\x88\xf0\xc1C\x99T\x17\xd4\xff\xbb\xd3s\xcd\ \xaa\xdaI\xff\xdb3\xe9\x95=S_\xdd5\xf5\xbf\xbb\ &\xbf\xbak\xcc\x8b;N^Xz\xef\xd7\xd5\xab\xcb\ qH:66\xf0D1\x9c\x0f\x0a\xa6a\xa4\xd9\x8d\ \xe3r\x9c1x\xd2\x0e\xf1\x14\x00\xe6\xafP\x9a\xcaq\ \xc1\x5c\xef\x98\xc6}*Z\x9c\xe8@E\xdcZ\xd1\xd8\ \xa2\x8c\xe1\x82\x12\x0d\x07\x19EMZ<c\xb1\x03\xe0\ \x84\xcd\x81\xe4\x15\x04\xb3\xe1a6\x10\xf9\x17\x0fR\xa4\ \xc0Q\xacq\x9fl\xe81\xb6*9xkJ1\xd4\ \x9a2h\xbaEW%\x8a\x1aQ$.\x1a\x932?\ \xc7\x0d\x0b\x18;F\xb7>\x91@\xb4\xa2\xbfG|\x86\ \xdb\xf9\x93a8\x9b\x16\xcfi\x80\xd8\x1e-x<\xfc\ \x07uV\x11\xd0\x0b\xbc\xda\xfeF\xb5\xd4\xa75\x1a\x0c\ \xfc\xc7\x92\x85\x1f \x86\xae\x8ef@E\xce\x1b\x9e\x90\ \x91\x8aCQ\x80x\x99\xaf\xb8\xb6\x004\xaf\xa4z\xc2\ \xa9DB\xb5L1\xfc\xbezT\x8b\xa8\xd1\x5c\xc9\x11\ \xc7\x03\xb6\xf9 E\x98\xaab\x84vC~G\xe1\xc6\ G\x88\x01|\xd1\x00\xaeH-\xaf\x16\xc4\xdf\xa8\x132\ XW\x04\xd7\xc2\x9f\xea\x8a\x09\xac#Ce\xc60\x1a\ k*\xc4rwa\x9e\x0eH\xb0\xcb\xaf,\xc8\xcfK\ \xb4\xe1\xfd\x1cX\x8f>V\x88\x98\xe8\xdf\xf1\xce\xf3\x9a\ t\xe3a\xd0\xe64\xfb\x06\xb9=\x88\x01\xd3C\xfe\xcd\ i\xd9p\xc06A\x8cQ\xcbX\x8a\x83\x9d3:\x9d\ /r\xff\xe2\x94Vq\x15\x00&\x1b2\xefa\xeb\x0c\ \x94E\xc6lv\xf3\xee_\x14\xe09\x86\xc8\xf8r\x9c\ \xc8\xc0\xca.\x060\xc3\x09\x9a\x1c\xb3\xfc\x87\x830\xd2\ \x86\xe6\xad\xaf.\xd8X\xbeue\xd5\xf6/\x1bJ\xb7\ \xeb\xaa9OQ\xf4\x12\xc43A<\x80\x10`r\xa0\ \x8f\xcd+\xdd\x0d\xa4\x84MV\x1e\x9c\x95\xe9T\xf8P\ L\xa2\xfb\xe1\xd1\xde\xcf\x95\x97\xd7\xf8S\xdc\xce\xb3\x87\ %\xe0\x856\xd7\xcf\xfb\x00\x86>w\x90sL\x06\x0f\ v\x5c\x89k\x17\x90-Y\xb6%\x9b\x19\x89o\xe8\x00\ C\x95R\x9c)Yb-b\xd0\xffc'\x0e\xe5\x1d\ \xc7\xfc\xa1\x820S\x12\xcd\x9c\xa40\xe3\x90D\xb4\x9f\ \xb1\xcar\x08\xcc\xc0`\x98x\xe8\x9a\xca6\x05\x16\x9d\ \xa9\xbf3-y\xc5\xe9\xe9__\x9c\xfa\xd5\xcf\x5cK\ NT\xdf<\xbc\xfa\xe3k\x03\x8d\x15x\x09\xa6\xbb\x08\ \xb1\xa7\xe4q\x9b\x1d\xeb%`\x1b\x05\x0b\x82\xecN\xcb\ 4\x97\xba\x1f\xb3Js\xfc\xb0\xd4\x17f\xa5\xf2\x89\xb8\ u\x9c\x84;\x1e\x0dg\x82h\xc1\x1c\x96\xce$\xf9\xc2\ q\x19\x8e\xe6\xc7\xbd\xc5\xbe\xde\x09\x96G|yj\x82\ \xddv\xdb\xe1in\x9cJ(\xce\x01\x8e\xa7\x00\xe8\xae\ \x81j\xcaH\x0cug`GL\xc6\x18\x899\xc4z\ 4\x18\xd3\x068\xd3\x9d\x90\x8c\x06N\x0d\x1cvB\x82\ ;\xa7M\x19\x9d\x9d\x1e\xb9\xc9\x0d\x0bn\xce\x11]\x0b\ j\x85\x8b\xbd\x8b/p|v\x06\xab\xf8V\xd2|\xb0\ \xd3L@\x9c\x9b3\xd8\x90R\xb1P\xfa`\x9e\x7f\xe5\ MZ\xe5\xb7\x86\x1e\x99\xb1\x83\x9c\xac'\x0e\xe1\xf1\x18\ \xf3c\xb4\xd8\x17\x06\x85B\xc9\x1c\x90/\xb6t;(\ \x00\x98\x82\xc6\xfc1\x83\xfe\xef\xb04\x19\x1b\xe8pi\ \xe1\xa6i\x9f\x00\xb2D\xcbwo\xa4_Ev\xfb`\ \xd1\xe3\xa6bb\xa6\xeb\xaa\xf1\xc9(\x01\xbd55L\ \xa0\x0c\x80\x89v*\x0c\x0a\xc5\xe8\x01)]\x11\xd8x\ \x0a\x80\xecH\x0c\x0e=\x1foy\x8b\x0d!\xd1%f\ o\xd8\xa65\xc54\xeep@\x92\xfb\xa8\x1c\x07\xcey\ \xcc\xbb\xf3\xc3\x01CfHc\xd2l\x87\xa6\xc7\xa2=\ \x9d`\xda\xb2\xc0\x96\xa7\xf5\xcf\x7fm\xafX-\xe1\xe4\ \xb1\x07\x07\x10\xd6uI\x97U\x9f\xb2\xe7mc\xd9%\ j\xe1\x0a\xb1#l\xec)\x83\x0c[\xb2\x1e\xb6\xf2u\ \x00\x8a\x8f;Kv\xa7\x9a\xab=\xcb\x95\x87\x0d\xba{\ Z\x96\xce\xc7\xcd\xf5'x\xae z\x05\x90\x14\xb7N\ \x1b0k\x90\x1d\xcdAo\x06\xabf\xc6\x919\xae\x9f\ \x8dK\x13!\x8d\xb7d\xc5\xf5\x1e\x00\x98\x91\xe1\xc7\xb3\ Q\x17JL1k\xc2bk\x1b\x183\x8c\xa6r\xe9\ \xd3\xb3\x82E_\xe8:\xb4q\xa2h\xecc\x9d\xf1g\ \xa3\xd3\xdd6\xac-r\xa9\xec\x08\xfe\xec\xb4\x0e\x8e\x92\ \xec\xd2=\xd3\x07%8pz7\xb1/\xae\xe0iT\ \xbfo\xdd\xfd\xca\xc6\xbfJ*\x0e6e\xd8@9\xb8\ \xfc\x83f\x0185\x1e\x04\xdeW/\xad\xbe\xce\xbf\xeb\ ]\x0d_\xd9\x15nT0\xc5.\x0d?\xd7|\x0a.\ 6\xf0\x0e\xb0\x9a4VwdA\x18{\xd2La\x94\ 0\xbb\xcc~uh\xc6_\x8fHJ\x10\x99(\xfa\xd9\ B{\x0f\x98\xf3\x0c\xc3\xd6U\x99\x8e\x88\x0cH\x04\x9b\ \xac\xfc\xdf\xd4A\xb9\x89\x0a\xda\x8e\xde\x97\xc1\xb8\xb5\xc2\ F\xf0\xf84\xfb_f\x0dL\xb0c\xd6\xe1\xa53\xce\ \x19(\xce\x02\xa0\xd8\xec\xec\xb0\xdb\x8c\xec\xf9\x10\xa5\x1d\ \x84\x14\xed5\xfcyJ\x8d\xd5\xd7k\xf8\xe0XdW\ \x85q\xc1K\xd2\xdc\xc1IwNJ\xb0\x851n\x90\ \x8b\x84,\xcb\xc6\xb5\x87\xa6O\xc3\xf9g\x10\xd3\x93\xf8\ \x02V]\xdf\xf9\xbam\xebS\xd8\xcf\xd3\x9c^mO\ \xc3\x13\x13\x81}X\xd9\x0d\xd4I\xeb\xfeO\xaf\xfc&\ \xfcz/\x1c'\x0f\x9eo\xd8\x9db=z\xe0\x94\x8a\ /\xfbd\xc5\xe6\xe2!\x8d\x7f\x9c\x84Is\x94\x80b\ \x1a\x17M\xce}\xe7\xf8\x01n\x9b\xcct\x99\xbf\x95\xb3\ o\xc3\x8b\x19s\xdbl}_\xcb\xfa\x07X\xf2\x0f\x1b\ \xe4z\xf6\x98\x1c\x97\x0d\xd6zc\xb2@\x08G&\x19\ \x10\xc2a)\x0e,\xa4\x1c\xb1/~\xc4U\x00x\xf0\ l\x8ed\xdbQ\x8f\xcb\x93o\xd4SF\x19L\x06)\ \xe3\x0a\x8b\x8a\xa0'\x0c\x86\xeb\x805|\x94\x0e\x22\x1d\ \xbe\xbc\xe5\xd2\xd2\x0b\xf5\x8aoy]\x0f\xe1\x1e\x85\x85\ \x19#\x97\x1e\x96{\xf5\xa4\x0cG\xf3\x04I\xfcd\xf8\ %\xcein\xe2\xef\xc6r\xc8\xc6\xe5cR/\x9f\x90\ \x8e&7\xde\xe0y\xf8)\xf5\xa0\xcf\xd8\xfe\x02.\x86\ \x87y\x1c8g\x81z\xe3\x87g\xf1ij\x1e|\xf8\ \x17.B\xc3\xb2\x0e\xd73&7{\x11\xf6)\x9b1\ \x8f\x82\xd8\x08\xa6OJ\x1d\xff\x13\xb1\xb5W\x80\xad\xa6\ )y\x19o.\xc8\x9d\x9e\x03\xb2\x04am\xfe\x8a\xf8\ *\xe3\x03O\x13\xf8\xe2\x8b\x91\xc3{\xb4\x0c\xb7\xd3\xd1\ S\xe1'\xda\x22K\xf2\x11\x83\x12\x1e\x9e=0I\xe1\ \x8f\xd3\x98\xc9\xdb\xc3\x98%\x19\x032<Yy\xf4\xa8\ !#2\x12\xb8\xd9\x8f\xbf\xc92\x89\xa7\x00\x9858\ (\xbc\xb2\xdd\xadL\xbc\xdev\xe2\x1b\xd2\x89\xef\xab\xd3\ \x1f\xf5\x8c\xb9Y\x9d\xfa\xa8q\xe2B\xfbI\xef\x18\xd3\ \x1e\x96pBS.\xbaP\xb1\x83b\xe1)\xd6\xbf\xb8\ V\xad\xde\x16]\xec\x83\xa5\xb8\xf5\x88\xcc\x7f\x1e9 \ \xcd\xc5\xdf\xb9\x8c\xad\x01\xee\x13\xc3%\x9d\xd7\xa8\xe1\x5c\ \x89\x8a\xf2\xd7Y\x03\xef\x9e\x99\x9d\xea\x8a\xe5e\xb4\x1d\ \x80'\x05\x1bQ\xfb\xf53R\xfd\xae\xf0\xeb\x14\x98\xba\ \xcd\x09l\x14}\x1a(\xfa\x92w\x18`\x9fA\xa7\xc8\ \x8e$i\xfa\x9f\xe1\xc7\xcc\xbe\x91\x02B\xac\xc3\xc7\xee\ v\x1e\xf9\xa0\xac\xc4:%g\x5c\x81l\x89\x113u\ P\xc2\xab'\x0e\xbeyb\x9a\x0d\xa7\x8d\xc6\x8bl)\ \x1e\xdd\x0c\xc3\xd9F!;\x85\x95.mA\x03cH\ \xc9\x90C\x89\xde\x82Y\x0fdg\x1f\x92\xf2\xcc\xfcA\ `<\xc2\xef}\xedB0wC\xb0\x8c1)\xf2\x7f\ O\xc8\x9b\x91\xed\x82\x15nU\xbb\x8a\xb8\xb6\x00\x103\ \xb4(\x03\xcc\x95a\xcf\x9a\xe8\x1esv\xea\xf4\xdf8\ \xc7\x9cm\x1b0Ar\x0f\xb0\x8f:\xc7\x98z\x97\xa4\ 8\xb0\x80Ck\x80?\xf3c4\x16\x18\xcb~\xae\x97\ \xae\x13\xb6;B\x1c6\xe5\xd4\xd1\xa9\x9f\x9c\x9a\xf7\xeb\ \x89iG\xa4+Y.\x19\x14\x1d\xc2`\x97\xa5L\x97\ mr:\xbbj\x5c\xf2\xa2S\x07\x9e?6\xcd\xce\xc7\ ~\x99\x98\xc7\xc6\x0f\xf4Pm\xaaN.|\x06\x0d\x05\ \x0b\xdf\x9ebX\xf0\x03Q\xa6i\xf2\xd6' \x12\xcc\ \xc9F:\x03o\xb7;\xd2F\xf9\x0f\xbf\xdf\xb0%\xf1\ \xe9\x90#C\xc6\xa6\x98a\x8c\xbdNN\x1d)6\xf5\ \x0ex\x84\x98H\xc9\x0e\xe5\xc6i\x83\x9e\x9c7p\xe6\ \x00{\xf3\x0b\xf5\xe2\x9ev\xe1\xa0\x0dp\xc1\xf9q\xc4\ AT@\x98\xd9\xc0\xc4x\xce\xe2BD\x8d\xc8[<\ #)\x8c\x1d;$\xf9\x95\xe3rF\xa78\xcd\xf1\x10\ =\x0a\xb3\xc9\xc6\xf1\xb9\xae\x17\x8e\xcf\x1f\x96\xe6\xe4\x81\ \xc4\x8dX$\xba\x86\xb8\x0b\xc0\x8f\x1c\x10d3\xbe\xf1\ \x03\x7f\xf6Q\x17\xe8\x87\xdf\xc5/\x09\xed=\xaf\xd01\ \xa9\xa9T_s\x93Z\xb3\x83\xb7\xb4#\x18\xda\xd8|\ \x16yh\x9a\xfb\xce\xe9\x03^>i\xc8;'\x0d~\ \xef\xa4\xbc\xf7N\xce\x7fgA\xee;'\xe5\xbd\xba`\ \xe8]\xb3sFg%\xb7$y\x97\xa1{\x8a60\ \x9f\xc7\x5c\xe6\xdf\x91b\x18u\xbb\x9aj\xc3|@\x9a\ G)c\xee\xd1\xa7\x19\xe3\xae\xc0x\xc4O\xe7\x12\x8a\ 7U\xb1i\x84\x1e4\x1er\xad}\xc2e\x92\x8c\xc2\ (v\xf7.0T6\xc6\xce\x18\x99\xfc\xc2q\x83o\ \x9f\x98\xe8\xc4\xc7\x04 nu^\x15\x0f\xe7r\xa3\x04\ \xbdF\xdf\xc1\x7f\xf8bs\x07\xda\x93\xedbW\x14\xf0\ P\x1aY\x8a\x86\x05\xa0\xb7\x829B,Z\x04\x9e,\ \xbc\x18\xcd\xceM\xfc\xcf\x89y\xc7\x0e\xc4\xd7\xbf\xf2N\ Xh\xeb\xf1\xf4\xefR\xf0\x0c\xd8\x0d\xc2O\xc8O\xca\ \xa4\xdf\x1c\x9a\xf6\xcfc\xf2F\xa6\xbb\xe4\xae4\xce-\ t\xc99\xd0,\x1d\x8c\xd8(\xc3\xc7\xe6\xb4\x8f\xfd\x99\ >\xf9wPG\x87\x0c\xc7\x8b\x04F4k\xd8k,\ \xbfT-\xff:|\xe3iz\x89/\xee\xe3\xc8\xb2\x9c\ \x95\xe4\x1a\x9b\x958=7yVN\xc2\x8c\xbc\x94\xf1\ Y\x09\x03\x92\x9c\x0a:\x11\xae\xc4\x91]\x02K\xf4o\ \xc5\xeb\xc1\xd3D|\x22\x88\x02\x8c\x85`\x83\xd2\xb4\xd3\ \x5c\xec\x18~5\xfc\xa2\x14\xbb}\xf2\xf5\x0c4\xd5\x9e\ bf\xa3\x8e\xc1\x87m\x0d\xd9\xb0'\x06&\xde\x9c2\ \xed\x06\xd9\xee\x94\xbb#r\xa2\xe1\x80P\xb1\xcc\x04\xc7\ \xaf\xa7\xe5\xad\xf9\xc9\x90KG%\x0fM\x02Q\x80+\ \x85]p-\xf0\xe9<\xba\x22\xc2T\x17\x5c2\x0c\x9b\ \xcc\xe6\x0f\xb2\xdd<%\xa3\xdc\xc7\xdf\x04\x8b[#\x06\ \xdb\x0dL\x1a\x86C\xcf\xe3\x1cT\x22j\xcc\xdc\x85\x80\ \x1ddlX\xaa\xe3\x99\x13\x86\xdf89='\x01\x1a\ j`\x8a\xa3(\xc4\x91\x81M\xf8\xe6\x11\x0ep\xaa\xc9\ Y\xceW\x8f\x1d\xf4\xbbi\x03\x93\xf9\x8c\xa5\xcd\x85\xb2\ k\x03\xd1\x1d\x22\xd3\x168\xab}\xfc\xe5\xfa\x11wC\ \x22\x08\x8b\x05\xc5\x18b\xc1S$}~\xad^\xfd\x03\ w\x15\x13]\x1dqm\x013\xa4\xf9\xaa\xc5J\x14p\ \xe5\x90\x0dUQ\x1b\x22\xb5\x12 }\xca\xb8K\xd8\xf1\ \xff5r\x8eF\x93\xd8\xa1A\x84\xca\xbf\x969U>\ \xf6\x0d\xf7\xa4\xab\xf8\xd4\xb2\xdd\x1dQ\x91\xc2K\x01\x06\ \x12b(/\xcd}\xff\xbc\xdc\xb7O\x1e\xf2\xf0\xcc\xac\ t\x07\x83\xf6\x00\x1fme:\x8c\x1b\xdc?L\x8f$\ E\xf9\xdb\x9c\x81/\x9e4d\x9fO\xf2\xaa\xb2\x19\x88\ (\x80P&\xdb\xe4\xdc\xb4$\xb1N\xf4.\xb03\x14\ \x0cP\xb2S\xb9\xe5\xf0\xac7\x17\xe4.\x18\xec\x82\xa6\ @\xa4\xc50\x0app\x80$\xa5:\xd9c3\xd3\xfe\ {b\xdeq\xc3R\xcdN\xc6\xee87\xa7g\x04\x00\ .\x9c)\x0e\xe7\xb8\x8b\xf4)wH\x8a\x13/\x96a\ \xeb\x18/\xdbS\xac\xad\xb8\x5c\xad\xda\x12\xed\xf3\x01=\ \x06\x06^\x8f\xfe\xb9\x5c\x1c\x1d\x85\x807ZDV\x06\ ;\xa6!>e\x9b-\xf3P\xfb\xfc\x7f\x07g>\xcd\ r\xe7\x06\xddC%\xc5\xcd\xa7\xbfF=\xc0\xe9\x88\x14\ \x97\xdf1X\xcf\x99\x1b\x9c\xf9\xb8s\xc1k\xf2\x80\xf1\ p\x08\xaf\xfa\x08\x7fz9B\x05 \xe309?\xc5\ \xf9\xf3C\xb3\xd6\xfft\xc4\x133S\x8e\xc9K\xcas\ I\x0eY\xc6\x91\xd3\x18{(\x7f\x10\x95\xf81\x17\xcc\ xm\x03w\x0c{\xcd\xb6\xbe\x18|\xc5=\x91\x92\xec\ lR\x86\xfd\xf6\xc3\xd2\xbf\xbfx\xe49\xa3S\x9c6\ e\xd1\xae\x1a\xcd\xf4/\x84o\x1d\x03\x89{X:s\ \xda\xf0\x8d\x95\xbd\x99^\x1e\xbc.\xc2\xac^\xf0B'\ CUjt\x86\xfb\xdf'\x0cy\xfb\x94\xbc\xf9\xb9\xae\ d\xbck\xc3's\xfc\xb1\xcf\x91g\xb3\xa8\xe0Y\xcc\ \xf4\x05\xbd\x84\xda\xef\xd0D\xe5\xec\x11)K\xcf\xc8\xff\ \xe9\x84A\x99n>\x8c\x05m2\x94\xe8n\xb2\xcc\xf1\ o>G\x84\xae\xfa\x82\xdb^c\xeb\xefff\x04\xf3\ \x98\x85K7\x92\x87J\xb3\xff\xa6\x0c8\x5c\xe6\xb5\xae\ \xde\x0f\xa6\xa9\xa1\xab\x1b\x1ee\x9b\xff.6E\x08\xcf\ \x10`\xa6\xdd\xf2\xdc'\xd8\xe0c\xd0\xa6Gb\x9b\x0f\ HG|-\xb3\xea\xa9\x90\xfd%\xba\xafN\xd6\x03:\ T\xfa\x99Sv\xa5\x04\xed\x03\x9d\xa9\xb9\x12\xda}\xc8\ h\x91\xf9\xdf\x0b\xc1(\x97\x0cU\x97\x8a\xea\xbc\xfb\x1a\ \xb5e;J\x97\xd6\xbb\xf6T{\xfd\x1aT\xe8\x00\x1e\ %p\x89<^\x0fD\xc4\x14\xecFG\xe0\x04\xea\x19\ \x98\xe9dY\x1a\x97\xa4\x9c\x9c\xaf\x1c5|\xe0\xd8t\ Gr\xf3\xd3\x82;K+\x8f]T\xe3Su\xdeL\ \x8d&GB8\x1f?\xc2y\xde\x94a\xb0\xdcq\xb4\ \x7fYT\x7f\xce\xa2\x92`KbF\x86~B~\xd2\ \xbf\x8e\xc9u\xe3h\x87hX\xbc\xb7\xfe\x17\x9f\x95\xa8\ Q\x9e]:&\xcf\xfd\xdf\x05\xf9\x1d_`\x07,\xdf\ _w\xde'\xa5f\xa2E\x06?f\xa6\xa3\xfa\xfd\x9f\ \xcfjY\x8d\x1aQ\x94\x0c\xc9\xa7j\xdfW\xfb\x16\xed\ \xa8|\xb3@+\xf2\xe0\xabgq\x07\xaf\xa1Gg\xa0\ \xf1H\xf4\xdcP\x18\x9b\x94\xc2~59k\xfa w\ ^\x0a\xbe\xa0\x00jp\xa6\x9bn\xa6\xa7\x05\xc0\xd0!\ >\x03[\x9ec\x9b\x1e\x92T/\x04\xc5L8(\x95\ rR\x9e1\xef\x19{\xe6\x04s[\xd4\xb9\xaa{\xe0\ \xb6\xc8\x08\xec|\x9d\xad\xbd\x15C\x8f\xff\x91\x05\xd8L\ \x08\xdd\x91\xae\xcf\x7f\xd5=\x10\xae:\xb2K\xe6G\x8b\ \xa4\x84e8\xd2\x1ca\xf5c\x8c\x22\xb8\x19\xbe\xcd*\ \x8c\xdcCy.^\xf0\x18\xc3\xcbi\xf9V5\xa3\xda\ \xe3\xfdjg\xf1f=\xed\x87j\x7f\xb97\xd0\xe4S\ }\xaa\x1aT\x9c\xa0\x13\xf8\x985\x9f8\x0a\x8a\x9fC\ \x96d\xd5\xef\xb6\xdb\x9c\x0e%\xc3\xe5\x98\x90\xe9\x9a\xe4\ \xf6\xcf\x1c\x9e\x99\x99\xe0\xb0\xc9\x8a\x88$\x1eu~U\ ;\xfb\xa3\xfd_U\x04\xf0\xce9/\xf6\x91eE\x08\ %c\xc9v\xf9\xbbs\x87$\xbbq\x02\x92p\x04 \ \x00Z\x03g\x8f\xe8<\xf8g\x9c8\x04\x05\xc0e\x8b\ <\x9c<:?\x01\x01XZ\x1a\x9d\x00\xc0\xe9\x8f\xcd\ s\xbf\xba`\x08W\xc8\x88N-\xe0\x02P\x12q\xc9\ 13\x02\x93f9j\xde\xfb\xf9LX\xe7I\x17M\ \x00L\xc0;\x8c\x00\xee'\x16dIj\xf4\xab\xab\xf6\ T\xbe_j|_\xd9T\x11`\xd5^\x15\x1a\xe9\xe8\ \x0aN\x82\xfb\xb9C\xfc\xe1\xc7\x00\xb0\xcf\xdc\xc0]p\ \xef$\xb7M\x1e\xe00\x06$9\xe6d*\xe7\x0cO\ \x1c\x99\x9df\xfb\xb1z\x0by\xc4\x92\x02 \xa2Z\xf5\ \x06\xb6\xbe\xac|\xfbgn\x980\x121\xda\xe0'e\ \x98|\xe4\xe3r\xc6\xa1\x98\x97cH\xd1n\x00/\x04\ \xaa\x0c\x95;\x95O\x8e\xc1\x1bH\xd81\x1da\x8ar\ 1\xd4SF\xdbN\xfe\xd0fsa\xde\x89\xf0\x921\ \xd7\x99\xf7T\xcc$\xe5G\x9b\xc6\xde\xf4\xc8\x8cY\x9e\ oqs/\x8f\xd2\xcei\xc9\xbb?^)\x968\xb8\ DhN\xea\xba\xeeW\xf5\xc6&oSP\xf3+n\ \xbf\x86]\x8a\x9a\xaeC#\xdf\xae\xc8.E\xb2\x05\xbc\ \x89N[\x82\xcb\xe9\xb4\xd9\x14~\x1f\x9c\xfb\x85\xc0\x92\ \xf0Y\x92>\xdcSw\xd9\xd22<\x95\xb9\xd5<C\ \xb8`4\x1bL\x9f\x95\xe5x\xf7\xf4\xe1f\x84w\x1c\ \xed\xab\x8a\x1a~\xba\xa8\x10\xe4\x0a\x09\xff<&\x86t\ |~\xe2S`\x86#\x17\x00\x8c<.\x00\xbf\x5c\x0a\ -\x80\x96\xab\x8f\x04C:zp\xc2\xcb\x0b\x86\xc0Y\ \xa3\xcaZ\xc6\xf2\xfd\xf5\x17..\xc6\xc5\xc8/\x1c\x98\ \xee\xac{\xfb\xe2\xe9p0>u\x11\x8b=\xe5qa\ ^\xc1\x019\x0c2\x8fQ\xdb\xe4+\xf0\xb2\x82z_\ Au\xfd\x9e&es\xbd\xbe\xb7\xd6_\x17\xc0\x82k\ \xba\xc2\x0cb\x1e\x82\x87\xb3L\xb7ml\x86c\xbc[\ \x1b\x92\xe6\x1a\x92\x9e8$A\xcaKv\x1d8\x0f\x0d\ w\x05g\x89*\xc2\xe2A\x0f\x0b\x80\x09\x0f\x83\xe1\xdf\ \xf4\x9c\xbc\xe9A>k&\x02\xc5\x0cc=)\x9f\x1d\ \xf3\x82\x92zH\x8f\xc5Px\x18\x98\x03\x18\xce\x00\xfa\ \xe9\xb9R\xc57\x98\x01#\x0c\xb0H\x88\x09\xd7\xd8\x0f\ \xff\x1d\xac\x81o\xbd\xfc\x92{?\xcdy\x9b\x7fcf\ \xe2f\x11\xfbp\x01X\xc6$\xe3\xdb\xf9\xcf\x01\x91\xdd\ R(\x9aT\xfd\xbc\x85{\xbf\xaa\xc2w\x97F\x93\x18\ \xdc#H\xc7;\xa7f\x5c3\x99O~\xdeY\xa2z\ \xfd\xfe\x92\x9a&\x11\xb6\x88\xe0\xce]\x0e\xfb\xa0\xd4D\ 3\xefE\x94\x7fx\x114<\x81`Em\xa3\xb8\xf8\ \xc8q;m\xd9i\xc9\xb0\x10E\xd6\x05\x9b\xeb\xf1\x07\ +j\x1a\xf8\xd9#>\x1cp\xd8\xe5\xbc\x8c\x140\xdd\ ]Tt \x8aP\x18\xcc\xec\x04!\xe435\x1b\xba\ \xae\x1arS \xe0\x0d\x04T\x15j\x198*\xd1n\ S\x5c.{\xa2\xc3\x0emIX\xe5\x15}\x11 ~\ 0\xd4\x10q\x88\x0f\xdf\xd0\xc3\xf4\x22\x01\xd0\xd5@\xe0\ \x87\xe7\x95\x0d\x0f\x1cp?\x00\xa2[6\x12\xf3\xe4\xb9\ \x8f+\x03\x0e\x87M\xbd$\xd6\xda\xd2\x12\x8d\xde\xca\xed\ \xf6ON6\xb4@\xb8A\xc5\x03\x9b\xab\xe9\xeel\xf9\ \xb4\xa5\x8a3\x91_{\xd7\xe4b+q`\xde\x86\xc8\ \x84\xb5\x03#\x94\x17f\xf1\x8f\xdb\xb1d\xa3{S\x0a\ \xe0X(\xd1\xbfXV\xfaY\xa1\x07o\x10cb\x1c\ xtx`\x08\x0c\x87b{{A\xf6\xf4\x1cs\x08\ P'\xa9j\x96\x053\x14\x11\x9d\xf1\xa0\x8b\x15Ka\ \xd3|8\x9e:\x8a\xc3\x01\xe1\x03\x1cj\xc6V\xa4\xfc\ x\xe5\xd1\x1d\x8e\xff<\x04@tW\xd0\x09\xdcw\xfc\ \xe2\xe9b\x9e\x8a\xaf\x1dD\xcbN\x93VNp\x05\x0e\ \xc7\xbc\xd7\x15A\x8c\x9c\x18\x1aJ\xf1\x03\xe2\x821Y\ \xb1\xbb\x1c\x13\x7f\xa5M\xbeC\x92\xed\x18OX,\xb1\ 9ex\x0a\x8c/~\x13\xac/\xe0\xb5\xec^\x0a\xbf\ \x04\xc4\x9d5J\x1bzV$E\x17.\x13\xaeK\x96\ \x14E\x9a\xf8[\xc5\x91\xc0\xbd\xe9-\x15\x84>\x0d\x8f\ I\x01_\x85\xff\x1f?\xb8\xb1\xf9\xff\xc7\xe4\x82\xa2\xc9\ Q\x0d\xed\xc1\xefj\x96\x14x\xb85\xc3\xbbt\xc2A\ $p;\xc0\x06\xba\xe4\xc3\x06\xba\xc1\x17\x0e\xdf\x11\x1a\ \xee\x06S?\xd23\xf2\x03\x05bS$\x88#c\xc8\ x\xe6\xf1<>\xa3\xf2\x81\x1f\xcd\xaf<\xba\xc3\x0f\x8c\ \x02\xb1-\xbe\x98\xd7&\xd2F`\x86\xb75\x07\x84\x03\ i\xe5\x84G/w!<\xediz\x85\x00\xb4 K\ \xcc1\xeeg\xda\xc4\x9b \x8a\xf8\x88\x0b(C\xd8\x08\ \x90\x1a\x0b\xd8\xd2\x9fiU\x9b\xb8\xab^\x0d\x04\xda>\ \xf1\x1a=)W\xacw\x06\x18\x1c\x1d\xdfQ\xa6I\xb9\ \xc7\xdb\x86/\xc0\x5cF\xf4\x10X4\x01C\xf2\xf9\x83\ \xf7\xac-\x7f\xfa\xfb\xda\xd6m\x86h\x80v\xc5]G\ \xa4\xdb\xf8\xe4$\x04\xd1\xdb\xe8\x15]@-4\x07\xc6\ \xf0mz\x06\xe7\xd3\xd7|f\x01\xc4\x8e \xd8\x95\x9c\ /\x1f\xfb\xaa\x92\x92\x8fb\xdck\x0d%\xbf\x06\xbf\xa7\ Z^v\x81T\xb3\x03\xbb\xb3p4\x0e\xaa\xd9\x81A\ \xe6\xae`\x9d\xefb\xb2\x94\x7f\x82m\xee\x93\xd0\x0a\x82\ ]\xbd\xf7\xd2\xfa#-Y\x8e\x7f\xe3o\xb9W\xbb\xf5\ \x8b\x92\x8f\xf77\xe1j\xb4ia@\xab\x8e\xb7\x1dF\ \xa58\x96\xffd\x88MQ(Y\x89^H\xef\xaa\x98\ \xa0]Gd\xc7\xb8K\xf4)\xbf\x83\xd0\xf1*\x18|\ t\xb4\x95\x8d\xfb\xb5%\xe7\xe9\x15\xdf\xfcX\x5c{!\ \xbc\x9c;\x12\xd2\x94y\xff\x0a\x0c\xbd\xc0P\x14\x088\ v\xf9\x1d\x1cfp\x87\xf7\x91t\xbb+8\xf1\x16e\ \xe6\x03\xb2\x82c\x03\x00\xb1\x9f\xe8F\xcc\xbe\xc5\xc6\xa0\ \xfa\xe4\x86\xaa\xd3>,\xf8\xa8\xa0\x89\xa7O\x0ci\xc1\ e\xdf&K\x17\x8fN\xb2A\x03Ol%\x88\xdeE\ \xefj\x01\xb4\x80\xa12$\xef\x96\xe7m\x1b\xeea\x9a\ n\xc8\x06\xd3dC\xc6\xf1\x95F\xca\x10\xdb\x89\x1f1\ W\x0a\x94OnU{\x97\xc5\xc4\xd8\xe4\xa3\xbat\x1c\ \xcam\xf8\xf6\xaf\xb0}w\x87\xd1P.I\x01\xbe\xd3\ T\x03\x08\xb3l(.)k2\x9b\xfag{\xfa\x08\ \xd8\xc3d\xea%\x88\x193\xf6Q]\x0f\xce\x15\x98\xa3\ 0\xbf@\xa2\x98\xf1\xcf\xdb^\xf0\xc3\x9f8\xf7\x04\xb4\ \xc5{k\xef\xfe\xa6\xbe\xca\x17\x94\xf0\xd1\x14\x1c\xc5e\ \x1e\x15\x9d$\xf3\x16\x00\xcbK\x90?<9;7\xda\ \x811\x04\xd1\xd5\xf4V\x010\xff4\x7f`\xf3\xbf\x94\ M\x0f\xe3\xa0nh\x17\x98\xc5\x9bIr\xcapi\xd6\ #\xca\xc0\xc3y\xb1\xea]\xe5\x0am\x87\xb9\xd0\x1c\xb1\ \x86\x1a\xf0Ul\xb15l\x91|\xd5\x9a\xbf\x11\x1f\xbf\ \x02\xf5r\x0d42\x8ep\xa4\xe5\xf3\xf9\x18\x102\x10\ \xb1\xc3\xe3\x5c/\xab\xac\x95\xa0=eN\x01\xa8\xc86\ E\x81\x05XS\xf8;\x0f4]\xf7\xf9\x03\x0dM\x9e\ \x8a&m\x9f\xea\xdaT\x15\xd8W\xe7[U\x1a\xac\xf1\ \x07[\xe6\xf1hI\x8b\x18\x04\x00\x8e\x95\x9e9*\xeb\ \xccC2\xc0\x03\xf3\x8f z\x1b\xbdT\x00L l\ \x10:\xdf\xb6W\x95o\xee\x905\xfep\x8c(I\x86\ \x94\x94\xcfN|\xd7\x960\xa0\x97\xdb\xcd\xe6\xe8\xc5o\ nL\xf8\x02^\x87\xf9\xa6x\xa8\xf5\xc3F\xb2\x0d\xf1\ A7\xb4:op\xda\x7f\xf7\xf8\xf11g\xcc'\x8a\ \xc4p\xd05\xb6\xba \xa21\xba!A4hO\xea\ F\x00*\xff\x86\xb9\x0bRE\x83\x9d\xe6~L\x96\x98\ \x81\x16\xc0\xfcL\xe5\xb53Ga\xb3\x8e\xd2\x97\xe8\xad\ \xf4\xf6n\x07(;\xaeQ\xe7\x18\x13\xae\xe3\x05\x18*\ \xcf\xb0\x81\x9b\xd4\xc6\x02\xe3\x933\xf5\xaaM|\xa2\xa6\ f3\xdb{A\x13\xc0\x1fN\x84\x05\xb8\x04\xf8\xc6\x89\ \xd8\xcc\xedD\xcc`{\xd1\xcc\x17/~_Wg0\ ?\xd4\x1b4\xc3\xab\x19\x8d\x9a\xee\x0d\xea\x1eMoR\ uO\xd0\xa8\x0f\x18\x8dAX\xd6\xfc\xfc\x85\x02\x90\x10\ 85\x04>\x98cN\x9e\x13\xeb8c<\x1e\xc2b\ \x18\x896\xdb\xf5\xd3s\xc0\xeb\xde\x9e1\x09k\xd3\xab\ \x05\x00\xeb\xc6`\xf4mN\xc7\xe4\xeb\x83S\xfe\x88%\ \x15\xca,\x14Y^v\x8d\xc6}\xfa\xca_i\x81J\ ^\xfa{)\xe6%pp,yk\xc4\x08a\xe1\x94\ \x88\x16\xa8\xb4\xab\x0cg\xfa\xf1\x04\xd9s\xdb\x1b\xf1N\ \x11\x87W\xfa\xf9\x17|\xe3\xc7\x5cl\x06\x97\xf8\x0eL\ \x08s\x85\xaf\xc7\x80\x0c\xad\x0a\x9c\xa2Z>ux\xd2\ \xac\xdcDSc\xc4>\x82\xe8}\xf4\xf6\x16\x80\x09\x14\ L\xd7\xf8\x0b\xd5\x897\xc3\x82\xc1\xf8DY\xba\xa4\xe8\ 6\xa3\xb1HZx\x92Q\xb6\x8ejZ\x96\xc6\x90l\ \x90'$\xe9\xf6e{\xcb\x9a|=8\xe72Z|\ \xdd8:\xdb\xf1\xf0\xec,\xa8\xa6@\x98zs\xed\x84\ \xfa\x86\x00\xa0\x04(N\xe7\xa4\xab\xd5C\x7f\xc7\xeb\ i\x0a\xef\xb7\xe5\x0dvO\x99\xf6\xe5\xadZS\x95\xc1\ \xa7\xd8\xc6\xb99\xc4\x02\x14<\xf1C\xf4w \x95\xd9\ \x97\xbb\xcb\xfeW\x100\x0c>\x97g\xf7\x02'\xe4y\ \x0e\xf2\xa9\x9c\xeebwL\xcfr\xdal\x98O\xb1\x05\ @\x10\xbd\x97\xbe\xd2\x02\xc0V:N\x161\xf1\x0a}\ \xc2\xf5\x06\xe3\xe35xs\x00k~\xf5\xbb\x8dO~\ \xa2\x95\x7f\x83E\x10,\x01\x96\x7f\xf8GE\x80_.\ \x0c\xb1\xf6\xed\x12\xbd\x1cU\xd7\x9f\xdf\x8d\xf7\xd5M\xc4\ \xd6\xee\x02\xce\xc7oM\x19C\xdd\xfa\x9d\x87gN\xcc\ J\xc0\x8dfH\xba=0\x04\x11>\xbdz\x14P[\ \xb0~\xaf\xa9\xbe\xcd\xcf\xda6\xdd\x07e\x8b\x1bx.\ \x04\xb0'q\x98q\xdc\x1bZ0\xd0\xb4\xf5\xedT}\ \x07\x93\xa5\xa6\xa4\xc3\x13F\x9fcs\xa7`c\x9cF\ \xd9\xf7_tC\xddZX}\xcc\xe2\x9a\x1e\xd3y\xec\ \xed\xc1\xbb\xcaO\xcc\x1b\xf4\x93Q\xa9xK\x01\xa7{\ \x14;\x09\xa2\xd7\xd2\xd7\x04\x80\x87\xd6\xd0\x83\x81\x0d\x7f\ \xb3mz\x9c?\xb2\x83m\x00\x03\x9b2\xba\xee\xca\x94\ \x03\x8d\x92\x11\x80u\x00K\xa0-A\xcd;\xd9>\xf1\ *%m\x14\x1e\x8d\xcf\xf7P\xb9\xec'\x98\xed;H\ goP_\xf0\xde\xbe\xefk\xfd\xdd\xff\x8a\x1b\xcc\x90\ f\xfec\xec\xe7\x87$\xdc93;\xc9i\xc3\x0d\xdd\ \xdf\x0c!\x88\xc8\xe9c\xf5b\xb3Y\xcdd\xbb}\xe2\ o\xfc\x93~\x87w\x83y\x97\x0f\xbe\x19\x1c.\xc6W\ \xc5\x8c\x00,@\xe1\x13\xe5/\xd8\xa4\xec}\xc3\xf8\xf4\ \x82@\xc1r.\x0aD?\xc2LQC\xfa\xe7\xa6\xaa\ \x1f\xea\xfc\x98\x17z\x00\x9e\xdd\x98\xf4\xabC\x5c\xf7\xcc\ \xc9A\xeb\xcf3_\x8f\x04\x85 \x22\xa5\xafv\x8c\xc8\ 6\xa7{\xf2U\xfa\xb8k\x0dY\x81\xc2\x86\xa6\x80\x97\ 9\xd1D\xe0\xdf\xb8\x80\xbdD\xcc\xf0U\xb0\xb5\xb7\xa8\ \x95\xdf\x9b\x0e\x89~\x02\xb7\xbd\xb5\x01\xf5\xfd\xbd\x8df\ \xb2\x9b\x9b\xbb\x03l}\xe0\x17f/I\xbfpD\xc2\ \xcd3\xc4\x8b\x18\x09\xa2\x0f\xd1\x87\xb3,4\x04\x1c\x93\ \xae\x09\x8c\xbd\xe1\xc7r\xdf\xa6\xee\x85\xb5B\xbc\x03\xc0\ \x98\xb7\x5c\xdb\xf4\x04\xb7\x19D\x7f\x01m\xb0\xfe\xca\xf7\ U?\xd4\x04\xf8\x13\x82\xdd\x99\x99q42\xd8\x7f\x99\ \xb1_\x8fM\xbewNN\x8a\xcbF\xb9\x8b\xe8s\xf4\ I\x01\x00{\xce\x91d{\x82\x9c>\x0a\xb7\x98\x9b\x9b\ K \xeek\x86;\xc4f\x80R\xb2$PS \xb6\ \x12\xfd\x01VP\xa7\xde\xff]-\x7f\xa8\xb7%\xf1\xbb\ \x03\xa8S\xe8\xb2\x94l\x93o?,\xf5\xb6\x19\xd9\x89\ v1\x8f7\xcfk\x04\xd1g\xe8\xdb\x8dV]\xd7\xb5\ \x1f\x9e\x87\x05~'\xaeC\xc0Fh>i\xff\xdbb\ \x95\xe8\xd3`\xe5[W5\xedo\xdf\x95\xfb5~'\ \xa8\x9b\xe0\x1d?`\xeb%6$\xd1\xfe\xe2q\xb9\xd7\ N\x19\xe8\xb4A3\x80\xec>\xd1'\xe9\xdb\x02\xc04\ \x9f\xcb\xb7\x13\x17x\xb1\xec\x00\xd3F\xe8u{\xccU\ \xa2O\xc3;\xdf\xd9\xa6\x8a\xa6\x0f\x0a\xfc\x98\xb6\x9d\xa5\ ~\xdc\xc0\x8e\x1f\xdd\xa9\xc8'\x0fu/<%g\xf6\ \xe0$\x9ch\x94\xa6\xf4 \xfa,}[\x00\xf8C^\ a\x15<0\x19`)\x0c\xa3e\xc6_\xa2/\xc3[\ |\x7f\xf9\xba\xa2\xc6\xa7ri\xef6\xe3k\x0cp;\ \x1e\x9e\x9e\xf6\xc4\xd19\xd9\xc9n2\xf9D_\xa7\x8f\ \xb7\x00l.\xc3\x9e\x06\x0b\x9dw\x01!\x86\xec\x1a$\ \x16\x89\xbe\x0c\x88\xf9\xbb\xdf\x17\xaf(\x0dH\x0c\xdf\x1d\ \x8dc\xbd\xba\x06\xf0\x97\xbf\xda\x05\x87\x95\xa59\xe5\x0b\ F:\x97\x9e\x96s\xee\xf8L\xb7\x1d\xc7\x9e\xf1\x0f\xa9\ \x00\xd1\x87\xe9\xe3\x02 \xdb\xd8\x88s\xf9bg&\x00\ $\x82\xc9\xc6\xe0\x13\xc5*\xd1\x97\xa9\x0fh\xff\xdc\x11\ \xc0\xa7\xaf\xb8\x09\xeeB+lN\x22\x22K\xe7\x0cO\ x\xfd\x84\x9c\x87\xe6\x0e\x19\x94\x02\x15\x7flu\x8an\ \x1f\xb2\xffD_\xa6o\x0b\x00\xc0\x0e9\xd7\xb0\xa7\xc0\ \x02\x8e\x04\xe7\xa3\xc1\x0f\x02ko\xb0\x99IZ\xdax\ g\xce\xe1b+\xd1'\xe1\x89,I\xab\x0a\xeb\xbf\xab\ \xf43sJ\xa88\x81\x9d\x84\xe0}s6\x82\x0f\x94\ \x8d\x0c\x87|\xe2\xe0\x84\x95\xa7\x0f\xfe\xc7\xfc\xdc\x89\x03\ \x12\x1c6\x94\x9c~Pj\x08\xc2\xa4\x8fM\x05q\x10\ \xdc\xb8\x1b\xda\x86\xc7\xa4M\x7f\x83\xa6:\xf6\x04\x98%\ \xb4\x05,\xd1\xfcm\x92J\x821\xe7o\x8e!'\xb0\ \x83\x1c\x10}\x07LL|\x7f\xaf>\xe3\x8d\xdd\x15>\ \xb4\xd21\xd6\xc0\xc1C\xa8\xce\x83G\xd8\x91\xc4\xbd\x93\ \xb1G\xc9\xb0\xdb\xf4\xf1\xe9\x09\x17\x8cL\x98\x95\x930\ *\xdd\xec\xef\x81|\xc3\x8b\x0a,\xe2\x0f\xb6\x00\xf0\x87\ \xfa2}]\x00t(\x8dj\xb0Q\xff\xfc\xd7\xac\ t\x95\xa4\xab\x07\x15K\x9c\x10\x14\x0a\xae\xcd\xad\x8e\xbe\ \xdc5\xe5\xb7\xb2\xc2\x9f\xd4'\xfa&(\xf6\xba\xf1\x87\ \xb5\xa5\xcf~_\xc7@\xee\xb1\xf7'\xd6\xd4\xc4\xdc\x8f\ \xdd\x83\xbaK\xb6%\xc9jN\x8asf\xb6\xfb\xfca\ \xaeq\xd9\xa98\xa7\x0f\x02\x19H\x97d\x10\x00\xca9\ D\x7f\xa3\xcf\xb7\x00\xcc_\xcd\xdf\xa0}\xff\x1c\xfb\xe1\ \x09I\x0b@9\x85?\xa8\xdca\xe16$-!G\ \x9at\x8b}\xc4\x19\xb2\xcd\x0e%\x98\x04\xa0\xef\x02r\ \xbe\xa3\xce\xf7\xd3E\x05\x85\x8d8\xf33&0OM\ \xf8\xe7\x89}`\xca\x8a\xed\x98G\x84\x03\x81\xe9\x08\xdc\ \xdb\x98<(\xc1>$I\x19f\xf3\x1d20uH\ \xaak\xb8K\x1b\x9a\x99\x94\xe4\xb4\x9b\xadD~\xa0\xe9\ \xdc\x5cj^&\x88\xfeB\xdf\x16\x80\x16\xf8xP\xc9\ \xdfPj\xdf\xfa\xa4T\xb4\xcc\xd0\x83Xz\xed\xc9\xde\ \xfcs\x13\xc6] ;Sd\x9a\x0e\xba\x1f`HA\ Mk\x0c\x1a*\xd8o\xfeR_U\xc3E\xc9`\x9a\ \x86wk\x03\x9a\xc6W\x0dU\xc7\xa1;\xd0\x5c\xc0\xf7\ \xb4@\xfd@3\x9dK\x9a\xae9\xed\x0e\xa7\xdd\x96\x9e\ \xe4Nr\xd8\x14I\x87/\x80\xfb\x8eP\xfd\x80\xb0\x14\ \xfdG\x00\xb0\x92\x86KP\xe8\x83\xbc/\x08\xcc\x82\x03\ [\xeeL\x01\x13\xd0\xfd\x13\x05\x13q\x07\xad8&\xb2\ \x98\xd3\xfb\xc0\x8c\xdb\x81\xd9nq\x06nZ\x9a\x8c\xe6\ \x1a\xff6\x17\x04$\x00\x84\xa5\xe8/\x02\x80\x7fXv\ \xc5\x90\xf0\x96k\xe2\x9b\xf0\x97\xcau\xdf\xc7LUP\ v\x9e\xaeQr@v\x87\xcc\xc2;\x0bq\x91\xb2\x08\ aE\xfa\x89\x00\x10\x04A\x10\x91B\x1d#\x04A\x10\ \x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\ \x00\x82 \x08\x8bB\x02@\x10\x04aQH\x00\x08\x82\ ,\x0a\x09\x00A\x10\x84E!\x01 \x08\x82\xb0(\ $\x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\ \x04AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04a\ QH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E!\x01\ \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80 \x08\ \xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\x8bB\ \x02@\x10\x04aQH\x00\x08\x82 ,\x0a\x09\x00A\ \x10\x84E!\x01 \x08\x82\xb0($\x00\x04A\x10\x16\ \x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\ \x82 \x08\x8bB\x02@\x10\x04aQH\x00\x08\x82 \ ,\x0a\x09\x00A\x10\x84E!\x01 \x08\x82\xb0($\ \x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04\ AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04aQ\ H\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E!\x01 \ \x08\x82\xb0(\xcc0\x0c\xb1H\x10]\xc3\x9a5k<\ \x1e\x8f\xdf\xefOLL\xcc\xc9\xc9\x19=z\xb4\xd8A\ \xf4n|>_K\xc2\xa5\xa6\xa6\x8a\xadD?\x82\x04\ \x80\xe8B^}\xf5\xd5_\xfd\xeaW`D\xc4:g\ \xfe\xfc\xf9K\x97.\x15+D\xaf\xe4\xcb/\xbf\xbc\xf9\ \xe6\x9bW\xaf^-\xd6%\x09\x04\xe0\xac\xb3\xcez\xfe\ \xf9\xe7\xc5:\xd1/\xa0. \xa2\x0b\xf9\xea\xab\xaf\x0e\ \xb2\xfe\xc0\xb2e\xcb\xbe\xf9\xe6\x1b\xb1B\xf4J\xdex\ \xe3\x8d\x03\xad?PWW\xf7\xc2\x0b/\xac[\xb7N\ \xac\x13\xfd\x02\x12\x00\x82 \x08\x8bB\x02@\x10\x04a\ QH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E!\x01\ \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80 \x08\ \xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\x8bB\ \x02@\x10\x04aQH\x00\x08\x82 ,\x0a\x09\x00A\ \x10\x07\x13\x0c\x06\xc5\x12\xd1\xaf!\x01 \x08\xe2`T\ U\x15KD\xbf\x86\x04\x80 \x88\x83!\x01\xb0\x08$\ \x00\x04A\x1c\x0cu\x01Y\x04\x12\x00\x82 \x0e\x86Z\ \x00\x16\x81\x04\x80 \x88\x83!\x01\xb0\x08\xf4JHK\ SPP\xb0u\xeb\xd6\x92\x92\x92\x8a\x8a\x8a\xb4\xb4\xb4\ \xc1\x83\x07\xe7\xe7\xe7\xc3wJJ\x8ap\x11\x1b7\xdc\ p\xc3\xdf\xfe\xf67\xb1r\x00_\x7f\xfd\xf5\x11G\x1c\ !V\xe2Daaa%\x07\xae\x05\xbe!c\x0f8\ \x80\xdc\xdc\x5c\xe1\xae\xbb(---..\x86\xb8\x85\ \xef\xba\xba:\x88\xd2\xd4\xd4T\xf3{\xe6\xcc\x99\xb2\xdc\ }u\xaf\xa6\xa6&\x08\x06\x84\xc7\xfc\xce\xce\xce6S\ 9//O\xb8h\xc39\xe7\x9c\xf3\xd6[o\x89\x95\ \x03\xf8\xea\xab\xaf\xa6M\x9b&V\xe2AMMM\xd9\ \x010\xc6\x06\x1e@VV\x96p\xd7\xed\xac_\xbf~\ \x1fGQ\x14\xc8?\xa3F\x8d\x8a{\x8e\xed\x15@9\ !\xc2\x072\xc4u\xd7]w\xec\xb1\xc7B\x19\x86\x0c\ 1o\xde\xbc\x9f\xfe\xf4\xa7\xcf?\xff\xbc\xd8\x1d-\xcf\ =\xf7\x1c\xf8\x03\xbe\x81\x9f3f\xcc\x00\xff\xe1,`\ \xd1\xc4\xee0x\xf1\xc5\x17\xcf<\xf3\xcc\xe9\xd3\xa7O\ \x9a4i\xf6\xec\xd9\x17]t\xd1\xca\x95+\xc5\xbe6\ \xec\xda\xb5\xeb\xee\xbb\xef\x1e9r\xa4\xc8\x04m\x98<\ y\xf2_\xfe\xf2\x17p&\x0e\x88\x96\xeb\xaf\xbf^\xf8\ \xd8\x1a\x10\x00\xe1\x226\xa0\xa2\xfa\xee\xbb\xef\xfe\xf2\x97\ \xbf\xcc\xc9\xc9\x11^\x87\x00\x04\x00\xe2\x04R\x0aRP\ \x1c\xdc\x05,[\xb6\x0c\x22v\xfe\xfc\xf9\x1d\xdbw\x9b\ \xcd6w\xee\xdc[n\xb9\xe5\x8b/\xbe\x10Gv\x01\ EEEO>\xf9\xe4\x09'\x9c \xce\xda\x86!C\ \x86\xdcv\xdbm\x1b7n\x14\x07\x1c\xc0\xe9\xa7\x9f.\ \x1c\xb5\x06\x04@\xb8\x88\x8dM\x9b6=\xf8\xe0\x83\x10\ Q\xc2\xdf\x10@N\xbe\xef\xbe\xfb\xc0\xb18,\x06\xa0\ \xa2s\xcd5\xd7\x80\x87\x90\xb7\xa7N\x9d\x0aE\xec\xfe\ \xfb\xef\x17\xfbZ\x03\x01\x1b1b\x84\x08\xc1\x01$%\ %]x\xe1\x85q\x09L\xef\x81\x04 2~\xf6\xb3\ \x9f\x89\xec\xd0\x9a\xaa\xaa*\xe1\x22r\xa0^&|i\ \xcd\xd9g\x9f-\x5c\x84\x81\xcb\xe5\x12\x8753g\xce\ \x1c\xb1\xef\x00\xc0\xe2\x9cz\xea\xa9\xc2E\x18\x1c\x7f\xfc\ \xf1;w\xee\x14\x07GN\xd7\x09\xc0\xf6\xed\xdb\xc1\xa0\ \xb7\xbd\xeap8\xed\xb4\xd3V\xadZ%<\x8a\x07\x8b\ \x17/\x06\xd3\xe0p8\xc4\x09\x22\xe1\x98c\x8ey\xfd\ \xf5\xd7\x85Gq\x02r\x14\xd4'\xc4\x09\xc2\x00\xb2\x04\ \xd8Gq0\xe7\xe4\x93O\x16\xfbZ\x13\xbb\x00,\x5c\ \xb80\x8a\xaa4\xd4\x8a>\xfe\xf8c\xe1ET\xb4+\ i\xabW\xaf\x16\xbb9k\xd7\xae\x1d=z\xb4\xd8\x17\ \x02\xa8\xa5\x09\xd7\xfd\x02\x12\x80\xc8\x00\xa3,2Bk\ \xa0e-\x5cD\xce\xfe\xfd\xfb\x85/\xad9\xee\xb8\xe3\ \x84\x8b0\x10\xc7\x1c\xc0\x981c\xc4\xbefn\xbd\xf5\ V\xb1/\x12\x92\x93\x93_~\xf9e\xe1E\x84t\x85\ \x00\xf8|\xbe\xdbo\xbf]x\x14\x03\xa0\x1f[\xb6l\ \x11\x9eF\x0b\xd4\x16'L\x98 |\x8c\x81K.\xb9\ D\xd7u\xe1il\xbc\xf0\xc2\x0b\x19\x19\x19\xc2\xdf\xb0\ Q\x14\xe5\xd1G\x1f\x15^\x18F\xa8vC,\x02\x00\ \x1as\xfe\xf9\xe7\x0b\x8f\xa2\x02T\xed \xa1\x0a\x9f\x13\ O<Q\xf8r\x00\x1f~\xf8\xa1\xd8m\x18\xd0\x94\x84\ \x96\x99\xd8\x11\x1ahk\x8a\x03\xfa\x05$\x00\x91\xd1\x87\ \x04 //O\xec3\x0c\xa8\xe9\xc4\xd8u{\xed\xb5\ \xd7\x0a\xbf\x22!\xee\x02\xf0\xce;\xeft\xda\xdb\x13\x11\ Q\xd7\xbe\xbf\xfc\xf2\xcb\xf8\xf6\x86O\x9d:u\xcf\x9e\ =\xc2\xf7h\xb9\xf2\xca+\x85wQ\x01M\x01\xbf\xdf\ \x0f\xfe@\xbbDljM\xd4\x02\xf0\xfc\xf3\xcf\x0b/\ b\x03\xda|\x9f|\xf2\x89\xf04\x12\xda\x15\x80\xd7^\ {\xcd\xdc\xfb\xdf\xff\xfeWl\xea\x8c\xac\xac,\xf3\x90\ \xfe\x01\x09@d\xf4!\x01HII1w\xbd\xf5\xd6\ [q\xb9\xe5\x18\xaa\xcf\xb4\x03\xe2+\x00P\xb7\x15\xc7\ \xc7\x95g\x9eyF\x9c l\x8a\x8a\x8a\x12\x12\x12\xc4\ \xf1\xf1\xe3\xf8\xe3\x8f\x17'\x88\x8a\xdf\xfe\xf6\xb7\xc2\xa3\ \x18\x18>|\xf8\xae]\xbb\xe6\xce\x9d+\xd6[\x13\x9d\ \x00<\xf4\xd0C\xe2\xf8x\x00\x99y\xe1\xc2\x85\xc2\xeb\ \xb0iW\x00^|\xf1E\xd8\x05\xed\x00\xb1\x1e\x06$\ \x00\x96\xa6\x0f\x09\x00c8\xc4\xeb\xb9\xe7\x9e\x13\xeb\xf1\ \xe0\x95W^1O\x17&q\x14\x80vG\x13\xc5\x8b\ \x07\x1f|P\x9c&<~\xf3\x9b\xdf\x88#;\x04\xaa\ \xab\x83\x06\x0d\x1a=zt~~\xfe\x80\x01\x03\xc4\xd6\ \x0e\xf9\xfd\xef\x7f/\xce\x11!\x7f\xf8\xc3\x1f\x84\x17\xf1\ T\xbfV\x14\x02\xd0i\xc7cjj*\xb4\xa5N\ 9\xe5\x94\xcb.\xbb\xec\xdcs\xcf\x05\x15\x1c3f\x8c\ \xd8\x17\x9a%K\x96\x88\x13\x84G(\x01\xf8\xf6\xdbo\ #\xd2\xf2\xcc\xccL\xe1c\xbf\x80\x04 2\xfa\x90\x00\ \x00\x8f?\xfe\xb8XjCRR\xd2%\x97\x5c\x02\x95\ \xfa\xf7\xde{o\xdd\xbau\xd0\x04\xbe\xf7\xde{\xa1\xec\ \x89\xdd!\x80\xcaWDc\x93\xe2%\x00\x9dv \xb8\ \xdd\xee\x0b.\xb8\x00\xae\xe2\xf3\xcf?\x87\x0alSS\ SYY\xd9\xc6\x8d\x1b?\xf9\xe4\x93\x07\x1ex`\xca\ \x94)\xc2]h\xe0@q\xb208\xf3\xcc3\xc5a\ m\x985k\xd6}\xf7\xdd\xb7x\xf1\xe2\x8a\x8a\x0a\xe1\ \xba\x99\xea\xea\xea\xf7\xdf\x7f\xff\xa2\x8b.\x12NC\x10\ \xc58%\xf0V\x1c\x1c\x82\xe4\xe4\xe4\x993gB~\ \x00\x97\xabW\xaf\x86\xe0A\xd2\x9ft\xd2Ibw\xd8\ D*\x00\x1d\xd7?.\xbc\xf0\xc2\xb7\xdf~[8m\ \xcd\xb6m\xdb\xce;\xef\xbc\x0enfL\x9a4I8\ \x0d\x8fP\x020n\xdc8\xb1\xd2\x06\xd0\xef\xb1c\xc7\ :\x9dN\xb1\xce\x01\xf7\xc2\xc7~\x01\x09@d\xf4-\ \x01h\x17\xa8\x93>\xfa\xe8\xa3\x81@@\x1c\xd9\x9a\x15\ +V\xfc\xe4'?\x11N\xdb\xe3\xba\xeb\xae\x13N\xc3\ .\x02PTT\x94\x96\x96&\x8el\x03T\xdf\xc0\ \xaei\x9a&\x5c\x87\xe0\xcb/\xbf\xecx\xf8\x13\x88\x84\ p\x1a\x06PQ\x15\x875\x03\x15C\xa8\xbco\xdf\xbe\ ]\xb8\xe8\x10\x10\xdd\xc3\x0e;L\x1c\xd9\x86\xdbo\xbf\ ]\xb8\x0b\x9b\xc9\x93'\x8b\x83\xdb\x03\xda7\xa0\x88\xc2\ ik@\xfb\xcf:\xeb,\xe1.\x0c\x22\x12\x80\xdd\xbb\ w'&&\x8a#[3|\xf8\xf07\xdexC\xb8\ \x0bMMM\xcd5\xd7\x5c#\x8ei\xc3-\xb7\xdc\x22\ \xdc\x85A\xbb\x02p\xe8\xa1\x87\x8a\xa5\xd6\xdct\xd3M\ \xd02\x10G\x1a\xc6\x96-[\x9ex\xe2\x09\xf3\xc68\ \xec\x12[\xfb\x05$\x00\x91\xd1\xd7\x05\x00Z\xd9\x8d\x8d\ \x8d\xe2\x98\xd0tP\xea\x800\xcd\x1c\x10\x17\x01\x80\x9a\ \xa08\xac\x0d\xe7\x9f\x7f~D\xf5\xe5\xc7\x1e{L\x1c\ \xd9\x1e\x10Z\xe1\xae3\x0e\xbc\xd7\x0am\xa9?\xfd\xe9\ O\xa1\x045\x14^\xafw\xfa\xf4\xe9\xc2\x8b\xd6@\xb5\ W8\x0a\x0f\xa8\xcb\x8b#\xdb\x00a;p\xa0K(\ ^z\xe9%q@gD$\x00\xa1\x14\x17\xc4\xaf\xb8\ \xb8X8\x0a\x030\xf4\xe2\xc86\xac_\xbf^8\xea\ \x8cv\x05\xa0-\x83\x07\x0f\xfe\xf4\xd3O\xc51m\x80\ T\x13K\xfd\x05\x12\x80\xc8\xe8\xd3\x02\xf0\xf7\xbf\xff]\ \xb8\x0e\x83\xab\xae\xbaJ\x1c\xd6\x86\xcb/\xbf\x5c8\xea\ \x8c\xd8\x05\xe0_\xff\xfa\x978\xa6\x0d\xa7\x9cr\x8ap\ \x14\x09\xaf\xbd\xf6\x9a8\xbe=\x0a\x0a\x0a\x84\xbb\x0e\xf9\ \xf8\xe3\x8f\xc1\xb1,\xcb\xd0\x1e*//\x17[#\x04\ \x8c\xa9y\xd2\xb6\x84o\xd7@\xce;x\xfe`\xcd\x9a\ 5\xc2]g\xac\x5c\xb92??_\x1c\x16\x9a\xf0\x05\ \xd4\xb8\x9a1c\xc6@\x93N8\x0a\x9b\xb6M.\ \x93\x1bn\xb8A\xb8\xe8\x8cp\x04`\xdc\xb8qQ\x0f\ 3\xed\xa3\x90\x00DF\xdf\x15\x80H\xef\xdf655\ A;]\x1c\xdc\x9a\xf0\xef\x83\xc5.\x00S\xa7N\x15\ \xc7\xb4&%%\xc5\x1c\xb0\x18\x05\x0f>\xf8\xa0\xf0\xa5\ \x0d\xf7\xde{\xafp\xd4\x19`\xc5B\xf5\xab\x84O\xa8\ :2\xc8\x9ep\xd1\x19\x1d\xdc\x1d\xb9\xef\xbe\xfb\x84\xa3\ \xf0\x80\xb6Tjj\xaa88\x04\xe1\x0b\x00\xe4^q\ Lk\xa2\x1b\xc4\xb9s\xe7NEQ\x84\x17\x07\x10~\ k\xa9S\x01\xc8\xcb\xcb\xdb\xbd{\xb7pm\x19h2\ 8K\x00m\xfc\x0b/\xbcP\xac\x84\x87\xdb\xed\x0e5\ z\xaf\xaa\xaaj\xc9\x92%b\xa5+Y\xbat)H\ \x85Xi\xcd\xdbo\xbf\x1d\xdd\x93\xb7\xc0-\xb7\xdc\x12\ j\x9c{\xf8\xc3\xd5sss!\x8a\xc4J\xb4\x84\xaa\ t\x87\xba\xea\xb6\xbc\xf1\xc6\x1bb\xa953g\xce\xbc\ \xed\xb6\xdb\xc4Jx\x0c\x192d\xfb\xf6\xedqy\xae\ \x0dZ0\xed\xe6\x90\xdbo\xbf\xbd\xd3\x81\x06\xed2r\ \xe4\xc8k\xaf\xbdV\xac\x1c@uuu\xf8\xfdW\x1d\ \xf3\xea\xab\xaf\x86\xaa\xf1\xf4cH\x00\xfa?O=\xf5\ \xd4\xc5\x17_,V\x22\x01\x9a;\xa1\xe6P[\xb8p\ \xa1X\xeaJB\x8d!\xb9\xf4\xd2K\x8f=\xf6X\xb1\ \x12\x15\xa1\xba\x95\xa1\x0e\xf8\xce;\xef\x88\x95\xae\xa7]\ \xa3\x06\x94\x95\x95\x89\xa5\x0e)))\xf9\xe8\xa3\x8f\xc4\ JkB\xf9\xdc1\x03\x07\x0e|\xeb\xad\xb7\x06\x0d\x1a\ $\xd6\xa3%T\xc2]q\xc5\x15b)rB\xe5\xe1\ \x15+V\x88\xa5\x18x\xec\xb1\xc7\xe6\xcd\x9b'V\xac\ \x04\x09@?\xe7\xd6[o\xed\xa07\xbfSB\x95\xba\ \x88\x9e\x9d\x89\x8e\xa2\xa2\x22\xa8\x94\x89\x95\xd6\xdc|\xf3\ \xcdb)Z\x16,X\x10j&\xb2\x95+W\x8a\xa5\ \xaeg\xfc\xf8\xf1\xed\x0eB\xaf\xad\xad\x15K\x1d\x12J\ \xab\xc6\x8e\x1d\x1bj\xd2\xaaN\x193f\x0cx\x1b\xe3\ \x93\x83\xed&\xdc\x05\x17\x5c\x10K\x15\xfb\x88#\x8e\x80\ \xe8\x12+\x07\xd0\xc1\xad\x9409\xef\xbc\xf3B\xf5U\ \xf6{H\x00\xfa33f\xccx\xe0\x81\x07\xc4JT\ \x84\x1a*\xbes\xe7\xce\xa6\xa6&\xb1\xd25\xac^\xbd\ Z,\xb5f\xda\xb4iq\xe9\xa6\x08\xd5\x17\xb1n\xdd\ :\xb1\xd4-\xb4;\xd4=L\x01\xd8\xb0a\x83Xj\ M\x073\x80\x86\xc3\xacY\xb3:\x1e\x06\xd61\xeb\xd7\ \xafo7\xfc\xa1\xba\xdd\xc2\xa7\xdd\xe796o\xde\x5c\ __/V\x22\xc7\xe9tvpO\xa8\xdfC\x02@\ t\x04\xb4\x8b\x0fz\x10\xa6\x85\xef\xbf\xff^,u\x0d\ _~\xf9\xa5XjM\xa8\xfb\xf0\x91\x12\xea.e\xec\ 5\xca\x88HOO\x17K\x07\x10\xa3\x00\x1c}\xf4\xd1\ b\xa9'\xf8\xe2\x8b/\xc4Rk@W\xc4R\xb4\x84\ zx\x22\x16\xcd\xbe\xff\xfe\xfb\x87\x0e\x1d*V\xac\x07\ \x09\x00\xd1\x11\x8c\xb1P}\xa3[\xb6l\x11K]\xc3\ \x9a5k\xc4Rk\x8e:\xea(\xb1\x14\x1b\xd0\x92h\ \xd7\xf8\x06\x83\xc1\xeel\x04\x843\x03e(\xbe\xfb\xee\ ;\xb1\xd4\x9axEQt\xb4+\x00\x90\x91B=u\ \x15>\xa1\xa6\xd3\xa8\xab\xab\x13K\x91\xf3\xcb_\xfeR\ ,Y\x12\x12\x00\xa2\x13B\xf5\xdbv\xa9\x00\x04\x02\x81\ P-\x80\xd8\xedH\x0b\xa1&\x7f\xefjm;\x90\xc6\ \xc6F\xb1\x14!\x9b6m\xf2\xfb\xfdb\xe5\x00\x86\x0c\ \x19\x12\xc5t\xd0q\xa4]\x01\x88K\x90B\xbd\xa8.\ \x16\x01\xb08\xfd\xea\x95\x90\xdb\xb6m\xbb\xec\xb2\xcb\xc2\ \x1cA\xd1\x16\xa8\xa4\xfc\xeaW\xbf\xea\xe0\xb1C \xd4\ \xab\xf2JKK\xa3\x1e;QPP\x00\x85V\xac\x1c\ \xc0q\xc7\x1d\xf7\xe9\xa7\x9f\x8a\x95\xce\x80\xc0\x8b\xa5\x03\ \x981cF\xa8zt\xf8\xdcs\xcf=w\xddu\x97\ X9\x80+\xae\xb8\xa2\x83\xa7\xb4L\xa2~%$\xd4\ m\xdbm\xef\x1fr\xc8!;v\xec\x10+1s\xc9\ %\x97\xb4;\x88\xf0\xd1G\x1f\x85\x90\x8b\x95.&7\ 7\xb7\xed\x1b\x81\x86\x0d\x1b\xb6g\xcf\x1e\xb1\x12\x82\x0f\ >\xf8\xa0\xdd\x97\x9c,X\xb0\xc0|N-\x16\xae\xbb\ \xee\xba\x7f\xfc\xe3\x1fb\xe5\x00\xbe\xea\xec\x95\x90\xa0\xdc\ \xed\xf6\x19\x8e\x1a5\xea\xc0\x97\x01\xb4\xcd\xae-[:\ \xb0H\xdb\xb7o\xff\xdf\xff\xfe'V\x0e\xe0\x91G\x1e\ \xb9\xf1\xc6\x1b\xc5J\x08 Z\x16/^,V\x0e\x00\ \x048\xd4|\x15\x96\x80?\x0d\xd0O\x88\xe5\xce\x95I\ \xa7s\xbd\xf6\xad\x07\xc1@\x00\xc4\xee\x18x\xf6\xd9g\ \x85w\xad9\xef\xbc\xf3\x84\x8b\xd0D\xfd \xd8\xb2e\ \xcb\x84\xd3\xd6\xc48a\xf2A\xfc\xe9O\x7f\x12\xfe\xb6\ \x06\x04O\xb8\x88\x8a\xea\xea\xea\x95+W>\xf5\xd4S\ \x90!\x8f>\xfa\xe8I\x93&\x99\x13\x82B\xeeJJ\ J\x0a\xe7\xf1\x85\xa1C\x87\x0a\xbfB\x13j\xfc{\xf8\ \x0f\xc7v@\xa8\xe9N;}\x10,\xea\xeaW,\x84\ \x93^\xa1\x1e\x04\x0bgf\x94~L\xbf\xea\x02\x8a\xbd\ nXYYYXX(V\x08N\xa8\x89\xd8\x1a\x1a\ \x1a\xc4R\x17\x10\xea.h|;7\xc0\x1c\x8b\xa5\xd6\ \x84y\x0f\xf6 \xa0\xb1u\xef\xbd\xf7\xce\x9b7\x0f\x02\ \x09\xdfW_}\xf5\x13O<\xb1|\xf9\xf2\x8d\x1b7\ B\xd5\x15\xday\x90\xbb\xc0\xdc@\x1dY\x1c\x10\x1b\xa1\ \xfa=\x92\x93\x93\xc5RO\x10]\xd4\x11=E\xbf\x12\ \x00\x9f\xcf'\x96b\x00\xaaob\x89\xe0\x842(\xb1\ \x8c\xbd\xeb\x94\x9a\x9a\x1a\xb1\xd4\x9avo\xdbFM\xbc\ \x04\xe0\xe5\x97_>\xfc\xf0\xc3g\xcd\x9au\xe7\x9dw\ ~\xfe\xf9\xe7bk\x17\x13*\x90\xa1:\xca\xbb\x87\x1e\ \x11\x80\x81\x03\x07\x8a%\x22B\xfa\x95\x00\xc4\xa5\xee\x93\ \x97\x97'\x96\x08N\xa8X\xd5u],u\x01\xa1\xec\ H\xf7\xb4\x00\xc2\xafI|\xf8\xe1\x87`\xfa/\xbe\xf8\ \xe2o\xbf\xfdVl\xea.BEQ\xcf\xf6h\xf7\x88\ \x00t\xfa&w\x22\x14\xfdJ\x00f\xcf\x9e-\x96\xa2\ e\xca\x94)\x99\x99\x99b\xa5=\xda\xbd\xd7\xda\xbf1\ B\xdc`\xe8`\x8e\xfe\xd8\x09\x06\x83b\xa95\xf1}\ \xfaLUU\xb1\xd4\x9a0/\xed\xf1\xc7\x1f?\xf5\xd4\ S\xbb\xdf\xf4\x9b\x84R\xa9.\xed\x9a\xeb\x94v\xa7l\ \xebR\xae\xbc\xf2\xcaP\x8ft\x10\x9d\xd2\xafF\x01\x01\ \x0f=\xf4PEE\x85X\x89\x9c3\xcf<\xb3c\x15\ 9\xf7\xdcs\xdf|\xf3M\xb1r\x00\xfdx\x14\xd0\xa2\ E\x8b\xda}\x1e\xd8|\x03\x97X\x09A\xd4\xa3\x80\x9e\ y\xe6\x99v\xdfo\x1e\xce\xd0\xa3\xf0y\xfa\xe9\xa7\xdb\ \x9d'\xe3\xb6\xdbn\xbb\xef\xbe\xfb\xc4J\x08n\xba\xe9\ \xa6G\x1eyD\xact\xc8\xfc\xf9\xf3AQR8f\ 2A\xa1\x0b\x04\x02~\xbf\x1f\xbe\x17.\x5c\xd8\xb63\ m\xe8\xd0\xa1{\xf7\xee\x15+!\xb8\xfd\xf6\xdb\xdb\x0d\ \xe4\x1dw\xdcq\xef\xbd\xf7\x8a\x95h\x89z\x14\x10\xa4\ l\xbb\x0e&M\x9a\xf4\xe7?\xffY\xacD\x08D\x1a\ \xcbr\xcb\xb7\xb9\x00\x1cr\xc8!a6\xd9i\x14\ P\xfb\xe0\x9d`\x22l\xce9\xe7\x1c\x11q\xad\xe9\xc7\ \xa3\x80\xda\x1dx\x07\x80\xe9\x14.B\x13\xf5(\xa0\xd7\ _\x7f]8m\x0d\xa8\x8ep\x11\x0f\x1e~\xf8a\xe1\ ok:}\xfd}\xc7\xca\x07\xb5\xe0\x8b/\xbe\xf8\xed\ \xb7\xdf\xae\xa9\xa9\x11\x07\x84\xa6\xdd\xd6F8\xa3\x80\xfe\ \xf2\x97\xbf\x08\xd7\xad\xe9\xd9Q@\xa1\x06b\x1cz\xe8\ \xa1\xc2E\x0fA\xa3\x80\xda\x85\x1e\x04\x8b\x0f\x9a\xa6\x89\ \xa5~G(}\xea\xd2.\xa0P\x9e\x17\x15\x15\x89\xa5\ xPYY)\x96Z\xd3\xf1\xa5\xed\xd9\xb3\xa7\x83\xf9\ \xf5@\xf3\xa0I\xf7\xd2K/\x9du\xd6Y\xe1DQ\ \xd4}&\xa1n\xf6\x16\x17\x17\x8b\xa5\x9e \xd4\x1b\x05\ \xba\xf4\x8e\x11\x115$\x00\x91\x01\xadN\xb1\xd4\x1a\xa8\ G\x88\xa5~G\xa89\x7f\xbat\xe8E(\xd3\xb9}\ \xfbv\xb1\x14\x0f6o\xde,\x96Z3x\xf0`\xb1\ \xd4\x1e\x8f?\xfex\xa8!\x98\xcf=\xf7\xdcc\x8f=\ \x96\x93\x93#\xd6\xc3 \xea\xa9 B\x09@|52\ RB\x09@\xbc\x06\xbf\x12\xf1\x85\x04 2\xda\x9d\xbc\ \x17\xe8\x9d\x02\x00M<\xb1\x14\x03\xa1\x04\xa0\xe3\x17\x91\ \xc7H\xa8\x8e\xdd\xb2\xb2\xb2P#D\xa3 \xd4\x94\x0f\ 3g\xce\x14K\xed\x11\xaa\xff\xe7\xaf\x7f\xfd\xebe\x97\ ]&V\xc2\xa6\xdd\x16@8\x097r\xe4H\xb1\xd4\ \x9a\x9e\x15\x00\x87\xc3\xd1n\xcd\xa0\x1b\xa6\x8f%\xa2\x80\ \x04 2B\xbd\x04\xaa\xbf\xb6\x00\xea\xeb\xeb\xd7\xae]\ +VZ\xd3\xa5\x02\x90\x9b\x9b\x1bj\x0e\xa2P\xe1\x89\ \x94\x8a\x8a\x8av\xa7[8\xf4\xd0C;\x18\x09\xf6\xde\ {\xef\xb5\x9d\xb9\x01\x98>}\xfao\x7f\xfb[\xb1\x12\ \x09Qw\x01\x85\x9a\x13{\xef\xde\xbd\xfb\xf6\xed\x13+\ =\xc1\x8c\x193\xc4RkBM]G\xf4 $\x00\ \x91\x11\xaa\x05\xb0u\xebV\xb1\x149q\xa9\xa7w\x11\ \x9f|\xf2\x89Xj\xcd\xd0\xa1C;\x1e/\x1b;\xa1\ F\x9b|\xf0\xc1\x07b)6B\xbdM\xa5\xe3Y\x8b\ \xb7m\xdb&\x96Z\x13\xea\xad\xe5\x9d\x12\xb5\x00\xa4\xa6\ \xa6\x86\x9a\xc78\xf6w\xda\xb0\x18\x86;\x87\x12\x80P\ \x93W\x13=\x08\x09@d\x84\x1a1\x16\xca.\x84C\ \xa8\xd7\xba\xc6\x0e\xd4pc\x9c\x1e#\x94\x00ti\xf5\ \xdf\x04\xea\xd4b\xa95\xf1z\x1bet\x02\x10\xea\x96\ x\xa8\xa9\xea;%j\x01\x00B5\x02b\x14\x80\xca\ \xca\xca\xf7\xdf\x7f_\xacDN(\x01\x085\xb2\x8b\xe8\ AH\x00\x22c\xe2\xc4\x89b\xa95_\x87\xfd\x16\xef\ \x83x\xed\xb5\xd7b\x7f\xc1a(***\xce<\xf3\ \xcc\xaa\xaa*\xb1\x1e!px\xa8\x99\xe0b|%o\ 8\x84j\x01\x80\x09\xee\xf4\xf9\x83N\xf9\xfe\xfb\xef\x17\ -Z$V\x0e\xc0\xe9tv\x5c\x97\x0f\xd5\xbb\x92\x9d\ \x9d-\x96\x22$\x16\x018\xf2\xc8#\xc5Rk^|\ \xf1\xc5X\x1e\x07\xbb\xf8\xe2\x8b;}\x0a\xa1\x03B\x09\ \xc0\xf2\xe5\xcb\xe3{\x0f\x9f\x88\x1d\x12\x80\xc8\x085\x19\ =\xd4\xb9JKK\xc5J\xd8,]\xba\xf4\xa7?\xfd\ \xa9X\xe9\x1a\xc0\xd2\x9d}\xf6\xd9\xd1\x0d\xc2{\xea\xa9\ \xa7BuOE\xdd\xe3\x11>\xf3\xe6\xcd\x0bU\xc3}\ \xe8\xa1\x87\xc4R\xb4\xfc\xe1\x0f\x7f\x10K\xad\xb9\xfc\xf2\ \xcbC\xcd\x0fa\xd2\xee\x14\xfc@\xd4*\x1b\x8b\x00\x9c\ r\xca)b\xa95\xc1`\xf0\xf9\xe7\x9f\x17+\x11r\ \xf5\xd5W\xb7+\x8d\xe1\x93\x9c\x9c\xdc\xee<\xd5\xc0\xd3\ O?-\x96\x88\xde\x01\x09@d\x8c\x1d;6\xd4}\ \xe0\x7f\xff\xfb\xdfb)<\xd6\xaf_\xdf\x0df\x14X\ \xb1bE\xa8\xe7\xd7:\x00\xea\x80\xa1\x9e\x93:\xf9\xe4\ \x93#\x1a\xe9\x185\x97^z\xa9Xj\xcd\xb7\xdf~\ \xfb\xc4\x13O\x88\x95\xc8y\xfd\xf5\xd7\xdf~\xfbm\xb1\ \xd2\x1a\x10\x00\xb1\x14\x82P/\xa5\x8a\xaen\xbb`\xc1\ \x82\x1f~\xf8A\xacD\xce\xa4I\x93\xc6\x8c\x19#V\ Z\xf3\xb7\xbf\xfd-\xd4D\x17\x1dp\xe3\x8d7\xfe\xf3\ \x9f\xff\x14+1\x10\xea5[\x8f<\xf2H\xbc\xee\xe1\ \x13\xf1\x01\xaaxDD\x84z\xa4\x10\xaar{\xf6\xec\ \x11\x8e:c\xf1\xe2\xc5\xa1FL\xb7\x10\xfb\x93\xc0\x07\ r\xec\xb1\xc7\x96\x97\x97\x0b\xd7a\x10\xaa\x12\x07@\xed\ R8\xea\x8c\xa8\x9f\x046\x09\xf5\xa0\x96\xc9\xfb\xef\xbf\ /\xdcE\x02\xc8\xa1\xddn\x17^\xb4&\x9c\x08\x0fu\ E\xd0\xcc\x12.\xc2\x06tT\x1c\xdc\x86!C\x86\x08\ G\x9d\xd1\xc1\xd0#\xd8%\x1c\x85\xc7%\x97\x5c\x22\x8e\ \x0cM\xa7O\x02\xb7\x10\xaa\xad\x0c\xf9\x10ZQ\xc2Q\ 7BO\x02\xb7\x0b\x09@\xc4<\xf9\xe4\x93\x22\xef\xb4\ \x012\xb7p\xd4!\x1d\xf8p \xf1\x15\x00\x00\x9a/\ \xdf}\xf7\x9d8\xa0C:0+\xa3F\x8d\x12\x8e\xc2\ F\x01\x00B5\x02\x80\xe4\xe4\xe4O>\xf9D\xb8\ \x0b\x8f\xff\xfd\xef\x7fYYY\xe2\xf86\x84c\xdd^\ ~\xf9e\xe1\xba\x0d\xabV\xad\x12\x8e:\xa3\xa4\xa4\xe4\ \xb4\xd3N\x13\x87\xb5G\xf8\x02\xd0q\x03\xe2\xb9\xe7\x9e\ \x13\xee:d\xeb\xd6\xadaN\xa4\x18\xbe\x00\xb4;\x07\ \x94\xc9\x91G\x1eYVV&\xdcE\x85\xcf\xe7{\xe5\ \x95W\xae\xb9\xe6\x9a\xf0\xa7`!\x01h\x17\x12\x80\x88\ \xe9\xf8A\x1b\xa8\x09\x16\x14\x14\x08\xa7m\xa8\xae\xae\xee\ \xc0\xa2\x1dD\xdc\x05\x00HII\x816\xbe8\xa6=\ \xa0vv\xe1\x85\x17\x0a\xd7\xed\x016T8\x0d\x83\xd8\ \x05\xa0\xa6\xa6\xa6\xe3\xe7ro\xbd\xf5V\xe1\xb43\xee\ \xb9\xe7\x1eqL{\xfc\xf9\xcf\x7f\x16\xee:\xa4\x83\xe9\ \x8e'O\x9e\x0c\x86I\xb8\x0b\xcdG\x1f}\xd4\xe9\xfc\ e\xe1\x0b\x00\xf0\xeb_\xffZ\x1c\xd6\x1e\x7f\xfd\xeb_\ \x85\xbb\x10\xfc\xe7?\xff\xe9\xb41\xdaB\xf8\x02\x00\xcc\ \x9b7O\x1c\xd6\x86\x09\x13&\xbc\xf6\xdak\xc2]$\ @\xce\xb9\xf9\xe6\x9b[\x9e\x82\xfe\xdd\xef~'vt\ \x06\x09@\xbb\x90\x00D\xc3\x15W\x5c!\xb2O{@\ q\xba\xfb\xee\xbb7o\xde,\x5cs`\xf5\xce;\xef\ \x0c5}B\xbbC\x0f\xbbB\x00L\xa6M\x9b\x062\ P\x5c\x5c,\x0e\xe6\xacY\xb3\xe6\xb6\xdbn\xeb\xd8\x16\ D\xda\xd1\x11\xbb\x00\x00\xedN\xbfz `Oo\xba\ \xe9\xa6P~\x82\xcd\xba\xe3\x8e;B\x0d\xdf29\xfa\ \xe8\xa3\x85\xeb0\xe8\xa0sl\xca\x94)\xe6\x9d\xf3v\ \xd9\xb0aC\x98\xaf\x1a\xce\xcf\xcf\x17\xc7\x84A\xa7\x8f\ }Ar\xff\xfb\xdf\xffnjj\x12\x07p \xf5\xa1\ }\x00\x17.\x1c\xb5\xa1\xdd\x9b\xe1\x11\x09\x00\xe4\xf9P\ \xbdm&\x0b\x16,x\xfb\xed\xb7\xeb\xea\xea\xc4\x01!\ \xf0x<\x1f~\xf8!\xb4J\xc7\x8f\x1f/\x8el\xe6\ \x8c3\xce\x10\x8e:\x83\x04\xa0]\xfa\xdbt\xd0\xdd\x03\ 4`G\x8d\x1a\xd5\xe9H;\xa8\xa7@\xedUQ\x14\ h4t\xf0\xa2\xb1\xfb\xee\xbb\x0f\xcaI\xdb\xc1\xa0 \ \x001N\x07\xdd)#G\x8e\x84@B\x09\xdc\xbf\x7f\ \x7f\xa7\xf7\x0cA\x1b@$\xc6\x8e\x1d+\xd6\xc3 \xea\ \xe9\xa0\x0f\x02*\xb9`X\xc5Jh@_\xb3\xb3\xb3\ \x07\x0d\x1a\x04\xdf\x90F%%%`\xe6:\x1d\x9f\x03\ W\xf4\xd6[o\xb55.\xa1\xd8\xb8qc\xc7\x8fA\ \x9c|\xf2\xc9s\xe7\xce\x05\xc9\x81\x90\xe8\xba\xbe\x9b\xb3\ b\xc5\x8aP\x0fU\xb4\x05\x04 \xd4\x03\x07\xedr\xff\ \xfd\xf7\xff\xfe\xf7\xbf\x17+\xa1\x81\xf8\x81\xb6\x85\xd7\xeb\ \xdd\xb3gO\xc7\x133\xfc\xf0\xc3\x0f\xc7\x1f\x7f|\xdb\ \xd7\xa3\x82\x00\x84\x1a\x9e\xdb.\xcf>\xfbl\xa8\x1b\xc2\ \x07r\xe4\x91GB\xfcgdd\xa4\xa7\xa7\xc37\x94\ ,\xc8\x93\xd0\xf83\x9fj\xee\xe0\x09\xb2c\x8f=v\ \xc9\x92%b\xa5Ch:\xe8\xf61u\x80\x88\x94\xbf\ \xff\xfd\xef\x22\x06c\xe3\x91G\x1e\x01\xdf\xda\x1do\x13\ \x97\x16\xc0\xbb\xef\xbe+\x96b\xc3\xedv\x83\x09\x13'\ \x0b\x9b\xb8\xb4\x00L:\x98\x803\x16f\xcc\x98\x01f\ N\x9c#l~\xf7\xbb\xdf\x89\xe3c\xe6\xe9\xa7\x9fn\ [M\x8e\xa8\x05`\xd2qGP\xf8@\x9d`\xe7\xce\ \x9d\xe0\xe1\xb8q\xe3\xc4\xa6\x03\x88\xa8\x05`\xd2\xe9\x9b\ \x15b\xe1\xfc\xf3\xcf\x17\xa7\xe9\x0cj\x01\xb4\x0b\x09@\ \xf4\xc4n\x05Z\x86\xd3\xb4[M\x8e\x8b\x00\xc0\xaeu\ \xeb\xd6\x8d\x181B\xacG\x85\xcdf\x83z\x96y\xa2\ \x88\x88\xa3\x00\x00w\xdey\xa78>N\x9cp\xc2\x09\ P\xd9\x14\xbeGH8cf:&99\x19Zx\ \xe0U\xdb\xa1\xa5Q\x08\x00\xd0\xf1\x8d\xe5p\x987o\ \x1eT\xbaM\xdf\xda}\x12;\x0a\x01\x00^}\xf5\xd5\ X\xe6\x96\xe8\x00\xc8\x12\xe2\x1c\x9dA\x02\xd0.$\x00\ 1q\xf5\xd5W\x8b|\x14!S\xa7N\x05\xbb,|\ 1\x8cv\x9f!\x88\x97\x00\x00\xe5\xe5\xe5\xe7\x9dw\x9e\ \xd8\x14!PG\xfe\xfc\xf3\xcfM\x7f\x22%\xbe\x02\x00\ <\xf1\xc4\x13\xb9\xb9\xb9\xc2\x97\x18\x80\x06\xcd\x03\x0f<\ <\x8d\x96\x8e\xef\x96w\xcc\xa9\xa7\x9e\xbae\xcb\x16\ \xd3\x9f\xb6\xaf\xb4\x8dN\x00\x800\xef1\xb4\xcb=\xf7\ \xdc#|\xe1\xb4\xfb\xb0wt\x02\x00\xacZ\xb5j\xca\ \x94)\xc2\x978q\xe6\x99g\x06\x83Aq\x82\xce \ \x01h\x17\x12\x80Xy\xe1\x85\x17\x22z\x19\xa4\xa2(\ \x7f\xfa\xd3\x9f\xc4\xc1\xcd\xb4;\x1dP\xec\x02 \xcb\xb2\ \xd8\xcd\xf9\xd7\xbf\xfe\xd5\xf1c\xaem9\xc8(DJ\ \xdc\x05\x00\x802\xff\xc7?\xfe\xd1\xe1p\x08\xbf\x22\xe7\ \xaa\xab\xae*))\x11\xde\xc5\x06Di\xa8\x07\x03C\ 1p\xe0\xc0\x83Fg\xb6\xadkG-\x00\xc0;\xef\ \xbc\x13\xe6[\x12[\x986mZ\xdb1\xacg\x9cq\ \x86\xd8}\x00Q\x0b\x80\xc9+\xaf\xbc\x12\xd1-\x84v\ ILL\xbc\xf4\xd2K\x17-Z$<\x0d\x0f\x12\x80\ v!\x01\x88\x03555P\xf3\x0a5}q\x0b\xe3\ \xc7\x8f\x7f\xf0\xc1\x07\xdb\x1d\x01\xbdm\xdb\xb6\xb6\x93k\ \x82\x99\x13\xbb\xc3@\x1c\xd3\x1a\xa7\xd3)v7\xb3g\ \xcf\x9e\x9bn\xba)\xd4`\xa4\x16\xc6\x8e\x1d\xfb\xfb\xdf\ \xff\xfe\x9bo\xbe\x11\x87E\xcb\x7f\xfe\xf3\x9f\xb6\xf61\ \xa2!7\xa1\x00\x0b~\xff\xfd\xf7\xb7[K\x0d\xc5\xdc\ \xb9s\xff\xf2\x97\xbf\xac_\xbf^x\x11'\xea\xea\xea\ \x1ez\xe8\xa1C\x0e9D\x9c&4\xb3f\xcdz\xf8\ \xe1\x87\xdb\xbe'\xb2\xed\xfd\xff+\xaf\xbcR\xec\x8b\x0a\ 0j\xcf>\xfbl8\x91\x03\xe6\xf8\xe5\x97_\x16\x87\ \xb5\xa6m\xdf\xfd\x84\x09\x13\xa2\xee1;\x10\xa8\xee\xfc\ \xe2\x17\xbf\x186l\x98\xf07<\xa0\x01q\xddu\xd7\ \xbd\xf7\xde{\xba\xae\x0b\x8f\x22\x01\xf2s\xdb\xb79\xfd\ \xe4'?\x11\xbb\xad\x0a\x8d\x02\x8a'\xabW\xaf\x86\x0c\ ZZZZYYYQQQ]]\x0dU9`\ \xf0\xe0\xc1`\xf5N8\xe1\x04\xe1.\x04P\x0b\x83\xa2\ \xab\xaa*\xd4\xd3'O\x9e\x1c\xd1;\x17\xdb\xedc\x85\ \xba\x12x(VZ\xf3\xfa\xeb\xaf\x7f\xf1\xc5\x17\x10T\ \x10$\xc0\x1c?\x93\x93\x93\x03\xdf\xd0\xf2\x88\xbd\x9av\ \x10-\x1e\x8f'\x10\x08@x\xc0\xff\x88\xc6\x11u\ \x0a\x5c\xe0\xa7\x9f~\x0aQ\x07\x86\xb5\xb6\xb6\xd6\xfc\x86\ 8\xcc:\x00\xb0\xce'\x9dtR\xa8Y\x1c\xe2\x054\ k\xbe\xe4@\xd2\x83\xa1\x04\x92\x93\x93!\xe9\xa1>\x0e\ \xdfP\xa1n\xdb\xd5\xd3BAA\x819`)==\ =777\xd4<\xcf\x91\xb2c\xc7\x8e\x8f>\xfa\x08\ <7),,\x04\xff\xa1\xaa1j\xd4(H\x85\x05\ \x0b\x16\xb4{\xa7\xb7\x05\xbf\xdf\xff\xc3\x0f?@\x0c\xdb\ l\xb6\x94\x94\x94\xf0\x07J\x85\x09x\xbet\xe9\xd2\xed\ \xdb\xb7C\xaa\xb5$\x1f\xd4Z222 \x90\xf0\x0d\ @\xf2\x1dv\xd8a`\xfd\xc3\x7fX!\x14\xd0v\x84\ 4jjj\xd24-!!\x01jl\x916\x95\xfa\ \x1f$\x00\xfd\x84v\x05\x00$\x04J\x94X!\x08\x82\ h\x0dM\x06\xd7\x9f\xe9\xf81\x1c\x82 ,\x0e\x09@\ \x7f\x86\x04\x80 \x88\x0e \x01\xe8\xcf\xd8l6\xb1D\ \x10\x04\xd1\x06\x12\x80\xfe\x0c\xb5\x00\x08\x82\xe8\x00\x12\x80\ \xfe\x0c\x09\x00A\x10\x1d@\x02\xd0\x9f!\x01 \x08\xa2\ \x03H\x00\xfa3t\x0f\x80 \x88\x0e \x01\xe8\xcfP\ \x0b\x80 \x88\x0e \x01\xe8\xcf\x90\x00\x10\x04\xd1\x01$\ \x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04\ AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04aQ\ H\x00\xfa\x09\xf7\xde{\xef\x84\x09\x13Z\xde\x93\x95\x94\ \x944}\xfa\xf4\xdbn\xbb\xcd\x5c%\x08\x82h\x0b\xbd\ \x0f\xa0\xbfQ[[\x0b2\x90\x90\x90 \xd6\x09\x82 \ B@\x02@\x10\x04aQ\xa8\x0b\x88 \x08\xc2\xa2\x90\ \x00\x10\x04AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\ \x04aQH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E\ !\x01 \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80\ \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\ \x8bB\x02@\x10\x04aQH\x00\x08\x82 ,\x0a\x09\ \x00A\x10\x84E!\x01 \x08\x82\xb0($\x00\x04A\ \x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\ \x12\x00\x82 \x08\x8bB\x02@\x10\x04aQH\x00\x08\ \x82 ,\x0a\x09\x00A\x10\x84E!\x01 \x08\x82\xb0\ ($\x00\x04A\x10\x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\ \x10\x04AX\x14\x12\x00\x82 \x08\x8bB\x02@\x10\x04\ aQH\x00\x08\x82 ,\x0a\x09\x00A\x10\x84E!\ \x01 \x08\x82\xb0($\x00\x04A\x10\x16\x85\x04\x80 \ \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\x00\x82 \x08\x8b\ B\x02@\x10\x04aQH\x00\x08\x82 ,\x0a\x09\x00\ A\x10\x84E!\x01 \x08\x82\xb0($\x00\x04A\x10\ \x16\x85\x04\x80 \x08\xc2\xa2\x90\x00\x10\x04AX\x14\x12\ \x00\x82 \x08\x8bB\x02@\x10\x04aQH\x00\x08\x82\ ,\x0a\x09\x00A\x10\x84%\x91\xa4\xff\x07\x17\x83m\ \x0c\x9a\x8fEd\x00\x00\x00\x00IEND\xaeB`\ \x82\ \x00\x00;\xb8\ \x89\ PNG\x0d\x0a\x1a\x0a\x00\x00\x00\x0dIHDR\x00\ \x00\x00\xbb\x00\x00\x00\xc2\x08\x02\x00\x00\x00\x9e\xdf$4\ \x00\x00\x00\x01sRGB\x00\xae\xce\x1c\xe9\x00\x00\x00\ \x04gAMA\x00\x00\xb1\x8f\x0b\xfca\x05\x00\x00\x00\ \x09pHYs\x00\x00\x0e\xc3\x00\x00\x0e\xc3\x01\xc7o\ \xa8d\x00\x00;MIDATx^\xed}\x07\x80\ \x1c\xc5\x95vUu\xf7\xa4\x9d\x99\x8d\xd2f\xadr\x02\ I \x84$$$!\x09\x01\xc6\x80\xc1\x18\xec\x03\x1b\ l\xc0>\x13\x0c\x18\xb0\x8f\xb31\xc6\x19\x1f8\x03>\ \x93\xc3\x19\xfb76&Ge\x94\x05\x08%\x10\xcaa\ \x836\xa7\xd9\x9d\x99\x0eU\xff{\xd5\xbd\xab\x80\xa4\x9d\ \xd9\x9d\xb4\xb0\xdf\xec\xcetWWw\xbdz\xef\xabW\ \xaf\xaa\x13\x15B\x90\x01\x0c f0\xe7w\x00\x03\x88\ \x0d\x03\x8c\x19@|\x18\xe8\x952\x0e\xd2\x22\xb6Q(\ |`\x95\xca\x15A(\x13DPN\xa1\x9dS;-\ \x0d\x18`L\xc6\xc1f\x8ce\x194\x5c':*\xa9\ \xdeA('Z\x90{K\x99\xbf\x981\x05\xf2\xd0\x01\ \xc6\x0c\xa0\x1b`\x11\xb3m\xbf\xb5\xf5a\xb5n\x05o\ \xdbK\x088\x15p-\x94g\x0d\xa7E\xa7\xf117\ {\xf2\x86\x0c0f\x00H\x14\xecs\x04\x89\x1c\xdc\xa0\ \xbcs\x15\x89\xb6\x1dM\x0bA8d\xc8*\xa7g=\ \xa6\xe4\x8d\x81m\x94\xa6!\x0c\x1d\x88|3\x08@\x02\ \xb3y\x07[s\x13\x8b\xb49\x91\xcca\x80p\x06Y\ \x12\xaa\xe4\x8b\xfeC\xd4\xbd\x8b\xeb\xe9\xc0\x00c2\x08\ \xdc\xe2\xe6\x86_\x93P\x15\x07\xef\xf2\x09>P\xec\x9e\ \xe0#h\xa4)\xfa\xe1\x93\x82\x9b\xce\x86\xd4b\x801\ \x19\x04\xf3\xe0\x0a\xad\xfa-\xd9\xdd\x1c3\xb6\x85!\x92\ \xdcF\x88V\xbd\xd8\xe8\xac\xb7SS\x8c\x01\xc6d\x0c\ \x04g;\x9fE\xef\x12\x0b\xac\xce\xaa\xf5\xcf9\xcb\xa9\ \xc5\x00c2\x05\xd0\xcb\x88\xce:\x18G\xf7L\x19!\ \xb8\x10\xe5b\x95\xb3\x9aZ\x0c0&S \x8a\xe1\ \x16\x84\xb30br\x92\x8e\x07\xd9c\x89H\xc8YM\ -\x06\x18\x93)\xa0\x8aFU\x17.\xf4\xece\x04P\ \xc6\xcc*r\xd6R\x8b\x01\xc6d\x0a(UX\xfeD\ X\xe8q~\x0c(\x05\xe1N\xa8\xe0b\x9c\xc2\xe1\xd0\ A\xa5tFm\x801\x19\x03J\xcdaW\x10\xd5\x8b\ \xfd\xd2\x09\x81\x13}\xcc\x1d(;\x15\x16\xd1\xdb\xa4\x16\ \x03\x8c\xc9 x\xf2FDF\x7f\x97\xc42\x93kE\ \x95\xc5\x97\x87\xdb\x9b$\xbd\x06|\xccg\x15\x10\xf6\xfa\ N\xfa\x8a(\x9d\x87K\xd8\xe5Hp\x82g\x07\xecE\ \x1c$\xc1:N\xca\xd0\xceJ\xba\xf2\x86Hk\x0d\xee\ (\xd3!\x83<Lr1p^)S`\x1b\x02y\ \xa1\xb7D\xd7\xfcX=\xf02\x0c\x9f\x80\x18\x98N\x1d\ \x1b\xc1\x8f\xdd\x099\xf3{\x82X\x81\x0a1\xff\xdf\xae\ @\x1e\xac\xe3\x04_\xf2;\xa9\x01\xc6d\x0a\xd0\x10]\ \x8c\xe0\xdc\x14\x95\x0b\xad}\xaf[\x8d[U\xb3\x15R\ :\xad<w\xc98\x8b(\xee}\xcf\x0ba\xd9\xf9\x80\ \xb0\x9f\x95\x7f\x1a9\xf3!W\xb0H\xee*\xd3\xe5\ R\x920\xc0\x98\xcc\x05\xf2Fo\xa3\x96\x09\xde\xc3\xe4\ \x9a\xea\xf1\x83\xad:\xd7\xfe\xd4\xb5\xfb)\xe9t`\xc0\ $\x83\x0aAx`\xa8y\xf6\x8b>\x7f6\x12\x0f\xe2\ \xa0d\x9e\xd3\x1e`L\xe6B\x9a\x06\x19`\xaf\xc1\x1f\ \x15T\x8fvXk\xefR+_d\x96%\x18n\x82\ l\x94\x0a\xab`\x1a\x99\xf9GW\xa0\x18{\xa6d:\ \x99\x01\xc6d.\xbaL\x03\xe6?\xc2F\x96\xa9G\xd7\ \xff\xdc\xb5\xf3)4\x9fs\xbd\x15$3\x9e{\xb2\x98\ \xf3\xb4\xdb\x9fk\xf7V\xb05\x19\xd4\x19\x18+e.\ \xd0\xe0\x88\xee\x05\x07\x8a\xaai\x93\xef4\xca/\x00\xdf\ \x02\x94Aj`\xe8\x22\x94\xc6\xcdd\xf97\xf5\xd6\x03\ \xe8\x98>\xc1\xb3Da\x801\xfd\x12\x9a\xdb\xab\xce\xb8\ O\x1fz\xb9<\xd7\xdd\xcd\x0c \xcd{t\xc5\xf5\xe1\ \xce\x16\xb9\x9ax\x07\x03\xf8\xec\xf6J]\x15\x97\xcdQ\ ~\x1dZE@\xfb\x95\xbf\x19\x8cHG\x9b\xb9\xeeN\ W\xe5k8c\xc3!\xe0qd\xe6y\x93\xc9\xac?\ \xbb\xfc\x85X\x17\x0c\x92\xbb6$\x02\x9f]\xc6X\x82\ \xdb\x0e\xd62Mat\x9a\x91fXV=\xb9\xd4\xe5\ U\x14W\xbf`\x0c\x17\xc2\xd4\xc3\xfa\xbb\xbf\xd2v?\ \xcd\xe4\xcc\x1e\xf6N\xd8\x18\xa8\x95\x7f*\x9f\xf5\xa8/\ \x90\x87\x89\x03\x8c\xe9\x0b\xb0\xdf\x975\x16F\x87\xd5\xb2\ \xa3c\xe7[\x9e\x83\xcf+\x1d\x07\xa5\x1e,\xa2\xb8\x98\ wp\xb4\xfc\x22\xf7\xb0\x0bYp\x04\xd5|r'\x07\ \x99F\x22\x9c\x02&\xc4\x8cv\xf2\x157i5K\x80\ @\x90(\x1d&\x83/+o<\x9d\xf3\x84+\x00\x9e\ \xc6AB\xe4\xff\xec1\x06a\x99u\x1b\xc9\xc6\xfbH\ \xe3\xfb\xc4\x0c\x1f\xa5G\x18\xc1r\xd0\xb7\xea\xe2\x05S\ \xd5\xd3~\xa0\xe4\x8eG#\xd8.?\xc3\x18c\x03\xea\ \x03C\xee\xe8\xfa_y\xf7\xfc\x15xr81pr\ o\xce\xe3n_6,CCa\x14\xefu\xea#>\ {\x8c\xb1Lk\xff\x1b|\xcd\xf7\x89\xd1N\x05\xc3[\ \x0c\x8f\xe4\x01(D\x8eN-\xd0\xb0`LL\xff\xb5\ 6\xfcR\xc1\x14\xbc\x0f1S\x19\x03\xdf\x86\xde\xa9\xaf\ \xbcC\xabz\x8dq\x8b0\xecomo*r\xc6\x93\ \xb3\xc0\xd3\xe0\xc54\x09\xb9[\xe5\xb32V\x02\xad\x0a\ \xc197\xcd\xedO\x89U\xb7\x13#\x04\xfa\x93\xd7\x99\ \x1c\x83\x04x\xd7\x10Qp\x92Cp\xfa\xde\xcf\xf5m\ \xff\xc8|5i.\x9f:\xfd\xd7\xfa\x90\xcb\x05U\x91\ (P\x039\xf7G[\xb7\x89\xd5\xb7G:[l\x0d\ \x00\xbd\x90G}\xc0g\xc5\xc7\xc8j\x0a\xb3\xf6]\xba\ \xf4Zn\xb4B\xb5\x9111\x0022\xca\xf8\xdc\xa7\ \xb5\xd2Y\x99\xe9c\xba\x01u\xd4#\xed\xc6\xaa\xef\xb9\ \xaa^\xc7s\x08\xe0\x0e\xa4\xbc@\x14+w4;\xeb\ \x1f.\x7f\x9e]\xef\xbe8\x9b\xcf\x8a\x8f\x01p=\xc2\ \xd7\xdf-t\xa4K\xec\xcd\x04\x03I\xce\xc5\xa6\xdf\xa2\ [\xcap\x08\xa2y\xfc\xda\x8c\xfb\xf5\xe1W\x12\xaaJ\ \x17c\x83*\xcd;\xc5\xaa\x1b\xa3\xe1v'\xa1\x0f\xf8\ \xf43\x06\xe9!\x09\xc2\xf7<\xcf\x9a?B\xbd\x82\xaf\ \x90gdb\x03\x84\xc2\x826mj\xdf\xb3\x5c^\x22\ )\x07\xb1\x19\x09<\x03I\x99\xdb\x1bpM\xb9\xcb,\ Y\x00\xa6\xb5\xab\x8e>\x85\x12\xa5n5[x\x89\xd9\ Q\x8f\xce\xd6\xe9\x9fz\xd3\xbd|\xfa\x19cw%f\ 4d\xed~N\xf6\xef\xf1\xf5,\xa8_8\x067\xdc\ \xbb\xfe\x02\xa3Y'5\xb3\xa1\xb9\xbd\xf4\x8c\xfb;*\ \xbe\x02\xce\x05\xfb_\xac\x04\x826\xef\xe2\xabn1\xf0\ &\x04y&\xaa+=.\xa4\x991\x92\xe8P\xa7C\ \xdf2\xad+]\xa2{\xc5\xc9\x15'`7\xf8\xa6\x1d\ \x95Jg\x0d\xc5k\x94\xe2\x03\xf4J\x14\xa7\xca(m\ \xde\xd6\x5c\xb7\x1f\x13\xfa\x03<\xde\x80w\xfa=\xd1\x92\ \xf3\xa0\xc2\xb6N\xf1\x9b\x0a\xa5v\x15}\xe3|\xbd\xb3\ \x05g\x10d\x14k\xc3\xd9-\x06d\x82\x8f\xb1\x1d$\ \x12\x02V\xd0]Z\x067un\xe9\xf2\xd2!\xbb6\ \xf8-+\x19G\xdd\x1c\xd8\x87\x0dU\x8ap\x93\xbc\x8e\ :\xce#`\xa0\x88\x17\x9cP+\x124\xb6;\x89\x99\ \x0d\xect)\xd3\x5c\x1e\x18=E\x87}\x15\xe3\x5cT\ /V\x05\xbb\xa8\xd0~\xbe\xf2;F\x04b\x1aL\xb4\ [T\xecH\xf3X\xa9\xabt\xf8\xa1<\xda\xd1\xb6w\ E\xe4\xa3\xbf\x15\xb0\x1d\xcc\x8c\x0a\x85\x85Iq\xe7\xa0\ s\x0b&\x5cF\xfd\xf9\x8a\x0c\xef!w\xdc\x03\x16\xa9\ \x14s\xd7\xf3|\xf5w\xf1\x19Ox\x04gK\x5c\x80\ \x1e\xc9\x9ar\xb7\xe7\xa4\xeb\xfa\x87\x93\xe9\xd2\xad\xa9G\ \xf5\x15\xb7\xaa\xd5\xaf1\xce\x0f\xcd\xd3\x80\xbb\xf1\x95\x92\ \xcf/Q\xdd>l\x89\xa8\xd4X\xab\x95\x1e\xc6\xd8\x85\ \x82\x90\xb0`\x99:\xd9\xf7\xb2\xb9\xf7uZ\xbb\x92Z\ \x1d\x90\xdceS\xa4\x11fa\x9a\x15\x1c\xa7\x94\xcf\xa7\ \xc3/S\x83\xa5\xb8\x05[\x8a\xfc\xc4\x00Y\x960v\ \xfe\x8b\xac\xbe\x035\x05\xfa\xea\x15e\xe00\xe6)?\ \xf0L\xf8VR\xafpC\x1c\xdf$q\xb5\x16\xbb\xe2\ \xf0\x17\x0dwX\xef\xfe\x5c\xdb\xf7\xff \x09\xf6\xefj\ 5\x94\x17\xcfd\xb3\xfe\xa2\xb8\xfd\xb2\xe3\xc2#\xc7r\ \xfct\xf6J \xa6\xde\xb2\x97\xbf\xb2\xc0Zy;\xab\ ^\xc8xX\xbaS)6\xfe\xd9\x0b\x94\x0aSm\xdd\ L7\xff\x8e\xbc:?\xbc\x7f%\x8cWd\x07\x05\xdf\ 1\x02\xaf\xaff\xaa[\xea#\xf6\xbd\x8e\x06\xc8\xc2\x5c\ \xd91\xd2\xb4/\xf8\x84\x88qT\xf5p\xa0\xfeP\x9f\ \x0a\x8c\x9e\xbc3~F\xca\xcf\x97\xc7\x86\x14\xd8\x04\x9b\ \x05\xadYA^\x9ec\xc0\x90[\x9e\x9f\x92L\xea\x19\ ic\x0c\x8c\xef\x8c\x03\xab\xc8\xd2\xaf\xd3\xb6\xbd\x0c\xa4\ \xefI\x5c\xac\x93\x11VW]gm}\x88\x83\x83\x8d\ \xc3r8\x83k\xb9\x0a\x84\xe2\x03\xf7\x12\x93V\x8e\x05\ \xdc182F\xb5\xf6\x1dP\x0cDt\xa6\xc5u\x0b\ \x06\xf5\x16,;\x1bz\x03A\x14M\x99\xf6\x0bR~\ .\xc7\xc6\xe6\x18\x1dtN\xc3\xf5|\xe5\xb7M\x88i\ \xc0\xfd\xc4VB\xaa{%\xdbU\xc2\xbf~`9]\ ~-\xe3:\xa6\x22\xe7{\x00\xec\x08\x99\xc0\xe4\xe0/\ \xf8\xe8\xab]\xa7\xdf-wB\x9fdg8\x1e\xb0D\ J\x8c\xb6\x83\xd6\xa2\xaf\xaam\xdb`\xf7X\x8a;\x1c\ \x14\x22\x18(\xd6\x95c]\xb0\xd2\x9d\xe5\xc7F\x9a\x04\ \x80\x9c@Gyh\xd1\x12\x8e>\xb5\xb5\xf9\x8dj\xb3\ \xb2\xdd\x0c[f\xb6K\x1b\x93\xe3\xfa\xf2P\xed\xf3\xa3\ \x07)\x10\x8bak\x01!8\xf8\x0f\xb9kO\x90J\ \xc7_\xa3\xd3\x5c\xfd=\xb2\xefE\xdb\xd3\xe0\x00\xd0\xfe\ \xca\x1eN\xceyUs{\x84\x8ct\x0ek\x15(\x0e\ \xac\x1c^\xe540\x06\xca3\x9b\xf7\x91e_\xa7\xed\ ;1\xa68L\x9a\x1eaKK\x99\x8bO\xfd\x95k\ \xc4\xa5\x9201\xb9I\xce\x0d\x0b\x94\xb5\xeby\xb0I\ \x1c\xe5\xd9%\xe2\xe9l\xda1\xf4\x1b\xd9\xb3\xefAS\ \xc5#p\xec@\xcd\x08\xb2\xa3\xa9\xe3/\x9b\x1a_8\ \x10m7p\xd4\xc8\xd0\xa4\xb8\xd5\xf6\xc1\xa5>\xf7\xb5\ \xe3|W\x8f\xcb\x09x\xdc\xb0G\x8cu\xef\x06\xf8u\ \x1em\xe3\xeb\xee\x14{_\x83U$\x87\xac\x0a\x96T\ z\x16;\xe3w\xaa7\x8f[Q\xb3\xa3\x8e\xe8a\xe6\ \xcdQ\xbc\xf9\xa0a\xc8\x08\xc0|\x12\xa9g\x0c7\xf4\ \xa8Xx\x19k\xdc(\x8d\x11\xa7\x01p\x1f4:\xf7\ \x15\x99\xe7\xbe\xe5\xf5g\xc7\x18\x87\xc2~F\xe3v\xf1\ \xfa\x02&\x0buRc\x01\x16H\x0cw\x8ez\xdeK\ Zv\x05,\xd3\xc3\xd4\x97@\x80\xd5\xde\xdc\xd7z\xdb\ ;\xb5\x8d\x11\xbb\x8dC\xcf\x0b\x05\x81\xa8\x1c+\x8c\xbd\ \x06\xfaW\x85\xf1\xe1\x01\xf7\xf3\x9f+\x19\x94\xe5bq\ \xcc\x5c#\x901p\x08\xb3\xc3\x5c~3\xafZ\x88\x17\ \xd1\xd8\xaa\x90\xcd\x90f\x0f3\x06\xcf\xa15o3#\ \x0c9\x09S\x89'h\x8e\xb8\xc65\xfa2M\xf3`\ 6\x09\xe5\x9e{\xa0\xdd\xa4\x0eh\xf1}/\xd2\x8f\x9f\ \x91\xee!N\xba\x00\xec]\xe0\xcf\xe8 -;\x95\xa1\ \x9f\xb7\x9fo\x1b\x0b\x14on\xd4\xe0J\xe3\xbb\xa8\x0e\ y$;\xfd\xc4\x90\x9e\x9b\x1a\xc3\xaf\xf0\x0c\xbf\x801\ 5n\x81c\x00\x9eS'b\xf1\xfe\xd0\xcd\xef\xd45\ E@AX\xbf\xee\x8a:\xbf\xce\x0a\x1a\xb91b\xbe\ [\x17>o\x88\xd7\xab\xa1\x9b\x902\xc6$\x15d\xc3\ \xfe\x85i\xa4x\xb6h\xddM\xdb\xf7`\x22\x96'\xff\ \xa2\xcd\xac\xe9\x03f\xb4Q+Ly\x84Z\x9d$\xd2\ \xa4\x1c\x5c\xc8j\xd7[\xc1\xe1\xd4[(\xcbI\xf6@\ \xf1\x93\x00\x17\xb3\xe7\xd5X\xa3\xac\x13B\xad_a6\ lqVb\x01\xa5\xda\xf8k\xac\xbc\x09\xd0[K\x0d\ \xf5\x0c\xe9\xd1\x04\xcb\x1b\xeb;\xed\xfb0\xdcrR\x13\ \x0eJ\xeaC\xe6\xcd\xcb\xea\x9a\xa2\x10\xdf#\x03\x9c\xf4\ \xe3\x006\xbf[\x1f\xbdv\x11^7(O[\xc4M\ b\xea\xc9Vg?H*.\x82E\x8e\xb68v\x89\ \xb0\x05'\xc9\xebV\x8b\xe57\x88\x86\xcdv9\xa9f\ \x0c\x8f4+\x8d\xef9+}\x82\x10V\x944nt\ \xd6b\x83\xe6\xcbQ\xce|\x80\x15LBm\xa0\xa2l\ J\xe0\xbf\xfc\xb3U\x87\x09\xb0`o$\x83N\xe3g\ =\xc3\x5c^\xfb\x08\xc9\x80n\xf2\x1f\xaf\xaaj\x88Z\ \xb2t\xe9KN\x08\x18ZZ\x84\xac\xae\x8d\xfek{\ \x13\xc6\xc1\xf1\x03\xca`\x8a\x8bM\xfd)/\x9cy\xa2\ g\xd6K\xcf\x06\xfd\x13\xe98h\xbd\xf7#\x81\xb3\xf0\ ]\x03\xad\x94\xc1\xaa{\x8f\xeax#q\xdfA\xa1.\ \xf5k\x9c\x95\x9e\x80\xfeTNN\xa8\xc1!\xec\xac\xc7\ \xc4\xb0\x0b\x1c\x82 ;\xec\x87\xcf\xc1\x9f\xcd\x1aT\xa8\ \xb4\x1e\xb3\x86|I\x9d\xf3\xa8\x9a5H\xee\xdb\x83!\ {\x8d]m\xfa;u0f\xc4;I\xf0\xcfI>\ \x01 \x92\xc2|/\xec\xe9\x0cY@\x9e\xf8`\xab\x02\ x\xc7\xc0\xd3\xcc{T\x0c\x9a\x22+},@!6\ \x9d \x86\xaa\xdfh}\xf8gPQ\xaa\x19\xd3\xb1g\ Q\x22z$\xa90Nx\xf3\x0eg=\x0e\x08\xe6\xc9\ \xd7f\xfe\x91\xcd~\xc2*\x9ak\xb9\xf2P\x87\x92=\ \x10\x22H\xa7\xabpw\x0e/\x99\xcbg\xfc\xd13\xeb\ 7\xd4\x9b\x0f#\x05\xf47\xc8\xad\xa4`sM{\xa3\ .\x03\xdb8\xb1\xb6.\xda\xd0a\x22\xef{\x09pV\ \x824}\xd0\xe3\xfe\x0c\xc3\x08!\x0e,\x12\x96\x95\xb2\ \xb1\x12F\xfePV\xed\xdf\xcf\xcf3\xb6\xa0\xbd\xfb\xd6\ bQlA\xa2Z^\xe0\x8a\xf8:\xa6\xc3\xeb\xcb\xcd\ \xa8\xe8\xac1\xda\xaa\x1bw\xae\x0a7\xee\x13\xdc\x0a\x14\ \x8e\xca\x1d=K\xcb*\x10\xbeRE\xd1 \x0f\x8aI\ \x1d-%\xdc\xcd`\x1d\x04\xf9\xce\x0b\xef\xfd\xbf\x86\x80\ ,\xc8N\x8e\x15@\xf4\xa7\xa7(\xe7L\x1c\x09\x1e\xd0\ I\x8a\x07P)\xabi\xb7xu.\x0ar\xc2\xaa\xd9\ \xd5\xe7\xderz\xces)\xf31v3\xe5\xf9\x1e\xe8\ \x92\xa0\xa6}\xa5)\xd4\x0f\xea\xc8\xe2\x7fn\xb6\xf4\xc9\ \x0e\x14\xcd\xa3\x04\x87z\xcaf\x94\x9eu\xc7\xc8K\xff\ 4\xea\xb2\x87\x8af\x7f\xd7]t\x1a\x0b\x0c\x05\xba\xd8\ ylU:\xcb\x09\x05\x9aA^\xdfW\xa9\x15\xa2F\ \xe2?<\xc4\xc9\x95J\x81\x1c\xed\xf4\x06P`\xb4u\ \x17\x0a\xd13P8a\x84\xadH{\xca\x18\xe3\xb4\xd1\ v\xdd\x0b\x85K\xff\xdfW\xc0\x11\xadDF\xeep<\ [\xaa\x04\xc8\x16;,.\xa2\x168`(=\x16\xcb\ \x1d\x01 Y\xc8\x846\xd3{\x81\x95\xac\x02\xf8\x8e\xb5\ `\xe6b\xaa;e\x8c\x91\x8d\x942#k\x98\x93\xd0\ G\xc8\x16\xafz\xf0N\x9c\xbe\xc0\x16K\x02\xa7*\x9c\ E\x09'G\x92\xa1*4[11\x8a\x89\xdf\xc9\xc0\ \x0e\xf9\xcc\x8c\xcdI\x1c\x03\xd8D\xfcCH\x0c\x13Z\ \xb6d\xcc\x97\xad\x06\x8bR\xc6\x18\x07\xd9\xc3\xce\xc0\xc0\ \xbb\xcf\xe6\x00\x97\x0e\xa1/\x0b\x969\xeb\xfd\x16@\xcd\ \xf1Y`;0[\xdc\x86W)\x1d\x15\x94\xa7q{\ \x0b\x977\x9b\xe4O\xeeq\x7f[2\x96?NI\xa1\ \x8fq@\x0b\xcf\x14\x8a\xbb\xd7\xcd\xa2\x1b\x18\xbbS\x85\ \x17Ls\xd6S\x0e\x19\x96\x1d\x82\xc5\xf1\xaa\xf1\xa3\xe0\ d=\x0e\xba\xdd\xd8\x8c\x11\x85\xae\xf8b\x11<4\x1c\ \xbf\xd4\xc7\x86\x0f\xce\xedC\xa7\x04\xd1\x81\x22\x86\x7f\x09\ 4\x891\xd5'\x05\x96\xb5\x00\xe0\x92'GL\xfe)\ x\xe2T3F\xcb\x1eb\x05\xc7\xcb*\xf7\x09\xe0\xa6\ \x04Q\x94\x82S\x9d\xf5t@\xd6\x01\x15\xfa\xfc\x87\x07\ \xaf]X}\xfd\xd2\x9amM\x11\x19\x8f\xc6\x0a\xd9\xfb\ \xd1\x09\xf9\xee\xa1A\x15\xa7>b\x03\xbaW9\xd9;\ \xa7\xc4\x93\xed\xee\xcb\x8d\xb1\xc05\xca\x86~\x81\x96\xcd\ \xc7^\xd1I<\x02\xd8\x1fP\xcaU/;\xe5\x87\x9a\ ;\x08)\xa9f\x0cQ\x5cj\xe9,P\x95\xb3\xda[\ \x80\xceh`\xa8V\x9c6\x1f\x03@?G\xc8s;\ \xdb\xbe\xbd\xaa\xf5\x95}\xa1\xe7w\xb6}\xe5\xcd\xaa\xaa\ \x90\x85\xe6\x8c\xa7I\x0c\xf2\xaa\x97\x0c\x0fR\xbce:\ F\xc0\xc1)t\x0f\x97W\xb85\xa5\x8f\x9a\x14\x8a\xcb\ \xab\xce~\x84\x97\x9duL\xca\xe2\x85\xe5\xccK\xc7_\ \xaf\x8c\xb8\xccvf)g\x0c\xe8y\xc4\x15\xc2\x93\x8f\ n\x10\xaf\xeeqR\xe3\x85P\xbc\xfa\xe4\x9fQ\xd5\xed\ \xac\xa7\x1c\xf6\x94RD7\xff\xbd\x03\xe7\x0b\xa42i\ u\x87y\xd1+\x07\xd6\x1d\x8c\xe2\xe0\x07\x81=\x95\xcc\ ~\x22@#\xbeqB\xde\xf9\xe5YxV\xc9\xbe\x1c\ \x0e\xe1l=\x0c\x90\xe6x\x02\x8d\xd1_N\xcb\x9fZ\ \x01#\x9d>0\x06w\x05\xc9)\x9e\x15>\xe3O|\ \xea}$\xf7dKN>\xa1 \xa8d\xcd*9[\ \x99\xf7\x84k\xd2\xad\xdd\xddQ\xcaf\xf0\x1cpy\xff\ G\xdb\xfeu\xbe\x95WS#\x82\xcfQ\xe8\x95\xbf\x11\ \xccE\xe7<\xae\x96\xccH\xe2\x09\xc2\x13B\xeaMt\ \xea\xd65oU.\xae\x0d\x83@`xL'd\xb0\ \x9b\xbdvQyy\xd0~-\x05\xd4\xaf\x87f)Y\ B\xebB\xd1\xab\x17\xd5|P\x1f\xc5\xe3\xcat\xfb\x80\ \x87\x80-\xcc\x04\xf3*\x8c\xfc\xe7\xd8\xc0\x8f\xa6\x97\xa8\ \xb0\x94\x08\xd8'4q\xcaC\x98F8\x1c=\xf8\x9e\ \xd5^\xaf\x15\x8c\xf4\x0c\x9e\xa0h\xd8\xeb\x1d^\x85T\ _\xed`k\xc3\x1d(28#u\xabP%\xbd!\ \x0c\xa1\xdc$\xfb\xdf\xb4\x82\xc3\x95\xdc\xd1NR:\xa0\ 2f\x09\xfe\xe6\x81\xcen\x97\x0eN\xa0\xc3\xe4\xaf\xed\ \x0bM\x1d\xec+\xcaR%cz\xa8!\x92\x8f\x0a\xbf\ K\x9dY\xe4\xe9\x88Fw\xb4Y0b\x06\x1c\xb5\xa3\ \x9dV\xe8Q\xef8%\xf7\xc6I\x05n5q\xa7\xba\ \xe0\xd0x,\x88Y\x14Es\xbbs+<E\xe3]\ 0\x90FFB\x85`\xdb\xa1\x82R\xedcl@\xa1\ \xdc2C\xeb\xee\xf5\xec~\x1cm\x8f\x22\xe3t\xc81\ \x80\x95\xb1\x17PPX\x94\xc1\x83\x93Uh^2\xfb\ Y\xb5\xf8TT^\xcc\x91cb\x01r=\xbb\xa5\xf6\ \x8eum&^}\x8c\xb0\xf5[\xecS\x17]R\x91\ \x07\xe1\x06\xae\xe2\x08\xf8\xc4\x12\xda\x86\x80\xef%{\x9a\ \xefZ\xdb\x5c\x15\xb1tNy\xd7\xc3\x9e\x15F\xbd\x0a\ =\xb5\xc0\xfd\xdb3r\x87\xe4\x05\x0e7a\x8a\x916\ \xc6\xc8 FD\xb7\xffK\xdd\xf4K\x81O\x14C1\ \x8e\xa9\x08GB\xf8R5J\x14aEd\xc8\x09\xfd\ \xafL\xd7\xfcd\xca\xcf\xb5\x91\x17\xc7z\xd1k\xc2\x01\ \xb5\xb0\xf8c\x1bk\x7f\xb2\xa1]>4JN\x05b\ \xff+\xca\xb3\xd4?\xcf-\x9eZ\xe4\xb3kw\xe2\xcb\ \x05\x0f7\x84n\x9a\xdb\x9b\x22[\x9a\x8d\x96\xb0i\x98\ \x86\xd7\xe5*\x0ah\x13r\xb5\x8a\x1c\xfbP\xd8>\x9c\ \xac)G\x1a\x19\x03\xcd\x0e\xabm\x1aQ\xb1\xea\x16q\ \xe0M\x22,\xfb$\xfe\x91\x80\xac\x90D\x89\xaf\xd0\x9a\ \xf9\x88&\xda\xc5\x92k\x88\x05q\x83\xdcfgW\xdc\ \xd6\xb9/z\x0aN\xc2\xa4\xd4\xa3K\x83\x7f\xdbTs\ \xeb\xbb!p\x9e]\xbe\x04\xa2\x03Z\x96\xa5\xae\xff\xf2\ 0\xec\x9c\x90J=xA\xe7H\x82X\x94+P3\ \xac\x9e\xdcS\x8e\xbeP98R\xc1<\x9f9\xc6t\ \xc3.\x1d\xbe\xad\xa6\xed\xb4vi\xa4n+i\xd8\xa8\ \x89\x16\xc2\xf1\xfeY\xa2\x06\xa3\xde\x11\x9e\xd2\xd3H\xfe\ \xa9\xb4\xe8LUUAu\xe6\xfe7\xc4\xea\xef\xc9g\ s\x1c\xea\xa0\xa8\xe2!3~\xab\x0e\xbd\x00\x12\xd3\xd5\ =\x19\xa6\xf5\xc8\x96\x86\x9f\xbc\xdb\xcaaLj\x8b&\ \xedZ\x9c\xa5<5\xbfdb\x81\x07\x8d\x8f\x13\xb4\x89\ \x8b?\xd2\x81\x8c`\x8c\x5c\x02m\x82\xa2\xe5\x07Y\x80\ \xea\xed\xfa\xb6\xd1\xadea4l\xa5\x0b/\xc7K}\ q#\xe6\xc1=\x15\xb78\xe7\x05W\xc1\xf8\xf4\x054\ \x18\xc7\xfc}\xcb\xc1\xef\xaei\xb7\x84\x09c\x1a\xdb\x03\ \x02\x85\xcb|\xca[_\x18Z\x80'aA\xe3\xc8\x19\ \xdc\xa1\x7f\x22=\xca=\x0a\xd8\xe6\xa4\x22e\xf8\x0b\x22\ \xd9\x0a\xed\xfe\xc6?\x1bv\x8a\x96\x7f\x12=\xe37B\ \x0b\x00\xbd\x04\xb5$\xc9\x08\xb1\xa2\xf4\xad/\xea;\x9f\ \x87h\x11\x88\x98\x8e\x96\x80r^:\xae\xf0\xc7\xa7e\ S\xa6`\x1b\xc0\x16\x01_\xbc\xb2\xc3X\xf0\xc2\xae\xf5\ \xb5\x1d\x98\x02\xbd\x8d\x03g\xb7\xfe\x854\xfb\x98^\x03\ \xc7Z\x8d\x9b\xc8\xdb_\x16f\x04\xd6\xb0!\xe3\xf4\xbc\ B5\x1f\x9f\xf5\x98\xabl\xbaM2'w\x0a\x01\xfa\ 4\xb9\xf5\xf4\x07\xd5\xff\xbd\xa1\x13\x02aFT\xd0\xb1\ \xbd\xa9\x22\x8b-\xbcxX\xd0\x03\x99\xf0\xa2\xb6~\xda\ =e\x84\x8f\xe9\x0d\xa0\xef\xc9\x9fD\xcf\xb8_@\x04\ \x03+xM<\x18\x80\x0b3D\x97_e\xeey\x15\ \x83\xc54Aa\xec\xeaSJ\x7f0)[!\x0c\x87\ \x84]\xd8\xdb\xc1\x17\xbc\xb8kCm\x04\x18dwX\ \xfd\x11\xfd\xd6\xc7\xc8\x0f\xc0\xac\xdf\x22\x16_\xce\xf4\x0e\ \xdb2\x18r\x82-T\x1f\x99\xfd\x17<\x81\x85s\xb1\ )\x1duw\xeb\xd3\xe2\xe2\x99M5w\xbe\xd7\x8e\x97\ \x9d#}\xe5X\x87\x88r\x9f\xfa\xca\x05e\xc5~\x0d\ WAV\xf8\xa4)\xf0\xea\x1d\xfa\xab\x8fA=K(\ \x05\xe3\xd8\xb4\xfb\xb8\x96\x85,\x82?4\x83 f\x07\ Yz\x0d>a$\xe5\xed\xc1\x11\x0b\x04c\xf4\x8a\x89\ \xc5wN\x08\xda\xcc\xc6MR\xec\x03\x1d\xd6\xe5oT\ ~\xd4\x18\x05Ye\xfc\xd5\xcf\xd0_}L7p\xd2\ \x0cF)\x0d[\xc4\xa2+\xa8\xd9f\xd3\x06\x8d\x03\xbf\ Z\x90\xcf|\xd03\xe4,\x991\xd5\x90g\x8b\x88a\ \xf1\xc7\xdf\xaf\xbc{S\x18\xd4l\xcf\xe0\xc1\x02pe\ l6}\xfa\xdc\xb2\xa1A7\xd2\xa8_\xf9\x98\xd4\x9f\ WJ0\xa05\x83%\x94\xac\x22\xe1\xaf \x95\x8b!\ \x22\x96\x0dW\x9a\x85\xeb\xec\xc0\x1b\x86\xa7\x90\xe5\x8c\x92\ 3\xb1)m\xd1\xd2\xd1P\x95\xb1)%Ab\xe8\xeb\ \x1bu\x0b\xd9\x0c\xe4\xe0 JcD,\xda\x17:\xa3\ \xd03\xc8\xa7\x22\xc1e\xbbM\xb1\x84\xbdC\x7f\xed\x95\ >\x09Vq\x9e\x98\xfb(q\xae\x7f\xc0\xb39\xc8\x19\ \xb3\x93\xbd\xffS\xb3z5r(}\xb8i\xf2\xa0[\ \xc7xeO\x0a\xa3kI\x0bJv\x87\xac\xef,\xab\ \xab\xed\xb4\xf0\x0a\xf7\xfe3h\xea\xf7>\xa6\x1b8\xd7\ \xeb/'\x81\x11\xe4\xe0rjE\xe54\x19\xc6\x9b\x8c\ \x87\xc9\xfeWy\xce8\x1a\x1c\x8a\x86JGS\xd6\x14\ efy\x8e\xd0#\xeb\xeaup1N\xf8BIC\ \xc4\x5c^\x19\x9aQ\xe2\xcf\xf7@x.e\x93\x9fL\ F\xbf\x8fc\x8e\x82\xe0\xa6Q\xb3\x96,\xbf\x86\x18x\ \x01\xa5\x1d\xd3\xc0\x17se[S\xefs\x0f;\x0f\xf2\ \xa4\xcb\xf9\x87\x0d\xf3\x81u\xfb\xff\xe7#\x1dc\x1a\x1c\ \xd59\x9eetP}fA\xe9\xd0lw7W2\ \xb9{\xfa\xf4\xf8\x98.P%PF|CD\xed\x0a\ j\xe9\xd2\xd3\xa0!\x04x\x9d\xaa7\xf1\x09\x8c\xd9C\ \xd3\x15i\xaa\x8c\xce(\xcf\x8evvlh\xe2\xce]\ \xf9\x92\x19\x0d\x11\xbe\xaa\xbac\xfe\x10\x7f\xd0\xadHz\ C\xf2g\x951\x8790;\x1e\xc59r9\xe2E\ #b\x8f\x0e\xcb\xce\x14\x8a\xd3\xe0\xfa\x0c<\x0a\xcb\x19\ M\xf3&\xf2\xca7\x08\xd7\x813P\x9e\x82\x11\x84E\ k\xdf\xb1\x02c\xd5\xeca\xb6Y\xec\x1dR\x07Y\xcb\ )E~+\xdc\xb1\xbe\xc1\xbe\xf0\x85\xe2iKB\xeb\ #\xd6\x9a\x83\x9ds\xcb\xb2\x82.<Ur\x18\x9d2\ \x0e\xc9\xed\x95\xe0\xe0xt\xc1\x1b\xc3\xe6\x9ev\xab#\ j\xedl\xeeh\x89\x18\x05>\xd7\xf0lO\x96G\x1d\ \x19T\x83n\x0d\xc61\xa8 {\x9f\x04\x01hh\xed\ {\x9d\xaf\xba]\x18\xed\x18\x06Kb\x02=\xa9\xa2\xf1\ \x19\x0f\xba*\xe6S\x86\xb7U\xa7\x01x\xc6\x92\xff\xe2\ \x9d\x03\x7f\xd9\x11\x8d\xc0\xd8I^Qco\x98\x98\xef\ yt^q\x85\xbc\xdc\x13\x90\x99\x94I:c\xda#\ \xd1?mi[V\x19\xfa\xa8\xd94,\xbcx\x04\xc2\ \x0b*@ITS\xf8)\xf9\xee9\xa5\x81\x1b&\xe6\ y\xb5\x04\x8f~q:DXb\xffb\xbe\xe66\x12\ m\xc3\x9abL\x89\xd5e\xae\x1c2\xf37\xea\x90s\ \x9c\xac\xa9\x85T\xb8\x08E\xcc\xdf\xaf\xaf\xfa\xd3\xc7\x11\ $\xb1\xac8&S1)\xcf\xf5\xf0\xfc\xd2aA\x17\ \xf0;\xc6\x07\xb6\xa5\x18\xc9b\x0c\x1c\x96\x0b\xbe\xae:\ \xf4\x9dw\xea\xf6\x85$O\xa4S>\x0aP\xb8\xa0\xe2\ \x94\x5c\xf7\x03s\x8aF\xe5\xb9@w\x09W\x93Y\xb5\ \x8c\xbfs\x13\xd5\xf1I\xfc\xc0\x19`*\x9eOP=\ |\xf6\xc3\xae\xe23qj6MS\x0c\xa0\xa2\x9f,\ \xdd\xfd\xf0n\xd3\xc0\xa9>\x80d\x8e \xa7\x16x\x9e\ >\xb7d\xb0\xd7\x19=ARFM\xf1%+\x8e1\ -\xfe\xc4\x87\xcdw\xacj\xa8\x0bc\xdb9\xae\x87\x85\ -\x9c\xd4F\xac\x97v\xb7\xe6y\xd4\x09\x05\xde\xe3e\ \xec=\xfc\xa54\xefd^\xb9\x88Y:\x88\x01\x1c\xc5\ \x02\xb8\xc1j\x16[Y\xa3\xd4\xdc\x11\xc7\xa4rj0\ \xa58+\x1a\x0a\xado\xc6f\x03\xa2\xa1\x13\xa4\xa4\xa6\ \xd3X}0<\xbd\xd0\x97\x8f\xa4\x01\x80J\xd2&\xe1\ '\x91`\xc6H\x97\x81q\xdbK\xbb[\xeeZ\xd7\xd8\ \xa6\xc3\xa2\x0ck\x8fK\x19l^\xb0\xa5\xd3\x22\xab\x0f\ v\x9c1\xc8S\x16p\xc1\xba\xcd2;G\x1f\x01\xa6\ \xa0\xc1\xa14o<\xaf^\xc6\xac0\x1c\xd4y\xcc\x92\ \x19e\xd5\xaf\x9b\x81\x93\xd26z\xa2\xd4\xa5\xa8s\x86\ \xe6\x98\x9dm\x1b\x1b\xb9\x89b!i\xe0\xaf\xa6\xd3\x5c\ w\xb0c\xc1\xd0\x80_\xb3\x9f\xa7)\xbb\x02\x9bP\xe9\ F\xa2}\x8c\xecfv\xb5F\xae_T\xd3\xa4\xe3h\ \x08\xc2\x93\x134\x12g\x9b\xdc\x1c1\xc5\xea\xda\xe8\xe7\ *|A\x17^S\x92(\xed\xc8\xa3S\xea/'9\ \xa3E\xe5\xdbB>r\x06RpD\x22\x04\xabY\xc2\ \x03\xc3\xa8\xf44\x89*1F@aR2zj\xa1\ ?\x14j{\xbf\xd9\xc2\x98F6\x16@}\x84\xaf\xae\ \xee\x98U\xe2\xcdqw\xdd\x90\x85\xf2\xa5T\xc2c\x22\ \xc1m\x0b*\xa4\x0b\xfe\xe3\xb5\x0d5:\xd6\x0d\xfb\x80\ X\x01\xe4b{\xda\xf5\x877\x81\xe6\x90vNr\xa2\ \x00\x01K\xe9|6\xef\x09\xe2\x09J1)\x04\x0fP\ yn\xb4\x92\x95\xb7X\xd5\xef\x92\xae{GR\x0f\xbf\ [\xf9\xd9Y\xc3\xbf1L\xd1\xf0\x09\xbb\xd0\xcc\x1c\xa5\ ml\x8c\xde\xbc\xac\xa69\xea\xdc]\x9b\x09t\x01$\ \xdc\x1b\xd3]\xcd\xd1\xf7\xeb\xc2\x0eWb\xaf\xa3T\x13\ \xf4\xe5o\x1e\x08\x1dl\x97\x8f\x9eO,dkf\x85\ \xd3\xd8\x8c\xdf\x11W@^\xb0\x82s!\x0c4\xc0\xa3\ d\xd97\x8d\xdd/\xe1\xed0qP<\x81@2\xdc\ 5k\xe8\xd7*\xc0\xc7\xd8\xefB\xc2\x7fH_[\x17\ \xbd\xf4\xd5\x03\x1f7\x83B`\xd5\x96\xad{!=H\ <c\xd6\xd7\xb47E\xf1*\x22\xf4\xad1S\x06\xb3\ \xe37\xab\x0a[\x1f\x1c\x84\xc1p\x82a\x8b\x83\xf7\xfc\ \x95-`s\x9f$\xee \x16&\xc0\xd3\xa0\xf6\xb9\xde\ L\xd7\xde\x11\xaeZ\x9f\x16O#\x05c~\x97z\xef\ \xbc\x11\xd7\x9e\x9c\x8d\x93\x8d\xd8;\xe1\x17l\xd8\xd2\x1c\ \xbdyiu\xd8\xc4\x8b'\x90\xd1(o:\x9dM\x82\ \x19\x03\x95ZQ\xd9nO\x81\xf7\x0e\xbaE^;\x10\ vV\x92\x00\x90P\x194\x99M\xbf\x9f\xa8^\x8e-\ \x9aA\xf4\x80\x93\xf3V\xd4\xb5\xf4*}\xdf\x1bN\xbe\ t\x00x\xf3\x83\xc9\x83\xae\x1a\x9fc\xaf\xc9o\x90\x98\ nh\x8c^\xfc\xca\xbe\xed\xe8idBz\x1c\xa1\x83\ \x043\x06\xda\xc1\xc66\x18x\xf4\xe9\xb0\xdb\x1a;\xed\ N*\x19\xc0\x89B\xa6(C?\xa7\xcc\xff\xab\xd0\xfc\ 2h\xb0/\x09\xa6\xc2\xea`\xabo5j\xd6Y\x1c\ \xefF\x90\xd9S\x07\xe9ih\xc0\xad\xdc;}\xf05\ c\xa5\xa7\xc1\xfb*\x81\xe2\xf8\x1a\xf7\x0f\x1a\x8do-\ \xa9\x0e[\xd0\x18Q0\xf45\xce,N\xaa\x91`\xc6\ Dt\xddtn\xef\xea=Z;\x8d\xae\x11p\x12A\ \x07MR\xa6\xfd\x8aj~\xd9\x96\xbb:P+L\x16\ ]\xc1\xf7\xbe~\xa8\x89\xa7\x1c\xc0\x8f\x1fN\xcd\xbfz\ l\xb6\xdd9\xc1\xaa\x9d\xbe\xad9z\xc9+\x95\xbb\xdb\ \x0cG\xb64\x09\x98`\xc3(8\x85\xda'\xba\x80\xe5\ \x14\x96\x0am`L3\xecBz\xce\xb3D\xf59o\ \xde\xc4>\x8b*\x5c'\xabo\x097\xa4\xed%\xa2 \ G@U\x7f5c\xf0\x97Fe#g\x08N\x07\xa0\ S\xc1\xee)r\xcb\xf2\x83\x11|6':$;\x7f\ \x8a\x91\xe0\xb3\x04\xe0F?\xf7\xf7\xad\x1bB\xf2B\xf9\ ^\x01,7\xb3@\xfd\xf7\x17F\xa4F#\xe0\xdc\xf9\ \x817\xad\x15\xb7\x123\x0c\xca\x90}\x14\xd0\x86\x11\xc5\ \xcdO\xbf\xd75\xeab9q\x93\x1e\xdbt\x18\xe6\x8f\ \xd64<\xb3\xad\x05\x05\x00;aCBI&\x0dr\ ?8\xabxd\xae\xf6qm\xdb\x9a&kGC\xc7\ \xc1\xa6\xd6\x88afy\x5c\xa5\xf9\xb9#\xf2\xbc3r\ \x05>\x1f\x0f+b\x1f)\x91\xf4J0c\xc0\xdew\ \xaf\xac\xfe\xdf\x0fC\xbd\x96\x10\x04\xbav\xa8\xf2\x8b\xb3\ G\xa5\xa6\x09a\xed\x057\xeb\xde\x13\x8b\xaf\xa4F\x18\ \xcf\x90\xe2\x95{ \x05g\xae\x809\xfb\xff<%\xa7\ B\xe0\xe3\xe4N-\xc04Q\xd3\xbasm\xf3_\xb7\ 5\xe0\xc5ax\xcaM*E\x90\xf1\xd9J\x9e\xdf\xb5\ \xbe6\xacs\xc2\xc1\xe5\xe0<)>t\xd3\x22\x5c\xa3\ \x02|\xe6%\x15\x9e\xfb\xe6U\xd8\x87I,c\x12}\ \x96\x00*i\xf0\xd7\xf6w\xe0U\xd0r\xce\xdbN\x8f\ \x05\xd8/\x08\x12\xd0\xe8\x0f\xa7\x17\x96\x05\xf0E\xa0\xce\ \x86\xe4\x02G\xb1\x8a\xbf\x88\x04G\x88\x9aex\xb9'\ j\x1f\xa4g\xcc\xd2\xd9\xfe\x7f\xf3\xe0\x18\x16\x1c\x86J\ \x97\xba\x97\xcd<5\x82!T\xc6\xe6\x97z\xeb;\xf9\ \xc6\xa6\xa8\x1c\x0c`X\x03j\xad\x8f\x88\xfd\xed\x102\ J\xe9\xa5\xa2\xa4Lr\x92\x89\x12]\xd0\x8d-\xc6\xdf\ >n-\xf2\xd0\xb1\xf9\xddomI\x0cm\x12\xddz\ \x04=\xa5\xd0W\xe6W\x880aY\xc4\xfe4@\xdb\ \x1a\x84\x8e\xcbf\xa3ra\xdc\x8b\xeb)\x00h\x11G\ O\x10\xd3T|\x9e\x9e\xf3O\xa2e\xa1\xd8\xf2\xf4\x0e\ \xa8\x9e\x03\x81V\xde\x12\xad^\x099\x1d\x81\x12\xea\x92\ O\x0c\x94\x8bRM\xa1\xf7\xce\x18|\xc5\xa8\x80T\x8f\ |\xef\x1a\xfc\xa1\x80RpL\xc4\x9cr\xab\xfdk\xf3\ \x82UvZ7\xae\xa8\xbb\xef\xbdzh\x88\x00\x1e\xc7\ \x93\x19O\x84\x84\xfb[Q\xe8\xd5.\x1a\x9a%\x8f\x0c\ v\x8f\xf9~DYMF\xf9\x97\xc7\xe4\xe5x5\x96\ \xfc\xb1\xd2Q\x00\x22\xa8\xb9c\xe9\xb4_\x12\xcd%\x9b\ \xb1\xbc\x88\x07\x86\xb6f'[z\x8d\xb1\xfbU\x01m\ \x00s\xd9\xf6H)\x804\xf7\xcd(\xbajt\x0e\x95\ C\xea\x989+\x0c\x93<\xb4\xa5\xe5\xf9\xed\xad\xc8\xaf\ \x04\xd9:\xc1\x86\x01\x923Fo;u\xd0\xd4B7\ \x1c\x1c\xbb\x99\x98\xa7Q!\xb3\x8b\xb2\xf1\xb9\xb0c\x8c\ {$\x12H\x04\xc6\x94\xe1\x17\xb3s^\x10*>\xc5\ [\xb2\x067P\x18=\xad\xbb3|`mZ\xe8b\ \x17\xa9\xa9\xf4\xa2\xe1~\x01\xcaE\xbe\xc4\xca\x19\xa8B\ \xa7!\xee^\xd7\xd0\x19\xc1\xf779\xa9}C\xe2\x19\ \x03\xf0j\xea\xef\xce(\x9c\x98\xa7\xc2:xxYF\ \x0f\x95\x04\xb7\x09\xba\x88pq\xeb\xaa\xfa\x0f\xea\xa2\xb6\ #u\xb6\xa5\x04(\xb7M\xf8\xbc\xb1\xca\x8c\xdfP\xc5\ \x83g\x9d\xb0\x0b\x90&3Z]+\xaeBO\xc3\x0d\ \x99=u\x90Z\x10QK<\xfd!\xbe\xb1M\x0e\xe7\ b\x22\xae\xac\x0e\xf6Q\xf5Q\xeb\xfa%\xd5V\x82f\ \xfc\x92\xe2\xfc\xc1\xd8\xa3\x0a\xbcO\x9d]46\x1b<\ <\xac\xf6\xccn\x9c\xaf\x97\xd8\xd6\xa4_\xf9v\xd5\xc6\ \xfa\x88\xbd\x9az\xe0\xd8\xba\xe2<z\xee?\x99\xea\xcc\ \x08\xdb\x9c\xa1\xa6A\xd6~\xdf8\xb0T\xe6J%\xa0\ xZ\xd3\x16}\xf3@\xd8~\x94-\xfe\xc7\x07\xba\xb6\ \xc1\xfa\xa8%1\x5cO\x0ac\x90\xdb\x94\x95eg\xbd\ ra\xf9\xd7\xc7\x04\x0a<]\x8f\xdf\x91\x9e\x03\x8d`\ /\xc8\x08\xb3\xc0M=x^\x07Xe7\x02Q\x1f\ \xb1n\x5cZ\xbd\x15\xdf\xb3\x89\xf7T\xcb\xc4\xd4\x01\xc6\ \xb0\x10\x08\xab\x05\x93\x943\xff@4\x9cCs\x18\x0f\ \x862\xdb\xe9\xca\x1b\x8c\xbdo\xe0\xbd\xba\xa9\x13\x0c\xf5\ \xf6\xcf\x1d\xcdQT\x97\x0cz\xe3\x85\x10\xad\xba\xb5\xad\ \xbe\xd3Y\xed-\xa4\xc5xR\x18\xd3\x8dl\x8f\xf2\x8b\ \x19E\xaf\x5cX\xf1\x93I\xde\xd1\xd9\x1e\xd0=\xfaJ\ \xb9\x89Q21\xdf}\xff)\xeeW/\x1c\xf2\xbdS\ \xf2\xf0\x11`\x18\xf7\x80ap,\xb0\xbd\xd5\xf8\xea\xdb\ U\x1f7%\xe1\xb2\x87\x98A\xcb\xe6\x93yO\x11\x97\ \x9f\xe1\xcc+J\x8d\x1e\xc7\xd2\xc9\xda;\xa2\xfb\x97\xc6\ o\xb7^\x02\x0a\xb28yzG\xbb\xd3\xa0\xe2\x07H\ \x0f\x8a\xdf\xb0\xb7\xcaY\xef\x13\xb0\x01%\xb1\xad )\ \xa1\xe7\xc5R\xa0\xe6\xa2\xb15\xb4\xb7\xae1\x1c\xd6\xb3\ |\xee\xe1\xc5\x83s\xfc\x1e\xd8\x08\xd1\xa5e\xf2\xfb\xdf\ o\xf8\xdf\x0f\xdb:\x9d\x93Rr\x17\xc2G\xe7x\x1e\ \x9aS4\xb1\x00_\xcf\x0d)\xf61S\x06\xf0p@\ \x14\xbe\xff-k\xcd\xf7i\xb4\x09k\x80\xc0Q\x1cW\ \xdc\xd6\xb4\xdf\xb8\x87\x9d\x0f\xce\x08\x92\xbaZAR\x00\ \xe6\xe9\x8c\xea\x13\x9f\xdd\xdfnZ\xbd+G\x1aX\x9c\ W\x91\xf5\xcc\x82r\x99\xd0KH\xaa\xe0\xdcC\x12\x19\ \x13;L\xd3\xfc\xf3\xe6\x96{\xdfo2\xb8\x09\xb12\ \xc4\x9b\xc8\x1cBF\xe5h\x8f\xcd/\x1e\x9b\x9b\x84+\ \xc6c\x83\xe0\xa6Y\xb3\x92.\xfd\xa6e\x85\xd1\xd5\x08\ p\x8e\xf8 P\xc5\x9dmM\xfb\xbd\xabb>\xe4I\ \xaal\xa0\x85\xb6\xb0~\xf2\xb3{\xa1Wr\x92\xe2\x06\ \x1a\xf9\x8c\x22\xcfK\x17\x0eu\x12z\x05Y~\xca\xa7\ =\x8e\x07EQ\xae\x9f\x98\xfb\xcd\x93\xb3]\x0a\x04=\ \xb2\xc7\x96-wG\x8by\xcd\xc2\x9a\x8f\xb0\x13O\x13\ \xb3\x99\xa2\x94\xcc\xa2\xb3\x1ed.\xf9\xec \xc1\xb9\xc0\ \xa7^Y\x91\x16\xf6\xce\x7f\x9a\xfb\x17\x12|\xacw2\ !\xf0\x95\x07\xd0\x86\xec&\xd4;\x80&\xb3\xd4\x98\xe7\ \xc6\x8e\x07\x94!9\x91o/\x00\xf4P\x18\xfb\xd1\xe9\ \x83\xfe\xeb\xd4\x5c\x97*\xf5c\x0bH\xc9\xce6\xfd\xba\ E5\x1f6\xe1C{b\x19v%\x1c\xe0BX\xf9\ <:\xe7Q\xe2\xceF\x1ac\xfc\x09\x89\x0cG\xda\xab\ \xbf\xab\xef~Y\x86\xc1\xbd7g\x0f\xa0\xc4\xadR\xbc\ #\xbb\xb7@\xc9(+\x14}\xbe\xb2Qv\xcc\x99\xc2\ \x184\x04\x0cQ\x18\xfb\xe6\xc9y\xdf\x1c\x97\xa3)h\ \x10H\x02\x8e@\x0f\xb5\xa3%r\xed\xc2\x1a\xf8\xc6|\ \xa9\x05\xca%ES\x8a\xa6\xb3i\xbf\x16xa\xb9d\ \x0cn\x13<\xdaJa\xc8\xbdg!\xb7\x926O#\ \x88\xca\x94)9\x10\xd9\xf5\xb6\xee\xa8C~rQ\x9e\ \xb3\xda7d\x0cc\xba\xe0V\xd9\x0f\xa6\x14\xdc1)\ WC/\x0a\x1e\x05\x07(\xf0\xbd\xab\xdd\xf8\xf6\x92\x83\ \xbbZ\xa32WZ\xc0\xe9\x90s\xd8\xcc?\x0a\xd5g\ \x9b\x0eO#\x00\x9b\xac\x88\xb2\xf6V}\xef\xdb2-\ I\x10\xdf=}\x90tm\xbd\x03\xf5jlB1r\ \xbd\xef\xc88\xc6@\xf54\x85\xdd8)\xffk\xa3\xf0\ \xf6.\xe9\x0a\xa57\x14ds\xa3\xfe\xf5\xb7\xaa\xb6a\ \xf7\x84p\xb2\xa7\x0a8\x1dLUV:W\x9d\xf5\x00\ \xce\xd3`\xcb\x05F\xe3#j\x84\xd1\xae\xac\xb9-\xbc\ g\x11\x87Q_\x12Nq@\xed\xc7\xe5\xfb'\xe4\xbb\ \xa0S\x965\x8f\xad\xee\x90\x8f\xe2\x9d<\x80a\xd9\xca\ \x84\xfc\x84\xbc\x88J$\xf7i \xbd\x00t\x00\x00\x88\ \xf5f\x97d\xe9&\xff\xa0!\x827~\xc9N\x00\xc4\ m\xd2\xc5\x86\xba\xc8\xec\xd2\xac\x1c\xe8\xd8S=z\xb2\ E\xc3\x9b\xe5D`(\xa9YF\xe5\x19\x03\x9c\xad\x06\ A\xb8\xa1\xd4,&\xfe\x91,g$d\xb1wH\x14\ @\x01\x8aBUa\xbd\x89\xd7\xcc\xa32\xba\x14\xd2\x03\ \xb0#\x830\x88\x91\x87f\x17\x0e\xcb\xe9\xf3xS\xee\ \x9d\x81>\xc6\x81K\xa5\xff5\xa5\xe0\xabc\x83@\x0d\ \x88h\xb0a\xc9\xe9\xd7\xf7\x1b\xc3\x97\xbcv`O\x9b\ \x81\xad<\xc6\xd6\x96P\x08\xa6\xb2\xa1\x17\xd09\x8f\x08\ -(\xe9\x022\xc8\xbbm\x8cv\xbe\xea\xa6H\xf5J\ {F8\x81\xb2\x81\xa1!\xf0\xbd|L\xde\xe9\x83\xbc\ h68\xf0\x09}\xac\xd4\x8c\xfc\xc6\x8b\x07\xe8uc\ \xfd3K\xfd1\x92\xecD\x90\xc7\xcd8\x1fs\x18\xf0\ \xb9\x0b\xb3\x8b}\x84\x8bw\x1b\x22\xe0\xede\x8d\xb1\x99\ \xb4\xe9\xe2\xfd\xfa\xf0\xec\x12_\xd0\x85\xd7\x88\xc8\xcc\xa9\ \x04\x16\xc9\xfc\xa5$g$\xa9^B,\xe0\xae\xec9\ 1L\xb7\x94\xaa\x85\x86\xabH\xc9\x1f\x8b\xd9\x12)\x1b\ ^\xfe|\xc1\x10\xcf\x86\xba\xce\xfd\x9d&\x98\x1fg9\ \x9dM\x9f\x84\x1c1\x80\x85)\xbd\xb0\xc2\xff\x83\xa9\x85\ >\xe7\xe5}}\x94\x87\xe2\xac\xe6\x09\xc9\x9a~\x80x\ \xbai}\x7fU\xdd\xdfw\xb4\xca'R\xca:\xa3\xcc\ \xb4\xd8\xa7\xbc|\xd1\x90\x8a\x80\xf3|\x9e\x14Cz\x11\ \xc2\xab\x97Y\xcbo`z;\x9e\xa1\xc7\x89\x13\x02\xee\ \x85\xa9\x1ec\xd63\xde\xb2\xa9=\xbeP).\xd8\x96\ jh\xed\xbc|q\xe3G\xcd\x9d\xa6|\x98\x9e\xbd\xe9\ (\xe0\x0b\x9a\x84\xa20\xfe\xc5\xe1\x81?\xce*a\xd8\ \x81'\x86\xbb\x96ee\xb2\x8fq\x00\x9e\xe6\xac\x12\x1f\ P\xfb\xfd\xfa(^E/\x07P\xf0\xd7i\x90\x0f\xea\ ;\xe7\x94\xf8\x03n\xd8\x98\xd0\xf6\x1c3\xa8\x7f\x08\xcd\ \x19)\xaa\x97`L\x83\x8d\x1a\x9a\xb5\xa0\x1c<\xcd\xab\ \x96\x7f\x04\xcb\x1d\x85y\x8ec\xd7\xde\xc1\xe7V\xffc\ t\xa0@#\x9b\x1a\xf5N\x88\xb3\x81\x1fPu\xd9\xea\ \xa1t[\x0b\xd0WN\xcew\xff\xf7\xe4\xec\x1b'\xe4\ k\xf2\xfd\xc6\x89\xd1\x0e\xf6J\x19s\x96\xa0G\x18\xa6\ \xf9\x83\xd5\xf5Omo\x92gO\xa1\xfeRn|u\ \x11{\xfe\xf3C\x86\x05\xdd\xe9\xa1\x0c\xc8\x01\x9e\xba\xfe\ }s\xd1U\xccl\x03\x91\xc0Z\xf2\xed\x15\x10\xbd\xab\ |\xc1?\xb5\x82S\xe4\xfb9\x13\x09\xdb\xbb\x99\x5c\xbc\ \xf2q\xd3\xef\xb7\xb6\xd6F\x04\x8c\x0e\xc0\x94\xe0\xd1\xbc\ \x8cA\xc4\xf2\xb3\xa9\x85S\x0a\xbdx'Pbu\x02\ \xbe\x8b\x83\xf7\xea'\x8c\x01\xbb@\xf7\xf4?\xef6=\ \xfcas\xd4\x0ei\x108\xab59_y|NA\ I~\xae\x93\x96Z\x80\x02q\xd0[\xf9\xb6\xb5\xea6\ \x1am\x97\x0e\x05\xbc\x0d\xc7~J\xf1\x91i\xbft\x8d\ \xfc\xa2\x9d3Q\xb0'\xbe%;\x91\x9b0\xa2l\xec\ \xd0#\x86\x95\xe5\xd6\xf2|\xaa\x0a\x0c\x95\xde\x0e\xb6\x02\ m\xed]\x12\x02\xe4\xa9\x95\xf1qL7l9\x0d\x8b\ \xdf\xb3\xba\xfe\xd1\x8f\xed\xa7:\xd9\x1fP\x1c\x1d\xec!\ \x0b\xcf/*\xca\x0d\xc8\xd5\x94\x0e\x00\xa5\x5c\xd0\x19\x11\ \xde\xb8\xcdz\xf3Rf\x86pJ\x06%\xc3+\xf8\xa8\ \xe21\xe6>\xef->\x09rA\x93\xb7w\xe9+\x90\ \x0e\xb2\x03\x94\xe34\xbbO\xea6#\x94\x81\x8b\xb8\x0d\ u\x91@\x80\x09,+\xf3\xe6c\x8e\x07p\xb0\x00P\ \xfa\x19EnhU\x1b\x9b \xa69\xa4\x8f\x0e\x8bl\ \xac\x8f\xcc-v\xf9\xdd\xa9\xee\x9e\xa04\x94\x0c\x8c\xe3\ \xc9#\xd9#x\xf5b\xc2\x0dd-\x1a\x0c\xaf,W\ \xf6\xff\xcb\xcc\x1a\xa6\xe0<M\x82\xa8\x8c\x859e\xca\ \xba\xc2?\xae\xdap\x16e\xa6\x84\x03\x8f\xdc_|\x8c\ \x0d\x90\x16|\xb2%\xf8\x7f\xaf\xa8{z;\xc4\x0d\x18\ 4H\x851\xd8\x00}\xf7\xdb\x9f/*\xce\x0d8\xb9\ S\x0bKX`%\xde\xb8\x95\xbc\xf5%bvb\xcf\ \xd1\x05\xca41\xe7\x19W\xf9Lg\xbd\xdf\x82\x8b\xfe\ \xe3c\x1c\xc8\xf6\xc3(;\xb3\xd8\x17\xd2\xf9\xe6\xa6\x88\ d\xbclM\x94\x84\x0d\xb2\xb9!<\xabP\xf3{\xf0\ \x85\x22\x00\xb9O\x8a \x1b<e\xde\x02\x11\x1c.\xaa\ \x97\xc2\xe8\x09%\x93\xa38\x8c<*_\xb1\xb2F\xd0\ l\xf9\xf8\xb4\xd4\x0a\x960H\xdf\xd2\xcf|L7\xc0\ \x06\x06\x17?^}\xf0\xb1\x8f\xda\xd10\xd8\x99K\xea\ \x08>,K}\xe9s\x85\x859\xe9y\xf1<\xea\x13\ \xbc`\xcbv\xf1\xe6\xa5Bow\xb8,CT\xa6\xf9\ \xcc9OyJ\xa6\xf6_\xc6\xc0\xa7\xbf\xf9\x98\xc3\xa0\ P2\xbd\xc8\xd7\x161>l\xd6M\xe4\x8a\xdd\x04h\ K\x94lj\x0c\xcf,te{\xf0\xb1\x9d\xb6\xc1R\ \x09\xf4*\x9e|\x9c\xa7\xa9ZJ\x85\x81Q\xaa\xf4?\ \x18\xd3\xec{\xc9\xf0\x95\xb3\xdc\xd1r\x1d\x93\xed]\xfa\ \x05d\x9b\x84\xda\xf4O\x1fc\xf3\x1d~aT\xf9\xb3\ \xb55\x8f|\xdc\x89\x8c\xc1\x89W\x8c\x1f\xc0\x1c\xe3s\ \xd4\xe7\x16\x14\x16\x04\xfd\xa9n\xd0R\x9d\xb6r\xad\xba\ \xf5d\xd1\xd7M3\x84\xf7\x19\xe19<\x10\x91QW\ \xc0\x98\xf9\xb0\xb7|\x06d\xeb_\xce\xc6\xa6J\xbf\xf5\ 1\xd8j\x11\x0a\xa5\xd3\x8a\xbc\x0d\x9d\xfa\xb6\x16\xd3\x92\ \x8f\xcf\xc4\x8d\x94\xd4F\xf9\xe6\x06}N\x91+\x00\x9e\ &\x95\x00\x0eH\xd9p\xd9WB\xb3\x87\x93\xaa\x85D\ \x98\xf8\xcc)\x94\x17\xba\xd3\xa8\xb2\xffe#k\x84\x9a\ ;\x22\xb1\xf3%)A\xff\xf51]\x90\xd3Y\xd4\xb4\ \xf8\x8f\xd7\xd4>\xfaQ+\x9a\x0b\x92\xa4\xc1\x14B\xc7\ \xe5\xd0\xbf\xce+,\xc9\x0fP\xf9\xf4\x95T\xc2Q,\ \x08\xd3\xbc\xd5z\xfbr\xa2\x87`\x0d\xd4\x0d\x8e\x10=\ \x8e\x16\xe0\xd3\x7f\xe7\x1ez\x0e\x88\xda\xe7\xeboS\x04\ Y\xa3\xfe\x1c\xc7\xd8\xc0v+/\xf7\x9c:\xc8\xb5\xb3\ \xc5\xdc\xddn\xc8\xe76B\x1a\xfe\xd4G\xe9\x07\x0d\x91\ 9\x85\xee\x80G\xc5\xa4\x14B\xca%\xe1\x1dD\xb3\xca\ \xe9\xc1\x95\x02\x1fs/?\x00Kg\x95\xaf\x9b\xc1Q\ Jp\x18\xbe\x80\xbf\x9f\x00\x09\xdf\xdf}L78\xb8\ {\xc3\xfa\xf9\x9a\xdaG\xb6\x870\xd4\x04\x93\x09\xbc\xd5\ \x18x21W\xfd\xdb\xd9E\x83\xb2\xfdN\xd6\xd4\x02\ \xbd HR\xbf\xc9\x5c\xfcUf\x80l\x90\x04b\xe1\ /we\xf3\xa9\xbf\xf3\x0e_ 3f:\x80*P\ \x97~\xefc\x0e\x87\xa6(\xa7\x17\xb9k\xdb\xccm\xad\ :^M\xe9\x80\xd6F\xac\x0d\xf5\xd1\x19\x83\xb5l\x9c\ \xa7\x91\xfe'\x95@\xaf\xc2HV1\xcd\x86\xd1\xd3\x12\ 6\x9e\xf2\x91\x5c\xa6V\x84V\xbdex\xca\x94\x9c\ \x11\x18\xb6w\xa1[\xf4#\x16\x11\xf6\x9eNb\x8a+\ \xe2\x84\x8e\x9f\x1a\x1fc\x03\xaa\xd3\xa9\x1b?\x5cY\xfb\ \xb7\xdd!\x0bBMH\xc2\xd6\x8c\xa7\x17\xa6\xe4k\x7f\ ]P\x94\xedK\xdf\xcdr\x96nA\xdf\xb4\xfcza\ tH\x17\x83b@LC\xdd9b\xfao5\xbcY\ \xce&\x04\xa6snE\x1bv\xb5\xed]EZ\xf7{\ \xbd\x22\xac\x96\x05*\xa6\xba\x07OP\x14\x9c\xd9\x91\xdd\ \x03\x02\x8f\x9bZd\x0cc@\x0e\xe0\xb0=b\x06\xd8\ \xaap\x1a\xa3\xa3IG='T\x93\x1d\x08\xb7E\xf4\ ;V\xd4\xbd\xb2\xbf\xc3\xe4\x18\x08\xc3a\x801\x82X\ \xd3\x07y\x1e\x9e[T\x14\xc0k\x1f\x9dx\x22\x85\xc0\ \xa7\xf3\x12\x8bW.\xe6+o\xa3z+\xac\xc0\x1f\x08\ \x81\x95c.2\xfb\x11\xb5l\x167\xa2\xbcq\x8b\xb9\ \xf39V\xfd\x1a\xd3;@\x1f\xe0y\xa4\x0ad\xe4\xce\ <\xa2\xf4l2\xf2r\xad`\x22q\x05\x18S\xf0\x08\ \xa88Y@J\x90)\x8c\x11\xf8\x5c\x22\xbb\xe6\xa8\x04\ \xc30:\x22Q\xaf\xcb\xe5vw\xbf\x9b\xcf\xd6\x0a\xf8\ \x8d\x13\x8eHem@\xd1a\xc3\xfc\xe9\x9a\xda'\xb6\ w\x00\x87\xba\xc8\x06\x9eF\x9c^\xa0=9op~\ \xc0\x97\xfa1\x8a\xa3j\xc1\xad\xda5b\xd9\xb7\x04^\ \x1aa7\x13\x94\x9ai\x012\xfe[V\xd5B\xd6\xb2\ \x8dXQH\xc2\xd8\x1d\x9a\x07\xee\x05\xf2\xdb\xdfX5\ FU\x917.2\xe9g\xfe\xd2S\x81\x878\x97\x99\ Bg\x931\x8c\x11<dX\xdb\x9b\xa2\xaf\xedl~\ \xbdZ\xdf\x1b2!\x85\x11V\x92\xa5^X\xeeZ0\ 4{|\x9e;\xe8\xd1\x04\xb1\xdf7\xd43\xa0^\xed\ \x91\xe8m\xcb\xeb^\xad\xb4/U\x03\xad\xe2y4P\ \xfb\xa4\x5c\xf6\xc4\xfc\xe2\xf2\x9c,;g\x8a\x01\x82\x81\ \xc6y\xe5\x22\xbe\xfa\x0e\x12i\x94\xad\x04M i\x8d\ WH\x80[<1\x03\xc0f\xf2\xe6?\x95L\xbe\x9d\ \x8e\xb9VQ]\xa9\x9c\xd796c\x0eO\x94,G\ \xbf.\xb0A\xc06h\x13(pB\x5cawA{\ \x1a;~\xf1~\xe3\x92\xeah\xc8\xe0XL\xf7?l\ \xa3$\xa0\xb1i\x05\xea\x8f\xa6\x15\x8f\xcd\x97/\x85\xc6\ ;\x91{(\xde>r[X\xff\xd5\xfa\xda'w\xd8\ W\xc5\xc2\x111 \x06\xfd\xce*r=\xbd\xa0\xdc\xeb\ \xb2\xaf\x80=\xf1\x91\x12\x0c\xa7\xca\xc2\xe25+\xade\ \xdfdF\x04'\xc5\xe4&\x80]\xe5\x133F\xe6\xc6\ \x5c\xf8\xd0\xe1\xe1\x97\xba\xa6\xfd\x922\x06,JM]\ \x8e?V\xb2\xe5\x02A\xf4\x0e\x1e\x0dq\x01:\xb7\xbd\ \x1f\xc6j\xb0\xe9\xc4\xb5\x8a\x1dp\xb8\xc5\xfb\xdb/}\ \xabfK\x93\x01#\x1c\xbb\xda\x87\xfe\xe5\x0f\xa4\xefn\ \xe7\xcf\xedn=%@\xcb\x82\x1e\xd0\x8f\xbd\xcd9\xc4\ \xb1 \xf7\xa4\x1eM\x9dV\xe4\xdd\xdb\xa2\xefl31\ ^\x94\xed\x176\xee\x0bq\xdd\xe0s\xcb\xc1\xcd`6\ g\x9f\x94@\xca\x05`4PA\xf3'\x09\xbc\x9e&\ \xea\xa4uU\xd9\xc9z\x1c\x1c\xca\xc5M\xd2\xb4\x09\xc8\ \xc7\x06O\xc5.7%u9\x1ec \xae\xd0\xf5m\ \x7f\x15[\x1f\xb2\xb6\xff\xd5\xda\xfe\xac\xd8\xfb2\xd4\x8d\ \x87\x1bD\xeeD\x9c\x99O\x9cpk\x0fv\xdc\xb0\xf4\ `+\xbe\x1f\x10\x94!?\xc7\x02\xa4\x02o\x16U\x85\ G\xf8\xe9\xe8|_\xec\x02\xb8\x18\x9dW\xe6\x85Xx\ S\x83\x81\x13\xbf\xe0\x22\xa5\xb2\x03\xd4\xfc\xd2\x98|\xa8\ i\x0a\xb4|\x5c\xf8\xcb[\xc2\x8a\xa7\xf1\x1dg\xb57\ \xa0\xa4\xe5cRz6\xf5\x14\xa0F\x92_\x97#z\ %\xb9\x0c}\x8e0\xda\xf7\xd3\x15\xffI\x9a\xb6I\x7f\ \x02\xbd\x11\xe6\xc1\x05\xf8\xcd\x1a\xac\xccy\x9c\xe6\x9f$\ E\x83\xde\xa1W\x22\xc2q1d\xa3\xfb\xdb\x22\xf3^\ \xaej\x0b\xe3\xa3\xf7m\x1c\xf7\x80\x8e/\x16%~\xd7\ \xf2K\xca\xf1-\xd9\xb1\x15mW0\xa4\xf3\x9b\x97\xd7\ \xbc\x06\xbe\x05V\xe5\x9e7\x8e\x0f\xde=\xbd\x08\x97\x93\ \xaf\xe5c\x02\x04\x13f\xd8|\xf3R\xda\xbc\xd5I\x8a\ \x17X3\xf9\x15\x1c\xaa\x5c\xf06S\xf0e\xbe\x98\x94\ L\x1c\x1d1A\xf1\xd1\xfaMb\xc9WI\xa3\xa4\x0b\ H ?\xf8\x85\xa4!\xb4\xa3\x8e/\xbd\xce\xaaY#\ sK\x91\xe3\x87\xdc\x0d\x0f\xf8\xc4\xd6\xa6\xf6\xb0|*\ \x82\xb4\x1c\x00\x97\x8f\x09\x14\x01\xcb\xaf\x0e\xe9w\xaf\xa9\ 3\xe5}(\xb1\xc0>\xac\xdf\xc5\xfe0\xbb\xe8\x86\xf1\ \xfe,M\xd1\x14\xfa\x1fC\xdd\xdf9%_\xf6\xae\xe9\ \x84\xa8[K\xdbv;+\xbd\x00\xea\x03>\x82\xb4\xef\ \xd7w\xbd\xe4$&\x19G\xfb\x18\xa3\xa3\x85,\xba\x88\ \xb4\xee\x95\xe2H\x1b\x1d\x0e\xc8\x0b\xb1#\x04\x04Z\x16\ \xfb\xfcB%P\xdc\xbb\xd7Sc\xf3\x12\x10\xfb\x91\xa9\ \xff\xd8Q\xd9\x01\xc4D\xdb\xd9\xa4<1`/\xc8\x95\ \xebV^\xb9\xa0|t\x8e;\xb6\x9d\x10\xb8\xa3\xfc\x82\ \xc0\x1a\xbe\xf1^\xcaC\xd3`\xb1\x1d\x22\xd1\xe0\x82G\ V\xffD\xdd\xf18F }\x01T\x83\x0b\xabt\x9e\ k\xee#T\xc1\x13\xf5I\xad\xd1\x11\xf6\xe6\x9c\x8b\x9d\ O\x91\xd6}r\x0d\xacx,\x17\xa2\xa0\xcd\xa8\x19\x82\ \xf8\x06\xa2\x1d'\xb1\x17\xa0\xf4\xbd\x9a\xb6Z|\xfec\ \x1cM\x1ds\x0a\xecb\xb6\xd4u\x00u\xb1s\x8b\x15\ \xb0+\xb0\x84\xb9T\xe6V\x19\x8c\xd1\xd1e\xa5\x8b,\ 6LC\xab|%\xf6\x0a\x1c\x0fh\x0f\xa8O\xeb\x0e\ \x11\xa9w\x92\x92\x89#=D\xb4ET\xbd\x852H\ m\xc2\x97\x93\xde\x0dL\x90\xa9\x90\x05\xdf\xfbp(\xfe\ \x88\x13x\x90\xd7\xb6\xd5\x18\xb2\xe3\x93%\xc5d;\xc8\ \x0d\xdf&\x17\xbb\xdb\xa1/\x8bi\x17\x1b\xb2\x0c\xdc\xd9\ \xf9u\xd6\xe5B\x9a\xa0\x87jy\xb4\x11e\xe8;\xc0\ oF\xea\x88\xde\x8e\x1ar\x92\x92\x85#\x18C\xa3\xf5\ Jg\x1d.\xd8\xeb'Fh\x9f\x89\x22\xf6\x0e\xd8\xb4\ \xb6\xd5\x87\xe2\xf0\x116\xc0\xad\xe0\xce\xa2\xaa\xd3J\x91\ \xa9\xa1\xc0.!e\xd7f\xbb6\xfcs\xbe\xa4H\xf1\ \x02\xf6\xa2\xa1}\x0cg){\xb3\xfb\x91\xc0\xe92E\ \x98\xcd\xb5\xd09\xc8AJ2q\xa4\x8f\xb1\xc2\xc4\xc4\ \xb0\xc2Y=>\xa0e(\x22\x1aiCz\xf5\x1a\x96\ '\x10#9\x0f\x01\x07\xc7\xa0\x13j\x9a\xf0\x03\xfb&\ [?\xa0\x0c\xb0\x07\xf4\xd6\x00\xbc\xcd\x05W\x08\xbe\x0a\ H\x16\x8c\xbf\xbd&\xae\xd1Z\x05u\xc1\xe7\x88\xf4\x0d\ \xd2W\xe24k\xa8\xb9\xdeYN&\x8ed\x0ce\x82\ \xa9\xb1\x94\x88\xed\x8aRU\xf39\xebq\x03\xcb\x18\xc2\ \xc2qW\x0e\x9b\x13\x0a\x98\xe7\x86\x10\xaaW\xad;^\ 8V\x15V\xb8\xad\xe9\xc0\xa6\xbam\xcb\x1b\xb7\xafn\ ?\xb8\x9d\x9bQ\xdcf\xff\xf7\x0a\x093-\xf2Vj\ \x14\x0fh\xaf$\x11G\xf6Jj\x80\xa9\x01Y`\x8f\ \xa5\x0a\x93\x04\xdd\xc1\x02g-n\xe0\xf1\xe7O\xa8\x90\ S<\x18\xea\xc7\xea\xdbQ/D\xa1bd\x16*\x1c\ ?I\x80-\x0c\xca$\xa5\xeb\xac\xdd\xac\xbfq1\xff\ \xf7\xe9\x81e\x17\xe5\xbe\xfb\xb5\xecuWz\x16\x9ek\ \xfesr\xd3\xeb7\xe9\xa1z9\xdc\x94\xf9\xe2\x84\x16\ ,\x95F\xeek\x15\xe4I\x06\xf8c\xde\x9c|{)\ \xa98\x821\xdc3\xd8\x0c\x8e\xb0\xcdyb`\xdf\x90\ 7\x86\xd0\xbe\x5ct-N\x1f\xe4\x86q2\x9aFN\ p;\xc9=\x01\xf2\xb9UetQn\xfc6\x8a\x09\ \xd2\xfe\x08n\x19V\xe5\x9b\xe17\xbf\xe2Z\xf4\x05Z\ \xbf\x81X\x11\xb4\x8eCZN\x8d\xf6`\xfd+\xe4\xe5\ \xd9\xd1\xe5\xb7[\x0d\x1b\xe4\xb9\xf78\x00|\xe7YC\ \xa4\x1e\xfb\xfc8`9x%L\xc9\x1f\xd4\xa7G\x82\ \xc7\x88#\x18\xc3\x5cYF\xc5\x97A\x80\x1em\x01~\ Zk\xff\xd8\xea\xec\xd3pn\x90\xdf;\xa7\xd8\x85\xa7\ \x15eH\x12\x0bP2A\xc6\xe4\xa8'\xe7&\xf1\x0e\ \x01\xd9_\x08}\xeb_\xf8;7h\xf5\xab\x88\x05F\ =Z@X\xc7\xd7\xe9\x98\x11e\xcf\xf3b\xc9\xd5f\ \xe52gC\xcc\xd0\x82\x85B\x0d\xd8\x17\xb1\xf7\x11\xc8\ Vo\x01\xf3f\xdb\xabI\xc5\x91q\x0c\x94;l\x01\ \x1du\x05\xa1\xf2\xc1\xdd\xc7wpx\xed\x1e\x8c\xaa\xde\ \xbe\xc4\xa8Z\xc1\xf1\xb1o\xbd\x98\x98\x01\xbb\xd0+G\ \xe7zU<\x0b\xd1\xa3W\x17\xd8s\xe1u\xbb~\x8d\ \xfctj\xa1\xcf\x95\xac+\xbd\xb1\x183\x1aY\x7f\xaf\ \xb2\xe97\xc4\xc41\xbc<\xc7g\xd3\xe8\x10\x80\xe4\x00\ y\x1a\x1fF\xb6md\xd5\xcd\xd1]/X\x16\x04\xc5\ \xb1\xaa\x82*\x1a\x19v\x19\x9eT\xed+0\xec5\xfd\ c\xb9\xab\x00dL\x8eV\x0e\xe1h\xc6(\xaaFO\ \xbdS\x14\xcd\x05\xdb\x9c\xa0h40|:\x0e\x8aU\ \xb7X8\xe3\x17\x9f\x98\xa8lY\xb3Ye\xfe\xbb'\ \xfaT`B\x0f\x84\xb1{k\xc6\x98\xb8\xe9\xe4\xdc\xd3\ \x8b\x9c\xb7e\xda\x07I,\x80\x06|\xe7?\xd4m\x7f\ \xc6\xae\x07- \x13\xe5\xa6\xc3!m\x83\x80m`1\ \xaa\xb7\x92\xf5?\xe4\x0d\xef\xc5>x\x82\xfdX\xd9\x5c\ \xa1\xb9\x9d\xf5\xde\x03\x8aT\x22E\xe7+\xaaGJ\x9a\ x\x9d\x1c\x8e#\x19#\xcbR]\x01u\xce\x03l\xd2\ wyp\x94\x90\xcfL\x00\x0a#I\x08\xe5\xbe2\x81\ \xee\x052\xcagL\xc2W\xb8\x8e,\xbe\x82\xd7o@\ \x8f$!\x0f\x14\x13l\x8d\x7f\xfd\xd4\x92\xeb'\xe6\xb9\ \xa0\x1dK7\x22\x0b\xc3/\xa7L;I^\x11\xe3b\ \xe2\xda1\xd9\xd7\x9e\x94\xeb\x981\xa1\xc0rd\x91\xdc\ \x88\x88\xedO\xe2bl\xb0\xf7\x83\xecTo\x13\x1f=\ \x8a\xf3\xe6R|\xf8wr\x1c\x1f\xb4`2\xcf\x9b\xd4\ u\x88\x98\x8b\xec\x82\xbd\x17h\xc3\xc8\x9d\x98=>\xc1\ \xcf5:\x1e\xc0\xfcG\x09j'\xa0(<\xdab\x85\ \xaa\xcc\xc6\xedzk\x95;X\xaa\xe4\x8fQ\xfd\x83\x8d\ \x03\xcb\xe9\xba\xff\x82\xa1\x92\xbd'\x88\x8b\x22\xfb\xcb\xc8\ \x9c\xc7\x95\xbc1\xc8\xf0x\xda=\x1c\x03J\xd2M\xfe\ \xf6\xee\xb6\xdb\xd75\xb6D@\xdd8\x96\x95\x87\x85-\ \xf8/_\x89)\xb2T\xe5\x97\xd3\xf2/\x1d\x9d\xe3\x82\ \x91\x92\xdc7\xae\x82z\x84\xecX\xe1\x80\xa2i\xf5\x1f\ \x02\xbb~\x0f\x01)\x9ey\x8a\x09P\x07[\x12!\x14\ \x95\xcfz\xc6S>\x03E\xef\xe9\xc4>\xd6]\x08\xbd\ e\x87\xf2\xda\xf9\xc2\xd21k\x9c'\x98\x9c\xf7\xf8i\ ^\xe5\xdc\x17\x95\xdc\xd1R%\xf1\x1d\xa1\x17\xf8$c\ \x10v\xa2$\x84\x0dH@#\xcaEj|\xfc4}\ \xffg\xd4\xb2/i\xb17\x08\xe2+f3\x1fP\x0a\ \xa7\xc4\xf5\x80I\xd4\x99,\x07\xba\x80}-\xd1'?\ n_s\xa0u_D\xb4\xe8x\x05\x97FI\xc0\xa5\ \x0c\xf1\x8a3\x8a\xb2\xae\x1c\xe3\x1f\x95\x1f\x80|1\x1b\ 2>\xd8U6;\x1b\xc5\xab\xb3H$$/P\x92\ \xf5\x8b\x1d2;/\x9a\xe9:\xfb\xff`wX\xee\x89\ 1\xb23\x16\xb4s\xdb\xbf\xb4\x0dwS\x1e\xc2\x96\x12\ \x178\xb6)>\xe1N\xd7\xc4o\xc9\xdb\xe4\xd2\xcd\x98\ #\x01\xa2`\x22\xfe[\xba\xbe\xe3\x1f\xca\xbbw\xc9:\ \xdb\xb5\x94_\x81\x0a:\xfbQ5o4t_\xd0\xb3\ bJO\x90\x05\xc1\xbfSQXm\xea\xd4\x1b\x22V\ \x9b\xceM\x18/\x0a+\xe0V\x07y\x94<\x9f&\x9b\ \x9f\xfcO\x8eR\xa4$\xbcu\xc7\xd2\xac5\xdf@C\ \xa2\xf5\xe2,H\xaa\xc7\xf2\x16[\xf3_\xf4\xe5\x16\xc1\ JO\x8c\xc1\xbaC.nE\x8dM\x0f\xb2-\xbf\x97\ \x15\x94!\xb5\x93\xe5\xd8\xb0\x88Pp_ %m\x1f\ yC\xf6\x94\x9b\xa9\xe6\xb6\xa5M\x1bcz\x84\x05\xe3\ \x89\xcd\xffK7\xdeO\x84\x01\xb2v\xf9\x09!\xb2\xca\ \xe8\x99\x7fT\x07M\xee\xfb\x9d\xa1 X\x0a\xea\xdf\x0d\ \xdb~\xe6\x96?\x93\x0f~\xed$\xc5\x09\xa9H\xc1U\ ?\x99\xfd\xb0\xbbt\xa64_L\xf2s\x01\xea\x14\xfc\ \xa3'\xc5\xe6\xdf\x0b\xa3Evg\xce\xa6c\x03\xd4\x0d\ e\xb9\xb2\x8c\xb17\xb8O\xfe\x16\x0eV\x92\xe3w\x8f\ \x89^\x96\x04\xbbi\xe3\xaf\xe5\xa7\xdd\x03j\x91t\x81\ j`\xa7E;\xaa\xc8;7\xf1\xa6\x8fd\xae>!\ \x95t\xb1\x01\xa4\xb1\x22M\xceJ/\x00M\x07L'\ L\xc5\x8c\xfb\x9d\x8e\xd0\x95+\xe3\xae\xa6\x0b\xfe&\x8a\ \xcf\x92\xad\xf8D\xfbC_d\xe5Oa\xf3\x9f\xf3N\ \xfc\xb6\xa2\xe2\xd3\xb8\x9c\x0d)A\xaf\xb9I\xa9\xe2r\ \x8f\xfb*?\xe5.\xa2\xb8\xa5\x8b\xc1\xa7\xc0!{:\ \xaa\xade\xd7\x9a\x8d[{;O\x936\xa0\xf0\xf8B\ \xbf^\x02\x0c-\x7f\xe10\xf1\xbd\xd0\x11\x9d\x0a\xe8\x93\ \xa9j\xfe\xc9\xda\xdc\xc7\x8d\xe9\x7f\xa1%\xb3\x0co\x05\ Q\xbc\xf2\xea\x0e$\x10\x0e\x07\x14O\xd4U\xc6\x8bg\ \x19\xd3\x1fp\x9f\xf7w6h<\xec\x82\xd1rj[\ V/{\xa5np3b|\xfcw\xfa\xfe=\xf6m\ \xf1\xa8/IC\x11\xa8 3\xfe\xa0\x0c\x9a\x1c\xeb\xfd\ E\xe9\x06\x9a\x05\xba\xda\x8d\xbf\xa3[\xfe\xe8$\xc5\x09\ \xd8\x1f\xfd\xa2\xe2e\xb3\x1e\xa4e\xf3\xa4\xc7\x89\xa3\xee\ \x87\x19\x02\xba)\xcb\xec\xa8g\xd1\x1a\x1eie\x5c\x97\ \x8f{u3O\xd0\xd0\x06\xbb\xb3K\x08\x12\xc5\xf6h\ i\xd0m\x9f\x19\x83\xaf6 \xfa\xd6\xc7\xe8\xe6\xfb\x88\ \x19\x86c\xd9\x95\x80\xf8\x8d\xf9K\xc5\xec\x87\xb5\xfc\x93\ \xec\xb4\xb4T/v\xa0\x1e`\xac\xbb\xf3\x1ft\xed\xf7\ Qz\xfc\x8fO`[\x93\xdc\x95\xcb\xe7>\xeb\x1d\x0c\ \xb5\x8e\xaf\xcaro\xc7\x16\xb0\x0c{B?/\xf7\xb7\ \x0fbo\xc2d\xf8\xb6#\x81\xde]2\xdbG\xf4\xb5\ H`:(F\x1b{%\x9fp\x87\xed]\xeddt\ 3\xa1*\xba\xe2F\xde\xb8\xc5VF\x7f\x00\xe59\xa7\ \xa19\xd0,q\xcb,\xcd(\x84\xa7@\xcb\x1d\x11\xff\ \xde\x0e?\x91d\xa8QX\xb2\x97\x1d\xd2\xc94\x5c\xb4\ \x13\xd8!U\xa7\x1a}f\x8c\x0d\xd5\xeb>\xf9:\xcb\ \x8ei\x10\xd8`\xb1\xd6m{\xf9\xf2\xeb\xad\xb6]2\ 1\xc3\x81\xed\xda\x95WA\xf2'\xa3\xeb\x8c\xdf\x1e8\ 1E\xa8R~\x8e\xaayl\xda\xc5\x0bP$~w\ )\x15\x17\xba\xd6\xe4\xa2\x5c\x90?v\xb2\xbdW\x8a\x91\ H\xb7\xa6\x8d\xfb\x9a5\xe16P\x1aVH69\xec\ \x80C\x07\xf8\xa2\xab\xac\xfa\xf7\xc1i\xdb~;S\x81\ \x06`\x8afN\xf9\x1f\xaa\xb8\xe3\xb0\x06\xb6\x0e\xf8\x82\ \xaaQ\xe2-f'\xdd\xe4\xa4\x7fJ\x91\x18\xc6 \xed\ )h\xdb\xe3\x9a\xf0-k\xd2]\x84i\xa8?\x0c\xf4\ \xb1\xad\x89\x8e\x03b\xc5w\x8c\xb6\x03\x99<t\x92U\ @x\x0bFY\x15\x97H\xd2\xc7\x08\xa8&\xd4\x8b\x11\ E!\x13nS\x5c>y\x98\x1eN\x11\xf4_$\xd2\ \xc7\x00\x18\xa1\xaeqWZ\x13nG\xae\xd8\x9e\x19\xcf\ /0\xf04t\xf1\x95V\xe3f\x99+\xa3\x01Bk\ \x13n\xe4\xfe\x12g\xbd'HW\xaaP\x18Q\x97,\ P\x87\x9d\x87\xfc\xf9T\xa3\xafc\xa5\xa3\xd0u4\x11\ \xd9\xfc0^_bE\xec\xb6\x0a\xa4\xc1M\x81r6\ \xffY%Xn7i\x993\xf3 \xeb\x10\xedhb\ K\xbeB\x9aw\xe0\xac\x81<\xd9\x0a\x89\x87\x8b\xect\ C2z\xc1\x17\xe3@\xf82\xeb!{\xa6\xfb\xd3\xea\ ]l$\xd8\xc7 \x11\x10\xcc5\xeej~\xca\x7f\xc1\ \xe1e\x08\x09\x7f\xe0\xb7\x05\x0d\xed\xb7\x16^\xce\xeb\xdf\ \x83eg\x87\x0c\x84\xb4\xb7\xcb\x97\xa3\xcc~D\xaf\xf8\ \x8a\x80\xbe\x06\xbd%0\xe3(\x99!\x1f\x9e\x17\xe1\x9a\ \xc7\x98\xf0=e\xfa\xaf\x99\x82\x17y\x01\x9c\xed\x9fR\ $\xd8\xc7t\x03\x0f+Hx\xeb\x13\xea\xc6\x9fR\x8b\ \x0b&\xa8\xc5\x04\x937p\x04\x87\xa8\xe7\xbeF=A\ \xd9\x18\xed\x96\x9aA@u\xc8\xb98\x8e\x0f\xb2\x11\x91\ \xfd\xcb\xd4\x0f\xee\x12\xedu\x84\xe8r\xa3M\x1f\x90\x99\ \x09\xc5C\x0a&\xd1)\xbf\xd0r\x87\xc3\x96\xb8N\xda\ \xf7_$\x8d1\xf6\xc7\x8a\xea[\x1eQ6\xdf\x8f7\ \xe8\xe29\x17\xdc\x02v`\xc1a\xe4\x8c\xdf*\x83a\ \x1c\x9bq\x8d\xb2\x9b\xc2\xdd\x9a\x11\xa6\x1e\xa9\xdf\xaa\xb6\ o%\x91&+\x1a\xc2y3\xa0\xbbg\xb0\xc8;\xcd\ \x95S.\xa7\xea\x11\x9fz\xefb#Y\x8c\xb1\x01\x07\ \x87\xc3G>~Vy\xef.\x86\xef\x9bGk\xe0\x07\ \xec\xe2/\xa7\xe7\xbe\xa0\xfa\x06e\xb8\xa2\xbb\xf4\xe3\x90\ \x1dO\xbc\xca\x15J\xec\xcb\xc5\xc1\xafd\x1c\xe9\x93\x8a\ \xa4;RP\xa6g\xd4\x97\xc4I7c\x88\x08n\x06\ \x13\xa4\x0dB\x07\xc4[\x17\xf3\xc6\xcd\xf6\xfd\x86\xc9\xe4\ mB\x80\x9c\x90\x93\xf2\xb0\x00U\x80o<\x0bh\xa7\ \x7f\xa6\x90\x5c\x1f\xd3\x0d\xe8\x94\xc2[\x9et}p\x0f\ \x07\xb5c\x81\x10\xd9\x00q8\xcb*'\xe7\xbf\xac\xb8\ \xf3\xf1\x0c\xc3gN\xf9\xfd\x12)\x0a\xd6\xd0\xd3\x8c\xbf\ \xc2\x9cp\x07z\x18\x8a!%\xe5D\xe1\xaa\x08U\x91\ W>'j\xd7\xdbn\x7f\x00\x99\x8f\x14\xf9\x18\xa7\xdb\ \xe1fd\xd3#\xda\xe6{q\xa0A\xe4{p\x81:\ \x10 \x07\x87\xb3s\xfe\xa5\xf8\xf2\xa4\x93\xc7\x9c\xd2\xed\ \xcb\x0e\x0cW\x06\xbcO\x06!e>\x06\xa7\xcd\xf1<\ \xc2\x84\xeb\xf8I\xb7\xc0P\xdbN\xc5\xb1*\xd0\xa1m\ \xb7x\xeb\x8bV\xdd{\xc8+\x18\xa6\x22\xb9\xe0_8\ \x97N\xc3O\xbf\xba2\xeb\xd3\x8d\x14\xf9\x98n\x08\x08\ t-3\xb2\xe5Qu\xf3\xafl\xff\x82\x14\xc1`X\ \xf0\xac\xa1\xe2\xec\xe7,C\xef\xdc\xf6|6\xdf\x01a\ f\xa7\x7f\xb2o\xf4\x97To\x10\xf2}Ff;2\ \x1f)g\x8c,NpC\xdf\xf8\x07u\xf3\x03x9\ \x16N\xa7\x02c\x80\x10\x9c{\xf2\x99\x1e\x22B\x87u\ \x00\xf6F\xaa\xcf,=_\x9b\xf0m%g\x14\xee\x0d\ \xcc\x19\xe8\xa3\xd2\x8aT3\xc6\x06\x14\x8a\xf76oy\ \x5c\xddr\xaf|\xa3\x19\x92\x06\xd3\x91%G\x12\xc2\xde\ \xea\x19\xc4\xa7\xffF+\x9f\x03\xdb\x07\x18\x93^\xa4\xcd\ \xd53\xd5\xed\x9d\xf4m>\xee&\xc1\x14\xa0\x00\xd2V\ 2\xc1qB]<\xb6\xe7\xe4E\xa4\x9e\xae\xfd\x9e\xd9\ \xf0\xa1\x9dq\x00iD:\x83\x03<a9\xf1F}\ \xec\xad\x87X\x00\xa1\xcd\x91.\x04\x07M\x18\xc5P\x1a\ \xae\xb36?\xe8\xd0j\x00\xe9Cz\x18\x03\x04\x90\xc0\ \x97\x86\xb3\x5c\x08Pl\x22@\x82C\x08\xdc\xd6\x05\x99\ \x11\x1d\x8dR\xb3Po>\xe0\xa4\x0e MH\xf3\x00\ \x84sn}\xf4\x04,\x80\x1b\xb1S\x8e\x0bpDV\ \x84\xec\x7f\xdeY\x1d@\x9a\x90f\xc6P+\xe2\x89\xec\ \xc4\x85\x9e\x02\x14\xdb\xfd\xf0\xd6=\xf6\xea\x00\xd2\x854\ 3F\xce\xce\xf5\xe4]$p\xcc\x84#\xbb>?5\ n\x00}C\xba}\x8c\xea\x11Z\x0e,\xf4\xdc+!\ \x04\xf3\x14:\x8b\x03H\x13\xd2\xcd\x18\xa6\xd2\xe1\x97\xc9\ \xc5\x9e\x86\xcd8`b\xa2\xec\x5cgu\x00iB\x9a\ \x19\x03\xa0#/\x13Z\x10\x16p\x0a\xe6X\xd3\x89\x02\ \x81\x81\x8c\x953\xde]<\xd9I\x1d@\x9a\x90~\xc6\ (\x9e<6\xf6Z\xbcC\xc5\x8eU\x8e\x82\xa4\x11\xf6\ XJ\x16\x9dp+K\xf9\x0bd\x07p\x14\xd2\xce\x18\ \xc9\x86\xf1\xd7\xf1\x92Y\x84\xa9\x9f\x0c\x82\xd1\xbdP\xca\ U\x8f1\xe6\x1bj\xd9\xbc\x98\xa2\x9d\x01$\x13\xe99\ \xaf\xd4\x8d\xae\xd2\x85\x15m\xb7>|\x8c~\xf4 \xb1\ t9\x92vNk\xc3\x9f\xe5+&\x13\xbf\xa7\x0d\xff\ \x02S5 \xcc\xe1\x93{\x03H=\xd2\xcc\x98n\xc8\ a6\x89\xb6\x1f\xd4\xb6=D\xaa\x96\x08n\x80XT\ \x0b\x84\xcb/\xf3\x8d\xfb\x0as\x07\xd9\xc0\xd5\x0e\x99\x81\ \x0cb\x0c\x08#\x978\xd0\x85p\x13\x5c\x89\xa0.|\ ;\x19U\xc0\xdb\xa4\xe5Y)\x03\xf8$2\x861\xf8\ q\xce\x1f\xd9\xeb\x0ed\x12\xfe\x0e\xf4E\x99\x81La\ \xcc\x00\xfa\x0b\x06\x5c\xfd\x00\xe2\xc3\x00c\x06\x10\x1f\x06\ \x183\x80\xf80\xc0\x98\x01\xc4\x87\x01\xc6\x0c \x1e\x10\ \xf2\xff\x01\xbb-E\x13\xd9\xd5\xbc\x12\x00\x00\x00\x00I\ END\xaeB`\x82\ \x00\x00\x01\xb7\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M6.1\ ,10L4,18V8H21A2,\ 2 0 0,0 19,6H12L\ 10,4H4A2,2 0 0,0\ 2,6V18A2,2 0 0,\ 0 4,20H19C19.9,2\ 0 20.7,19.4 20.9\ ,18.5L23.2,10H6.\ 1M19,18H6L7.6,12\ H20.6L19,18Z\x22 />\ </svg>\ \x00\x00\x04\xa1\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22>\x0d\x0a<svg xm\ lns=\x22http://www.\ w3.org/2000/svg\x22\ \x0d\x0a version=\x22\ 1.1\x22 id=\x22mdi-set\ tings\x22 width=\x2224\ \x22 height=\x2224\x22 vi\ ewBox=\x220 0 24 24\ \x22><path d=\x22M12,1\ 5.5A3.5,3.5 0 0,\ 1 8.5,12A3.5,3.5\ 0 0,1 12,8.5A3.\ 5,3.5 0 0,1 15.5\ ,12A3.5,3.5 0 0,\ 1 12,15.5M19.43,\ 12.97C19.47,12.6\ 5 19.5,12.33 19.\ 5,12C19.5,11.67 \ 19.47,11.34 19.4\ 3,11L21.54,9.37C\ 21.73,9.22 21.78\ ,8.95 21.66,8.73\ L19.66,5.27C19.5\ 4,5.05 19.27,4.9\ 6 19.05,5.05L16.\ 56,6.05C16.04,5.\ 66 15.5,5.32 14.\ 87,5.07L14.5,2.4\ 2C14.46,2.18 14.\ 25,2 14,2H10C9.7\ 5,2 9.54,2.18 9.\ 5,2.42L9.13,5.07\ C8.5,5.32 7.96,5\ .66 7.44,6.05L4.\ 95,5.05C4.73,4.9\ 6 4.46,5.05 4.34\ ,5.27L2.34,8.73C\ 2.21,8.95 2.27,9\ .22 2.46,9.37L4.\ 57,11C4.53,11.34\ 4.5,11.67 4.5,1\ 2C4.5,12.33 4.53\ ,12.65 4.57,12.9\ 7L2.46,14.63C2.2\ 7,14.78 2.21,15.\ 05 2.34,15.27L4.\ 34,18.73C4.46,18\ .95 4.73,19.03 4\ .95,18.95L7.44,1\ 7.94C7.96,18.34 \ 8.5,18.68 9.13,1\ 8.93L9.5,21.58C9\ .54,21.82 9.75,2\ 2 10,22H14C14.25\ ,22 14.46,21.82 \ 14.5,21.58L14.87\ ,18.93C15.5,18.6\ 7 16.04,18.34 16\ .56,17.94L19.05,\ 18.95C19.27,19.0\ 3 19.54,18.95 19\ .66,18.73L21.66,\ 15.27C21.78,15.0\ 5 21.73,14.78 21\ .54,14.63L19.43,\ 12.97Z\x22 /></svg>\ \ \x00\x00\x01\xb6\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M2 1\ 2H4V17H20V12H22V\ 17C22 18.11 21.1\ 1 19 20 19H4C2.9\ 19 2 18.11 2 17\ V12M12 2L6.46 7.\ 46L7.88 8.88L11 \ 5.75V15H13V5.75L\ 16.13 8.88L17.55\ 7.45L12 2Z\x22 /><\ /svg>\ \x00\x00\x01!\ \x89\ PNG\x0d\x0a\x1a\x0a\x00\x00\x00\x0dIHDR\x00\ \x00\x00\x18\x00\x00\x00\x18\x08\x06\x00\x00\x00\xe0w=\xf8\ \x00\x00\x00\x09pHYs\x00\x00\x0e\xc3\x00\x00\x0e\xc3\ \x01\xc7o\xa8d\x00\x00\x00\x19tEXtSof\ tware\x00www.inksca\ pe.org\x9b\xee<\x1a\x00\x00\x00\xaeID\ ATH\x89\xed\x94\xcd\x09\xc2@\x14\x84\xbf\xb7XF\ \xacDm@\x0d\x12\xc4\x9f\x0a\xd2S:P<h\xc8\ \xc1\x02\x226\x92\xd4\x91\xe7!\x046\xba k<\xe9\ \xcei\xf7cv\x86\xbd\x0c\x04\xbd\x91\xd8\x97\xe5z;\ 7\x0d\x99\x22\xd1gaZ\xabJZ\x5c\x8e\xd7\x8e\x19\ \xdb0$\x1c@\x91\x08!\xb3\xd9\xe8\xc5\x00\x98F&\ y~\xb8\xfb\x84\xc7\xc9f\x0a\xa6\x04\xc667.\xb3\ o8@q>\xdd\x5c\xdcY\xf0M\x85\x82P\x10\x0a\ ~\xb6\xa0\x1d.?-V\xfb\x99\x8b\xf7\xd6T\xd0\xba\ ]TS\xc6\xc9\xce\xb3B\xbbCe\xd3\xde\x0fT%\ }6x\xaa\x12L:\xe0\xfd?\xea\x01\x9e\x91)\xfb\ ]\xf2X\xf2\x00\x00\x00\x00IEND\xaeB`\x82\ \ \x00\x00\x04]\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M4,2\ H6V4C6,5.44 6.68\ ,6.61 7.88,7.78C\ 8.74,8.61 9.89,9\ .41 11.09,10.2L9\ .26,11.39C8.27,1\ 0.72 7.31,10 6.5\ ,9.21C5.07,7.82 \ 4,6.1 4,4V2M18,2\ H20V4C20,6.1 18.\ 93,7.82 17.5,9.2\ 1C16.09,10.59 14\ .29,11.73 12.54,\ 12.84C10.79,13.9\ 6 9.09,15.05 7.8\ 8,16.22C6.68,17.\ 39 6,18.56 6,20V\ 22H4V20C4,17.9 5\ .07,16.18 6.5,14\ .79C7.91,13.41 9\ .71,12.27 11.46,\ 11.16C13.21,10.0\ 4 14.91,8.95 16.\ 12,7.78C17.32,6.\ 61 18,5.44 18,4V\ 2M14.74,12.61C15\ .73,13.28 16.69,\ 14 17.5,14.79C18\ .93,16.18 20,17.\ 9 20,20V22H18V20\ C18,18.56 17.32,\ 17.39 16.12,16.2\ 2C15.26,15.39 14\ .11,14.59 12.91,\ 13.8L14.74,12.61\ M7,3H17V4L16.94,\ 4.5H7.06L7,4V3M7\ .68,6H16.32C16.0\ 8,6.34 15.8,6.69\ 15.42,7.06L14.9\ 1,7.5H9.07L8.58,\ 7.06C8.2,6.69 7.\ 92,6.34 7.68,6M9\ .09,16.5H14.93L1\ 5.42,16.94C15.8,\ 17.31 16.08,17.6\ 6 16.32,18H7.68C\ 7.92,17.66 8.2,1\ 7.31 8.58,16.94L\ 9.09,16.5M7.06,1\ 9.5H16.94L17,20V\ 21H7V20L7.06,19.\ 5Z\x22 /></svg>\ \x00\x00\x02\x04\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22>\x0d\x0a<svg xm\ lns=\x22http://www.\ w3.org/2000/svg\x22\ \x0d\x0a version=\x22\ 1.1\x22 id=\x22mdi-goo\ gle-spreadsheet\x22\ width=\x2224\x22 heig\ ht=\x2224\x22 viewBox=\ \x220 0 24 24\x22><pat\ h d=\x22M19,11V9H11\ V5H9V9H5V11H9V19\ H11V11H19M19,3C1\ 9.5,3 20,3.2 20.\ 39,3.61C20.8,4 2\ 1,4.5 21,5V19C21\ ,19.5 20.8,20 20\ .39,20.39C20,20.\ 8 19.5,21 19,21H\ 5C4.5,21 4,20.8 \ 3.61,20.39C3.2,2\ 0 3,19.5 3,19V5C\ 3,4.5 3.2,4 3.61\ ,3.61C4,3.2 4.5,\ 3 5,3H19Z\x22 /></s\ vg>\ \x00\x003\xf2\ \x00\ \x02IFx\x9c\xed}\xf7\x9b\x1c\xe5\x95.\xbb\x7f\xc0\ \xfd\xe1*\x00\x1ai\x04\x02\x05r\xc6\x16\x06\x13l\xb0\ \xc1`\x0cw\x17\x1b\xd6k\x1bgl/\x18{\x1d\xf0\ \xda`s\xb9\xbb\x06/\x98\xe5\xc1\xf8A\x1a\xc4\xe4\x91\ F\x9a\x19\xe54\xca9\xc7\xd1L\xf7\xe4\x9cCO\x87\ \xca\xdd\xe7~\xe7\x0bU\xd5-\x8d4\x9d\xa6\xab[\xf5\ \xea)\xf5\xc4\x9e\xaa\xafN\x9d\xef=\xf9\x8a+\xfe\x81\ \xfc\xab\xdds\x05\xf9\xff\x9a+n\x7f\xfc\x1f\xaf\x98y\ \xc5\x15W,\x22G-9\xde\xfa\x07\xf6u\x84\xe7\xde\ \x7f\xbcb\xef\xff\xba\x82\x1e\x02\xe0\xc2\x85\x0b\x17.\x5c\ \xb8p\xe1\xc2\x85\x0b\x17.\x5c\xb8p\xe1\xc2\xc5\x04\x88\ \x90\x7f\x10\xc1\x83}L\xffEl\x87\xf9\xfdH\xa6O\ \xd5\x85\x8b\xa4\x10\xb6\xc9\xb5\x11\x8e\x80\xac\x87!\xa0\x92\ C\x8b\x80J>\xd6\xc3a\xfa\xbdp$\x9c\xe9Su\ \xe1\x22)\xa0\xac\x87\x89<\xf7\xf8$\xf8\x8fC:\xdc\ W\xa5\xc3=+5r\xa8\xf0\xc8j\x0d\xfe~|\x1c\ \xc6$\x9d\xfe\x8c\x0b\x17\xd9\x06\xa1\xc7Q\xce}\xb2\x0e\ \xef\x1eW`^\xa9\x0a\xb3\x8b\xc8Q\xac\xc2\x9cb\x8d\ \xbe\xd2\xcf\x0bU\xb8\xa7R\x82J\x8f\x04\xaa\xa6\x93\xdf\ cz\xde\x85\x8bl\x00\xd3\xe5\x11\x18W4\xf8\x97Z\ \x05\xae)ar\x8d2\x1e{\xa0\xac\xe3\xf7\xe6\x97)\ \xf0\xe7\xa3A\xd0t\xc3\x95u\x17Y\x83\x08\xe7-\xef\ \x9e\xd0 \xbfH\xe3\xfa\xfb|9\xa7\x07\xf9~\x1e\x91\ w|]Ht\xff\xe66\x05\x22aW\xd6]d\x07\ P\xd6\x87B\x1a\xdcX\xae0}\x8e\xb2^4\x81\xac\ \x17k\xb6\x9fQ\xe1\xcbke\xd0\x0d#\xd3\x97\xe0\xc2\ \xc5\xa4\x80\xb2\xfe\xd1i\xce\xc7'\xe0.\x13\xf1\x99|\ \xf2z\xb8+\x94\xe9Kp\xe1bR\xd0\x8c0<\x5c\ c\xe9\xf4I\xcb:\x7f6\xde<(e\xfa\x12\x5c\xb8\ \x98\x14\xfc\x8a\x01wT\x0a\xbd>99\xb7s\x99\x9f\ \xee\x0cf\xfa\x12\x5c\xb8\x98\x14\x82j\x18n_\xa9\xc5\ -\xebs\xb8\x1d\xfb\xf2.\x97\xc3\xb8\xc8\x0e\x18\xc4\xb6\ \xfc\xa7u\xfe\x849\xcc_O\xca\x99\xbe\x04\x17.&\ \x05\xb4M74\x87\xe2\xe70\xc5*\x5c_\xaaB\xc7\ \x88+\xeb.\x9c\x0b\x11+\xc5\x1c\xae\xa0j\xc0\xfbG\ \x03\xa6\x9e\x9e\xcd\xe58\x8a\x97\xc7\xf2\xf4b\xe6c\xff\ \xf1v\x89\xc6N]\xb8p*\xcc\xbc\x80p\x18\xde>\ *\xf1X\xa9f\xc9\xba\xcd\xffh\x97u\xfa\x0c\xf0\xd8\ \xe9]\xc4\x9e=;\xa8\xbay`.\x1c\x0d\x94sY\ 3\xe0o'\x82D\x86\x15\xd3\xced\xf2~\xbe.\x8f\ \xb5G\xefX\xa1\xc1\x8e\x0e\x85\xca\xb9\xab\xd7]8\x11\ B\x9f\xabF\x18\xde9.C~\xa1n\xc6@Q\xc6\ \xf3\xf9\xc7\xb7\xad\x90\xe1\xba\x12\xc1U,\x9bua\xb9\ \x06\xaf\xee\x96\xc0;\xaa\x9b\xf9b\xae\xac\xbbp\x220\ \xc7K\xd5\xc2D\x9f\x87\xb8\xdfE\xe4\xbf0Y\xcf#\ \x1f\x7fn\x95\x0c\xbd>\x09\xda\x87\x02\xf0\xde\xde\x01x\ i\xb7\x02/\xedUa\xe9\xc91\x18\x0b1\xce\x12v\ s`\x5c8\x1cXo\xf1\xdf\xc7\x14\xc6\xbb\xb9\xac\x0b\ \xfe\x8d\xf2\xfe@\xb5\x06G\xfaY~:\xca3=\xb8\ \xfe\xc6\x8f#aV\xa7\x11vs\x1b]8\x10\x11\xce\ 30\x17\x00\xf59\xca\xf7\x5c\x94q\x9e\xcf(\xf2\x1a\ ?_\xad\xc2PP7k\x93\xc2\xbc\xfe\xc8\xaa\xbf\x03\ \x9e\xbf\xeb\xca\xb9\x0bg\x02eV\xd6\x0d\xf8\x9f\x13\x12\ \xe4\x17[9^(\xdf\xc8\xd7\xa9Oe\xa5\x06\x87\xfa\ \xdc\x5ct\x17\xd9\x0d\xe4#\x9f\x9c\x0eR}\x9e_\xa4\ \x10\xf9\xd6\x98>'r?\xabX\x81\xcfTi0(\ k\xccw\xe8\xca\xba\x8b,\x83\xe0\x1d\x0a\xe1\xe7\xef\x9d\ \x90\xad\xf8\xbf\xcd\xa7\x88\xc7b\xc2[N\x0f\x11}\xce\ k\xa6sI\xd6\xed\xfd\x0f\xc2\x91\x0b\xf4D0\xfdH\ \x11w?\xcbb\xd0\xfbG\xe4\xb7\xb4>\x04\xd7\x94(\ L\x8f\x17b\xde\xaeb\xfa\x18\x17W\xa90\x10\xd4r\ \xf6>G\xa2\xec\x0e\xb6\x1ea\x9b\x1d\xa2\x91W\xdd&\ \xef.\xb2\x0bBwa\xaf\x8b\x0fN\xca\xb4\x96\x82\xd6\ \xd3\xa1\x8c\x0b\xeeB\xe4\xfc\x81\x1a\x1d\x8e\xf6\xeb9t\ \x9f\xadk\xb0\xebmE\x0b\xc3\xf1\x9e\x10\x14{Tx\ \xff\x94\x0ao\x1d\x96\xe0?\x8f(\xf0\xe1\x19\x05*\x1b\ \x15h\x19\x96\xa9\xcd\x1eq\xf5{\xd6\x81\xe9\xf3\x08|\ r&\x04s1\x16\xc4yy\xfe\xa7V\x5c\xf4\xbe\x1a\ \x15|R\xd8\xf4\xa9\xe4\xc6\xdd\xb5\xe44\xc2\xfd\xa5;\ \xda$r\xad\x1a\x5c\x8f\xd7_\xcc\xe2b\xac\x1f\x82F\ ?\x9eK\x8e\x85e\x0a<\xb7E\x81\xa6\x11\xd5\xf4\xab\ \xba\xc8\x0eH\x9a\x01K\xcf\xca\xdc\x7f\xae\x98\xb5\xa3\x22\ Vt\xef*\x05\x8e\x0c\x189\xe9\x1fg\xfd\x9b\xc2\xd0\ \xe63\xe07\xbb\x024\xc7\xc7\x8c\x1d\x5c(\x7f\x93|\ =\x9f\xd7\x8a\xcf/'\xba\xfed\x10Fe\xb7v\xd6\ \xc9\xb0\xf3\xd2\xbf\x9f\x96\xe0\xfa\x12\xd5\xecm1\x9b\xdf\ O\xbc\xe7\xf7\x11~\xde\xee\xcb%\xdeb\x81\x99\xd5\x11\ \x18\x0e\xc8\xf0\xc5\xb5Dg\x17\xf1=M\xc4\xcb.\x90\ \xdfc\xaf\x11\xa71\x07\xa2\x1f~\xb0S\x05\xdd\xc6i\ \x5c8\x0b(\xbb\xa8\xcf\xffvJ6\xf3Y\x84.\x17\ z\xfd\xfej\x0dN\x0ej\x99>\xd5\xb4\x01m\xcdV\ \xa2\xcf?S\xa5L\xbe\xae*F\xeeq\xed\xf2\xc8\xf3\ \xf1\x83\x1d*\x8c)n\x8f''\xc2\x88\x10\xdeB\xf4\ \xf9\xbc\x12\x9b\x0d\xcauY\x1e\xf9|q\x95\x06\x1d>\ \x95\xee\xef\xb9\x0a|\xde\x7f\xb9Wf:<\x01Y\xa7\ }o\x8at*\xef\xd7\x94\xaaP\xd1\xa0\xba\xdc\xdd!\ \x10\x9c\x05\xf3r?\x22r>\xb7\xd8\xda\xabgS\x19\ g\x9f\x7fv\x95\x0a\xc7\x06\xf4\x9c\xdc\x93\xed\xbe\x13\xcf\ `\x08\xe6\x95^$\x17\xf9\x92z\xdd\xcaW\xc65\xbc\ \xa7R\x85\x80l\xcf\xe5\xcc\xad\xb5\xcb&\x88{\xb0\xe4\ \xb4L\xeb\xe0\xcc\xb8\xbf\xc9O\xb1\xef\xa8\x06\x8dcz\ \xce\xde'!\x83\xaan\xc0W7\xb2\xe7\x9b\xc6\x0f\x12\ \xd1\xeb\x17\xa8\xa3\xfd\xafc\xaa+\xeb\x19\x84\xd0c!\ -\x0c\x1f\x9f\x91\x99\xddi\xca\xb8\xa5\xdb\x1f\xa8\xd1h\ <\x94\xe6&\xe6\xe8}\x12\xfe\xd5\xc6\x01\x09f\x97\xa8\ \x9c\xb3%\xa6\xd7\xa3d\xbdP\xa3\xdc\xef\xe1\x1a\x09$\ U\xa3\x7f#\x15kh\xfaD\xed\xf1[\xb3\xc7\xbd\xb8\ \xb7\x96\x9dm\xe6\xdaE g\xef\xe1\xc5 r\xc7\x97\ \x9cQ`~\x99\xd0\xe5\x8aM\xa7+\xf0Y\xa2\xcf\xcf\ \x0c\xa99\x1f\xff\x8ep\x7fx\x85W1kJ\xe2\xea\ \xf9q\x09\xbd~K\x85\x0cg\xfbC\xa9Y\xc3\x98\xdc\ \x04\x9a\x93a\xcb\x8f>/\xc6\x1b\xf5\x1a\xbe,\xeb\x1c\ 1\x1e\xfai\x9dr\x1e?g\x87\x0e\x9f\xab\xd6\xe0X\ \xbfF\xd72\xd7!\xfc\xe9\x7f9\x91\xbc.?\xbfv\ \x5c\x83\x05e\x1a\xd4v\x1af^s\xb2\xe7j\xcep\ \x88\xb0\xb9\x0d\xdd>\x15\x8e\x12[j[\x87\x0a\x1b\xdb\ tX\xdf\xa6Bm\xbb\x0a\x87\xfbuh\x1b\x91h\xed\ \x98\x90\xf3\x5c\xd5W\xb1\xb0?\xf3\xc8[\xae+e\xfe\ sa\x7f\x8a\xfc\xdc{\x89\x1dZ7\xac\xd3\xfb\x9f\x8b\ \xb1\xa2X\xe0\x9a\xe8a\x03\xde8$\xa7X\xd6\xd9+\ \xae\xf3\xdaV-)\xbe\x1e\x9bk\x86\xb6\xc5\x86\xc6 \ <_\xab\xc1\xfd+C\xf4y\xca/\xb1\xea\xd9\xf3\x8b\ q\xbf\xd6`\xf1\xf2\x00\xbcP\xab\xc0jo\x80>\x17\ \xb1u\x04\xb9\x8a0\xcfW,:'\xd3\xbd\xda\x8c\x11\ \x15Z\xb6\xe8\xbd+58;ty\xc5\xfc\x84^\x7f\ \xf3Hje}\x0e\xef\x17\x826\xff\x866-)\x9b\ \xc7\xe0z\x19sn\xce\x0ci\xf0/[esn\xc3\ \xe4\x9e;\x05\x9eY'\xc1Y\xa2\xff\xf1=\x8c\x1c\xe6\ \xa4\x88p\xc4\x80\xbf\x9dV\xe0\xdabK\xb6\xed\xbd\xa3\ \xd1\xafx\x84\xf0\x16#\x8d\xfe`'\xae\xaf\xd8\xeb>\ <\x93J9\xb7x\xe1\xa22\x19\xf6t\xc9I\xe9u\ \x94u\xec\xd9\xbd\xf4\xac\x067\x95\xb3xG<\xbd\xd5\ D\x8d\xe4\x22\xf2\xbb\x1f\x9c\x08\x82\xa2\xe9Y\xb7g\xc7\ \xeem\xf6\xdeD\xd6\xe7a\xaa\xcf\x8b\xeby\xbe\xa2\xcd\ \x0e\x15\xb5\xfd\xf7W\xab\xe0\x19fy\xb9F\x82\xbb\x9b\ \xbd\x1f\x80\xbdo\x8c\x1c\x18\x82\xe0P\x03\xf8{\x0eC\ \xb0g?\x8cw\x1f\x82\xd0\xd09P\x02\x83\xe4\xfbF\ \xc6m_\xc1}\xb7y\xc7,.W\x9c\x02\xee\xces\ \xfc\xef\xa9\x94\xa1w$\x18\xf75\xda\xedJ\xac\x07\xfb\ \xef\xe3\xb2M7\xa9\xe6\xbe1)\xbd^,\x93\xe7C\ \xe3u6\x1a\xfcy\xff\x08H\xaa\x11\xc5i\x9c\x0e\xbb\ <G\xd5\x0d\x84m\xf6\x08\xf9\xb8\xb0^\xa1\xbcqN\ \x94o\x91\xddW\x9c\xd7u\xb8WO:\xbe'l}\ <\x17\x94\x9d\xd0h\x1b\xf4\x1f|\x03\xbc\x95_\x02O\ \xf1\x22h(\xb8\x0a\x1a\x96N'\xafWCc\xf1\x02\ \xf0\xae\xfc\x22t\xef\xfd\x03\xf8\x06=\xac\x06;\x83\xb2\ \x8e\xc7PP\x83E+8\xafK\x91\xac\xe3\xf1\x8d-\ \x0a\xdd+\xe3\x95\xa90}\x06\xc3tm*\xcf\x05`\ Q\xa9\x92\x9ag\x90\xe7\xf7|x\x82\xe5!;\xdd\xef\ /|\xa6t=\xc8:\x0eK\x06\x9c\x1eTaeC\ \x08\xde?4\x06+\xeb%83\xa0B\x7f\x88\xac\x93\ \x87\xf1\x16\xb3\x06\xba\x90\xe53\xcd\xa1u\x16\x1a4P\ }\x9e\x1a\x1f\x01\xca\xba\xae\xc90\xe2YK\xe4y>\ x\x96\xcc\x80\xfa\x8f\xa7C\xfd\x12\x22\xe3\xf81\xca\xfa\ \x12\xfc|\x1a\xfd\x1e~\xadq\xd950r\xae\x1c\xc2\ \xbaj\xf3\x0fO\xdd\xda\xdb}s\xef\xed\x1f\xa6r\x9e\ \xa8\x7f}6\xe7.B\x9f\x5cC\xe4so\x87d\xca\ m\xdcz\x9d\xc8\xb9gD\xa6y\x1b4\x0fo\xa2\x99\ <q=\x83\x0a=\x16\x94+\xd00\xe2\xec\x1c>\xb4\ \xa0\x85\x8cc\x0f\xc5wO\xe9\xb4\xc7\xff\xdc\x12[\xbe\ \x16\xb9&\xe4+\x8bW)\xd4>\xb7\xaf\xbf\xc8\xbd\xfe\ \xcc*\x0d\x8e\x0f\x18)\xab/`\xf7F'\xba\xfcO\ \xe0)\x98E\xe4y\x06\x95k\xaa\xcb\x972Y\xf7\x90\ \xcf\x85\x8c{\xcc\xef\xcd\xa0\xba\xbeg\xefk\xa0J\xbe\ \x94\xc5\x5c\xe29o\xe1\x97\xc0\xf5|x5\xf3O%\ $\xebE\xf6\xfc_\x96\xef\x88\xfd\x8b-N\x19\xdfy\ \x8d\xcb:\xfc\x9fM\xb2\xad\xde1yY7\xdf\x87\x1c\ O\xadq\xf6\x5c6\xa1\x83\xc6d\x0d\xbe\xb3U\x99\xf4\ \x1aP{\xa6\x90\xdd\xc7;+Uh\x19M]\xdd\x9c\ \xd0A]\xc7\x96\x12=~\x15\x95q\x0f\x91\xe1\xc9\x1c\ L\xc7O\x83\xfa\x82\x99\xd0\x7f\xf4]\xfa>\x99Z{\ \xe4^E\xf5Zr2\xc5y\xfa\x8d\x15\x0a\x1c\xeeO\ |\x8d\xf1\xf7vvip}Yj\xfd\xfev\x99\x9f\ Kda\xf9\xe9\x11G\xcb:\xda\xd1\xbf9\xc0jb\ .\xda71V\xe7\x90\xe3\xee\x95\x1a\xec\xed\xd6R\xe2\ ?\xb7\xef\x09\xe3\x1d\xbb\xc0\xbbl\x8e\xa9\xc7'+\xeb\ \x0d&\xc7!\xc7\xa7\xd7\x12\xfb\xf5\x90\x15\x07\x9f\xe2{\ @mi\xcd\x80\xffwX\xe2k\xab\xdar\x1e'\xea\ M\x19\x9dC\x849\x8e7U\xa8P\xd3\xacP\x9d\x1e\ N\xd0\xd6G>\xf8\xfaA\x96o\x9a\x0eYg\x9cV\ \x81om\x09\x81\x843f\xc1y\x1d{\xf0\xfe\xaf\xf3\ \xfa)\xe7\x8a+\x1f\x8f\x5c\x17\xf6o\xe9\xf2\xa5.\xff\ \xdc\x8cG\x13\x19i\xaf\xfd\x81\xc9Q&+\xe7Q2\ \xcf\xb9MK\xcdc\xa0)\xf1\xfb-Ru=\xe2\x9a\ >8%\xc1\xb5e\xb19\xfc\x17\xaa\xd5\xc0\x9a$\x85\ \xfa7\xb0\x8f\xc8\xcd+4\xd8\xde\xa1&}.\x0a\xe1\ \x167\x94+)\xc9C\x9bX\xf7)po\x95\x0a\x9d\ Ag\xfac\x02\x8a\x01\xdf\xdc*\xc7\xdf\xd3\x9f\xdc\x87\ \x176\x87h\x9dp\xaa \xe4\x228\xdeO}+\xf5\ T\x9f\x8b#NY\xe7\xdc\xdd\xfb\xe9\x5c\x08\xf6\x1d\xca\ \x98\xac\x0b\x0e\x85\xb1\xc9M\x9d:<\xb71d\xb3\x85\ &\xa8K*\xc4\x5c\x00\x05~\xbe[\x86\xe3\x83zJ\ r\xfcq\xf6\x02\xdaVyi\xd2\xebb\x1f\xc2\xd9\xe2\ g\x87\x9c\x99\xaf\xdd\xe2\x8b\x10\x9bS\xf4\xc5\x8d\xef\xda\ \xee*\x97\xc0\xaf\xa6V\xd6\xf1\x18i^\xc7lL\x94\ \xd7\x8fg\xc4\xc5\xd7\xa3e\x9d\xf0\x99\xa5WA\xd7\xc9\ b3\x8fo\xaaa\xe7O\x22\x07\xf2To\x00\x9e\xda\ \xa8\xd0\x18h\xbe\xc9g\x18\x87D\xdd\xfb\xd2\x86a\xe8\ \xf7+\x94\xaf\x88\xbc\xc3d\xcfao\xa7j\xd6\xb2\xa6\ E\xd69G\xc3k\xd8X\xdf\xefHY?5\x1c\x86\ \x9b\x97\xab\xa6\xff6\x1e~6\xbf\x5c\x83\xd6\xa1\x14\xce\ V\xe4\xebs\xa2\xe6\xe7\x09s\x17\x8b\xc3p\xceN\xde\ \xa7s\xdf\xeb\xd4\x0f\xea\x84\xbc<\xa1\xeb\xa9? \xa4\ \xc3\xa9\x8e\x11\xd8Edc\x8fg\x10<\xbd\xe3\x94\xdf\ \xdb\xf3\x0aS\xf57\x97\xd7\x05\xadx_\x1ad\xdd.\ ?%\x07\xda\x1c\xc9\xd7O\x0c\x85aQ\xb9F}*\ \xf1\xd6\x8a]_\xa6A]\xaf?e\xe7\x22tAC\ \xd5\xb3\x8c\xab'!\xeb\xc8{\xd0'\x83\xf2\xde\xbd\xfd\ g\xd8\xb4\xc2\x11ql\xea\xe3\x85\x08\x7f\xf6\xac\x1cq\ =\x12\xdd\x0fF|/%\x7f\x93\xbcQ\xc5\xd9@z\ e\xdd&\xf3\xc5\xfb\xdbRs\xe2)F\xc3h\x18\xee\ \xaa\xd4b|\x04\x93\xe307\x95\xcb0\x18L\xadm\ \x8a\xc7\x91\xb2\x7fN\x88\xb7\x5cH\xaf\xa3O\xa6{\xf7\ \xcf\x9d\xb3\xa7\xc6(<*\xd7 j!x;\xbf\x14\ \xe79\xe0\xfb\xec\xe9\x14sJ\xf4\xf4p\x18\x91\xfbG\ >\xdeT?\xe4\x9c\xf5\xb6a8d\xc0\xe3\xeb\xe5\xb8\ \x9fy\xfc\xf9'\xd6\xcad\xcfM\xe1\xf5\xf0\xf5\xa9\xdf\ \xfa\x87\xb8\xfc\x8c\x13\xe9u\xc1a\xfa\x8e\xfe%%\xb1\ \xdclF\xc7\xa8L\xb8t\x9c9^q\xe9s\x16\x87\ \x9dW\x22C\xfd\x88\x91\xd2})U@\x9f\xed_O\ *\x10]_1\xf1\xfeD\xe3\xd6\xbcW\xd5\xdbGe\ \xd0SX\x7f!rq\xc6:\x0f\x12Y\xbd:*N\ \x9a\x98\xcf\x91\xfc^\xc1\xd50\xec\xa9J\xae\xb6\xc0\xd4\ Q\x11[\xad\x9a\xad.\xcda\xf7\xf4B@\x7f\xd9\x9d\ +X\xff\xcct\xc9:\xad\x9b_\xa9@W\xc0p\xa4\ ^\xc7\xf3\x19\x0d\xa8p\xffj{}\xe8\xc43\xb8\xe6\ |\xca\xaekQ\xb9\x02\xdd\xe3ZJ\xed=\xd3n\xd3\ 5\xf0\x94\xdd\xce\xe3\xfe\xd3\x12\x94u\xf6\x8c`\x1e\x8d\ <\xea\xe5\xdc \x89\xf3\xe2\xe7f\xe8\x0a\x18\x9a\x0ca\ \x9e\xe7\x94\xc9\xb8l<\xa0\xb1\xa4\xfd\xa1\x84\xf3\x15.\ \xbd\xcfk\x90Gd\xe3_\xb7\xca4\x0f\xd6\xa9\xb2\x8e\ \xebp\xa2O\x86\xfb\xaa\x15\xab/\xd5D\xb2\x8ey\xa5\ +5\xd8\xd7\xa3\xd1\x18xJ\xfb\x95\xd8xj\xf7\xfe\ 7\xcc\x1c\x98\x84\xfc\xeb\x9c\xabw\xed|\x99\xc8gr\ 6\x05\xb5\x1d\x95q\xe8\xda\xffG\xea\xaf\xf7,\xbb\x1a\ Zk_\x06%4\x925\xb5hx\x8e;\xba\x0dZ\ s\x94.YG\x7fc\xa5Gr\xecz\x98\xf9\x8dF\ \x04Z\xc7Txt\x9df\xab\xbb\xe0~S\x11\xcfF\ \x9f\xfa*\xd6\x1b\xdd\x88\xa4\xa1\x1e%b\xe5\x09J\xbe\ vh*\x99g\xf3\x1dN\x9f\x94\xbd*\xe2\xa5(\xe7\ \xcd\xe5\xb7\x82*\xfb\x92\xca!\xa1\xf9\xe8\x8a\x1f\xba\xb6\ ~\x87\xec\x133Y\xfe\x01\x8fS\xf5\x10\xd9\xcf\x96\x9e\ \x08x\x8e\xd8\xe7\xe1\xcb\xebR\xac\xd7\x85\xaf\x91\xf6\x89\ Pi|6\x1b\xd6C\xccu\xf9\xb4N\x86\xefmW\ \xe1ib\xb3~\xa1J\x82\xa7\x88\x0d\xfa\xe26\x05\x96\ \xd6\xa9\xa0i\xda\x94\xe82|\x7f\x7f\xef\x01\xf0\x96\xdc\ \xc8\xe2JK&\xc9\xdb\x970\xbb\xd4S\xb8\x00\x86\xbc\ k\x93:W\x94cE\x0e@\xe7\xee\xdf\x90\xbf\x7f\x15\ =\x8fz\xd3\x9fI\xf8\xd1\xea\xe7L]\xe1t\xd0g\ \x96\xf0\xad\xbf\x1f\x1eJ\xb1>g\xb5:\x98\xb3S?\ \xacg\x0d\xa7\x8bpN\x82\x87A\xf8\xa8_6`(\ \xa0\xd3\x5cPM\xd7m\xfd\xf1\xd3\xaf\xcb\x04\xb7\xf2\xb5\ \xd7\x82\x87\xf0\x06\xaa\xd3'\xe3\x87\xc4\x9f)\x9c\x0bC\ \xcd;\x08wQ\xe3\xce\xf3\xb6C\x95\xfd\xd0\xb9\xe5\xfb\ \xe4}\xaf\x84\x86\x82\x994\xaf\xcc<\x0fr\xf4\xed{\ \xdd\xd1\xf9\xdav\xe09\xee\x22\x1c\xe6\xe6\xe5\xa9\xe6.\ \xac\x86\xef\xed\x13\x1ahF$\xab8]\xd4\x01L\xd7\ \xd3\xbe7\x22R-xF\x9a\xa3b\xf6\xf3\x08\x0c\xb5\ @\xfb\x96\xef\x12\xae|-\xe3(K\x99\xac\x09\x9f\xa2\ \xd0\xf9\xde\xc2k\xa1e\xfd7@\x1a:\xc7\xea\xa8\xd0\ sMm\xe7\xf8j\x1ah\xeeg\xc8\x07\x1d\xbb\x88>\ G9\xa71\xa9\x19p\x8e\xfc\x8dsK\x18\x7fiY\ \xf7,\xcd)\xa3\xf7\xd6q\x11B\x06q-z\x98\xc9\ 9\xf6\xee\x99C\xe7\xaei6\x9f\xdb\xe4s[\xed>\ \x17\xe1\xc7\xc0<\x87\xdf\x1c\x90\xa9>\x07\x888z=\ \xb2\x01\xf8li\xaa\x1f\x02}\xc7\xa0s\xef\x9b\xd0\xbc\ b1x>\xb9\x1a\xea\x0b\xae\xa2u\x1c\xcd\x15wC\ \xe7\xbe?\xc1x\xf7\x01P$\x1f\x7f>\x13\xf3\x0f\xa1\ l\xc8\xd28tl\xfe\x1e\xab\xf3\xe3:\x1c\xb9\x8b\x97\ \xbcz\x89\xbc7\xad\xfa\x02\x04\xc7:S|\x95\xa9\x87\ \xf0\xe1\xd6v\xeap\xdb\x0a{\xad\xa4\xd5G\x96\xf56\ \xd1\xe3\x92u\x1a\x8f\x22?\x7f-y\x8f\x0f\xcfj \ gam\xb5S!\xe6\xf5Z\xb1\xf3\x08\xa8\xaa\x0a\xfe\ \xf1Q\xc2\xa7e\xdb\xbc\xa1\xe8\x18\xfb\xa4\xc1\xfd\x91\x11\ \xfaL\x85\xa0\x8b\xf0s\x0f\xb5C\xad\xf8+\xf3\x7f\xce\ \x84\xe6\x9a\xaf\x90\xbf;D\xedr\xa7\xc2\xb4\xa9\xc9\x9a\ \xed\xeb\xd5\xe1\xc6\x0a\xcd\xaco2\xfb\xbe\x13\xdeqc\ \xb9\x06\xef\x9f\xd2\xe0\xc9\x0d2\xaf\xe1\x88\xeeSf\xf5\ N\xb5lPZ\x0bX\xac\xc3\xe3\xeb58\xd6\xe5\xa7\ \xeb\xacg\x09\x8f\xcb\x06D\x04\x87\xb2\xd5\x8f\x8a\x7fv\ ?el\x9cg\xb2\x10\xcf\x08\xee\x09\x9d[\xbfO\xf5\ 9\x95m\xca\xcf\xa7Q;\x00\xf9K\xcb\xea'!4\ \xde\x9b\x92\xdc\xc3t\x02u9\xe6\xd7l!\xfa\x1cm\ F\x8b\xb3Xug\xf3JT\x9a\x17\xaf\x11\xee\x11P\ t8\xd8\xa3\xc2/v\x87`Q\x85\xcc\xe5]1\x7f\ /\x9f\xd7\x87/ |\xe5';\x15\xd8\xd5\xa5\xc0\xa8\ \x1c\xa6qD\x11as\xf0r\xb8\xb0\x81\xee\x13\x84{\ wn\x7f\x85p\x95\x99\x9c\x9fOg\xb5\xdbh\x1b\x10\ \xbb\xb4\xa5\xe6\x09\x08\x05\xc7\xb2B\x7f\xa9\x06\xfa\xd2u\ XXn\xf5_\x9b\xc3\x0f\x8c%\xddL\xf4\xfc\xd66\ \xab\xb7\xa6}?\xc4\xbe1\x9e\xfe\x10,;\xd4\x07\x7f\ \xdc\xd2\x07o\xec\xf1\xc1\xd2\x03}p\xba'\x08\x11]\ 7\xfd\xcc\xd9\xe2ku\xc1\x10\xe1v\xb7\x22\x8fCW\ \xed\xcf\xac>\x1c6\xdeRO\xf5\xf9W!0\xd2\xee\ h\x9f\x8b}\xdf\xdb\xd4\x815H6]n\xab\x83\x9a\ _\xaa@\xb5\x97\xf5e<\xcf\x17\x11\xdb+%\xe6\x80\ \x0b|\xcf\x91\x88s_\xbf\x1c@\xfb\xbc\xa9\x12\xe1\xe7\ \xbf\xa3\xbe\x15\xf4\x9b{\x97\xda};3\xa8\x1d\x1a\xf0\ \x0df\xfaT/\x890\xf7\xb7\x1c\xef\x0f\x139Wi\ \xaf{\xd6\x9f\xcb\xb2Eo\xaaP`kK\x0a\xeb\x0d\ \x1c\x0a\x11KpaA\x91\xfc\xd0\xb5\xf9\xdb\x84\xa3\x5c\ \xc5s\xdd\xa7\x99z\x1de\xbe\xb9\xe6I\x08\x8cvg\ \x85\x8f\x01\xcf\xb1\xa6\xc5 v\xa8n\xf3\x09\x0a\x0e\xa3\ \xc2\xb5%\x1a\xacoUh\xbf\xa2\x5c\x87\xc8\xab\xba\xdc\ !\xf6_\xcc\x91\xc1\x1e2\xf5\x22\xa7\x80\xd6\xfa1y\ \xf7,\x99\x09\xde\xca\x07\xc17\xdc\xcd\xea\xf6\x1d*\xea\ \x11\x1bw>3l\x10\xbb\xd2\xf2\x99\xd3\xbaR.\xeb\ \xa8\xe77\xb7\xca\x9cs\xe4~/\xd9l\x89e\xa7\x1b\ \xd8\xa3\x8f\xc6Cw\xbeB\xf3}=<N$8K\ C\xc14h]\xfb\x0c\x04}\xbd\x99>\xd5K\x82\xc9\ \xb9\x01\xeb\xdb\x0d\xea3d\xba\xdc\xaa\x1d\xc6\xcf\xaf#\ \xfa|]\x93L\xf9\xcd\xe5\x82\xcbZ\xd6m\xbeJ\x9d\ \xf0\xf3\xce=\xaf\x99\xb5\x1c^\x9b-J\xf9\xf9\xca\x07\ \xc1O\xf8\xb9\x93\xed.q=x?\x0f\xf5\xeb\xb0\xa8\ \x0ccA2\xef[\xa70\x9dN>F\xff\xf9\xbe\x8e\ \x14\xcd\xe0\xc8\x22\x5cn\xd7k\x87\x90\x0bM\x09\x10;\ \xf4\xd7Dw\xe7\x99\xf9[fN\xd9\xc7\xd3\xa1i\xf5\ \xd3\xe0\x1f\xed\xe0\xfe|gB\xf8K\x90\x8fR\x7fK\ \x99\x8d\x97\xdbc\xf7D\x9fW7\xca\xd4F3c\x11\ \x97\x09.g\xbd\xcer\x96U\xe8\xde\xfb[\xab^i\ )\xcf\xab\x119\x8be\xf7\xc0\xf8h\x9f\xe3\xf5\x81\x88\ \x87\xd6\x8f\xe8\xcc\xafX\x18\xed?\xc7\x1eU\x0b\xcb4\ 8\xd4\xed\xdc<\xf2\xb4#2\xf5=\xde2\x0b+\x8e\ \xaa+A\xe8E9'\xfc\xfc\x1c\xcfm7\xf59\xb1\ C\x9bW?C\xf4y\x8f\xa3\xe7C\xd8\xe3\xfe\xeb\xda\ D\x9c\x88\xcd\xb8\xc8+\xb4\xf89\xf2\xf6\xf5-lF\ \xb2S\xaf%\xed\xb8\xccd]\xf8X5C\x83\xee]\ \xbf\xa4\xfe\x16\x0f\xef_]/\xf4:\xea\xf3\x92\x9b\x88\ \x9c;\xdf\x0e\xc59\xe0\x18\x0f83\xa4\xc3\x02\xca\xcf\ \xf5\x98\x5cE\xe6W\xdc\xd7\x99|?\xbc\x5c@\xc4\x01\ \xbd\x80\xa6\x0a\x22\xee\xdf\xb3\xffuh\xf8d\xd6y\xf5\ M\xa8\xd3\x9bj\xbe\x0a\xfe1\xe7\xf3\x16\x04>\xb7\x1b\ \xdaQ\x9f\xdbr\xb8\x0a-\xff\xf9\xf5\xa5\x1a\xac\xc5\xfe\ \xa6\xae_\x99\xed\x81\x97\xc1:\x98{\xbd\xaeB\xdf\x9e\ \xd7h.\xae5\x97`\x86U\xdbD\xf4\xb9o\xa4\x0f\ \xd0\x15\x97h\xef\xdb\xb4\xc3\x96\xd3V?,\xecP\x9d\ \xce.\x98U,\xfa\x9ec\xdc_\x83\xb3=!\xd3n\ M\xffy\x81\xc9\x0fA\xf4\xb3\x01\xf1\xa1=\xaf\x00\xac\ <\xb9)\x5cbj\x9b\xea\xfa\xd4\xfd\xc1\x0c\x81\xc6\xfd\ 5\x19\xfa\x0f\xfc\x89\xf9[luD\xa2\xb6\xa3\xb1\xea\ K0:\xdc\x93\xe9S\xbd$\x04\x0f\xa3\xfe\x96\x0a>\ \xc7\x84\xe7\xd6\xe6\xf1Y&\x0bKU\xa8j\x9aZ\xde\ b\xe5P\xf3^\x0a\x84_\xe9\x9a\x04\xa1\xb1.\xf0\x0f\ 6\x81<\xde\x0b\xba\xae\x10\xddj\x98\xf5eS\xeb\x17\ !\x7f\xd3\xc8\xfd\x98\x99\xaek\xd0\xbb\xe7\xdf\xa1\xae\xc0\ \xb2A\xeb9O\xc7\xfe\x1b\xde\xb2\xdb@\xf6\xb5d\x05\ o\xc1\x9c\xfdc\xfd\x06\xdcXf\xcf\xaf\xb5z\xdf#\ ??\xde\xaf\xd0\xfe\x1dSz^\x116\xffL'\xb6\ \xd0\xc0\xb9jh\xac\xfc<x\x8b\x17\xd2ZI\xef\xb2\ |\xf0\x16^\x03M\xc57@s\xf5\x17!\xd0}\x90\ \xec\xb1\xfa\xd4\xaew\x84\xf5=\xcaE\x08\x7f*\xc6\x89\ \xfa\x0e\xff\x17x>\xc9\xe3v\xa8\x95\xaf\xe8]2\x13\ \xda\xd6|\x05t\x7f7\xdfS\x9d\xc9\xe7\x04\x0fA}\ \xbe\xb1C\xa7s\x15\xf3m5Dy4o\x91\xcd\xaf\ F\x7f\xcbT\xf9\xce\xed\xf9\x08\xd8#g\xe0\xd4Rh\ $\xf2,\xfa6\xd0z\xf6%3\xf8L\x9f\xe8\xba\xe0\ 6\xb2\x8f\x0e\xd4/\x07U\x0e\xc2\x94\xf4_\xcbqY\ \xd75\x05zv\xbd\xcc\xf2\xcf\x97N\x03\xd1\xb7]\xc8\ zS\xe9-\xa0\xfa\xdb3}\xaa\x97\x84\x90\xf5\x9d\xdd\ :\x9b\x0dP\xc8fk\xb0\xde\xb2:\x8d\x8b\xa2\x1dz\ \xb0#\x94\xda\xde<\x93:\xaf\x08\x04\xc7\x06\xa1\xa3\xea\ !\xc2\x0f\xaf\xe4\xeb|\xf1Z\xf7s\xfc\x19\xc0z\xc9\ \xae\x1d?!<zjzj\xe4\x92\xac\xdbm \x94\ \xf3\xfeC\xffEkP\xcf-\x9d\xce\xedQ\xa6\xd7Q\ \xce\xdbV?\x06\xcax\xbb#zRO\x08q-a\ \xcc?'r^!r\xb7X\xfe\xb9\xf01\xdeD\xbe\ \xbe\xd2#\xf3\xfe\x08\xe9\xbd\x1e\xfb\x1a\xe3\xbe)\x05\xc7\ \xa0e\xcd\xd7L;\xdf3\x09Y\x17\xfa\xa6\x9e\xfb\x05\ \xba\xb6~\x17\xb4\xd0`z\xe3\xf84v\x98;\xb6\xa9\ X\x7f\xf4\xb7\xf4\xecc\xf5\xfet&\xd8\x92if,\ \xb4\x9e\xdc\x8b\xe6\x8a\xdbAE~\xee\xf08y\x84\xcb\ \xee\xde\x1e\x9d\xce\x92\x16u\x9eyEVN\x17\xd6\xcd\ \xed\xe9\x96\xac\xf9\xa6S\xd4\xcbA\xacs{\xcd\xe3Q\ =a'\xd3\x8f\xad\xc1\xb6\xbf\xb29)WB\xd7\xbe\ 7\xd2\x1a\xb7\xcb\xb5\x9c^j\x1f\xa1>?\xf8&\xf5\ \xb7\x88\x99\xa7\xf6\xfe`m\xd5\x8f\x82:\xde\x96\x15v\ (\xf6VY\xdb\xaa\xb3^\xf8\x5c\xc6\xc5\xdc\xd8\xd9<\ \x8fkc\xcb\xd4\xd6\xeb\x0b?5\xce\xde\x1c\xac+\xb1\ \xd5\x98'\xd0C\x99\xf3z6\x7f9\x0f|]{\xd3\ \xb6\xcfF 7\xf2a\xec\xfe\xf3\x9e\xbd\xbf\xa1y\xb9\ \xf6\xdaP\x9a\xbb\x88\xfd\xedV|\x0e\xa4\x91\x06G\xe7\ \xbc\xd9\xfd\xd1\xb5\x9d\x06\xd1\xe7\x96\x8fe\x16\xb7A\xf1\ c\xaci\xde\xd6*\x81n\xf0\x9e7Su~\xdc\x16\ \xd5\x95\x00\xb4T}!:\xb7\x22\x01Y?W`\xf9\ ~\x9b\xab\x1f\x02E\x0d\xa5\xef\xbc\x1dz\xcf\xe3\x81\xd8\ O\x07\x8e\xbd\x07\xdee\xb3LNn\x8f\x87\xb6U=\ \x0c\xf2XS\xa6O\xf5\x92\xc0}\x5c\x0b\x1b\xb0\xbd\xcb\ \x80\x85\xdc\xaf\x18[7\x87\xfa|}sf\xea\xe6\x04\ \x17\x18\xeb9B\xd6vf\xfc\xf2}\x91\xbe\xb2tf\ [\xef\xa1\xb4\x9d\xb7S\xf5[<\xc0\x99\xed\xfd\x07\xdf\ v\xe8,3\xc7E\xf0G\xda\xaft\xf9gx\x0f\ j\xe7_+\x9e\xe3\xfa6\xc2\xcf\x97k\x17\xa8\x9bc\ \xb3\xc3\xb6\xb6\xc8\xa0f\xc8\xa7 8L\xe3\xea\xe7\x93\ \x9e_e?\xd8\xfe0\x13z\x8f/M\xdfyg\xc1\ \xfd\xbf Dm\x02\xd1\x81\x03'>\x00\xcf\xb2Y\xa6\ \x1f\x00{\xb8P[\x94\x1c\xad\x95\x8b!4\x5cG\xed\ p\xa7\xeea\xc2FF\x9d~\xa0\x17\xe5\x9c\xf5\x9d3\ {\x10q\xff\x22\xfa[\xaa\x9b\x14r)F\xc6lj\ \xfc\xbb:\xd9C\xbd\x85\xf3S0\xe7$\xda^\xf5\x92\ \xd7\xba\x0d\xaf\xa4\xe7\xbc\xb3X\xd6#\xdc\x87\xd4\x7f\xe8\ \x0d\x1a'\xaa\xe7=\xd9\xe9\x1c\x02\xae\xdb\x9bW|\x16\ \xe4\x91s\x8e\xbf\xc6\x08\xefy\xb8\xb5\xc3\xa0\xfd\xb8\xac\ \xfe-\xcc\xaf\x88s|\xe7\x15k\xb0\xb5]\xcbx\xdd\ \x1c\xae\xa5&\xfb\xc9\x9a\xe7\xa7X\xd6\x99\x9d\xda\xb9\xf1\ \x1b\xe99o\x98\xa2\xbc\xa04@\xd7u\x18\x22\xfa\xdc\ [0\x8b\xe6\x9f7\xf0\xf8\x84\xe0\xe8\xcd\xa8\xcfG\xea\ \x1d/\xe7\x08\xd6GT\x83\xf9\xa5:\xf3\x9d\x17\x8b\xb8\ ?\xfb\xfc\xa6\xe5:\xd4\xb6)\x0e\xd1M\x11P%\x1f\ Y\xf7\xd9)\x93sK\xd6\x89]\xb5\xf6\xa94\x9dv\ \xe6\xd7N\xe4\x01\x89\xe7\x0ey\xc6\xf8\xd8\x10\xb4z\x8f\ \x82\xe7\xd4\x1ehk<\x0e\xa1\xc0\x985\x17\x0e\xc2\xb4\ \xde\xbf\xff\xf0\x7fB\xe3\xa7\xf9\xe0\xb1\xc98\xe5}\xc8\ [V\xde\x07\xd2\xe0)\x87\xc8\xc6\xc4\x10\xbd\x9a7\xb7\ \x1bp\xebr\x1b_1\xf3\xcfu\xb8\x8e\xf0\xf3\xd5M\ \x12\xedk\xe1\x8cx\x00\xcb\x8bnZ6/\xa5|\xdd\ \xc3}f\x9d[\xbf\x99\x9e\xb3v\x00\x87\xa5\x1d\x1b\x0c\ \x0d\x94\xb1fh?\xfc\x11\xb4U\xdeg\xc6\xd3\x04\x8f\ \xc3\xfe\xa0\xadU\x8f@\xfb\xd1b\x90\xc7Z\xa9>\xf7\ \x14\x5c\xc9mP\xab\x7f\x0b\xd5\xe7\x15\xb7\x81<\x9c-\ \xfa<\x0cG\xfbU3_\x91\xf5\x02\xd5M\xbf\x22\xda\ \xa7\xdb\xdadp\xd2\xa5\xe0\xb9\x84\xc9\xfdj(\xb9)\ \xe9\x19\x9c\xb1z\x1d\x9f\x9d\xd3\x9b^K\xd3yg^\ \xef\xe9\xb2\x0f\xba\xf7\xfc\x0a\xbc\xc5\x8b\xac\xb8\x82\xc8Y\ \xb1}\xcc\xea\xfbg\x92\x9f\xbb\x01\xbc\xcb\xe6\xd09\xed\ \xf5Q~\xc5i\x84\x9f\xdfG\xe4\xfcl\xd6\xf4\xb6\xdf\ \x8c\xfc\xbc\x9c\xe9\xf1<\xdeCt\x0e\x8f\xfb\xa3>\xdf\ \xd4\xa6\x99\xf1P\xa7@\xec\xc3\xdd\x87\xdeN-_\xa7\ \xf7{&\x0c7T\xa6\xf1\xbc\xa7~\x1d\xc5\xdf\x0d\xfa\ \xba\xa0\xb5\xf2~6\x87\x85\xf7\xa7\x98\xec\xdaxcz\ C7\x96,\x02i\xe0\xa4\xa3\xe4\xe2B\x08\xf3\x99R\ \xa7\x86tXX\xa6\xb2x?\xea\xf1b\xab\xef9\xd6\ _l!\xfc\xdc\x91\xe0\xfa14\xd2B\xf6\xd6<\xb3\ \xff_*8LS\xd1|\x90\x86\xcfd\xfa\x0aS\x8e\ \xe0\xa0\x17\x9aW=\x02f\xde\xa1\xd8\xc7\xe2\xd0\x03\xe2\ h]u?\x84\xfa\x8f;b\x9f\xba\x14\x0c\xea?7\ h^b\x1e\xed\xdf\xc2k\xfd\xb9\x0f\xfdZ\xf2\xf5u\ \xcd*\xed#\xeaDX\xb90\x12tl\xf9.\xd5\xc5\ \x93\x9e\xe5sQ\x0eC\xee\xe3\x86o\x82\x9e#\xf9Y\ V\x0e\xa2\x04-k\x9e0c>l\xb6\xd6d\xf2\xe3\ b\xd7\x87\xfcN\xc1l\x08\xf4\x1e\xa39~\x197\xdb\ &\x80\xc8\xc9B\x19\xa9\x1b6\xa8>\x9fm\xda\xa0\x8a\ \x19/\xc2\xfc\xae\xbd\xedAG?\xb3\xa2\x86\x0e\xf7\xa7\ @\xd7.\xb2\xfe\xd8G\xe7\xfc\xfc\xf4\xc9\xca7{N\ f\xd0\xbcwE\x0a8z^C<\xa03y\x0d\x03\ z\x0f\xbf\xc5\xeb\x82\xa6'i\xcbO\xa7\xb9\xe9\xc3\xe7\ \x8a\x99\xef\xd4\xa1\xcb\xc4\xea,\x22P\xd5\xac\xc3\xf5\xe5\ V\xac\x9f\xf5>g\xfd\x5c\xae+Q\xa9>\xc7>\xff\ \x8e\xees-r\x8f4\x05\xfav\xff\xc2\x8cS'\xa2\ \xd7\xa9\xac\x93\xe7\x04s\x03F=+\xa7\xa6fc\x8a\ \x80\xb1e%\xd0\x03M+\xee\x85:\xbe>\xe7\x12\xb5\ o\xcc\xba\x17\xc2\xd5\xab\xbf\x02*\xce\xd0\x9d\xc2\x1c\xa8\ x\x10\xa1\xfa\x5c\xa3\xfd\x88h=Q!\x9bK!z\ \xa1/(\xd5aw\x97|^n\xb8\x13\xc1j\x94U\ \xe8\xde\xf5ofn\xae\xa8=JDWa?\x87\x81\ \xa3o\xb3\xfaS\x9c\x01\x1e\xc9\x8dz\x0a\xbc\x9e\xd1\xe6\ \x0d4oE\xf8X\x92\xe5xt\x0f\xfct\x1e\xf8\xba\ \x0f9\xcc7\xc7\xeb\xe6\x22\xacO\x11\xd6\xcd1\x1bT\ \xe8tf\x87^G\xf8\xb9\xdd\xdf\x228B\xe6\xfd\xe8\ \xd10\x9fA\xec\xc5\xb0\xfbU\x1e\x9f\x9en\xf6\xd5\x11\ z\xc7^\xe3h\xc6=DM\x9e-\x1f\x12ykS\ \xd9\xed0\xe2\xad\x89z\xc6\x1du\x13\x93\x00\xde\xbf\xee\ }\xbfO\xda\x8e9\xdfN\x9d\x09\x8d[_s\xd6:\ \xf1\xdc\xd7#}\x1a\x9d\xab(\xea\xe6\x04O\xc7\xb8?\ \xe6\xa5\x1f\xeb\x96\x1c\xab\xc3\xed`v\x96\x0c\x83{_\ \xa1\xf1\xba\xfa\x8f\xa7G\xd7\xdb\x11YnZ\xf1 t\ lx\x06<\xcbf\xd3\x19;\xecgX\x9eK#\xe6\ D\xd2X\xc8UD\x97\xcf\x83\xb6m\xbf\x8c\xda\xc3r\ \x85\xbb\x08\xa0\x9e;[\xfeX\x8a\xe3m3x\x0f\xa3\ \xa7\x1d\xb5^x\xff6v\x18\xb0\xa0\x8c\xdb\xa0\xbcw\ \x8b\xe8\x81\x8e~\x98U^\x85\xd5\xe28L\x87_\x08\ \xba\x16\x84\xbe=\xafZz\xfa\xe3\x19Q\xfbr\xdb\xea\ /\x822\xe2\xa5{\xf7X_=\xb1\xa1\x96A\xc7\xee\ \xd7\xe0L\xcd\x8bp\xa2\xf2y\xa8[\xfb}\xe8\xd9\xff\ \x06\x8c6\x14\xc3H\x7f\xb3\x99\x07\x1f\xceQY\xc7\x1e\ -\xcd\xe5w\xa64\xde&t\xca\xe9\xc2{\x1ca\x9b\ R\x7f\x1c9N\x0e\xeb\xb4\xae?\xb6\xef\xdc\x9c\xa20\ \xcc#r\xee\x1dLOMB*Am\xc50\xcbW\ \xef\xdd\xf9\x12\xe5)gE<[\xd4\x08\x14\xcc\x80\xf6\ 5_!\xdcFr\xd6\xbe\x9aa(r\x10Z\xca\xef\ H\xb9\xac\xe3\xfb\x9d^v\xa7#\xd6\x1a\xfd(5-\ \xbc_\xae\x99\xdf\x22\xe6\x96k\xac\xcf\x7f\x8b\xe6\x88s\ \xbd\x14hN\xa3:\x0e}{\x7fm\xf6\xc1\xb0\xebr\ \x8c\x87\xb4\xd5<F\xf3;\xc0\xce\xb9]\xd0\x9eO\xcd\ \xcb\xefI=\x87A\x9bh\xf9\xc3\x8e\xe8Kw\xb4\xdf\ <\x5c\xb1\xe5oY\x07\xf2\x96\xc3\xbdh\x87NM\ \x8f\x88d\x81\xfa\x1c\xe5\x1ck\x9f\xebM\xdb\xd3\xe2\xe8\ \xad\xab\x1e\x00]\x0f\x99\xb3\x1eY\x1d\x87\xf3\xafk*\ \x80k\xd1\xb4\xf1\xfbQ\xeb\x95*Yo\xaf})c\ \xf2#\xfc\xe7\x1b\xdb\x18?\xa7\xbd\xb8D\x0d4\xef\xb5\ \xb8\xb0\x14\xfb\x1494\xeeo\x83\xe9?W\xfc\xd0\xb7\ \xffW`\x8fk3\xbe\xc8>o\xa9~\x94\xf0\xf3\x86\ h\x1f\x8a\x0b\x0bd9\xfa\xcf\x14\xa6\xb4N\x91\xca:\ \xfah=\xeb2\xb7\xde\xe4\xef\x1e\xe9\xe7s\xcfyN\ \x0b\xd6\xcb\xd1\x98\x11\xd6Y\xa0>\xef\x0e\x82a8_\ \x1eD\xbf\xc8\xbe\x03\xbfe=G\x96\xda\xfd\x8a<\x96\ _\xf5(\x18Z\x08\xec\xb3O]D\x03\x97$4t\ \x0e\xbcE\x8b\xac\xba\x8a\x04}\xec\xe6\xfc\x0br4\x96\ \xdf\x03\xb2\xafkJ\xd7\x5c\xf4\xa4E;tk'\xca\ \xb9UK\x94Wd\xf5r\xb9\xa5\x1c\xeb\xe6T\x9aw\ \xefd\x99\x10\xbe?\x8d\xe8so\xed\xafX_.\xd1\ sd\xa9U\xab\xde\xba\xea!\xaa\xcf\xa9\xefD\xfc.\ 8\xc2-\xe0(\x84iO=\x05:7\xbd\x105\x7f\ +1\xdf\x8b\xe8\xa73\x0dz\xf7\xfe\x9e\xf6+\x99R\ Y\x8f0\xff\xc4\xfe>\x83\xd5\x87\x8a\x1e\xa2\x85\xa2G\ \xb4B\xed\xd0C=,N\xe4t\x84\xb9\x0f\xb0\xb1\xf6\ \xd7|\xb6k\xb4\x1e\xa2r\xbe\xfa+\xa0\xcb#\xae.\ \x9f\x04\xc2\xdcv\x91F\x9b\xc0St\x0b\x97\xf3\xc4\xb8\ ;\xb5G?\x9e\x06-\x15w\x83<\xde\xc9do\x0a\ \xb5\x0b\xd6\xcd\xa1>G\xbf\x22\xab\x0b\x15\xba\x9c\xd5D\ \xd3\xbe\x16\xadD\xce!\x9c\x15\xfes\xac!m\xda\xfe\ \x1f<\xbe)jv\xad\xf5n\xad\xfa\x02\xc8c-Q\ \xfc<\x1b\xae+S`\xfb>\xd3\x89#\x8d\xd5\xb4O\ \xb1\xbd\xbf\xc5\xa5\xe2\xa9V\xdd\x06\xe7?\x85\xf3\xc0\xcf\ s\x03\x22\xfc\xfd\xd3\x8d0\x95\xdd0\xec\xe8\x16us\ :\xe3\xe7f\xfe9\xcb\xcb\xdd\xda\xaad\x81>\x8f\xf0\ \xf8n\x18Zv\xbfE\xd6x6\xed\x19%\xe2E\x82\ g6\xaf|\x08\xb4`7\xe3\xe7\xe27]Y\x9f4\ \xb0ft\xe8\xcc\x12\x22\xef\xd7\xf3u\xe5yD\x17\xd5\ \xe7\xd3i?\x0c&\xe7\x0ba\xa4\xbed\xcak\xc3\xd1\ \x7f\xbe\xb3\x9b\xd5\x13\x89Z9\xda/\x97\xe7+b\xdd\ \xdc\xe6\x16\xc5\xa1Yh\xd1@yU\x88>o\xde\xfe\ :\x93q\xb2\xc7\xd6\xc7\xf0\x96\xb6\x9aGi<\xd4\xf5\ \xb7$\x0ef\xef\x1b\xe0\xeb:@l\xcb\xbb\xcc\xb9\xb9\ \x97\xe2\xe8x?ZW}\x01\xa4\xa1\x06\xfa\xfbS\xdd\ gwW\xb7\x0e7W\xf0x\xbf\xadn\x8e\xcem)\ Sag\x97\xcc\xfb\xce9_&\xf0\x1e4\xefx\xc3\ \xec\x09\x18UsA\xf4O{\xf5c\xa0\x85\xfa\xcd\x18\ \xbe\x93\xf3/\xb3\x01\xb8~\xb24\x0e\x03G\xde\xa2>\ [\xef'\xb3\xc9\x9a_i\xfaX\x18\xb7\xb9\x12\xea\xc9\ \xd7[V?\x09\x83G\xff\x02\x86\x1a\x9aR\xfdB\xfd\ \x13D\x9fon\xe3\xfd[\xccZ\x22>W\x91\xe8\xf3\ \x9b\xcbu\xd8\xd2\xe1\xfc\x9a\x1a!\xaf\xaa\xe4\x87\xd6\xdd\ o\xb2:\xf6X\xff9Yw\xec\xd5\xa8\xf0Y#\x8e\ \xee\xdb\x9dE0s_1\xb7=8\x08\xa1\xbe\xc3\xe0\ o\xdf\x01}'\x97@\xe7\xc1w\xa0\xff\xf42\xf0\xb5\ \xef\x82`\xffQP\xa41\xca-#S\xec\xe5\xc2\xbf\ \x87\xf3\xe60/\x97\xe9r^\xef\xcfs\x00\xd0\x7f\xbe\ \x89\xf0s\xcd\xf1\xfc\x9c\xf9}\xb1\x86\xab\x8d\xc89\xd6\ \x8b6\x14Lc\xf5\x16\xa6M\x84\xbd\x8cQ\xce\xdb\x5c\ \xde\x92b\xc4\xd6'\xe0\xb2\x86\xc3\xa2\x17\x16\x08\xf3\x89\ ~`\xfaZ\xa6\x22\xcfY\xc4\x0f\xc9I\xec&\xfc\xdc\ \x9c\xabX\xcc\xea+D\xbd\xff-\xc4>\xddF\xfb\x14\ \x81\xa3eB\xc4\x034%H\xf4\xf9\xffe\xf1\xbc%\ \xacF\x9d\xc6\xb1\x85\xff\xbc\xea\x11Pq\x06C8l\ \xc5\x04\x1c|].\x92\x87x\xf6j\x9aY\x9c\x08\xfd\ \x89\xf9\x85\xd1us\xd8\xe7\x7f\x0b\xd1\xe7\x99\xee;7\ \x19\x88\xb8W\xeb\x9e7\x19?\xbf@\xec\xbf\xb9\xf2A\ \x22\xe7\x1d zR9\xf9\xd9u\x91:\xa0\xfc\xee\xeb\ VaA\xb9\xad^\xae\xd0\xaa\x9b\xc3>\xa2\x1b[\xe4\ \xac\xd9\xe3i=\xfb\xbew\xb8\x8cO\xb3\xf9\xbe\x98\xac\ cozy\xb4\x95\xc6?\x5c\xeery\x80\xf6\xd0#\ zm]\xab\xce\xfaq\x15\xb1\x1a\xd1\xbc\x22\x11\xff\xc7\ \xb9-\x1al\xa6}D\x9d\x9d\x0fb\x97\xd9\xae\x03\x7f\ \xa65A\xde\x8fg\x80\x98\x91\xe6\xe1\xb6?\xeasy\ \xd4\x13\xe5o\xc9\x027\x92\x8b$\x81\xf2\xbb\xbbG\xa5\ }\xce\x19/\x8f\xae\x9b\xbb\x89\xc8\xff\xaeN5+|\ ox\x8e8\xa7\xa2m\xdf\xdb6\xdbsZ\x14oi\ E\xde\x12\xe8\x05\xdd\xd5\xe5\x97\x1dP\x9f\xe3|s\x16\ \xf7Wh\x1eW^!\xd3\xef\xf8\xf5\x15^\x99\xf6)\ \xca\x06\x99@Y\xef:\xf86\xd4\x7f\xc2\xf8\xb9\xd7^\ \x0b\x8d\xf5\xa1\xe5\xf7Bp\xe0\x0c\xf5\x81\x89\x9a\xb8l\ \xb8.\x17\x89\x83\xf6\x9d\x0b\x1bpj@c\xf1P^\ \xeb?;jn\x8b\x0a;;\xe4L\x9f\xea%!\xe6\ \xa4\xe3<\xf4\xb6\xc3\x1f\xb1:\xe8%\xd1\xba\xdcK>\ o\xc4\xd9Q\xbe\x96\x0c\xccJw\x91I oY\xd7\ \xca\xfbZ\xd8\xe7\xb6p[\x14\xfd-\xeb[\xd5\xac\xf1\ \xb7\x84\x0d\x05z\x0e\xbdg\xf6\xb0\xf0\xc4\xf8[\x9a\x88\ \x9c\x87\x06\xcfq9we=\xd7!\xfc\xcdx\x9c!\ \xfa|Q\x85\x15\xeb\x9fc\x9b\xad\x88\xfd\xe8vu\xa9\ \xbc\xe7\xa83\xe3\x87Q1\x0aC\x87\xee\x13\xcbh\x1f\ o\xb3\xce\xc2\xd6\xf3\xb8\xa9\xec\x0e\x90\xc7{h\xdc\xc0\ \xb0\xd5\xf2\xbb\xc8]\xb0|m\x03\xd6P~\xcelP\ V7\xc7\xeb\x89x_\xe8\xb5-\xce\xcf?\x17q|\ \x9c\xa9\xd3{\xe4\xaf4Nt~]\x00\xf9\xbc\xe2\x1e\ \x08\xf4\x1dg\xbd\x06];\xf4\xb2\x01\xca\xef\x89A\xec\ Sd\x9b7\x87>E\xecYD^\x17\x91\xaf\xef\xef\ T\xb2B&h\x9f\x22C\x83\x8ec\x054\xc6\x7f\x8e\ \xf0qoL\xfdzc\xf9\xdd\xa0K>\xeao\x09S\ \x9d\x9e\x1d\xf6\xf5\xe5\x0a\xbbO\xcc\xf4\x05Ct\xec\xe3\ BG\xec\xef\xe3\xbd\xde\xd0\xa63;\xb4H5g\x9e\ \x8bc>\xd1\xe7+\x1b\x15S\x1e\x9c*\x13f\x1d4\ \xf6\x9d;\xfe\x919s\xc8\xcc\xf3\xe7\xbc\xc5[v\x1b\ \xf8\xfb\x0eG]\x8f\x1b\xfbw6\xa2d\x98\xeb&\xa1\ \xa3\x84\xec\xdb_\xed=\xa5\xec97g\x06Us\xee\ \xf9\x9cB[\xec\x1f\xeb\xe6\x88\x9c\x9f\x1dT\xb2\x82\xc3\ \x8a\xeb\xeb9SNsA\xeb>\x8e\xae\x9b\xf3\x16L\ \x87\xe6\xf2\xdb\xc1\x08\x0dZk\xe1L\x93\xc3E\x0c,\ \xbd\x1c\x86\x86!\x0d\xfe^\xa7\xc3\xcfv\x04\xe0\x81\xca\ \x00\xdc\xb2B\x81\x1b+\x14\xb8\x81\x1c\x9fY\xa9\xc0\x0b\ \xeb\xc7\xe1\xdd\x93*\xec\xe8TA\xd343\x8fk}\ \x9b\x067\x94\xdb\xe6o\x15Z\xfa\x1c\xfb\xb7\xacm\xd5\ \xb2&O\x9b\xceH;\xf97:[\xa7a\xa9U\xaf\ %d\xbd\xa9\xfcV\x08\x0d\x9cd\xfc\xdc\xe5-\x8e\x87\ \xb0\xbb\x84\x9c\xcb\xaa\x06?\xdc\xa9\xc2\xdcb\x85\xc75\ Y\xef\xdb\xd8\x03}\xe4\xac\xdeY\x85{V\xaap\xaa\ W\x82\xbd=,Nd\xfa\xce\x0b\xf9,Q\xec\xc7E\ \xbe^7\xe4|\xff\xb9\xc8\xa7\xc4\xa3\xe7T\x19\xd1\xdd\ 3bzZ\xcc\xa05\xbc\xcd\xe5\xb7\x11\xf9\xd6\xcc\x9a\ \xe9\xf8\xfe\x86}\xef\xd4\xcd:a\xb0\xed\x8d\xd8;:\ \x1c\xb1\xbe\xe6>C\xc9\x83\xe5\xe7\x85\xa9\x0f|y\x83\ \x04\xb7\xae\xd0\xa2zj]H\xceMy/\xb2\xd7\x80\ 2\xbd\x1d='\x97\x1d\xe8W_\xd5\xa8f\xc5L\x07\ <ECW\xa0\xe7\xd8\xdf\x00\xe7\xb7\xd4\xf3Y\x80V\ \xdd\xdc4\xda\xeb9\xd8{\x90\xe7+\x86\xe3\xaa\xb5\x88\ \xb5u4M\x01\x7f\xff\x19\x18i(\x87\x81\x13\x1fA\ \xf7\x91\xff\x81\xc1SK`\xace-\x04G\xdb\xa2x\ \xa3\x8b\xe4@\xd7[7\xe0\xcd\xfd>\xb8\xa6D1m\ \xc99\x97\x90s\xbb\xbe\x17>DV7\x17c\x87\x96\ +p\xb4[\xe1\xf5|\xce\xbf_\xb8\x1e}gJ\x89\ >\xbf*jf\x94\x99\xafXv+\xe8\xd2\x10\xcb?\ O Vd\xe7\x89\xc3\xcd\xdb\xc8s\xf3Y\xf0|2\ \x87\xf0\xa4+y\x0d\x13\xef\xa5N\xfe\xbe\xf7\xd3k\xa1\ }\xdd\xb30>\xdc\xee\xcaz\x12\x10\xb6\x94fD\xe0\ \x83\x132\xe5\x22\x13\xf1\x95D\x0e\x94{\xec;\xb7\xae\ U\xcd\xf4\xa5^\x12\xa6\xadM\xe4o\xa8\xbe\x9c\xce\xcd\ \x89\xae\xc3e<\xbd\x89\xd8\xa1R\xff1\xf2\xb3\xf1?\ \xb7l\xa6\x05{.B\xa3-\xd0\xb6\xfde\x22\xe3\xb3\ /\xde\xc3\x87\xef)\x8d\xc57\xc2\xc0\xc9\x0fi\x1d\xc8\ T\xe8\x0b\x91w\x1c[\x07{\xbe?\x22;\x0cq\xa1\ \x8f\x0e\xb6\xfbh\x0f\x96\xfcB^\xf3\x96\x22YG\x7f\ \xcb\xee.-{\xe2\xfe\x11^S\xb4\xf69:w'\ Z\xfe\xa6Sy\x93}\xed\x09\xc7\xfd\xc3\x5c6d\x7f\ \x1f\xb4\xd6<\xc6\xe3Q3.!\xeb3\xa0n\xc94\ 6\xfb\xaa\xe0J\xe8\xda\xf7:\x8d\xdb\xa6\x1b\x96<\x87\ i\x1d\xf0xP\x81\x1d\xcd\xe3\xb0\xe2\xe4(\x1c\xef\x0a\ \x80\xa6\xea\xe6\xcfd\x03P\xcft\x8f\xabp\xdb\x0a\x16\ \xd3\xcc+N\xad^G\x1f\xfa\xde\x8e@v\xac\x07\xb7\ GU%\x04\xed\xeb\xbfn\xd6\xcb\x89\xd7\xe6\xf2; \ \xd4s\x90\xc6\x81\x13\x8d\x89\xd2x\x94*A\xc7\xe6\xef\ \xd0^H\xf5f_\xb5\x89\xe7~\xa1\x1d\xec]j\xfb\ \xb9\x82\xaba\xf0lQt\x1e|\xca\x96 \xda\x8e\x18\ \x0f\xa9\xb0\xc2#\xc1\xe3\xeb\x88\x1e,\x91\x19'-f\ 5\x92h\x9b\xbdX\x1b\x82\xfd\xbdFV\xd4\x9d`\x9e\ \xe9;'\xb4K\xda\x9f\xc9p\x98\x87jT\x904\xe7\ \xefs\xc2'BgVx\x97s]\xcb\xe6\x155\x11\ ~\xae\xfa\xbb\xe8z%\x93\x93\x8b\xcf\xc8\xc0\xb1\xbf\x9c\ \xcf\x8f\xe2\xe8\xc3F\x8f\x82<\x18h;\x94r{\xd5\ \xe2)8K1\x02\xdf\xda\xa6@~\xb12\xe1\xbd\x9d\ Cg\xae\xa9\xf0\xdeIb\x8b9<\x07\x1be\xf0\xa9\ u\x12\xcd\x1dO\x87\xac\xe3^q\x0dY\xab\x9d\x1d\xd9\ \xd2':lr\xf6\xde\xbaU\xd0\xb5\xe5\xdb\xd0\xb3\xe3\ \x87 \x0d\xd7\xb394I\xc6x\xd5`?4U\xde\ Gg\x1b%\xda_S\xcc\x05\xeb\xdc\xf5\x0b\x1a\xcbM\ \xb5\xac\xe3\x11T4\xf8\xee\x0e\x16\xff\xa3>\xe7\x89d\ \x9d\xcfa\x9b_&C\x85Wst_x\xcf\x90\x02\ \xd7\x95]\xda\xaf\x98\xb0^\xe7s\xd1\xdf;\x9e\x0d\xb3\ \x5c\xce\xcf{\x08\xd3y\x05\xe1\x0b~/\x91\xf7\xf6w\ \xd4\x12\x9d<+\xe1y\xeb\x0d\xa2/\x1b\xce\xe0,\xbf\ \x8b<;\x83)\x95u\xc1\xbf\xcb<\x0a\xcc-\xb6b\ \x81\x13\xc9\xba\xdd\xd7v3\xe1\xc1>\xc9\xb9>\x88\x82\ :\x99\xc7\x83\xd2\xa0\xd3\x8b\xd9Z\xcc\x22\xef\xfd\xf5-\ \xce\x8f\x1f\xa5\x0b\xf6\xe7\xa3i\xc7\x9f\xcc\xfa\xd3\x84\xf4\ \xba\xfd(\xb8\x0a\xfa[\x0f\xa5T\x97R\xdf3y\xbf\ g6\xaa\xf1\xf9(\xb8\xbc\xbf\xbcOs\xacm\xf6\xbd\ u\xfd<_%M\x1c\x86\xcb\xfb\xe7V:_\xaf\xa7\ \x0bv_z\xf3\xba\xaf\xb3\xb9\xca)\x90u\xccW\xe8\ =U\x92\xd2~ax\x9eCA\xa2\xd3K\xe2\xdc\xeb\ \xb9\x8e\x7ft\x8dBx\xb1s\xfa\xb2\xd1\xd6G|\xaf\ zd\xf9HZe]\xecs7\x97\x8c\xa7\xe7Zb\ |\xbf\x90$\xd7H\xe79\x1a\x9a\x0a\xed\xab\xbf\xccb\ EI\xceWf\xb2>\x0d\x06O\xbc\x9fb\xbe\x0ep\ \xbc_\x9eT\xbc\xfcBz\xfd\xde\x0a\x09Z\x87\x9d\xb5\ \x87\xb3\x1c\xad0,\xae\xe6\xe7Y\x98.\x9d\xce\xebH\ K\x82i\xbb\x8e\x0b\xf2l\x07\xcaz\xd80\xa0s\xdd\ \xd3\x9c\xbf\xa4@\xd6\x09g\x1f=S\x90R\xce\x80o\ \xb5\xa5%h\xc6\xbf\xe3\x95\xf5\xdb\xcb$h\xe8w\xce\ \x1e.v<\x5c\xff/\x97\xf7\x99u\x9f\xe9\xe2/\x98\ 'sk\x99?\xe9\xf3\x8e\x95gY3\xa0\xd5\xa7\xc3\ \xbeN\x09\xde\xd9\xd1\x05?]\xd9\x08?^\xe1\x85w\ w\xf5\xc2\x91^\x19Z\xc7\x0dPu\xdd\xf6;\xd1\xef\ 3\x95\x10~\xeb\xa3\xd5?5\xe7\xaa'\xaf\xd7gB\ \xe3\xc9\xcd)\xe7\xeb\xcd\xc3\xba9\x9b'\xae\xfbL\x8e\ \xc5\x952t\xfb\x9c\xe9s{u\xdb\x08\xcf)O\x97\ \xcf\x91\xf9b\x1e\xa9\x96\x92>W\x91K\x88\xfbQm\ \x9b\x06\xdf$\xf6\xee\xad\xe5\x92\xd5/\x8c\xf78\xc5\xcf\ \xe7\x92\xbf}\xebr\x05\xfe\xb56\x04\xab\x9b\xe4(\xdf\ o&\xf5}\xff\xd9\xe5\xd0P0\x93\xce\xa3NV\xd6\ \xbd\x85\xd7\x814\xd6\x9ab\xbd\x1e\x81\xa0\xac\xc3\xbc\xd2\ I\xe4B\xd9\x8fB\xd6w\xf3\xc9\xf5*h\xba3c\ )8\x1f\xf1b~\xa5\x94\xe8u\xf2\xfa\xa3\x9d\xc9?\ \xebx\x1f\x06C\x06\xfcd\x8f\xc2{\xca\x88\xb9I*\ \x8di\x08]$\xfe&\x9d\xabT\xc4r|^\xd9\x19\ \x82\xa1P\x98\xda\x87\x99Dh\xf8\x1cx\x8ao\x88\xca\ \x0fN\xf4h\xdf\xf8\x02h\x9a?\xe5>\xc7\x10\xd9/\ \x9f\xdd\x18\xdf^/\xb8\xc1_O;\xd7\x0f\xd3\x17\xc0\ \x9e\xfeJZe\x1de\xed\xd3\xba\xc4d\xdd\xca\xe3\x8e\ @\xbbO\x83\xa76h<?\xcd\xd2\xe3B\xa7\xb3\xcf\ \xf9^\xc29\xa45\xcbZ\x81\xa7\xc9\xfd\xeb\xf7\xab\x19\ \xcd\x8d\xd55\x05Z7\xff\x90\xf3\x98\xe9\x93\x9csb\ \x1d\xf5\xdc&E\x0e4\xd8P\x99\x92\xbeJ\xa6}C\ \xd4\xc0\xa8d\xc0\xf3\x9b\x15\xaa'&\xb3g\xcf\x11\xba\ \x86\xbc>I\xd6W\xd1\x9d\x1b;\xc5s\xfb\xf6\xd6P\ \x1a9\x8cJ\xeb6N\xf4%'\xeb\xc3A\xdc\x1fQ\ G+\xf1\xed\xad1|\xf2\xd1\xb5:\x04\x95\xcc\xcd\xb3\ \xa6\xf3\x07\xfc}\xd0\x84\xf38\x13\x98QHs\x16\x08\ Oo[\xf3U\x96\xaf\x10I\xfeZD\xed\x82\xa2\x85\ \xe1g\xbb\xc4~xi\xdb\x94\xd5\xe1(\xb4\xf7&\xd6\ \xe9\x9c\x1cp\xf6\x9cp\xdc\xd3+\x9bt3\xd6\x9b\x0e\ Y\x7fq\x8bDs\xe3\x13\x81\xa8\xe9{\xe7H\x90r\ \xf0\xb8}aQ\x07\xd3?\xaf\xef\xf7\x83fd\x86\xcb\ \xb0>1\x06\x8c\xb7o\x85\xc6\xe2\x05\xe7\xf5\xf1\x9d\x8c\ \xef\xa5u\xf5\x13 \x07\x87\xa2rk\x93\x01\xae\xf1\x08\ \xd1\xe7\xdf\xdb\xc6\xe6i2}\xad_\x92\xc3\xcc\xe6\xb9\ \x82\xff\xb4I\x87\xa3\x03\x1a\xeb\xe1\xe7\xe0\x1c\x01\x5c'\ IQ\xe1+\xebY\xbf\x96\xbc\xc2\xe4\xb9\xbbU\xab\xa1\ \xc1\xdd+\x15\x18\xf5\x07\x13\x8ew\xe0}\xf0\x0c\xe1\xf9\ (I\xe7\x1a\x8b\x9a\xc0\x9b+B\xd0:\xa6&T7\ \x97*`\xde\xc1X\xdbF\xc2\xdd\x17\x99\xf2.\xfa)\ \x89\x19yt\x86\x18\xbe~\xccf\x154\x14\x5c\x05\xad\ +\xee\x03\xc5\xdfOd*\xf9g\x95q9\x03\x02Z\ \x04^\xdc\xa6\x99={\xec\x1c\x05e\xff\xc1\xd5:\xbc\ ~@\x82\xc5U2\xdc\xbe\x5c\x85[\x09\xe7\xbds\x85\ \x02\x0f\xadQ\xa0\xb8.\x08\xb2\xa6\xa5`E\xd2\x0f\x91\ \xef\xd46\xa2\xc2\x83\xd5*\xb3\xf7\x92\x96u\xd1\xb3\x0e\ {\x90*tfz\xa2\xba\x07\xf3\xa6_\xd9\xc7uM\ q\xf2\xe7\x85z\x08\xe5\xfd\x0f\xbb\xc62\xca\xdbYN\ e\x18\x02C\x1eh\xdf\xfe*x\x0b\xaf\xe5\x9c\xe6\x7f\ \x9b\xbd\xdb\x1bD\xce:9ZJo\x83\xbe\xe3\x1f\xd0\ \x99W\xa9\xaaI\xc7\xf7\x18\x93\x0d\xf8\xd1n\xcd\xea\xc3\ \x86=\xd9L\x8e\xa2\xc27k\x15\x18\x0ajT'H\ \xc4fm\x1fS\xa0a@\x81>\xbfF\xfd-z8\ \xf3\xf6\xfeda\xcf\xe1\xc4\x9e\xa1\xf3J\x14\xcb\xceK\ H\xc6\xd9\xbaa\xde\xd0\xeb\x87d\xba\xaf\x19\x89\xe6J\ \x91\x7f\xbd\xe3*\xf5W\xce*L\xec\x9c\xce;\xe8\xde\ \xac\xc0-\x15*\xf8\x03R\xc6jj\x22\x22o\x96\xdb\ \xdd\xa1\xd0(t\x1f\xf9+\xb4\xaf\xfe\x124\x97\xdc\x04\ \x9e\xc2y\xd0R~\x07tm\xfc\x06\xf4\xd7\xd7@\xd8\ \xd0x=\x13\xab\xf9K\xb4f7b\xab\x19\x0f\xaaa\ xi\x17\xe6+\xa9\xa6\xef\xca\xb4\xf9\xc9\xfd{p\x8d\ \x0aAY\xb5\xe5wB\x14o\xb2\xe2\x16\xd9!\xeb\x02\ \x22\x8e\xba\xafS\x85\x07\xaa\x95\xa8\xde\xb9\x93\x96\xf5B\ \xf6{8#\xe3\x83\xd3*\xe1\xc4\x89\xdb*b-\xf7\ w+pCY\x1a\xfc\xa2\xe4\xbd\x8e\xf4\xc8\x8e\x98=\ g\x97\x19\xcc\xab\xc4\x1e\x1dX\xd7\xad\x1b:\xcf\xb3L\ u\x1ec\x98\xf1\xf3\x1d\xccWk\xf7\xb9\xcc\xe63\xc1\ \x9f\xdf*\xc1hH3\xfb\x80\xe4\x12\xcc\xf5&\xffZ\ F\x15\xb8\xbf\x1a\xed\xb8\xf8|\x1e\xb4wz\x99\x06{\ \xdaC\x84\xb7\x18\xacgg\x82\xf7\xc9\xe0qN\xec\x03\ 9\xb7XMx\xaf\x99X\xd65(o\x90\x1d\xe17\ \x988\xc7!\xf5\xfd\xd0\xf0\xbd\xd0\xf7\xf6\xe3\x9dLw\ \xcc\x15\xfeY\x9b\x0e\xb8\xa3R\x83\xf1\x10\x9bO%\xea\ \xc6s\x09\x96\xac\xb3g_\xd6t\xea\x9fy\xb1V\x86\ \x05\xa5\x8a\xd9\x17\xc0\xf2Y[\xfb\xdd\xb5%*<\xb1\ F\x82wO\xa8\xd0\xe9\xd3\xa2\xf7\xb8\x04\x87\xa4\x88\xfb\ ]\xd9\xa8F\xfb\xd3S\xa8\xd7\xff~:\xe4\x88\xfbh\ \xe7\x15\x13\x1d\xc9\xbd\xbf\x15\xa3\x18\x0f\xc9\xf0\xab\xddA\ \xc8/V\xa3\xe2\x0f,\xde\xac\x11}\xae\x80_\xd6\xc0\ >\xb7\xcc\x09\xfa ]\x88\xd5-\x01Y\x87M\x0d#\ \xf0\xad\xea\x1eX\xbcZ\x85;W\xe9pw\x8d\x06O\ \xd5\x8c\xc3\x87\xfb\xfa``\x5c\xa7\xbd0\x22\x11\xeb\xbe\ %\x7f\x12\xec}V5\xa1l\xf2^L\xa9\xe40\xc4\ \xeeZZ'\xc1\xe50\xb0\xc8\xcc\x1fR4\xf8\xd1\xb6\ ]\xcf<\x9b\x8f\x19\xebG\xd1f\xbfw\x95\x06!\ \x22\xe7`\xd3U\xb9\x8e\x08\xd7\x01\xd4v\xb2\xd5\x10#\ oC\xdb[!\xf6\xb8\xa2\x1b\xcc\x06\x8fy.R\xd5\ S_\xbc\xef\xd6v\x0d\xae+\xb6\xe5\xbb\xa4\x90\xc3l\ \xeaHm\xfd\x9aS\x81\xd7\xe8\x0bJ\xf0\xdb=\xe3D\ \x9f3\xbfr\xac\xdexa\x8b\x04>I6{\x8f\x89\ \x1ep\xb9\x0e;w\x8f~\x85\x0b\xe6\x87G\xef\xc1\xa9\ \xd1\xeb\xe2o\xd6\x0fcLNcz(\x05\xfeu1\ g\x0f\xfb\x1b\xb4\xf4\x0e\xe5\xf4\xfd\x14\x5c\x12\xe3'?\ \xda\xe6\xa79DV,\xce\xe20\x9f#\xb6\xd9HP\ 1c\xb0N\xca\xfd\xbf\x5c\x80k\x8d\xdc\xf1\xab\x1bd\ \xcb\xff\x9b$G\x17\xbc\xff\xa9\x0d\xe8\xe7\xc8\x8e\x1ed\ \x89\x02\xaf\xcd\x17\x92\xe0\xd7\xbb|\xcc\xbe\x8f\xf1\xdbb\ \xdd\xf4\xd77\xcb0.)Q\xfb\xb3\x8b\xa9\x85]\xb7\ \x944\x88\x1c\x8d\xe4y\x8b\xb0\xa5\xd7{\xc6s~\xae\ \x0b\xfa\xc6\xbf\xbf-D\xd6N\x8f\xf2+\x88\xf5xx\ \x8d\xc8\x85\x13\xfe\x9e\xcb\x83\xb78\x15\xb8\xf6!\xc5\x80\ /\xad\xb5\xe5\xec&X/8\x9b\xe7y<Mt\xba\ _\xcd\xcd{*\xb8\xb6\x9f\xe8\xea\xdf\xee\x0dP}\x9e\ o\xdb\x0f\x85\xcc?G\xf4\xb9/$\x9b\xf6\x98\x8b\xcc\ C\xec\xab\x87\xfb0\x8fE\xa7<3a\xfdN\xe3\xb9\ :\xec\xee\xca]\x9b\x14\xaf\x0b9\xc9\x0f\xb7\x05X\x1f\ \xb7\x18]\x8e\x9f?\xbcV\x82!\xbfb\xf6\xc1qe\ \xdd!\xe0\xb2\x8e>\x9f\x82z\x0d\x16\x94\x0a\xdf\xe3\xc5\ e\xde\xec\xc5f\xeba\x82\xb3\xb0K\xeb\xe4\x9c\xf3\xa7\ \xd9\xf9\x1e\xfa\xcf\xff}\x97\x1f\xe6\x0a\x19\xb7\xe5\xf7\xa3\ O\xfd\x1bD\x9f\x8f\x04\xa5(\xffZ\xa21\x10\x17\xe9\ \x83nD\xe8|\xd4\x1b\xf8\xfc\xa5\xfc\x8b\xd4\x0e\x8a:\ \x99Y<\xb7\x11\xefsE\xa3BsyS\x91\xf3\xed\ $\x98\xf9-\x8a\x0a?\xde\x1e\xa0=\xe9b\xfd\xe7\xb8\ \x1e\x8f\xadQ\xa0g\x5c\xc9\xaa~\xa3\x97+\x22\x5c\xbf\ \x1f\xebS\xe1\xeb\x9bU\x98\x7f\x91~e\xf9|\x86*\ \xf64{~\x8b\x0cg\x87u\xd3\x16\x0d\x87s\xcb\x06\ \xc3k\x09\xc8\x0a\xfcn\xcf8\x8b\xf9_\xc0g\xf5\xb5\ \x0d2\x8c\x06eS\xce\x9d:C\xd6\x05C\xc4\xd6C\ 6\xa0\x86\xe1\xc4\x80\x06\x7f\xde?\x06\xf7V\xc94F\ 2\xa7\x88\xe5\xec]S\xaa\x92\xaf\xe9\xf0\xd6\x91 \x9c\ \x1e\xd2i>_\xaa\xe3\xeeN\x00\x8b\xdb\x19T\xce\xbf\ ]+\x915\xd0\xa3\xfc\xaa\xe8\xbb\xc2\xbd\xef\xcbke\ \xe8#\xfa<\xd7\xae\xffr\x81\xe8\xf3Ok\xda4\x1d\ \xc6B\x1a\xb4\x8eH\xf4\xc0\x8f\x15\xde+\x03r\xf8\x9e\ \xa2\x8f0Dx\xcb\xef\xf71\xde\x12\x9b\x1f\x87\xfe\x97\ \xa7\xd7\x13;t<\x18\x95\xa3\xe4\x22\xbb\x10\x9d\xb3\x13\ 1g@D}-\xc7e}\x9cp\x92\x97v0\xbf\ \x22\xb5\xd5cb\xffO\xafW\xa0\x17\xf9yX\xc4\xfd\ ]Yw\x91=\x10\xfa\x19\xed\xd0_\xed\x09\xd0\xfc\xf3\ <\xbb\xff\xbc\x90\xe9\xf3\xc7\xd7)0\x12\x08\xb9|\xc5\ E\xd6\x82\xca\xb9\xac\x12}\x1e\xb2\xc5\xfdmv(\xf9\ \xfc\x19\xc2[\xda\xc7d\xb3\x9f\x86\xebsq\xe1D\xd8\ \xb9\xd7\x85\xeci\x85\xd8&\xbf\xdf\x1f4\xeb\x9e\xcd\xde\ 8\x18\x1f%\xc7\x13\xebd\x18\xf6Kf\xdc\xdf\xe5-\ .\x9c\x8a\xe8\x9a$\x96#M\xebL\x89\x8e\xc6\xb8\xff\ \xcb;C\xac_t\xa1\x1e\x15K\xc8#\xaf\xcfn\x94\ \xa1{\x5c\xbel\xf2q]d70~0\x22\x1b0\ \x10\x0c\xc3\x80\x14\x86\xd1\xa0J\xe3_\xd8_\xe7\xf7\xfb\ \xfc\xc0\xeaV\xd4(\x8e\x8e:\x1ek\x84Yo\x12\xab\ \xf6\xc0\x85\x0b' \xd6w\x84u\xe8\x1b\xdbu\xf8\xb7\ \xdd\x0a<R\x1d\x82\xbb*Uz<Z1\x06/\xef\ \x96\xe0\xd5\xdd\x01\xb3\x17\x94\xf0\xa1\x8b\xbeE_[/\ C\xafO\xb6\xf2\x14s\xdc\xff\xe4\x22\xbb`\xfaG\xb9\ \x1e\xff\xd9\xce\x909\x9f(\x9f\xd7\x93\xd8g\x11E\xd7\ YX\xfa\xfc\xfej\x19\x86\xc6%\x1eo\x08\xbb\x22\xee\ \xc2q\x88p\x9e1*\xe9\xf0\x9d\xed\xaa\x99\x9fo\xca\ 3\xfa\xca\x0b\xad\xcf\xf3\x8al\xf5\xed\x5c\xd6\xffy\xa3\ \x04\xddc\xd2y=[\x5c\xb8p\x12D\xfd\xeeGG\ Y\x0d\xe8\xe4s\xf0\x99\xee_\xbcJ%\xfa<\xe4\xe6\ q\xb9p<P\xd6\x1b\x87$\x9e\xbf\x16\x9f\xac__\ \xaaB\xd7\xa8\x14\xd5C\xc6\x85\x0b\xa7\x02u\xfa\x1f\x8f\ \x5c8'\xf1Ru\x84\xe8\x8bY{f\xc0\xdc\x1b\x5c\ Yw\xe1d\xa0\x1f\xf1\xd9MJ\xdc\xb3\xaaD\xec\xe8\ \xbb;\x9c3\x7f\xcb\x85\x8b\x8b!\xa4\x1a\xf0\xe4\xba\x04\ g+\x16b|\xd4\xb9\xf3\xa1]\xb8\xb0\x03{*>\ \xbd\x81\xcf\xda\x8c\xab?\xacJ{\xc4>W\x9b\x1d=\ \xcf]\xb8\xc0Y!?\xd8\xa9\xf1\x9a\xd8\xf8\xf4:>\ \x1b\xefok\xc9\xf4%\xb8p1)\xa0=\xb9\xb3\x8b\ \xcd\x1a\x8aW\xd6\x17\x96j\xb4n\xd4\x85\x8bl\x00\xad\ \x9d\xd5\x0dx\xb1V\xe2\xf3*'\xe6\xedy\xdc\xff\x22\ z#c>\xafSg\x88\xbap\x11\x0b\x91\xcf\xe8\x19\ \xd6\xe0\xb3U\x0a\xcd[\x9c\xb8\xc7\x93\x88\xa5jp\xdb\ \x0a\x05\x0e\xf7\x1b\x8e\x9e\xc1\xe5\xc2\x85\x1d\x22G\x00\xfd\ \xe3\xc7z$\xb8q\xf9E\xfc1T\xa7+4VZ\ ?\xac\xa5\xac\x17\xb2\x0b\x17S\x0d\x94{\x9cS\xfc\xbb\ \xfd2|\xbeF\x8f\x9e9\x8fs\x8b\xaad\xf8\xc5~\ \x05\xda\x87%7n\xe4\x22\xab!\xf8\x0c\xf6>\xe8\xf4\ \x1bp\xa0O\x87\xda\x0e\x03\xb6t\x1ap\xb0\xcf\xa0\xb3\ Jt\xc3\xea\x17\xed\xc2E\xb6\x22\x12\x93\xa7\x88\xff\x04\ O1s^\xc2\xd6\x0c)\x17.\x5c\xb8p\xe1\xc2\x85\ \x0b\x17\xe9\xc1\x15.\x5c\xb8p\xe1\xc2\x85\x0b\x17.\x5c\ \xb8p\xe1\xc2\x85\x0b\x17.\x5c\xd8\xf0\xff\x01S\xd9\xcc\ \xc2\ \x00\x00\x01\xe9\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22?><!DOCTYPE s\ vg PUBLIC \x22-//W3\ C//DTD SVG 1.1//\ EN\x22 \x22http://www.\ w3.org/Graphics/\ SVG/1.1/DTD/svg1\ 1.dtd\x22><svg xmln\ s=\x22http://www.w3\ .org/2000/svg\x22 x\ mlns:xlink=\x22http\ ://www.w3.org/19\ 99/xlink\x22 versio\ n=\x221.1\x22 width=\x222\ 4\x22 height=\x2224\x22 v\ iewBox=\x220 0 24 2\ 4\x22><path d=\x22M17 \ 3H5C3.89 3 3 3.9\ 3 5V19C3 20.1 3\ .89 21 5 21H19C2\ 0.1 21 21 20.1 2\ 1 19V7L17 3M19 1\ 9H5V5H16.17L19 7\ .83V19M12 12C10.\ 34 12 9 13.34 9 \ 15S10.34 18 12 1\ 8 15 16.66 15 15\ 13.66 12 12 12M\ 6 6H15V10H6V6Z\x22 \ /></svg>\ \x00\x00\x0d\x03\ <\ ?xml version=\x221.\ 0\x22 encoding=\x22UTF\ -8\x22 standalone=\x22\ no\x22?>\x0a<svg\x0a ve\ rsion=\x221.1\x22\x0a w\ idth=\x2224\x22\x0a hei\ ght=\x2224\x22\x0a view\ Box=\x220 0 24 24\x22\x0a\ id=\x22svg4\x22\x0a \ sodipodi:docname\ =\x22folder-open-ou\ tline-dna_mdi.sv\ g\x22\x0a inkscape:v\ ersion=\x221.1.1 (3\ bf5ae0d25, 2021-\ 09-20)\x22\x0a xmlns\ :inkscape=\x22http:\ //www.inkscape.o\ rg/namespaces/in\ kscape\x22\x0a xmlns\ :sodipodi=\x22http:\ //sodipodi.sourc\ eforge.net/DTD/s\ odipodi-0.dtd\x22\x0a \ xmlns=\x22http://\ www.w3.org/2000/\ svg\x22\x0a xmlns:sv\ g=\x22http://www.w3\ .org/2000/svg\x22>\x0a\ <defs\x0a id=\ \x22defs8\x22 />\x0a <so\ dipodi:namedview\ \x0a id=\x22namedv\ iew6\x22\x0a pagec\ olor=\x22#ffffff\x22\x0a \ bordercolor=\ \x22#666666\x22\x0a b\ orderopacity=\x221.\ 0\x22\x0a inkscape\ :pageshadow=\x222\x22\x0a\ inkscape:pa\ geopacity=\x220.0\x22\x0a\ inkscape:pa\ gecheckerboard=\x22\ 0\x22\x0a showgrid\ =\x22false\x22\x0a in\ kscape:zoom=\x2234.\ 666667\x22\x0a ink\ scape:cx=\x227.6298\ 077\x22\x0a inksca\ pe:cy=\x2214.4375\x22\x0a\ inkscape:wi\ ndow-width=\x221920\ \x22\x0a inkscape:\ window-height=\x221\ 017\x22\x0a inksca\ pe:window-x=\x22-8\x22\ \x0a inkscape:w\ indow-y=\x22-8\x22\x0a \ inkscape:windo\ w-maximized=\x221\x22\x0a\ inkscape:cu\ rrent-layer=\x22svg\ 4\x22 />\x0a <path\x0a \ id=\x22path2\x22\x0a \ d=\x22M 3.019531\ 2 1.1054688 A 2 \ 2 0 0 0 1.019531\ 2 3.1054688 L 1.\ 0195312 15.10546\ 9 A 2 2 0 0 0 3.\ 0195312 17.10546\ 9 L 9.1445312 17\ .105469 L 9.1445\ 312 15.105469 L \ 5.0195312 15.105\ 469 L 6.6191406 \ 9.1054688 L 19.6\ 19141 9.1054688 \ L 18.939453 11.6\ 54297 L 20.98828\ 1 11.654297 L 22\ .21875 7.1054688\ L 5.1191406 7.1\ 054688 L 3.01953\ 12 15.105469 L 3\ .0195312 5.10546\ 88 L 20.019531 5\ .1054688 A 2 2 0\ 0 0 18.019531 3\ .1054688 L 11.01\ 9531 3.1054688 L\ 9.0195312 1.105\ 4688 L 3.0195312\ 1.1054688 z \x22 /\ >\x0a <path\x0a i\ d=\x22path18\x22\x0a \ style=\x22stroke-wi\ dth:0.603887\x22\x0a \ d=\x22m 10.05859\ 4,12.818359 v 1.\ 00586 c 0,1.0549\ 12 0.776328,1.91\ 8934 1.814453,2.\ 617187 0.588029,\ 0.396848 1.28520\ 6,0.757182 2.003\ 906,1.09375 L 15\ .205078,16.9375 \ c -0.871154,-0.3\ 96849 -1.8288,-0\ .885793 -2.45312\ 5,-1.302734 0,0 \ 0.04152,0.0293 0\ ,0 h -0.02344 v \ -0.01758 c -0.77\ 673,-0.553241 -1\ .05543,-1.11269 \ -1.045673,-1.792\ 969 l 0.01443,-1\ .00586 z m 9.931\ 339,0 1e-6,1.034\ 706 c 0,0.687102\ -0.218218,1.223\ 105 -1.009465,1.\ 781701 h -0.002 \ c -0.04135,0.029\ 17 -0.07983,0.05\ 873 -0.123047,0.\ 08789 -0.878413,\ 0.587737 -2.1123\ 8,1.134646 -3.38\ 2813,1.697266 -1\ .270432,0.557597\ -2.576002,1.130\ 991 -3.599609,1.\ 824219 -1.038125\ ,0.698253 -1.814\ 453,1.562273 -1.\ 814453,2.617187 \ v 1.003906 l 1.7\ 2521,-0.01442 4e\ -6,-0.97506 c 3e\ -6,-0.723369 0.2\ 20042,-1.327079 \ 1.091196,-1.9148\ 13 0.878413,-0.5\ 87742 2.112379,-\ 1.134646 3.38281\ 2,-1.697266 1.27\ 0434,-0.557598 2\ .576004,-1.12903\ 7 3.59961,-1.822\ 266 1.038124,-0.\ 698253 1.816406,\ -1.562275 1.8164\ 06,-2.617187 v -\ 1.00586 z m -2.1\ 34464,5.330079 -\ 1.330078,0.59765\ 6 c 0.871154,0.3\ 96848 1.705751,0\ .797899 2.330078\ ,1.214844 0.0289\ 3,0.01952 0.0539\ 3,0.03906 0.0820\ 3,0.05859 h 0.06\ 445 v 0.04687 c \ 0.776774,0.55339\ 6 0.892039,0.869\ 797 0.915866,1.5\ 4973 l -3e-6,1.2\ 49098 1.756009,-\ 3e-6 V 21.86132 \ c 0,-1.054914 -0\ .778282,-1.91893\ 4 -1.816406,-2.6\ 17187 -0.58803,-\ 0.396847 -1.2832\ 51,-0.759138 -2.\ 001953,-1.095703\ z\x22\x0a sodipod\ i:nodetypes=\x22csc\ ccccccscccsccccc\ sccscccscccccccc\ sccsccc\x22 />\x0a <r\ ect\x0a style=\x22\ fill:#000000\x22\x0a \ id=\x22rect2651-\ 5\x22\x0a width=\x226\ .4471159\x22\x0a h\ eight=\x221.3990391\ \x22\x0a x=\x2212.612\ 979\x22\x0a y=\x2220.\ 646629\x22\x0a ry=\ \x220.57692307\x22 />\x0a\ <rect\x0a sty\ le=\x22fill:#000000\ \x22\x0a id=\x22rect2\ 651-5-8\x22\x0a wi\ dth=\x226.4471159\x22\x0a\ height=\x221.3\ 990391\x22\x0a x=\x22\ 12.526443\x22\x0a \ y=\x2213.492789\x22\x0a \ ry=\x220.5769230\ 7\x22\x0a rx=\x220.57\ 692307\x22 />\x0a</svg\ >\x0a\ \x00\x00\x01\x8c\ \x89\ PNG\x0d\x0a\x1a\x0a\x00\x00\x00\x0dIHDR\x00\ \x00\x00\x18\x00\x00\x00\x18\x08\x06\x00\x00\x00\xe0w=\xf8\ \x00\x00\x00\x09pHYs\x00\x00\x0e\xc3\x00\x00\x0e\xc3\ \x01\xc7o\xa8d\x00\x00\x00\x19tEXtSof\ tware\x00www.inksca\ pe.org\x9b\xee<\x1a\x00\x00\x01\x19ID\ ATH\x89\xed\x94=N\xc3@\x10F\xdfDH)\ S\xd2\x84\xce\x1c$-\x07\xa0@\x89 \x8d\x8f@*\ \xc8\x9a[\xa4\x85\xc6\x90\x03p\x13\x0ab*6\x0d%\ \xb5\xf1P,X\xd8\xf87\xb6h\xc8\xd7yw\xf6=\ \xcdx\xb5\xb0OM$\xf3e^N\x90d\x05\x8cw\ \xe4Y\x10\x9f\xa5\xf7\xf8\xbd0\xc8\xea:\xc1qgu\ \xf5sa\xf0\xbb\xa0s\x8e\xaa\x04\xbd\xa7\xbb@5D\ \x19\x01w\xfd\x0bTC\xd8\xce0\xc7\xef\xa0IQ\xc9\ A\x07\xfc\xda\xc1'1f\x13\x00\xe7EE\xbbv\xb0\ F\xedY\x0a\x17\xae\xcb\x0a\x8b\x05\xca-\xca\xc8\x8d \ \xbf\xa7aSx\xb9\x00I\xdc\x5c\xb7\xb3\x9c\xe4>3\ \x96\x1a8\x94\xfd\x03\xd1\x0bL\xf4\x86\xf1\x16<\xe8\x94\ \xa7(F\x18\xa2v\x8a\x99\xc4\x04\x1b\x03\xf5\xf0\x8a\x0e\ \x00\xd1K\x82\xe8\x86S\xf9\x00;O\xc7\xe2\xe0\xcb&\ \xf0\xf2\x0e\xd2\xe8\x15&\x1ab\xbc\x05@[x\x03\x01\ _\x9d<\x1f\x82$\xc0\xbc\x0d\xbc\x99\xc0Y\x0a\xefx\ \x93\xfc\xf9[d{`\xbeV\x08\xc4\xcf\x17\xb4\x86\xab\ \xf8\x1d\xce\xff\xc7|\x02\xfeFi\xc4QH\xcf\xf9\x00\ \x00\x00\x00IEND\xaeB`\x82\ \x00\x00\x01`\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M20 3h-1V2c\ 0-.55-.45-1-1-1s\ -1 .45-1 1v1H7V2\ c0-.55-.45-1-1-1\ s-1 .45-1 1v1H4c\ -1.1 0-2 .9-2 2v\ 16c0 1.1.9 2 2 2\ h16c1.1 0 2-.9 2\ -2V5c0-1.1-.9-2-\ 2-2zm-1 18H5c-.5\ 5 0-1-.45-1-1V8h\ 16v12c0 .55-.45 \ 1-1 1z\x22/></svg>\ \x00\x00\x01\x0c\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M8.71 12.29\ L11.3 9.7c.39-.3\ 9 1.02-.39 1.41 \ 0l2.59 2.59c.63.\ 63.18 1.71-.71 1\ .71H9.41c-.89 0-\ 1.33-1.08-.7-1.7\ 1z\x22/></svg>\ \x00\x00\x01\x0c\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M8.71 11.71\ l2.59 2.59c.39.3\ 9 1.02.39 1.41 0\ l2.59-2.59c.63-.\ 63.18-1.71-.71-1\ .71H9.41c-.89 0-\ 1.33 1.08-.7 1.7\ 1z\x22/></svg>\ \x00\x00\x01Q\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#4\ d5157\x22><g><rect \ fill=\x22none\x22 fill\ -rule=\x22evenodd\x22 \ height=\x2224\x22 widt\ h=\x2224\x22/><g><path\ d=\x22M13,5v14c0,0\ .6-0.4,1-1,1l0,0\ c-0.6,0-1-0.4-1-\ 1V5c0-0.5,0.4-1,\ 1-1l0,0C12.6,4,1\ 3,4.4,13,5z\x22 fil\ l-rule=\x22evenodd\x22\ /></g></g></svg>\ \ \x00\x00\x02\xc9\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#4\ d5157\x22><g><path \ d=\x22M0,0h24v24H0V\ 0z\x22 fill=\x22none\x22/\ ></g><g><g><path\ d=\x22M12,5V3.21c0\ -0.45-0.54-0.67-\ 0.85-0.35L8.35,5\ .65c-0.2,0.2-0.2\ ,0.51,0,0.71l2.7\ 9,2.79C11.46,9.4\ 6,12,9.24,12,8.7\ 9V7 c3.31,0,6,2.\ 69,6,6c0,2.72-1.\ 83,5.02-4.31,5.7\ 5C13.27,18.87,13\ ,19.27,13,19.7v0\ c0,0.65,0.62,1.1\ 6,1.25,0.97 C17.\ 57,19.7,20,16.64\ ,20,13C20,8.58,1\ 6.42,5,12,5z\x22/><\ path d=\x22M6,13c0-\ 1.34,0.44-2.58,1\ .19-3.59c0.3-0.4\ ,0.26-0.95-0.09-\ 1.31l0,0C6.68,7.\ 68,5.96,7.72,5.6\ ,8.2C4.6,9.54,4,\ 11.2,4,13 c0,3.6\ 4,2.43,6.7,5.75,\ 7.67C10.38,20.86\ ,11,20.35,11,19.\ 7v0c0-0.43-0.27-\ 0.83-0.69-0.95C7\ .83,18.02,6,15.7\ 2,6,13z\x22/></g></\ g></svg>\ \x00\x00\x00\xfb\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M4.41 22H20\ c1.1 0 2-.9 2-2V\ 4.41c0-.89-1.08-\ 1.34-1.71-.71L3.\ 71 20.29c-.63.63\ -.19 1.71.7 1.71\ z\x22/></svg>\ \x00\x00\x02\x98\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#4\ d5157\x22><g><rect \ fill=\x22none\x22 heig\ ht=\x2224\x22 width=\x222\ 4\x22/><rect fill=\x22\ none\x22 height=\x2224\ \x22 width=\x2224\x22/></\ g><g><g><path d=\ \x22M20.08,11.42l-4\ .04-5.65C15.7,5.\ 29,15.15,5,14.56\ ,5h0c-1.49,0-2.3\ 5,1.68-1.49,2.89\ L16,12l-2.93,4.1\ 1 c-0.87,1.21,0,\ 2.89,1.49,2.89h0\ c0.59,0,1.15-0.2\ 9,1.49-0.77l4.04\ -5.65C20.33,12.2\ 3,20.33,11.77,20\ .08,11.42z\x22/><pa\ th d=\x22M13.08,11.\ 42L9.05,5.77C8.7\ ,5.29,8.15,5,7.5\ 6,5h0C6.07,5,5.2\ ,6.68,6.07,7.89L\ 9,12l-2.93,4.11C\ 5.2,17.32,6.07,1\ 9,7.56,19h0 c0.5\ 9,0,1.15-0.29,1.\ 49-0.77l4.04-5.6\ 5C13.33,12.23,13\ .33,11.77,13.08,\ 11.42z\x22/></g></g\ ></svg>\ \x00\x00\x00\xfb\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M4.41 22H20\ c1.1 0 2-.9 2-2V\ 4.41c0-.89-1.08-\ 1.34-1.71-.71L3.\ 71 20.29c-.63.63\ -.19 1.71.7 1.71\ z\x22/></svg>\ \x00\x00\x01Z\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#b\ abdc2\x22><g><rect \ fill=\x22none\x22 fill\ -rule=\x22evenodd\x22 \ height=\x2224\x22 widt\ h=\x2224\x22/><g><path\ d=\x22M19,13H5c-0.\ 55,0-1-0.45-1-1v\ 0c0-0.55,0.45-1,\ 1-1h14c0.55,0,1,\ 0.45,1,1v0 C20,1\ 2.55,19.55,13,19\ ,13z\x22 fill-rule=\ \x22evenodd\x22/></g><\ /g></svg>\ \x00\x00\x01\xc2\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M11 18c0 1.\ 1-.9 2-2 2s-2-.9\ -2-2 .9-2 2-2 2 \ .9 2 2zm-2-8c-1.\ 1 0-2 .9-2 2s.9 \ 2 2 2 2-.9 2-2-.\ 9-2-2-2zm0-6c-1.\ 1 0-2 .9-2 2s.9 \ 2 2 2 2-.9 2-2-.\ 9-2-2-2zm6 4c1.1\ 0 2-.9 2-2s-.9-\ 2-2-2-2 .9-2 2 .\ 9 2 2 2zm0 2c-1.\ 1 0-2 .9-2 2s.9 \ 2 2 2 2-.9 2-2-.\ 9-2-2-2zm0 6c-1.\ 1 0-2 .9-2 2s.9 \ 2 2 2 2-.9 2-2-.\ 9-2-2-2z\x22/></svg\ >\ \x00\x00\x01D\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M9 16.17L5.\ 53 12.7c-.39-.39\ -1.02-.39-1.41 0\ -.39.39-.39 1.02\ 0 1.41l4.18 4.1\ 8c.39.39 1.02.39\ 1.41 0L20.29 7.\ 71c.39-.39.39-1.\ 02 0-1.41-.39-.3\ 9-1.02-.39-1.41 \ 0L9 16.17z\x22/></s\ vg>\ \x00\x00\x01D\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M11.29 8.71\ L6.7 13.3c-.39.3\ 9-.39 1.02 0 1.4\ 1.39.39 1.02.39 \ 1.41 0L12 10.83l\ 3.88 3.88c.39.39\ 1.02.39 1.41 0 \ .39-.39.39-1.02 \ 0-1.41L12.7 8.71\ c-.38-.39-1.02-.\ 39-1.41 0z\x22/></s\ vg>\ \x00\x00\x01h\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M20 6h-8l-1\ .41-1.41C10.21 4\ .21 9.7 4 9.17 4\ H4c-1.1 0-1.99.9\ -1.99 2L2 18c0 1\ .1.9 2 2 2h16c1.\ 1 0 2-.9 2-2V8c0\ -1.1-.9-2-2-2zm-\ 1 12H5c-.55 0-1-\ .45-1-1V9c0-.55.\ 45-1 1-1h14c.55 \ 0 1 .45 1 1v8c0 \ .55-.45 1-1 1z\x22/\ ></svg>\ \x00\x00\x01\x1b\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M12 2C6.48 \ 2 2 6.48 2 12s4.\ 48 10 10 10 10-4\ .48 10-10S17.52 \ 2 12 2zm0 18c-4.\ 42 0-8-3.58-8-8s\ 3.58-8 8-8 8 3.5\ 8 8 8-3.58 8-8 8\ z\x22/></svg>\ \x00\x00\x01\xcc\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#eff\ 1f1\x22><path d=\x22M1\ 8,13c1.1,0,2,0.9\ ,2,2s-0.9,2-2,2s\ -2-0.9-2-2S16.9,\ 13,18,13z M10,15\ c0,1.1,0.9,2,2,2\ s2-0.9,2-2s-0.9-\ 2-2-2S10,13.9,10\ ,15z M4,15c0,1.1\ ,0.9,2,2,2s2-0.9\ ,2-2s-0.9-2-2-2S\ 4,13.9,4,15z M8,\ 9c0-1.1-0.9-2-2-\ 2S4,7.9,4,9s0.9,\ 2,2,2S8,10.1,8,9\ z M10,9 c0,1.1,0\ .9,2,2,2s2-0.9,2\ -2s-0.9-2-2-2S10\ ,7.9,10,9z M16,9\ c0,1.1,0.9,2,2,2\ s2-0.9,2-2s-0.9-\ 2-2-2S16,7.9,16,\ 9z\x22/></svg>\ \x00\x00\x01d\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M5 13h11.17\ l-4.88 4.88c-.39\ .39-.39 1.03 0 1\ .42.39.39 1.02.3\ 9 1.41 0l6.59-6.\ 59c.39-.39.39-1.\ 02 0-1.41l-6.58-\ 6.6c-.39-.39-1.0\ 2-.39-1.41 0-.39\ .39-.39 1.02 0 1\ .41L16.17 11H5c-\ .55 0-1 .45-1 1s\ .45 1 1 1z\x22/></s\ vg>\ \x00\x00\x01a\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M13 19V7.83\ l4.88 4.88c.39.3\ 9 1.03.39 1.42 0\ .39-.39.39-1.02\ 0-1.41l-6.59-6.\ 59c-.39-.39-1.02\ -.39-1.41 0l-6.6\ 6.58c-.39.39-.3\ 9 1.02 0 1.41.39\ .39 1.02.39 1.41\ 0L11 7.83V19c0 \ .55.45 1 1 1s1-.\ 45 1-1z\x22/></svg>\ \ \x00\x00\x01<\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M18 19H6c-.\ 55 0-1-.45-1-1V6\ c0-.55.45-1 1-1h\ 12c.55 0 1 .45 1\ 1v12c0 .55-.45 \ 1-1 1zm1-16H5c-1\ .1 0-2 .9-2 2v14\ c0 1.1.9 2 2 2h1\ 4c1.1 0 2-.9 2-2\ V5c0-1.1-.9-2-2-\ 2z\x22/></svg>\ \x00\x00\x01D\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M11.29 8.71\ L6.7 13.3c-.39.3\ 9-.39 1.02 0 1.4\ 1.39.39 1.02.39 \ 1.41 0L12 10.83l\ 3.88 3.88c.39.39\ 1.02.39 1.41 0 \ .39-.39.39-1.02 \ 0-1.41L12.7 8.71\ c-.38-.39-1.02-.\ 39-1.41 0z\x22/></s\ vg>\ \x00\x00\x01\x95\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M19 3H5c-1.\ 1 0-2 .9-2 2v14c\ 0 1.1.9 2 2 2h14\ c1.1 0 2-.9 2-2V\ 5c0-1.1-.9-2-2-2\ zm-8.29 13.29c-.\ 39.39-1.02.39-1.\ 41 0L5.71 12.7c-\ .39-.39-.39-1.02\ 0-1.41.39-.39 1\ .02-.39 1.41 0L1\ 0 14.17l6.88-6.8\ 8c.39-.39 1.02-.\ 39 1.41 0 .39.39\ .39 1.02 0 1.41l\ -7.58 7.59z\x22/></\ svg>\ \x00\x00\x01\x89\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M12 2C6.48 \ 2 2 6.48 2 12s4.\ 48 10 10 10 10-4\ .48 10-10S17.52 \ 2 12 2zM9.29 16.\ 29L5.7 12.7c-.39\ -.39-.39-1.02 0-\ 1.41.39-.39 1.02\ -.39 1.41 0L10 1\ 4.17l6.88-6.88c.\ 39-.39 1.02-.39 \ 1.41 0 .39.39.39\ 1.02 0 1.41l-7.\ 59 7.59c-.38.39-\ 1.02.39-1.41 0z\x22\ /></svg>\ \x00\x00\x01\xcc\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M1\ 8,13c1.1,0,2,0.9\ ,2,2s-0.9,2-2,2s\ -2-0.9-2-2S16.9,\ 13,18,13z M10,15\ c0,1.1,0.9,2,2,2\ s2-0.9,2-2s-0.9-\ 2-2-2S10,13.9,10\ ,15z M4,15c0,1.1\ ,0.9,2,2,2s2-0.9\ ,2-2s-0.9-2-2-2S\ 4,13.9,4,15z M8,\ 9c0-1.1-0.9-2-2-\ 2S4,7.9,4,9s0.9,\ 2,2,2S8,10.1,8,9\ z M10,9 c0,1.1,0\ .9,2,2,2s2-0.9,2\ -2s-0.9-2-2-2S10\ ,7.9,10,9z M16,9\ c0,1.1,0.9,2,2,2\ s2-0.9,2-2s-0.9-\ 2-2-2S16,7.9,16,\ 9z\x22/></svg>\ \x00\x00\x02\x98\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#b\ abdc2\x22><g><rect \ fill=\x22none\x22 heig\ ht=\x2224\x22 width=\x222\ 4\x22/><rect fill=\x22\ none\x22 height=\x2224\ \x22 width=\x2224\x22/></\ g><g><g><path d=\ \x22M20.08,11.42l-4\ .04-5.65C15.7,5.\ 29,15.15,5,14.56\ ,5h0c-1.49,0-2.3\ 5,1.68-1.49,2.89\ L16,12l-2.93,4.1\ 1 c-0.87,1.21,0,\ 2.89,1.49,2.89h0\ c0.59,0,1.15-0.2\ 9,1.49-0.77l4.04\ -5.65C20.33,12.2\ 3,20.33,11.77,20\ .08,11.42z\x22/><pa\ th d=\x22M13.08,11.\ 42L9.05,5.77C8.7\ ,5.29,8.15,5,7.5\ 6,5h0C6.07,5,5.2\ ,6.68,6.07,7.89L\ 9,12l-2.93,4.11C\ 5.2,17.32,6.07,1\ 9,7.56,19h0 c0.5\ 9,0,1.15-0.29,1.\ 49-0.77l4.04-5.6\ 5C13.33,12.23,13\ .33,11.77,13.08,\ 11.42z\x22/></g></g\ ></svg>\ \x00\x00\x01\x18\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#4\ d5157\x22><g><rect \ fill=\x22none\x22 fill\ -rule=\x22evenodd\x22 \ height=\x2224\x22 widt\ h=\x2224\x22/><g><rect\ fill-rule=\x22even\ odd\x22 height=\x2224\x22\ width=\x221\x22 x=\x2211\ \x22 y=\x220\x22/></g></g\ ></svg>\ \x00\x00\x00\xfb\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#2e4\ 65e\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M4.41 22H20\ c1.1 0 2-.9 2-2V\ 4.41c0-.89-1.08-\ 1.34-1.71-.71L3.\ 71 20.29c-.63.63\ -.19 1.71.7 1.71\ z\x22/></svg>\ \x00\x00\x029\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M12 2C6.48 \ 2 2 6.48 2 12s4.\ 48 10 10 10 10-4\ .48 10-10S17.52 \ 2 12 2zm1 17h-2v\ -2h2v2zm2.07-7.7\ 5l-.9.92c-.5.51-\ .86.97-1.04 1.69\ -.08.32-.13.68-.\ 13 1.14h-2v-.5c0\ -.46.08-.9.22-1.\ 31.2-.58.53-1.1.\ 95-1.52l1.24-1.2\ 6c.46-.44.68-1.1\ .55-1.8-.13-.72-\ .69-1.33-1.39-1.\ 53-1.11-.31-2.14\ .32-2.47 1.27-.1\ 2.37-.43.65-.82.\ 65h-.3C8.4 9 8 8\ .44 8.16 7.88c.4\ 3-1.47 1.68-2.59\ 3.23-2.83 1.52-\ .24 2.97.55 3.87\ 1.8 1.18 1.63.8\ 3 3.38-.19 4.4z\x22\ /></svg>\ \x00\x00\x01\xae\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M18.3 5.71c\ -.39-.39-1.02-.3\ 9-1.41 0L12 10.5\ 9 7.11 5.7c-.39-\ .39-1.02-.39-1.4\ 1 0-.39.39-.39 1\ .02 0 1.41L10.59\ 12 5.7 16.89c-.\ 39.39-.39 1.02 0\ 1.41.39.39 1.02\ .39 1.41 0L12 13\ .41l4.89 4.89c.3\ 9.39 1.02.39 1.4\ 1 0 .39-.39.39-1\ .02 0-1.41L13.41\ 12l4.89-4.89c.3\ 8-.38.38-1.02 0-\ 1.4z\x22/></svg>\ \x00\x00\x01\xc2\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#eff\ 1f1\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M11 18c0 1.\ 1-.9 2-2 2s-2-.9\ -2-2 .9-2 2-2 2 \ .9 2 2zm-2-8c-1.\ 1 0-2 .9-2 2s.9 \ 2 2 2 2-.9 2-2-.\ 9-2-2-2zm0-6c-1.\ 1 0-2 .9-2 2s.9 \ 2 2 2 2-.9 2-2-.\ 9-2-2-2zm6 4c1.1\ 0 2-.9 2-2s-.9-\ 2-2-2-2 .9-2 2 .\ 9 2 2 2zm0 2c-1.\ 1 0-2 .9-2 2s.9 \ 2 2 2 2-.9 2-2-.\ 9-2-2-2zm0 6c-1.\ 1 0-2 .9-2 2s.9 \ 2 2 2 2-.9 2-2-.\ 9-2-2-2z\x22/></svg\ >\ \x00\x00\x01`\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M20 3h-1V2c\ 0-.55-.45-1-1-1s\ -1 .45-1 1v1H7V2\ c0-.55-.45-1-1-1\ s-1 .45-1 1v1H4c\ -1.1 0-2 .9-2 2v\ 16c0 1.1.9 2 2 2\ h16c1.1 0 2-.9 2\ -2V5c0-1.1-.9-2-\ 2-2zm-1 18H5c-.5\ 5 0-1-.45-1-1V8h\ 16v12c0 .55-.45 \ 1-1 1z\x22/></svg>\ \x00\x00\x01\x18\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#b\ abdc2\x22><g><rect \ fill=\x22none\x22 fill\ -rule=\x22evenodd\x22 \ height=\x2224\x22 widt\ h=\x2224\x22/><g><rect\ fill-rule=\x22even\ odd\x22 height=\x2224\x22\ width=\x221\x22 x=\x2211\ \x22 y=\x220\x22/></g></g\ ></svg>\ \x00\x00\x01:\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M12 2C6.48 \ 2 2 6.48 2 12s4.\ 48 10 10 10 10-4\ .48 10-10S17.52 \ 2 12 2zm0 18c-4.\ 42 0-8-3.58-8-8s\ 3.58-8 8-8 8 3.5\ 8 8 8-3.58 8-8 8\ z\x22/><circle cx=\x22\ 12\x22 cy=\x2212\x22 r=\x225\ \x22/></svg>\ \x00\x00\x01\x0c\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M8.71 11.71\ l2.59 2.59c.39.3\ 9 1.02.39 1.41 0\ l2.59-2.59c.63-.\ 63.18-1.71-.71-1\ .71H9.41c-.89 0-\ 1.33 1.08-.7 1.7\ 1z\x22/></svg>\ \x00\x00\x01d\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M19 11H7.83\ l4.88-4.88c.39-.\ 39.39-1.03 0-1.4\ 2-.39-.39-1.02-.\ 39-1.41 0l-6.59 \ 6.59c-.39.39-.39\ 1.02 0 1.41l6.5\ 9 6.59c.39.39 1.\ 02.39 1.41 0 .39\ -.39.39-1.02 0-1\ .41L7.83 13H19c.\ 55 0 1-.45 1-1s-\ .45-1-1-1z\x22/></s\ vg>\ \x00\x00\x01d\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M19 11H7.83\ l4.88-4.88c.39-.\ 39.39-1.03 0-1.4\ 2-.39-.39-1.02-.\ 39-1.41 0l-6.59 \ 6.59c-.39.39-.39\ 1.02 0 1.41l6.5\ 9 6.59c.39.39 1.\ 02.39 1.41 0 .39\ -.39.39-1.02 0-1\ .41L7.83 13H19c.\ 55 0 1-.45 1-1s-\ .45-1-1-1z\x22/></s\ vg>\ \x00\x00\x01W\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M2\ 4 24H0V0h24v24z\x22\ fill=\x22none\x22 opa\ city=\x22.87\x22/><pat\ h d=\x22M15.88 9.29\ L12 13.17 8.12 9\ .29c-.39-.39-1.0\ 2-.39-1.41 0-.39\ .39-.39 1.02 0 1\ .41l4.59 4.59c.3\ 9.39 1.02.39 1.4\ 1 0l4.59-4.59c.3\ 9-.39.39-1.02 0-\ 1.41-.39-.38-1.0\ 3-.39-1.42 0z\x22/>\ </svg>\ \x00\x00\x01*\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#008\ 1db\x22><path d=\x22M0\ 0h24v24H0z\x22 fil\ l=\x22none\x22/><path \ d=\x22M19 3H5c-1.1 \ 0-2 .9-2 2v14c0 \ 1.1.9 2 2 2h14c1\ .1 0 2-.9 2-2V5c\ 0-1.1-.9-2-2-2zm\ -3 10H8c-.55 0-1\ -.45-1-1s.45-1 1\ -1h8c.55 0 1 .45\ 1 1s-.45 1-1 1z\ \x22/></svg>\ \x00\x00\x01*\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0z\x22 fil\ l=\x22none\x22/><path \ d=\x22M19 3H5c-1.1 \ 0-2 .9-2 2v14c0 \ 1.1.9 2 2 2h14c1\ .1 0 2-.9 2-2V5c\ 0-1.1-.9-2-2-2zm\ -3 10H8c-.55 0-1\ -.45-1-1s.45-1 1\ -1h8c.55 0 1 .45\ 1 1s-.45 1-1 1z\ \x22/></svg>\ \x00\x00\x01a\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M13 19V7.83\ l4.88 4.88c.39.3\ 9 1.03.39 1.42 0\ .39-.39.39-1.02\ 0-1.41l-6.59-6.\ 59c-.39-.39-1.02\ -.39-1.41 0l-6.6\ 6.58c-.39.39-.3\ 9 1.02 0 1.41.39\ .39 1.02.39 1.41\ 0L11 7.83V19c0 \ .55.45 1 1 1s1-.\ 45 1-1z\x22/></svg>\ \ \x00\x00\x01G\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M9.29 6.71c\ -.39.39-.39 1.02\ 0 1.41L13.17 12\ l-3.88 3.88c-.39\ .39-.39 1.02 0 1\ .41.39.39 1.02.3\ 9 1.41 0l4.59-4.\ 59c.39-.39.39-1.\ 02 0-1.41L10.7 6\ .7c-.38-.38-1.02\ -.38-1.41.01z\x22/>\ </svg>\ \x00\x00\x01W\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M2\ 4 24H0V0h24v24z\x22\ fill=\x22none\x22 opa\ city=\x22.87\x22/><pat\ h d=\x22M15.88 9.29\ L12 13.17 8.12 9\ .29c-.39-.39-1.0\ 2-.39-1.41 0-.39\ .39-.39 1.02 0 1\ .41l4.59 4.59c.3\ 9.39 1.02.39 1.4\ 1 0l4.59-4.59c.3\ 9-.39.39-1.02 0-\ 1.41-.39-.38-1.0\ 3-.39-1.42 0z\x22/>\ </svg>\ \x00\x00\x02\x98\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#b\ abdc2\x22><g><rect \ fill=\x22none\x22 heig\ ht=\x2224\x22 width=\x222\ 4\x22/><rect fill=\x22\ none\x22 height=\x2224\ \x22 width=\x2224\x22/></\ g><g><g><path d=\ \x22M20.08,11.42l-4\ .04-5.65C15.7,5.\ 29,15.15,5,14.56\ ,5h0c-1.49,0-2.3\ 5,1.68-1.49,2.89\ L16,12l-2.93,4.1\ 1 c-0.87,1.21,0,\ 2.89,1.49,2.89h0\ c0.59,0,1.15-0.2\ 9,1.49-0.77l4.04\ -5.65C20.33,12.2\ 3,20.33,11.77,20\ .08,11.42z\x22/><pa\ th d=\x22M13.08,11.\ 42L9.05,5.77C8.7\ ,5.29,8.15,5,7.5\ 6,5h0C6.07,5,5.2\ ,6.68,6.07,7.89L\ 9,12l-2.93,4.11C\ 5.2,17.32,6.07,1\ 9,7.56,19h0 c0.5\ 9,0,1.15-0.29,1.\ 49-0.77l4.04-5.6\ 5C13.33,12.23,13\ .33,11.77,13.08,\ 11.42z\x22/></g></g\ ></svg>\ \x00\x00\x01Z\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#4\ d5157\x22><g><rect \ fill=\x22none\x22 fill\ -rule=\x22evenodd\x22 \ height=\x2224\x22 widt\ h=\x2224\x22/><g><path\ d=\x22M19,13H5c-0.\ 55,0-1-0.45-1-1v\ 0c0-0.55,0.45-1,\ 1-1h14c0.55,0,1,\ 0.45,1,1v0 C20,1\ 2.55,19.55,13,19\ ,13z\x22 fill-rule=\ \x22evenodd\x22/></g><\ /g></svg>\ \x00\x00\x01\xb0\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M3 13h2v-2H\ 3v2zm0 4h2v-2H3v\ 2zm2 4v-2H3c0 1.\ 1.89 2 2 2zM3 9h\ 2V7H3v2zm12 12h2\ v-2h-2v2zm4-18H9\ c-1.11 0-2 .9-2 \ 2v10c0 1.1.89 2 \ 2 2h10c1.1 0 2-.\ 9 2-2V5c0-1.1-.9\ -2-2-2zm-1 12h-8\ c-.55 0-1-.45-1-\ 1V6c0-.55.45-1 1\ -1h8c.55 0 1 .45\ 1 1v8c0 .55-.45\ 1-1 1zm-7 6h2v-\ 2h-2v2zm-4 0h2v-\ 2H7v2z\x22/></svg>\ \x00\x00\x01<\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M18 19H6c-.\ 55 0-1-.45-1-1V6\ c0-.55.45-1 1-1h\ 12c.55 0 1 .45 1\ 1v12c0 .55-.45 \ 1-1 1zm1-16H5c-1\ .1 0-2 .9-2 2v14\ c0 1.1.9 2 2 2h1\ 4c1.1 0 2-.9 2-2\ V5c0-1.1-.9-2-2-\ 2z\x22/></svg>\ \x00\x00\x01\xa0\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M20 6h-8l-1\ .41-1.41C10.21 4\ .21 9.7 4 9.17 4\ H4c-1.11 0-1.99.\ 89-1.99 2L2 18c0\ 1.11.89 2 2 2h1\ 6c1.1 0 2-.9 2-2\ V8c0-1.1-.9-2-2-\ 2zm-2 8h-2v2c0 .\ 55-.45 1-1 1s-1-\ .45-1-1v-2h-2c-.\ 55 0-1-.45-1-1s.\ 45-1 1-1h2v-2c0-\ .55.45-1 1-1s1 .\ 45 1 1v2h2c.55 0\ 1 .45 1 1s-.45 \ 1-1 1z\x22/></svg>\ \x00\x00\x01\x95\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#008\ 1db\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M19 3H5c-1.\ 1 0-2 .9-2 2v14c\ 0 1.1.9 2 2 2h14\ c1.1 0 2-.9 2-2V\ 5c0-1.1-.9-2-2-2\ zm-8.29 13.29c-.\ 39.39-1.02.39-1.\ 41 0L5.71 12.7c-\ .39-.39-.39-1.02\ 0-1.41.39-.39 1\ .02-.39 1.41 0L1\ 0 14.17l6.88-6.8\ 8c.39-.39 1.02-.\ 39 1.41 0 .39.39\ .39 1.02 0 1.41l\ -7.58 7.59z\x22/></\ svg>\ \x00\x00\x01:\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#008\ 1db\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M12 2C6.48 \ 2 2 6.48 2 12s4.\ 48 10 10 10 10-4\ .48 10-10S17.52 \ 2 12 2zm0 18c-4.\ 42 0-8-3.58-8-8s\ 3.58-8 8-8 8 3.5\ 8 8 8-3.58 8-8 8\ z\x22/><circle cx=\x22\ 12\x22 cy=\x2212\x22 r=\x225\ \x22/></svg>\ \x00\x00\x01Q\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#b\ abdc2\x22><g><rect \ fill=\x22none\x22 fill\ -rule=\x22evenodd\x22 \ height=\x2224\x22 widt\ h=\x2224\x22/><g><path\ d=\x22M13,5v14c0,0\ .6-0.4,1-1,1l0,0\ c-0.6,0-1-0.4-1-\ 1V5c0-0.5,0.4-1,\ 1-1l0,0C12.6,4,1\ 3,4.4,13,5z\x22 fil\ l-rule=\x22evenodd\x22\ /></g></g></svg>\ \ \x00\x00\x01\xae\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M18.3 5.71c\ -.39-.39-1.02-.3\ 9-1.41 0L12 10.5\ 9 7.11 5.7c-.39-\ .39-1.02-.39-1.4\ 1 0-.39.39-.39 1\ .02 0 1.41L10.59\ 12 5.7 16.89c-.\ 39.39-.39 1.02 0\ 1.41.39.39 1.02\ .39 1.41 0L12 13\ .41l4.89 4.89c.3\ 9.39 1.02.39 1.4\ 1 0 .39-.39.39-1\ .02 0-1.41L13.41\ 12l4.89-4.89c.3\ 8-.38.38-1.02 0-\ 1.4z\x22/></svg>\ \x00\x00\x01\xc7\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#0\ 081db\x22><rect fil\ l=\x22none\x22 height=\ \x2224\x22 width=\x2224\x22/\ ><path d=\x22M14.29\ ,5.71L14.29,5.71\ c-0.39,0.39-0.39\ ,1.02,0,1.41L18.\ 17,11H3c-0.55,0-\ 1,0.45-1,1v0c0,0\ .55,0.45,1,1,1h1\ 5.18l-3.88,3.88 \ c-0.39,0.39-0.39\ ,1.02,0,1.41l0,0\ c0.39,0.39,1.02,\ 0.39,1.41,0l5.59\ -5.59c0.39-0.39,\ 0.39-1.02,0-1.41\ L15.7,5.71 C15.3\ 2,5.32,14.68,5.3\ 2,14.29,5.71z\x22/>\ </svg>\ \x00\x00\x01\x0c\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M8.71 12.29\ L11.3 9.7c.39-.3\ 9 1.02-.39 1.41 \ 0l2.59 2.59c.63.\ 63.18 1.71-.71 1\ .71H9.41c-.89 0-\ 1.33-1.08-.7-1.7\ 1z\x22/></svg>\ \x00\x00\x01\xb0\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M3 13h2v-2H\ 3v2zm0 4h2v-2H3v\ 2zm2 4v-2H3c0 1.\ 1.89 2 2 2zM3 9h\ 2V7H3v2zm12 12h2\ v-2h-2v2zm4-18H9\ c-1.11 0-2 .9-2 \ 2v10c0 1.1.89 2 \ 2 2h10c1.1 0 2-.\ 9 2-2V5c0-1.1-.9\ -2-2-2zm-1 12h-8\ c-.55 0-1-.45-1-\ 1V6c0-.55.45-1 1\ -1h8c.55 0 1 .45\ 1 1v8c0 .55-.45\ 1-1 1zm-7 6h2v-\ 2h-2v2zm-4 0h2v-\ 2H7v2z\x22/></svg>\ \x00\x00\x01\x1b\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#bab\ dc2\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M12 2C6.48 \ 2 2 6.48 2 12s4.\ 48 10 10 10 10-4\ .48 10-10S17.52 \ 2 12 2zm0 18c-4.\ 42 0-8-3.58-8-8s\ 3.58-8 8-8 8 3.5\ 8 8 8-3.58 8-8 8\ z\x22/></svg>\ \x00\x00\x01G\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M9.29 6.71c\ -.39.39-.39 1.02\ 0 1.41L13.17 12\ l-3.88 3.88c-.39\ .39-.39 1.02 0 1\ .41.39.39 1.02.3\ 9 1.41 0l4.59-4.\ 59c.39-.39.39-1.\ 02 0-1.41L10.7 6\ .7c-.38-.38-1.02\ -.38-1.41.01z\x22/>\ </svg>\ \x00\x00\x01p\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M6 19c0 1.1\ .9 2 2 2h8c1.1 0\ 2-.9 2-2V9c0-1.\ 1-.9-2-2-2H8c-1.\ 1 0-2 .9-2 2v10z\ M18 4h-2.5l-.71-\ .71c-.18-.18-.44\ -.29-.7-.29H9.91\ c-.26 0-.52.11-.\ 7.29L8.5 4H6c-.5\ 5 0-1 .45-1 1s.4\ 5 1 1 1h12c.55 0\ 1-.45 1-1s-.45-\ 1-1-1z\x22/></svg>\ \x00\x00\x01\xc7\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#0\ e527c\x22><rect fil\ l=\x22none\x22 height=\ \x2224\x22 width=\x2224\x22/\ ><path d=\x22M14.29\ ,5.71L14.29,5.71\ c-0.39,0.39-0.39\ ,1.02,0,1.41L18.\ 17,11H3c-0.55,0-\ 1,0.45-1,1v0c0,0\ .55,0.45,1,1,1h1\ 5.18l-3.88,3.88 \ c-0.39,0.39-0.39\ ,1.02,0,1.41l0,0\ c0.39,0.39,1.02,\ 0.39,1.41,0l5.59\ -5.59c0.39-0.39,\ 0.39-1.02,0-1.41\ L15.7,5.71 C15.3\ 2,5.32,14.68,5.3\ 2,14.29,5.71z\x22/>\ </svg>\ \x00\x00\x01\xfe\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22 opaci\ ty=\x22.87\x22/><path \ d=\x22M12 2C6.47 2 \ 2 6.47 2 12s4.47\ 10 10 10 10-4.4\ 7 10-10S17.53 2 \ 12 2zm4.3 14.3c-\ .39.39-1.02.39-1\ .41 0L12 13.41 9\ .11 16.3c-.39.39\ -1.02.39-1.41 0-\ .39-.39-.39-1.02\ 0-1.41L10.59 12\ 7.7 9.11c-.39-.\ 39-.39-1.02 0-1.\ 41.39-.39 1.02-.\ 39 1.41 0L12 10.\ 59l2.89-2.89c.39\ -.39 1.02-.39 1.\ 41 0 .39.39.39 1\ .02 0 1.41L13.41\ 12l2.89 2.89c.3\ 8.38.38 1.02 0 1\ .41z\x22/></svg>\ \x00\x00\x01\x9a\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M17.59 3.59\ c-.38-.38-.89-.5\ 9-1.42-.59H5c-1.\ 11 0-2 .9-2 2v14\ c0 1.1.9 2 2 2h1\ 4c1.1 0 2-.9 2-2\ V7.83c0-.53-.21-\ 1.04-.59-1.41l-2\ .82-2.83zM12 19c\ -1.66 0-3-1.34-3\ -3s1.34-3 3-3 3 \ 1.34 3 3-1.34 3-\ 3 3zm1-10H7c-1.1\ 0-2-.9-2-2s.9-2\ 2-2h6c1.1 0 2 .\ 9 2 2s-.9 2-2 2z\ \x22/></svg>\ \x00\x00\x02\x98\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 enable-ba\ ckground=\x22new 0 \ 0 24 24\x22 height=\ \x2224px\x22 viewBox=\x22\ 0 0 24 24\x22 width\ =\x2224px\x22 fill=\x22#4\ d5157\x22><g><rect \ fill=\x22none\x22 heig\ ht=\x2224\x22 width=\x222\ 4\x22/><rect fill=\x22\ none\x22 height=\x2224\ \x22 width=\x2224\x22/></\ g><g><g><path d=\ \x22M20.08,11.42l-4\ .04-5.65C15.7,5.\ 29,15.15,5,14.56\ ,5h0c-1.49,0-2.3\ 5,1.68-1.49,2.89\ L16,12l-2.93,4.1\ 1 c-0.87,1.21,0,\ 2.89,1.49,2.89h0\ c0.59,0,1.15-0.2\ 9,1.49-0.77l4.04\ -5.65C20.33,12.2\ 3,20.33,11.77,20\ .08,11.42z\x22/><pa\ th d=\x22M13.08,11.\ 42L9.05,5.77C8.7\ ,5.29,8.15,5,7.5\ 6,5h0C6.07,5,5.2\ ,6.68,6.07,7.89L\ 9,12l-2.93,4.11C\ 5.2,17.32,6.07,1\ 9,7.56,19h0 c0.5\ 9,0,1.15-0.29,1.\ 49-0.77l4.04-5.6\ 5C13.33,12.23,13\ .33,11.77,13.08,\ 11.42z\x22/></g></g\ ></svg>\ \x00\x00\x01d\ <\ svg xmlns=\x22http:\ //www.w3.org/200\ 0/svg\x22 height=\x222\ 4px\x22 viewBox=\x220 \ 0 24 24\x22 width=\x22\ 24px\x22 fill=\x22#4d5\ 157\x22><path d=\x22M0\ 0h24v24H0V0z\x22 f\ ill=\x22none\x22/><pat\ h d=\x22M5 13h11.17\ l-4.88 4.88c-.39\ .39-.39 1.03 0 1\ .42.39.39 1.02.3\ 9 1.41 0l6.59-6.\ 59c.39-.39.39-1.\ 02 0-1.41l-6.58-\ 6.6c-.39-.39-1.0\ 2-.39-1.41 0-.39\ .39-.39 1.02 0 1\ .41L16.17 11H5c-\ .55 0-1 .45-1 1s\ .45 1 1 1z\x22/></s\ vg>\ " qt_resource_name = b"\ \x00\x05\ \x00o\xa6S\ \x00i\ \x00c\x00o\x00n\x00s\ \x00\x03\ \x00\x00p7\ \x00i\ \x00m\x00g\ \x00\x09\ \x07i\x93X\ \x00i\ \x00m\x00p\x00o\x00r\x00t\x00-\x00f\x00x\ \x00\x0d\ \x0e\x8f\x9f\xc5\ \x00f\ \x00i\x00l\x00t\x00e\x00r\x00-\x00r\x00e\x00m\x00o\x00v\x00e\ \x00\x06\ \x07\x05\xcc\x94\ \x00i\ \x00n\x00v\x00e\x00r\x00t\ \x00\x05\ \x00j6\x95\ \x00c\ \x00l\x00o\x00s\x00e\ \x00\x03\ \x00\x00jW\ \x00c\ \x00o\x00g\ \x00\x08\ \x08\xaa\x8d\x18\ \x00p\ \x00a\x00s\x00t\x00e\x00-\x00f\x00x\ \x00\x0a\ \x06\x992F\ \x00i\ \x00m\x00p\x00o\x00r\x00t\x00-\x00c\x00s\x00v\ \x00\x05\ \x00j+\x82\ \x00c\ \x00l\x00e\x00a\x00r\ \x00\x05\ \x00zy%\ \x00t\ \x00a\x00b\x00l\x00e\ \x00\x09\ \x06\x99\x17\xc9\ \x00c\ \x00l\x00o\x00s\x00e\x00-\x00m\x00d\x00i\ \x00\x06\ \x06\xd0:\xc2\ \x00f\ \x00i\x00l\x00t\x00e\x00r\ \x00\x0a\ \x06\xae\xa5\xfe\ \x00a\ \x00r\x00r\x00o\x00w\x00-\x00d\x00o\x00w\x00n\ \x00\x0e\ \x00R\x02\x87\ \x00g\ \x00m\x00s\x00-\x00s\x00p\x00l\x00a\x00s\x00h\x00.\x00p\x00n\x00g\ \x00\x0c\ \x0cf\x9f'\ \x00G\ \x00M\x00S\x00-\x00l\x00o\x00g\x00o\x00.\x00p\x00n\x00g\ \x00\x0f\ \x0f\xda\xd2\xf3\ \x00S\ \x00t\x00y\x00l\x00e\x00L\x00i\x00g\x00h\x00t\x00I\x00c\x00o\x00n\x00s\ \x00\x08\ \x07O\xafS\ \x00A\ \x00p\x00p\x00I\x00c\x00o\x00n\x00s\ \x00\x1b\ \x05\xed\x82\x07\ \x00f\ \x00o\x00l\x00d\x00e\x00r\x00-\x00o\x00p\x00e\x00n\x00-\x00o\x00u\x00t\x00l\x00i\ \x00n\x00e\x00_\x00m\x00d\x00i\x00.\x00s\x00v\x00g\ \x00\x0c\ \x0b\xdf,\xc7\ \x00s\ \x00e\x00t\x00t\x00i\x00n\x00g\x00s\x00.\x00s\x00v\x00g\ \x00\x15\ \x0d\xf0,\x07\ \x00t\ \x00r\x00a\x00y\x00-\x00a\x00r\x00r\x00o\x00w\x00-\x00u\x00p\x00_\x00m\x00d\x00i\ \x00.\x00s\x00v\x00g\ \x00 \ \x0b9J\x87\ \x00c\ \x00h\x00e\x00c\x00k\x00_\x00b\x00o\x00x\x00_\x00o\x00u\x00t\x00l\x00i\x00n\x00e\ \x00_\x00b\x00l\x00a\x00n\x00k\x00_\x002\x004\x00d\x00p\x00.\x00p\x00n\x00g\ \x00\x0b\ \x03~\xd4\x07\ \x00d\ \x00n\x00a\x00_\x00m\x00d\x00i\x00.\x00s\x00v\x00g\ \x00\x16\ \x02\x1e2\x07\ \x00g\ \x00o\x00o\x00g\x00l\x00e\x00-\x00s\x00p\x00r\x00e\x00a\x00d\x00s\x00h\x00e\x00e\ \x00t\x00.\x00s\x00v\x00g\ \x00\x0c\ \x0cf\x87\xff\ \x00G\ \x00M\x00S\x00-\x00l\x00o\x00g\x00o\x00.\x00i\x00c\x00o\ \x00\x1c\ \x00\xed\x17\xc7\ \x00c\ \x00o\x00n\x00t\x00e\x00n\x00t\x00-\x00s\x00a\x00v\x00e\x00-\x00o\x00u\x00t\x00l\ \x00i\x00n\x00e\x00_\x00m\x00d\x00i\x00.\x00s\x00v\x00g\ \x00\x1f\ \x05\xd2W\x87\ \x00f\ \x00o\x00l\x00d\x00e\x00r\x00-\x00o\x00p\x00e\x00n\x00-\x00o\x00u\x00t\x00l\x00i\ \x00n\x00e\x00-\x00d\x00n\x00a\x00_\x00m\x00d\x00i\x00.\x00s\x00v\x00g\ \x00\x12\ \x0b\xe4\xfb\x87\ \x00c\ \x00h\x00e\x00c\x00k\x00_\x00b\x00o\x00x\x00_\x002\x004\x00d\x00p\x00.\x00p\x00n\ \x00g\ \x00 \ \x04\x16\x94g\ \x00c\ \x00a\x00l\x00e\x00n\x00d\x00a\x00r\x00_\x00t\x00o\x00d\x00a\x00y\x00_\x00d\x00i\ \x00s\x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x16\ \x0d\x097G\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00d\x00r\x00o\x00p\x00_\x00u\x00p\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00\x18\ \x0e\xc6\xf1\xc7\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00d\x00r\x00o\x00p\x00_\x00d\x00o\x00w\x00n\x00_\x002\ \x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x16\ \x02jS\x07\ \x00v\ \x00e\x00r\x00t\x00i\x00c\x00a\x00l\x00_\x00r\x00u\x00l\x00e\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00\x14\ \x02\x1e\x04\x87\ \x00r\ \x00e\x00s\x00t\x00a\x00r\x00t\x00_\x00a\x00l\x00t\x00_\x002\x004\x00d\x00p\x00.\ \x00s\x00v\x00g\ \x00\x16\ \x0e\xd3\x0d\xe7\ \x00c\ \x00o\x00r\x00n\x00e\x00r\x00_\x00b\x00u\x00t\x00t\x00o\x00n\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00\x1d\ \x05\xe3\xd5G\ \x00d\ \x00o\x00u\x00b\x00l\x00e\x00_\x00a\x00r\x00r\x00o\x00w\x00_\x00s\x00i\x00d\x00e\ \x00b\x00a\x00r\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1f\ \x03\xf9#\x87\ \x00c\ \x00o\x00r\x00n\x00e\x00r\x00_\x00b\x00u\x00t\x00t\x00o\x00n\x00_\x00d\x00i\x00s\ \x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00!\ \x07\xa7\xcd\xa7\ \x00h\ \x00o\x00r\x00i\x00z\x00o\x00n\x00t\x00a\x00l\x00_\x00r\x00u\x00l\x00e\x00_\x00d\ \x00i\x00s\x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \ \x00\x22\ \x05Vv'\ \x00d\ \x00r\x00a\x00g\x00_\x00i\x00n\x00d\x00i\x00c\x00a\x00t\x00o\x00r\x00_\x00h\x00o\ \x00r\x00i\x00z\x00o\x00n\x00t\x00a\x00l\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\ \x00g\ \x00\x0e\ \x07\xc7d\x07\ \x00c\ \x00h\x00e\x00c\x00k\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x14\ \x02\x87a\x87\ \x00e\ \x00x\x00p\x00a\x00n\x00d\x00_\x00l\x00e\x00s\x00s\x00_\x002\x004\x00d\x00p\x00.\ \x00s\x00v\x00g\ \x00\x14\ \x0e0\xe6G\ \x00f\ \x00o\x00l\x00d\x00e\x00r\x00_\x00o\x00p\x00e\x00n\x00_\x002\x004\x00d\x00p\x00.\ \x00s\x00v\x00g\ \x00\x1f\ \x0dx\xe8'\ \x00r\ \x00a\x00d\x00i\x00o\x00_\x00b\x00u\x00t\x00t\x00o\x00n\x00_\x00u\x00n\x00c\x00h\ \x00e\x00c\x00k\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00(\ \x09e\x9a\xe7\ \x00d\ \x00r\x00a\x00g\x00_\x00i\x00n\x00d\x00i\x00c\x00a\x00t\x00o\x00r\x00_\x00v\x00e\ \x00r\x00t\x00i\x00c\x00a\x00l\x00_\x00s\x00i\x00d\x00e\x00b\x00a\x00r\x00_\x002\ \x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1f\ \x03P'\x07\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00f\x00o\x00r\x00w\x00a\x00r\x00d\x00_\x00d\x00i\x00s\ \x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x15\ \x04\x1bkG\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00u\x00p\x00w\x00a\x00r\x00d\x00_\x002\x004\x00d\x00p\ \x00.\x00s\x00v\x00g\ \x00 \ \x0b9G\x07\ \x00c\ \x00h\x00e\x00c\x00k\x00_\x00b\x00o\x00x\x00_\x00o\x00u\x00t\x00l\x00i\x00n\x00e\ \x00_\x00b\x00l\x00a\x00n\x00k\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1d\ \x0ev6\x87\ \x00e\ \x00x\x00p\x00a\x00n\x00d\x00_\x00l\x00e\x00s\x00s\x00_\x00d\x00i\x00s\x00a\x00b\ \x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1b\ \x0c\x84\xeeg\ \x00c\ \x00h\x00e\x00c\x00k\x00_\x00b\x00o\x00x\x00_\x00d\x00i\x00s\x00a\x00b\x00l\x00e\ \x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x15\ \x05p\x09\x07\ \x00c\ \x00h\x00e\x00c\x00k\x00_\x00c\x00i\x00r\x00c\x00l\x00e\x00_\x002\x004\x00d\x00p\ \x00.\x00s\x00v\x00g\ \x00 \ \x0490g\ \x00d\ \x00r\x00a\x00g\x00_\x00i\x00n\x00d\x00i\x00c\x00a\x00t\x00o\x00r\x00_\x00v\x00e\ \x00r\x00t\x00i\x00c\x00a\x00l\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00&\ \x04\xc0\xfa\x87\ \x00d\ \x00o\x00u\x00b\x00l\x00e\x00_\x00a\x00r\x00r\x00o\x00w\x00_\x00d\x00i\x00s\x00a\ \x00b\x00l\x00e\x00d\x00_\x00s\x00i\x00d\x00e\x00b\x00a\x00r\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00\x16\ \x04\xe6P'\ \x00v\ \x00e\x00r\x00t\x00i\x00c\x00a\x00l\x00_\x00l\x00i\x00n\x00e\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00\x1e\ \x0f9l\xc7\ \x00c\ \x00o\x00r\x00n\x00e\x00r\x00_\x00b\x00u\x00t\x00t\x00o\x00n\x00_\x00p\x00r\x00e\ \x00s\x00s\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x0d\ \x075\x9f\xc7\ \x00h\ \x00e\x00l\x00p\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x17\ \x06\xd9\x8fG\ \x00c\ \x00l\x00o\x00s\x00e\x00_\x00d\x00i\x00s\x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\ \x00d\x00p\x00.\x00s\x00v\x00g\ \x00*\ \x07\xe9\x1fG\ \x00d\ \x00r\x00a\x00g\x00_\x00i\x00n\x00d\x00i\x00c\x00a\x00t\x00o\x00r\x00_\x00h\x00o\ \x00r\x00i\x00z\x00o\x00n\x00t\x00a\x00l\x00_\x00s\x00i\x00d\x00e\x00b\x00a\x00r\ \x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x17\ \x00\x0c\xb2\x87\ \x00c\ \x00a\x00l\x00e\x00n\x00d\x00a\x00r\x00_\x00t\x00o\x00d\x00a\x00y\x00_\x002\x004\ \x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1f\ \x08C\xae\xc7\ \x00v\ \x00e\x00r\x00t\x00i\x00c\x00a\x00l\x00_\x00l\x00i\x00n\x00e\x00_\x00d\x00i\x00s\ \x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00&\ \x03\x16\xa1\xa7\ \x00r\ \x00a\x00d\x00i\x00o\x00_\x00b\x00u\x00t\x00t\x00o\x00n\x00_\x00c\x00h\x00e\x00c\ \x00k\x00e\x00d\x00_\x00d\x00i\x00s\x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00!\ \x0c}&\xe7\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00d\x00r\x00o\x00p\x00_\x00d\x00o\x00w\x00n\x00_\x00d\ \x00i\x00s\x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \ \x00\x13\ \x08*\xc8G\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00b\x00a\x00c\x00k\x00_\x002\x004\x00d\x00p\x00.\x00s\ \x00v\x00g\ \x00\x1c\ \x0bL\x9d\xe7\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00b\x00a\x00c\x00k\x00_\x00d\x00i\x00s\x00a\x00b\x00l\ \x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x14\ \x07\xc3!\x87\ \x00e\ \x00x\x00p\x00a\x00n\x00d\x00_\x00m\x00o\x00r\x00e\x00_\x002\x004\x00d\x00p\x00.\ \x00s\x00v\x00g\ \x00 \ \x01\x0a:'\ \x00i\ \x00n\x00d\x00e\x00t\x00e\x00r\x00m\x00i\x00n\x00a\x00t\x00e\x00_\x00c\x00h\x00e\ \x00c\x00k\x00_\x00b\x00o\x00x\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00)\ \x05\x01U\xc7\ \x00i\ \x00n\x00d\x00e\x00t\x00e\x00r\x00m\x00i\x00n\x00a\x00t\x00e\x00_\x00c\x00h\x00e\ \x00c\x00k\x00_\x00b\x00o\x00x\x00_\x00d\x00i\x00s\x00a\x00b\x00l\x00e\x00d\x00_\ \x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1e\ \x0f/\x91\x87\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00u\x00p\x00w\x00a\x00r\x00d\x00_\x00d\x00i\x00s\x00a\ \x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1f\ \x0c\x02\xce\x87\ \x00c\ \x00h\x00e\x00v\x00r\x00o\x00n\x00_\x00r\x00i\x00g\x00h\x00t\x00_\x00d\x00i\x00s\ \x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1d\ \x06w\xe7\x87\ \x00e\ \x00x\x00p\x00a\x00n\x00d\x00_\x00m\x00o\x00r\x00e\x00_\x00d\x00i\x00s\x00a\x00b\ \x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1e\ \x0c'\xe2\x87\ \x00d\ \x00o\x00u\x00b\x00l\x00e\x00_\x00a\x00r\x00r\x00o\x00w\x00_\x00d\x00i\x00s\x00a\ \x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x18\ \x07k\xaf\x07\ \x00h\ \x00o\x00r\x00i\x00z\x00o\x00n\x00t\x00a\x00l\x00_\x00r\x00u\x00l\x00e\x00_\x002\ \x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x16\ \x01\xe2Q'\ \x00f\ \x00l\x00i\x00p\x00_\x00t\x00o\x00_\x00f\x00r\x00o\x00n\x00t\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00)\ \x0a\xa4\xd9\x07\ \x00c\ \x00h\x00e\x00c\x00k\x00_\x00b\x00o\x00x\x00_\x00o\x00u\x00t\x00l\x00i\x00n\x00e\ \x00_\x00b\x00l\x00a\x00n\x00k\x00_\x00d\x00i\x00s\x00a\x00b\x00l\x00e\x00d\x00_\ \x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1a\ \x09\x1b\x9bG\ \x00c\ \x00r\x00e\x00a\x00t\x00e\x00_\x00n\x00e\x00w\x00_\x00f\x00o\x00l\x00d\x00e\x00r\ \x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x12\ \x0b\xe4\xf6\x07\ \x00c\ \x00h\x00e\x00c\x00k\x00_\x00b\x00o\x00x\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\ \x00g\ \x00\x1d\ \x00\xdb\x8a\x87\ \x00r\ \x00a\x00d\x00i\x00o\x00_\x00b\x00u\x00t\x00t\x00o\x00n\x00_\x00c\x00h\x00e\x00c\ \x00k\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1f\ \x08&\x0d\xc7\ \x00v\ \x00e\x00r\x00t\x00i\x00c\x00a\x00l\x00_\x00r\x00u\x00l\x00e\x00_\x00d\x00i\x00s\ \x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x0e\ \x0c`$\xe7\ \x00c\ \x00l\x00o\x00s\x00e\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x0d\ \x0d\xa6\x1f\xc7\ \x00e\ \x00a\x00s\x00t\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1f\ \x04\xacU\x07\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00d\x00r\x00o\x00p\x00_\x00u\x00p\x00_\x00d\x00i\x00s\ \x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x1f\ \x08ao\xc7\ \x00f\ \x00l\x00i\x00p\x00_\x00t\x00o\x00_\x00f\x00r\x00o\x00n\x00t\x00_\x00d\x00i\x00s\ \x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00(\ \x0f@Ig\ \x00r\ \x00a\x00d\x00i\x00o\x00_\x00b\x00u\x00t\x00t\x00o\x00n\x00_\x00u\x00n\x00c\x00h\ \x00e\x00c\x00k\x00e\x00d\x00_\x00d\x00i\x00s\x00a\x00b\x00l\x00e\x00d\x00_\x002\ \x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x16\ \x03gr'\ \x00c\ \x00h\x00e\x00v\x00r\x00o\x00n\x00_\x00r\x00i\x00g\x00h\x00t\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00\x0f\ \x0f'\x95g\ \x00d\ \x00e\x00l\x00e\x00t\x00e\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x16\ \x01\xbc~\xc7\ \x00e\ \x00a\x00s\x00t\x00_\x00d\x00i\x00s\x00a\x00b\x00l\x00e\x00d\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ \x00\x0f\ \x03\x8a\xb8g\ \x00c\ \x00a\x00n\x00c\x00e\x00l\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x0d\ \x09\xaa?\x87\ \x00s\ \x00a\x00v\x00e\x00_\x002\x004\x00d\x00p\x00.\x00s\x00v\x00g\ \x00\x15\ \x07\xd7s\x07\ \x00d\ \x00o\x00u\x00b\x00l\x00e\x00_\x00a\x00r\x00r\x00o\x00w\x00_\x002\x004\x00d\x00p\ \x00.\x00s\x00v\x00g\ \x00\x16\ \x0a\xc1\x08\xa7\ \x00a\ \x00r\x00r\x00o\x00w\x00_\x00f\x00o\x00r\x00w\x00a\x00r\x00d\x00_\x002\x004\x00d\ \x00p\x00.\x00s\x00v\x00g\ " qt_resource_struct = b"\ \x00\x00\x00\x00\x00\x02\x00\x00\x00\x0e\x00\x00\x00\x01\ \x00\x00\x00\x00\x00\x00\x00\x00\ \x00\x00\x00v\x00\x00\x00\x00\x00\x01\x00\x00\x19\xaf\ \x00\x00\x01|\x92\x17g\xb0\ \x00\x00\x00\x10\x00\x02\x00\x00\x00\x02\x00\x00\x00W\ \x00\x00\x00\x00\x00\x00\x00\x00\ \x00\x00\x00\xb2\x00\x00\x00\x00\x00\x01\x00\x005\x10\ \x00\x00\x01|\x92\x17g\xb1\ \x00\x00\x00f\x00\x00\x00\x00\x00\x01\x00\x00\x18\x17\ \x00\x00\x01|\x92\x17g\xb0\ \x00\x00\x00\x00\x00\x02\x00\x00\x00\x02\x00\x00\x00\x0f\ \x00\x00\x00\x00\x00\x00\x00\x00\ \x00\x00\x00\xc2\x00\x00\x00\x00\x00\x01\x00\x006\x9e\ \x00\x00\x01|\x92\x17g\xb5\ \x00\x00\x00\xd2\x00\x00\x00\x00\x00\x01\x00\x008T\ \x00\x00\x01|\x92\x17g\xaf\ \x00\x00\x00\x98\x00\x00\x00\x00\x00\x01\x00\x00):\ \x00\x00\x01|\x92\x17g\xb4\ \x00\x00\x00\xfc\x00\x00\x00\x00\x00\x01\x00\x00<\xa5\ \x00\x00\x01|\x92\x17g\xaf\ \x00\x00\x00\xea\x00\x00\x00\x00\x00\x01\x00\x00::\ \x00\x00\x01|\x92\x17g\xb2\ \x00\x00\x00T\x00\x00\x00\x00\x00\x01\x00\x00\x0e\x16\ \x00\x00\x01|\x92\x17g\xb5\ \x00\x00\x00\x1c\x00\x00\x00\x00\x00\x01\x00\x00\x00\x00\ \x00\x00\x01|x\x8d\xb7\x92\ \x00\x00\x00\x82\x00\x00\x00\x00\x00\x01\x00\x00\x1f\xab\ \x00\x00\x01|x\x8f\x19P\ \x00\x00\x004\x00\x00\x00\x00\x00\x01\x00\x00\x0b\x1f\ \x00\x00\x01|\x92\x17g\xb3\ \x00\x00\x01z\x00\x02\x00\x00\x00\x0a\x00\x00\x00M\ \x00\x00\x00\x00\x00\x00\x00\x00\ \x00\x00\x01V\x00\x02\x00\x00\x00<\x00\x00\x00\x11\ \x00\x00\x00\x00\x00\x00\x00\x00\ \x00\x00\x0a\x06\x00\x00\x00\x00\x00\x01\x00\x01\x8b\xc1\ \x00\x00\x01|K\xb5\xff\xdb\ \x00\x00\x0ez\x00\x00\x00\x00\x00\x01\x00\x01\xa5L\ \x00\x00\x01|K\xb5\xff\xf3\ \x00\x00\x0b\xb0\x00\x00\x00\x00\x00\x01\x00\x01\x94\xba\ \x00\x00\x01|K\xb5\xff\xf2\ \x00\x00\x10t\x00\x00\x00\x00\x00\x01\x00\x01\xb1\xfe\ \x00\x00\x01|K\xb5\xff\xeb\ \x00\x00\x0d\x8c\x00\x00\x00\x00\x00\x01\x00\x01\x9f\x1b\ \x00\x00\x01|K\xb5\xff\xee\ \x00\x00\x04X\x00\x00\x00\x00\x00\x01\x00\x01e\x94\ \x00\x00\x01|K\xb5\xff\xf6\ \x00\x00\x04&\x00\x00\x00\x00\x00\x01\x00\x01d?\ \x00\x00\x01|K\xb5\xff\xf8\ \x00\x00\x05\xf0\x00\x00\x00\x00\x00\x01\x00\x01qg\ \x00\x00\x01|K\xb5\xff\xeb\ \x00\x00\x0a~\x00\x00\x00\x00\x00\x01\x00\x01\x8eA\ \x00\x00\x01|K\xb5\xff\xf4\ \x00\x00\x06\xe6\x00\x00\x00\x00\x00\x01\x00\x01w\x0a\ \x00\x00\x01|K\xb5\xff\xd9\ \x00\x00\x10\x1e\x00\x00\x00\x00\x00\x01\x00\x01\xaf?\ \x00\x00\x01|K\xb5\xff\xe0\ \x00\x00\x10\xa6\x00\x00\x00\x00\x00\x01\x00\x01\xb3\xc9\ \x00\x00\x01|K\xb5\xff\xdc\ \x00\x00\x04\xf8\x00\x00\x00\x00\x00\x01\x00\x01k\xfc\ \x00\x00\x01|K\xb5\xff\xe3\ \x00\x00\x03x\x00\x00\x00\x00\x00\x01\x00\x01`\xbb\ \x00\x00\x01|K\xb5\xff\xdc\ \x00\x00\x07*\x00\x00\x00\x00\x00\x01\x00\x01xr\ \x00\x00\x01|K\xb5\xff\xda\ \x00\x00\x08L\x00\x00\x00\x00\x00\x01\x00\x01\x7f\x85\ \x00\x00\x01|K\xb5\xff\xe9\ \x00\x00\x0f@\x00\x00\x00\x00\x00\x01\x00\x01\xab\x5c\ \x00\x00\x01|K\xb5\xff\xd8\ \x00\x00\x08\x92\x00\x00\x00\x00\x00\x01\x00\x01\x81U\ \x00\x00\x01|K\xb5\xff\xe7\ \x00\x00\x08\xe4\x00\x00\x00\x00\x00\x01\x00\x01\x83\xf1\ \x00\x00\x01|K\xb5\xff\xf7\ \x00\x00\x0b\xf6\x00\x00\x00\x00\x00\x01\x00\x01\x95\xe8\ \x00\x00\x01|K\xb5\xff\xf3\ \x00\x00\x05\x84\x00\x00\x00\x00\x00\x01\x00\x01nY\ \x00\x00\x01|K\xb5\xff\xe8\ \x00\x00\x08\x1c\x00\x00\x00\x00\x00\x01\x00\x01}\xf8\ \x00\x00\x01|K\xb5\xff\xdf\ \x00\x00\x04\xb8\x00\x00\x00\x00\x00\x01\x00\x01i`\ \x00\x00\x01|K\xb5\xff\xe7\ \x00\x00\x0c\xd4\x00\x00\x00\x00\x00\x01\x00\x01\x99\xc6\ \x00\x00\x01|K\xb5\xff\xed\ \x00\x00\x09x\x00\x00\x00\x00\x00\x01\x00\x01\x88I\ \x00\x00\x01|K\xb5\xff\xe2\ \x00\x00\x09X\x00\x00\x00\x00\x00\x01\x00\x01\x86\x0c\ \x00\x00\x01|K\xb5\xff\xf1\ \x00\x00\x0dV\x00\x00\x00\x00\x00\x01\x00\x01\x9d\xbd\ \x00\x00\x01|K\xb5\xff\xf1\ \x00\x00\x05<\x00\x00\x00\x00\x00\x01\x00\x01l\xfb\ \x00\x00\x01|K\xb5\xff\xf2\ \x00\x00\x0b\x82\x00\x00\x00\x00\x00\x01\x00\x01\x93_\ \x00\x00\x01|K\xb5\xff\xed\ \x00\x00\x05\xce\x00\x00\x00\x00\x00\x01\x00\x01p\x1f\ \x00\x00\x01|K\xb5\xff\xdd\ \x00\x00\x10\xea\x00\x00\x00\x00\x00\x01\x00\x01\xb7i\ \x00\x00\x01|K\xb5\xff\xe5\ \x00\x00\x09\xac\x00\x00\x00\x00\x00\x01\x00\x01\x89\xfb\ \x00\x00\x01|K\xb5\xff\xe8\ \x00\x00\x0e\xba\x00\x00\x00\x00\x00\x01\x00\x01\xa6\x8a\ \x00\x00\x01|K\xb5\xff\xf9\ \x00\x00\x0b\x18\x00\x00\x00\x00\x00\x01\x00\x01\x90\x8f\ \x00\x00\x01|K\xb5\xff\xd3\ \x00\x00\x0a:\x00\x00\x00\x00\x00\x01\x00\x01\x8d%\ \x00\x00\x01|K\xb5\xff\xf7\ \x00\x00\x0f\x84\x00\x00\x00\x00\x00\x01\x00\x01\xacl\ \x00\x00\x01|K\xb5\xff\xee\ \x00\x00\x0e\x16\x00\x00\x00\x00\x00\x01\x00\x01\xa2\x0f\ \x00\x00\x01|K\xb5\xff\xe4\ \x00\x00\x06\x90\x00\x00\x00\x00\x00\x01\x00\x01u:\ \x00\x00\x01|K\xb5\xff\xea\ \x00\x00\x10\xca\x00\x00\x00\x00\x00\x01\x00\x01\xb5\xcb\ \x00\x00\x01|K\xb5\xff\xf6\ \x00\x00\x0d\xbe\x00\x00\x00\x00\x00\x01\x00\x01\xa0\xcf\ \x00\x00\x01|K\xb5\xff\xdf\ \x00\x00\x11\x1a\x00\x00\x00\x00\x00\x01\x00\x01\xba\x05\ \x00\x00\x01|K\xb5\xff\xd9\ \x00\x00\x07Z\x00\x00\x00\x00\x00\x01\x00\x01y\xd7\ \x00\x00\x01|K\xb5\xff\xde\ \x00\x00\x0bD\x00\x00\x00\x00\x00\x01\x00\x01\x91\xf7\ \x00\x00\x01|K\xb5\xff\xd4\ \x00\x00\x0eP\x00\x00\x00\x00\x00\x01\x00\x01\xa3\xb3\ \x00\x00\x01|K\xb5\xff\xdd\ \x00\x00\x0c\x90\x00\x00\x00\x00\x00\x01\x00\x01\x98{\ \x00\x00\x01|K\xb5\xff\xe0\ \x00\x00\x0d\x14\x00\x00\x00\x00\x00\x01\x00\x01\x9b!\ \x00\x00\x01|K\xb5\xff\xe6\ \x00\x00\x0e\xfe\x00\x00\x00\x00\x00\x01\x00\x01\xa7\xdf\ \x00\x00\x01|K\xb5\xff\xe1\ \x00\x00\x0a\xd0\x00\x00\x00\x00\x00\x01\x00\x01\x8f\x7f\ \x00\x00\x01|K\xb5\xff\xd6\ \x00\x00\x07\xe0\x00\x00\x00\x00\x00\x01\x00\x01|_\ \x00\x00\x01|K\xb5\xff\xde\ \x00\x00\x03\xbe\x00\x00\x00\x00\x00\x01\x00\x01b\x1f\ \x00\x00\x01|K\xb5\xff\xd6\ \x00\x00\x06L\x00\x00\x00\x00\x00\x01\x00\x01t\x1b\ \x00\x00\x01|K\xb5\xff\xf4\ \x00\x00\x0f \x00\x00\x00\x00\x00\x01\x00\x01\xa9\x91\ \x00\x00\x01|K\xb5\xff\xea\ \x00\x00\x06\x1e\x00\x00\x00\x00\x00\x01\x00\x01r\xaf\ \x00\x00\x01|K\xb5\xff\xf0\ \x00\x00\x07\xa0\x00\x00\x00\x00\x00\x01\x00\x01{\x17\ \x00\x00\x01|K\xb5\xff\xec\ \x00\x00\x03\xf0\x00\x00\x00\x00\x00\x01\x00\x01c/\ \x00\x00\x01|K\xb5\xff\xd5\ \x00\x00\x04\x86\x00\x00\x00\x00\x00\x01\x00\x01ha\ \x00\x00\x01|K\xb5\xff\xe2\ \x00\x00\x10P\x00\x00\x00\x00\x00\x01\x00\x01\xb0\x8a\ \x00\x00\x01|K\xb5\xff\xe4\ \x00\x00\x0cN\x00\x00\x00\x00\x00\x01\x00\x01\x97\x16\ \x00\x00\x01|K\xb5\xff\xda\ \x00\x00\x09\x16\x00\x00\x00\x00\x00\x01\x00\x01\x85\x0d\ \x00\x00\x01|K\xb5\xff\xe3\ \x00\x00\x0f\xc8\x00\x00\x00\x00\x00\x01\x00\x01\xae \ \x00\x00\x01|K\xb5\xff\xf5\ \x00\x00\x02\xcc\x00\x00\x00\x00\x00\x01\x00\x01P7\ \x00\x00\x01|\x92\x17g\xb1\ \x00\x00\x02|\x00\x00\x00\x00\x00\x01\x00\x01\x1a9\ \x00\x00\x01|\x92\x17g\xb4\ \x00\x00\x02`\x00\x00\x00\x00\x00\x01\x00\x01\x15\xd8\ \x00\x00\x01|\x92\x17g\xb2\ \x00\x00\x03\x0a\x00\x00\x00\x00\x00\x01\x00\x01R$\ \x00\x00\x01|z\x1a\x99\xb6\ \x00\x00\x01\x90\x00\x00\x00\x00\x00\x01\x00\x01\x0c\x99\ \x00\x00\x01|\x92\x17g\xb3\ \x00\x00\x02\x1a\x00\x00\x00\x00\x00\x01\x00\x01\x14\xb3\ \x00\x00\x01|x\xc9R\x9c\ \x00\x00\x01\xcc\x00\x00\x00\x00\x00\x01\x00\x01\x0eT\ \x00\x00\x01|\x92\x17g\xb5\ \x00\x00\x03N\x00\x00\x00\x00\x00\x01\x00\x01_+\ \x00\x00\x01|x\xc8\xac\x18\ \x00\x00\x02\xae\x00\x01\x00\x00\x00\x01\x00\x01\x1cA\ \x00\x00\x01|\x92\x17g\xae\ \x00\x00\x01\xea\x00\x00\x00\x00\x00\x01\x00\x01\x12\xf9\ \x00\x00\x01|\x92\x17g\xb6\ \x00\x00\x01\x16\x00\x00\x00\x00\x00\x01\x00\x00>\x17\ \x00\x00\x01|\x92\x17g\xb7\ \x00\x00\x018\x00\x00\x00\x00\x00\x01\x00\x00\xd0\xdd\ \x00\x00\x01|\x92\x17g\xb7\ " def qInitResources(): QtCore.qRegisterResourceData(0x03, qt_resource_struct, qt_resource_name, qt_resource_data) def qCleanupResources(): QtCore.qUnregisterResourceData(0x03, qt_resource_struct, qt_resource_name, qt_resource_data) qInitResources()
py
1a2f3cea35bbeab0fac5b0a6fd23d61f79be9b2c
class Encapsulada: atributo_visible = 'soy visible' _atributo_protegido = 'soy protegido' __atributo_privado = 'soy un atributo privado' def get_atributo_privado(self): return self.__atributo_privado def set_atributo_privado(self, cambio): self.__atributo_privado = cambio
py
1a2f3edc14828893e52f591f241c6f4d3c54fce3
# coding: utf-8 """ NetBox API API to access NetBox # noqa: E501 OpenAPI spec version: 2.8 Generated by: https://github.com/swagger-api/swagger-codegen.git """ from __future__ import absolute_import import unittest import netbox_client from netbox_client.models.writable_front_port import WritableFrontPort # noqa: E501 from netbox_client.rest import ApiException class TestWritableFrontPort(unittest.TestCase): """WritableFrontPort unit test stubs""" def setUp(self): pass def tearDown(self): pass def testWritableFrontPort(self): """Test WritableFrontPort""" # FIXME: construct object with mandatory attributes with example values # model = netbox_client.models.writable_front_port.WritableFrontPort() # noqa: E501 pass if __name__ == '__main__': unittest.main()
py
1a2f3fdd43f2a1b0b52665cf073448048e5e40fb
from operator import attrgetter import pyangbind.lib.xpathhelper as xpathhelper from pyangbind.lib.yangtypes import RestrictedPrecisionDecimalType, RestrictedClassType, TypedListType from pyangbind.lib.yangtypes import YANGBool, YANGListType, YANGDynClass, ReferenceType from pyangbind.lib.base import PybindBase from decimal import Decimal from bitarray import bitarray import __builtin__ class lsp_admin_group_include_all(PybindBase): """ This class was auto-generated by the PythonClass plugin for PYANG from YANG module brocade-mpls - based on the path /brocade_mpls_rpc/show-mpls-lsp-debug/output/lsp/show-mpls-lsp-debug-info/show-mpls-lsp-common-info/lsp-config-frr-admin-groups/lsp-admin-group/lsp-admin-group-include-all. Each member element of the container is represented as a class variable - with a specific YANG type. """ __slots__ = ('_pybind_generated_by', '_path_helper', '_yang_name', '_rest_name', '_extmethods', '__lsp_admin_group_include_all_group_id',) _yang_name = 'lsp-admin-group-include-all' _rest_name = 'lsp-admin-group-include-all' _pybind_generated_by = 'container' def __init__(self, *args, **kwargs): path_helper_ = kwargs.pop("path_helper", None) if path_helper_ is False: self._path_helper = False elif path_helper_ is not None and isinstance(path_helper_, xpathhelper.YANGPathHelper): self._path_helper = path_helper_ elif hasattr(self, "_parent"): path_helper_ = getattr(self._parent, "_path_helper", False) self._path_helper = path_helper_ else: self._path_helper = False extmethods = kwargs.pop("extmethods", None) if extmethods is False: self._extmethods = False elif extmethods is not None and isinstance(extmethods, dict): self._extmethods = extmethods elif hasattr(self, "_parent"): extmethods = getattr(self._parent, "_extmethods", None) self._extmethods = extmethods else: self._extmethods = False self.__lsp_admin_group_include_all_group_id = YANGDynClass(base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name="lsp-admin-group-include-all-group-id", rest_name="lsp-admin-group-include-all-group-id", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, is_keyval=True, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True) load = kwargs.pop("load", None) if args: if len(args) > 1: raise TypeError("cannot create a YANG container with >1 argument") all_attr = True for e in self._pyangbind_elements: if not hasattr(args[0], e): all_attr = False break if not all_attr: raise ValueError("Supplied object did not have the correct attributes") for e in self._pyangbind_elements: nobj = getattr(args[0], e) if nobj._changed() is False: continue setmethod = getattr(self, "_set_%s" % e) if load is None: setmethod(getattr(args[0], e)) else: setmethod(getattr(args[0], e), load=load) def _path(self): if hasattr(self, "_parent"): return self._parent._path()+[self._yang_name] else: return [u'brocade_mpls_rpc', u'show-mpls-lsp-debug', u'output', u'lsp', u'show-mpls-lsp-debug-info', u'show-mpls-lsp-common-info', u'lsp-config-frr-admin-groups', u'lsp-admin-group', u'lsp-admin-group-include-all'] def _rest_path(self): if hasattr(self, "_parent"): if self._rest_name: return self._parent._rest_path()+[self._rest_name] else: return self._parent._rest_path() else: return [u'show-mpls-lsp-debug', u'output', u'lsp', u'lsp-config-frr-admin-groups', u'lsp-admin-group-include-all'] def _get_lsp_admin_group_include_all_group_id(self): """ Getter method for lsp_admin_group_include_all_group_id, mapped from YANG variable /brocade_mpls_rpc/show_mpls_lsp_debug/output/lsp/show_mpls_lsp_debug_info/show_mpls_lsp_common_info/lsp_config_frr_admin_groups/lsp_admin_group/lsp_admin_group_include_all/lsp_admin_group_include_all_group_id (uint32) YANG Description: Include all admin group id """ return self.__lsp_admin_group_include_all_group_id def _set_lsp_admin_group_include_all_group_id(self, v, load=False): """ Setter method for lsp_admin_group_include_all_group_id, mapped from YANG variable /brocade_mpls_rpc/show_mpls_lsp_debug/output/lsp/show_mpls_lsp_debug_info/show_mpls_lsp_common_info/lsp_config_frr_admin_groups/lsp_admin_group/lsp_admin_group_include_all/lsp_admin_group_include_all_group_id (uint32) If this variable is read-only (config: false) in the source YANG file, then _set_lsp_admin_group_include_all_group_id is considered as a private method. Backends looking to populate this variable should do so via calling thisObj._set_lsp_admin_group_include_all_group_id() directly. YANG Description: Include all admin group id """ parent = getattr(self, "_parent", None) if parent is not None and load is False: raise AttributeError("Cannot set keys directly when" + " within an instantiated list") if hasattr(v, "_utype"): v = v._utype(v) try: t = YANGDynClass(v,base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name="lsp-admin-group-include-all-group-id", rest_name="lsp-admin-group-include-all-group-id", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, is_keyval=True, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True) except (TypeError, ValueError): raise ValueError({ 'error-string': """lsp_admin_group_include_all_group_id must be of a type compatible with uint32""", 'defined-type': "uint32", 'generated-type': """YANGDynClass(base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name="lsp-admin-group-include-all-group-id", rest_name="lsp-admin-group-include-all-group-id", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, is_keyval=True, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True)""", }) self.__lsp_admin_group_include_all_group_id = t if hasattr(self, '_set'): self._set() def _unset_lsp_admin_group_include_all_group_id(self): self.__lsp_admin_group_include_all_group_id = YANGDynClass(base=RestrictedClassType(base_type=long, restriction_dict={'range': ['0..4294967295']}, int_size=32), is_leaf=True, yang_name="lsp-admin-group-include-all-group-id", rest_name="lsp-admin-group-include-all-group-id", parent=self, path_helper=self._path_helper, extmethods=self._extmethods, register_paths=False, is_keyval=True, namespace='urn:brocade.com:mgmt:brocade-mpls', defining_module='brocade-mpls', yang_type='uint32', is_config=True) lsp_admin_group_include_all_group_id = __builtin__.property(_get_lsp_admin_group_include_all_group_id, _set_lsp_admin_group_include_all_group_id) _pyangbind_elements = {'lsp_admin_group_include_all_group_id': lsp_admin_group_include_all_group_id, }
py
1a2f40bea97426641a515b43d3e4286b9d5483b1
from __future__ import absolute_import import pynndescent from ann_benchmarks.algorithms.base import BaseANN class PyNNDescent(BaseANN): def __init__(self, metric, n_neighbors=10, n_trees=8, leaf_size=20): self._n_neighbors = int(n_neighbors) self._n_trees = int(n_trees) self._leaf_size = int(leaf_size) self._queue_size=None self._pynnd_metric = {'angular': 'cosine', 'euclidean': 'euclidean', 'hamming': 'hamming', 'jaccard': 'jaccard'}[metric] def fit(self, X): self._index = pynndescent.NNDescent(X, n_neighbors=self._n_neighbors, n_trees=self._n_trees, leaf_size=self._leaf_size, metric=self._pynnd_metric) def set_query_arguments(self, queue_size): self._queue_size = float(queue_size) def query(self, v, n): ind, dist = self._index.query(v.reshape(1, -1).astype('float32'), k=n, queue_size=self._queue_size) return ind[0] def use_threads(self): return False def __str__(self): return 'PyNNDescent(n_neighbors=%d, n_trees=%d, leaf_size=%d, queue_size=%.2f)' % (self._n_neighbors, self._n_trees, self._leaf_size, self._queue_size)
py
1a2f410f822d032450aa752eb50cb91c79df1f19
import os import json import uuid import boto3 from datasources.stac.query import STACQuery from datasources.sources.base import Datasource client = boto3.client('s3') bucket = 'usgs-lidar-public' class USGS3DEP(Datasource): stac_compliant = False tags = ['Elevation', 'Raster'] def __init__(self, manifest): super().__init__(manifest) def search(self, spatial, temporal=None, properties=None, limit=10, **kwargs): from db import Database names = [] stac_query = STACQuery(spatial, temporal, properties) # projects = stac_query.check_spatial(self.__class__.__name__)[:limit] with Database.load(read_only=True, deployed=True) as db: projects = db.spatial_query({"type": "Feature", "geometry": stac_query.spatial}) searches = 0 for item in projects: if item['name'] not in names: if temporal and item['year']: if stac_query.temporal[0].year != item['year'] or stac_query.temporal[1].year != item['year']: continue if properties: item.update({'properties': stac_query}) if searches < limit: self.manifest.searches.append([self, item]) searches+=1 # searches = 0 # for item in projects: # if item['name'] not in names: # # Temporal check by checking year of start/end date # if temporal and item['year']: # if stac_query.temporal[0].year == item['year'] or stac_query.temporal[1].year == item['year']: # if properties: # item.update({'properties': stac_query}) # self.manifest.searches.append([self, item]) # names.append(item['name']) # else: # self.manifest.searches.append([self, item]) # names.append(item['name']) def execute(self, query): # Download metadata from query item response = client.get_object(Bucket=bucket, Key=os.path.join(query['name'], 'ept.json')) metadata = json.loads(response['Body'].read().decode('utf-8')) xvals = [x[0] for x in query['geometry']['coordinates'][0]] yvals = [y[1] for y in query['geometry']['coordinates'][0]] stac_item = { 'id': str(uuid.uuid4()), 'type': 'Feature', 'bbox': [min(xvals), min(yvals), max(xvals), max(yvals)], 'geometry': query['geometry'], 'properties': { 'datetime': f"{query['year']}-01-01T00:00:00.00Z", 'eo:epsg': metadata['srs']['horizontal'], 'pc:count': metadata['points'], 'pc:type': 'lidar', 'pc:encoding': metadata['dataType'], 'pc:schema': metadata['schema'], 'legacy:span': metadata['span'], 'legacy:version': metadata['version'], }, 'assets': { 's3path': { 'href': f"s3://{bucket}/{query['name']}", 'title': 'EPT data' } }, } if "properties" in list(query): if query['properties'].check_properties(stac_item['properties']): return [stac_item] else: return [stac_item]
py
1a2f42594b5ecd2ed04908d58cdc7e79376638b5
import re import collections from enum import Enum from ydk._core._dm_meta_info import _MetaInfoClassMember, _MetaInfoClass, _MetaInfoEnum from ydk.types import Empty, YList, YLeafList, DELETE, Decimal64, FixedBitsDict from ydk._core._dm_meta_info import ATTRIBUTE, REFERENCE_CLASS, REFERENCE_LIST, REFERENCE_LEAFLIST, REFERENCE_IDENTITY_CLASS, REFERENCE_ENUM_CLASS, REFERENCE_BITS, REFERENCE_UNION from ydk.errors import YPYError, YPYModelError from ydk.providers._importer import _yang_ns _meta_table = { 'HistRecordEnum' : _MetaInfoEnum('HistRecordEnum', 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', { 'cfghist-bag-record-all':'cfghist_bag_record_all', 'cfghist-bag-record-alarm':'cfghist_bag_record_alarm', 'cfghist-bag-record-cfs-check':'cfghist_bag_record_cfs_check', 'cfghist-bag-record-commit':'cfghist_bag_record_commit', 'cfghist-bag-record-oir':'cfghist_bag_record_oir', 'cfghist-bag-record-shutdown':'cfghist_bag_record_shutdown', 'cfghist-bag-record-startup':'cfghist_bag_record_startup', 'cfghist-bag-record-backup':'cfghist_bag_record_backup', 'cfghist-bag-record-rebase':'cfghist_bag_record_rebase', 'cfghist-bag-record-last':'cfghist_bag_record_last', }, 'Cisco-IOS-XR-config-cfgmgr-exec-oper', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper']), 'CfgHistGl.RecordType.Record.Info.AlarmInfo' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record.Info.AlarmInfo', False, [ _MetaInfoClassMember('state', ATTRIBUTE, 'str' , None, None, [], [], ''' State ''', 'state', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('where', ATTRIBUTE, 'str' , None, None, [], [], ''' Where ''', 'where', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'alarm-info', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType.Record.Info.CfscheckInfo' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record.Info.CfscheckInfo', False, [ _MetaInfoClassMember('line', ATTRIBUTE, 'str' , None, None, [], [], ''' Line ''', 'line', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('user-id', ATTRIBUTE, 'str' , None, None, [], [], ''' UserId ''', 'user_id', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'cfscheck-info', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType.Record.Info.CommitInfo' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record.Info.CommitInfo', False, [ _MetaInfoClassMember('client-name', ATTRIBUTE, 'str' , None, None, [], [], ''' Client name ''', 'client_name', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('comment', ATTRIBUTE, 'str' , None, None, [], [], ''' Comment ''', 'comment', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('commit-id', ATTRIBUTE, 'str' , None, None, [], [], ''' CommitId ''', 'commit_id', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('label', ATTRIBUTE, 'str' , None, None, [], [], ''' Label ''', 'label', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('line', ATTRIBUTE, 'str' , None, None, [], [], ''' Line ''', 'line', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('user-id', ATTRIBUTE, 'str' , None, None, [], [], ''' UserId ''', 'user_id', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'commit-info', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType.Record.Info.OirInfo' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record.Info.OirInfo', False, [ _MetaInfoClassMember('config-name', ATTRIBUTE, 'str' , None, None, [], [], ''' Config Name ''', 'config_name', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('config-type', ATTRIBUTE, 'str' , None, None, [], [], ''' Config Type ''', 'config_type', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('operation', ATTRIBUTE, 'str' , None, None, [], [], ''' Operation ''', 'operation', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'oir-info', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType.Record.Info.ShutdownInfo' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record.Info.ShutdownInfo', False, [ _MetaInfoClassMember('comment', ATTRIBUTE, 'str' , None, None, [], [], ''' Comment ''', 'comment', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'shutdown-info', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType.Record.Info.StartupInfo' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record.Info.StartupInfo', False, [ _MetaInfoClassMember('boot-path', ATTRIBUTE, 'str' , None, None, [], [], ''' Boot Path ''', 'boot_path', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('how-booted', ATTRIBUTE, 'str' , None, None, [], [], ''' How Booted ''', 'how_booted', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'startup-info', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType.Record.Info.BackupInfo' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record.Info.BackupInfo', False, [ _MetaInfoClassMember('comment', ATTRIBUTE, 'str' , None, None, [], [], ''' Comment ''', 'comment', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'backup-info', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType.Record.Info' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record.Info', False, [ _MetaInfoClassMember('a', ATTRIBUTE, 'int' , None, None, [('0', '4294967295')], [], ''' B ''', 'a', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('alarm-info', REFERENCE_CLASS, 'AlarmInfo' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record.Info.AlarmInfo', [], [], ''' alarm info ''', 'alarm_info', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('backup-info', REFERENCE_CLASS, 'BackupInfo' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record.Info.BackupInfo', [], [], ''' backup info ''', 'backup_info', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('cfscheck-info', REFERENCE_CLASS, 'CfscheckInfo' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record.Info.CfscheckInfo', [], [], ''' cfscheck info ''', 'cfscheck_info', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('commit-info', REFERENCE_CLASS, 'CommitInfo' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record.Info.CommitInfo', [], [], ''' commit info ''', 'commit_info', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('oir-info', REFERENCE_CLASS, 'OirInfo' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record.Info.OirInfo', [], [], ''' oir info ''', 'oir_info', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('shutdown-info', REFERENCE_CLASS, 'ShutdownInfo' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record.Info.ShutdownInfo', [], [], ''' shutdown info ''', 'shutdown_info', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('startup-info', REFERENCE_CLASS, 'StartupInfo' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record.Info.StartupInfo', [], [], ''' startup info ''', 'startup_info', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('type', REFERENCE_ENUM_CLASS, 'HistRecordEnum' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'HistRecordEnum', [], [], ''' type ''', 'type', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'info', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType.Record' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType.Record', False, [ _MetaInfoClassMember('record', ATTRIBUTE, 'int' , None, None, [('-2147483648', '2147483647')], [], ''' Record ''', 'record', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', True), _MetaInfoClassMember('info', REFERENCE_CLASS, 'Info' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record.Info', [], [], ''' Content of the history ''', 'info', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('record-type', REFERENCE_ENUM_CLASS, 'HistRecordEnum' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'HistRecordEnum', [], [], ''' Record type ''', 'record_type', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), _MetaInfoClassMember('timestamp', ATTRIBUTE, 'int' , None, None, [('0', '4294967295')], [], ''' Time stamp for the history ''', 'timestamp', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'record', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl.RecordType' : { 'meta_info' : _MetaInfoClass('CfgHistGl.RecordType', False, [ _MetaInfoClassMember('record-type', ATTRIBUTE, 'str' , None, None, [], ['[\\w\\-\\.:,_@#%$\\+=\\|;]+'], ''' Record type ''', 'record_type', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', True), _MetaInfoClassMember('record', REFERENCE_LIST, 'Record' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType.Record', [], [], ''' History summary information for a specific type of history ''', 'record', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'record-type', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, 'CfgHistGl' : { 'meta_info' : _MetaInfoClass('CfgHistGl', False, [ _MetaInfoClassMember('record-type', REFERENCE_LIST, 'RecordType' , 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper', 'CfgHistGl.RecordType', [], [], ''' History summary information for a specific type of history ''', 'record_type', 'Cisco-IOS-XR-config-cfgmgr-exec-oper', False), ], 'Cisco-IOS-XR-config-cfgmgr-exec-oper', 'cfg-hist-gl', _yang_ns._namespaces['Cisco-IOS-XR-config-cfgmgr-exec-oper'], 'ydk.models.cisco_ios_xr.Cisco_IOS_XR_config_cfgmgr_exec_oper' ), }, } _meta_table['CfgHistGl.RecordType.Record.Info.AlarmInfo']['meta_info'].parent =_meta_table['CfgHistGl.RecordType.Record.Info']['meta_info'] _meta_table['CfgHistGl.RecordType.Record.Info.CfscheckInfo']['meta_info'].parent =_meta_table['CfgHistGl.RecordType.Record.Info']['meta_info'] _meta_table['CfgHistGl.RecordType.Record.Info.CommitInfo']['meta_info'].parent =_meta_table['CfgHistGl.RecordType.Record.Info']['meta_info'] _meta_table['CfgHistGl.RecordType.Record.Info.OirInfo']['meta_info'].parent =_meta_table['CfgHistGl.RecordType.Record.Info']['meta_info'] _meta_table['CfgHistGl.RecordType.Record.Info.ShutdownInfo']['meta_info'].parent =_meta_table['CfgHistGl.RecordType.Record.Info']['meta_info'] _meta_table['CfgHistGl.RecordType.Record.Info.StartupInfo']['meta_info'].parent =_meta_table['CfgHistGl.RecordType.Record.Info']['meta_info'] _meta_table['CfgHistGl.RecordType.Record.Info.BackupInfo']['meta_info'].parent =_meta_table['CfgHistGl.RecordType.Record.Info']['meta_info'] _meta_table['CfgHistGl.RecordType.Record.Info']['meta_info'].parent =_meta_table['CfgHistGl.RecordType.Record']['meta_info'] _meta_table['CfgHistGl.RecordType.Record']['meta_info'].parent =_meta_table['CfgHistGl.RecordType']['meta_info'] _meta_table['CfgHistGl.RecordType']['meta_info'].parent =_meta_table['CfgHistGl']['meta_info']
py
1a2f4274d7daac60ae99f57319b933b10c1b479f
from __future__ import absolute_import # Copyright (c) 2010-2018 openpyxl """Write the shared string table.""" from io import BytesIO # package imports from openpyxl2.xml.constants import SHEET_MAIN_NS from openpyxl2.xml.functions import Element, xmlfile, SubElement PRESERVE_SPACE = '{%s}space' % "http://www.w3.org/XML/1998/namespace" def write_string_table(string_table): """Write the string table xml.""" out = BytesIO() with xmlfile(out) as xf: with xf.element("sst", xmlns=SHEET_MAIN_NS, uniqueCount="%d" % len(string_table)): for key in string_table: el = Element('si') text = SubElement(el, 't') text.text = key if key.strip() != key: text.set(PRESERVE_SPACE, 'preserve') xf.write(el) return out.getvalue()
py
1a2f43964d70308be0f37bdf681a3903a890565b
""" Building Block ============== """ from __future__ import annotations import logging import os import typing import warnings from collections.abc import Collection from functools import partial import numpy as np import rdkit.Chem.AllChem as rdkit import vabene from ...utilities import OneOrMany, flatten, remake from ..atoms import Atom from ..bonds import Bond from ..functional_groups import FunctionalGroup, FunctionalGroupFactory from .molecule import Molecule logger = logging.getLogger(__name__) class BuildingBlock(Molecule): """ Represents a building block of a :class:`.ConstructedMolecule`. A :class:`BuildingBlock` can represent either an entire molecule or a molecular fragments used to construct a :class:`.ConstructedMolecule`. The building block uses :class:`.FunctionalGroup` instances to identify which atoms are modified during construction. """ # Maps file extensions to functions which can be used to # create an rdkit molecule from that file type. _init_funcs = { '.mol': partial( rdkit.MolFromMolFile, sanitize=False, removeHs=False ), '.sdf': partial( rdkit.MolFromMolFile, sanitize=False, removeHs=False ), '.pdb': partial( rdkit.MolFromPDBFile, sanitize=False, removeHs=False, proximityBonding=False, ), } _placer_ids: frozenset[int] _core_ids: frozenset[int] def __init__( self, smiles: str, functional_groups: typing.Iterable[ typing.Union[FunctionalGroup, FunctionalGroupFactory] ] = (), placer_ids: typing.Optional[tuple[int, ...]] = None, position_matrix: typing.Optional[np.ndarray] = None, ) -> None: """ Initialize a :class:`.BuildingBlock`. Notes: The molecule is given 3D coordinates with :func:`rdkit.ETKDGv2`. Parameters: smiles: A SMILES string of the molecule. functional_groups: An :class:`iterable` of :class:`.FunctionalGroup` or :class:`.FunctionalGroupFactory` or both. :class:`.FunctionalGroup` instances are added to the building block and :class:`.FunctionalGroupFactory` instances are used to create :class:`.FunctionalGroup` instances the building block should hold. :class:`.FunctionalGroup` instances are used to identify which atoms are modified during :class:`.ConstructedMolecule` construction. placer_ids: The ids of *placer* atoms. These are the atoms which should be used for calculating the position of the building block. Depending on the values passed to `placer_ids`, and the functional groups in the building block, different *placer* ids will be used by the building block. #. `placer_ids` is passed to the initializer: the passed *placer* ids will be used by the building block. #. `placer_ids` is ``None`` and the building block has functional groups: The *placer* ids of the functional groups will be used as the *placer* ids of the building block. #. `placer_ids` is ``None`` and `functional_groups` is empty. All atoms of the molecule will be used for *placer* ids. position_matrix: The position matrix the building block should use. If ``None``, :func:`rdkit.ETKDGv2` will be used to calculate it. Raises: :class:`RuntimeError` If embedding the molecule fails. """ molecule = rdkit.AddHs(rdkit.MolFromSmiles(smiles)) if position_matrix is None: params = rdkit.ETKDGv2() random_seed = 4 params.randomSeed = random_seed if rdkit.EmbedMolecule(molecule, params) == -1: raise RuntimeError( f'Embedding with seed value of {random_seed} ' 'failed.' ) rdkit.Kekulize(molecule) else: # Make sure the position matrix always holds floats. position_matrix = np.array( position_matrix, dtype=np.float64, ) conformer = rdkit.Conformer(molecule.GetNumAtoms()) for atom_id, position in enumerate(position_matrix): conformer.SetAtomPosition(atom_id, position) molecule.AddConformer(conformer) self._init_from_rdkit_mol( molecule=molecule, functional_groups=functional_groups, placer_ids=placer_ids, ) @classmethod def init_from_molecule( cls, molecule: Molecule, functional_groups: typing.Iterable[ typing.Union[FunctionalGroup, FunctionalGroupFactory] ] = (), placer_ids: typing.Optional[tuple[int, ...]] = None, ) -> BuildingBlock: """ Initialize from a :class:`.Molecule`. Parameters: molecule: The molecule to initialize from. functional_groups: An :class:`iterable` of :class:`.FunctionalGroup` or :class:`.FunctionalGroupFactory` or both. :class:`.FunctionalGroup` instances are added to the building block and :class:`.FunctionalGroupFactory` instances are used to create :class:`.FunctionalGroup` instances the building block should hold. :class:`.FunctionalGroup` instances are used to identify which atoms are modified during :class:`.ConstructedMolecule` construction. placer_ids: The ids of *placer* atoms. These are the atoms which should be used for calculating the position of the building block. Depending on the values passed to `placer_ids`, and the functional groups in the building block, different *placer* ids will be used by the building block. #. `placer_ids` is passed to the initializer: the passed *placer* ids will be used by the building block. #. `placer_ids` is ``None`` and the building block has functional groups: The *placer* ids of the functional groups will be used as the *placer* ids of the building block. #. `placer_ids` is ``None`` and `functional_groups` is empty. All atoms of the molecule will be used for *placer* ids. Returns: The building block. It will have the same atoms, bonds and atomic positions as `molecule`. """ return cls.init( atoms=tuple(molecule.get_atoms()), bonds=tuple(molecule.get_bonds()), position_matrix=molecule.get_position_matrix(), functional_groups=functional_groups, placer_ids=placer_ids, ) @classmethod def init_from_vabene_molecule( cls, molecule: vabene.Molecule, functional_groups: typing.Iterable[ typing.Union[FunctionalGroup, FunctionalGroupFactory] ] = (), placer_ids: typing.Optional[tuple[int, ...]] = None, position_matrix: typing.Optional[np.ndarray] = None, ) -> BuildingBlock: """ Initialize from a :mod:`vabene.Molecule`. Notes: The molecule is given 3D coordinates with :func:`rdkit.ETKDGv2()`. Parameters: molecule: The :class:`vabene.Molecule` from which to initialize. functional_groups: An :class:`iterable` of :class:`.FunctionalGroup` or :class:`.FunctionalGroupFactory` or both. :class:`.FunctionalGroup` instances are added to the building block and :class:`.FunctionalGroupFactory` instances are used to create :class:`.FunctionalGroup` instances the building block should hold. :class:`.FunctionalGroup` instances are used to identify which atoms are modified during :class:`.ConstructedMolecule` construction. placer_ids: The ids of *placer* atoms. These are the atoms which should be used for calculating the position of the building block. Depending on the values passed to `placer_ids`, and the functional groups in the building block, different *placer* ids will be used by the building block. #. `placer_ids` is passed to the initializer: the passed *placer* ids will be used by the building block. #. `placer_ids` is ``None`` and the building block has functional groups: The *placer* ids of the functional groups will be used as the *placer* ids of the building block. #. `placer_ids` is ``None`` and `functional_groups` is empty. All atoms of the molecule will be used for *placer* ids. position_matrix: The position matrix the building block should use. If ``None``, :func:`rdkit.ETKDGv2` will be used to calculate it. Returns: The building block. Raises: :class:`RuntimeError` If embedding the molecule fails. """ editable = rdkit.EditableMol(rdkit.Mol()) for atom in molecule.get_atoms(): rdkit_atom = rdkit.Atom(atom.get_atomic_number()) rdkit_atom.SetFormalCharge(atom.get_charge()) editable.AddAtom(rdkit_atom) for bond in molecule.get_bonds(): editable.AddBond( beginAtomIdx=bond.get_atom1_id(), endAtomIdx=bond.get_atom2_id(), order=rdkit.BondType(bond.get_order()), ) rdkit_molecule = editable.GetMol() rdkit.SanitizeMol(rdkit_molecule) rdkit_molecule = rdkit.AddHs(rdkit_molecule) if position_matrix is None: params = rdkit.ETKDGv2() random_seed = 4 params.randomSeed = random_seed if rdkit.EmbedMolecule(rdkit_molecule, params) == -1: raise RuntimeError( f'Embedding with seed value of {random_seed} ' 'failed.' ) else: # Make sure the position matrix always holds floats. position_matrix = np.array( position_matrix, dtype=np.float64, ) conformer = rdkit.Conformer(rdkit_molecule.GetNumAtoms()) for atom_id, position in enumerate(position_matrix): conformer.SetAtomPosition(atom_id, position) rdkit_molecule.AddConformer(conformer) rdkit.Kekulize(rdkit_molecule) return cls.init_from_rdkit_mol( molecule=rdkit_molecule, functional_groups=functional_groups, placer_ids=placer_ids, ) @classmethod def init( cls, atoms: tuple[Atom, ...], bonds: tuple[Bond, ...], position_matrix: np.ndarray, functional_groups: typing.Iterable[ typing.Union[FunctionalGroup, FunctionalGroupFactory] ] = (), placer_ids: typing.Optional[tuple[int, ...]] = None, ) -> BuildingBlock: """ Initialize a :class:`.BuildingBlock` from its components. Parameters: atoms: The atoms of the building block. bonds: The bonds of the building block. position_matrix: An ``(n, 3)`` position matrix of the building block. functional_groups: An :class:`iterable` holding the :class:`.FunctionalGroup` instances the building block should have, and / or :class:`.FunctionalGroupFactory` instances used for creating them. placer_ids: The ids of *placer* atoms. These are the atoms which should be used for calculating the position of the building block. Depending on the values passed to `placer_ids`, and the functional groups in the building block, different *placer* ids will be used by the building block. #. `placer_ids` is passed to the initializer: the passed *placer* ids will be used by the building block. #. `placer_ids` is ``None`` and the building block has functional groups: The *placer* ids of the functional groups will be used as the *placer* ids of the building block. #. `placer_ids` is ``None`` and `functional_groups` is empty. All atoms of the molecule will be used for *placer* ids. Returns: The building block. """ building_block = cls.__new__(cls) Molecule.__init__( self=building_block, atoms=atoms, bonds=bonds, position_matrix=position_matrix, ) functional_groups = building_block._extract_functional_groups( functional_groups=functional_groups, ) building_block._with_functional_groups(functional_groups) building_block._placer_ids = ( building_block._normalize_placer_ids( placer_ids=placer_ids, functional_groups=building_block._functional_groups, ) ) building_block._core_ids = frozenset( building_block._get_core_ids( functional_groups=building_block._functional_groups, ) ) return building_block @classmethod def init_from_file( cls, path: str, functional_groups: typing.Iterable[ typing.Union[FunctionalGroup, FunctionalGroupFactory] ] = (), placer_ids: typing.Optional[tuple[int, ...]] = None, ): """ Initialize from a file. Parameters: path: The path to a molecular structure file. Supported file types are: #. ``.mol``, ``.sdf`` - MDL V3000 MOL file functional_groups: An :class:`iterable` of :class:`.FunctionalGroup` or :class:`.FunctionalGroupFactory` or both. :class:`.FunctionalGroup` instances are added to the building block and :class:`.FunctionalGroupFactory` instances are used to create :class:`.FunctionalGroup` instances the building block should hold. :class:`.FunctionalGroup` instances are used to identify which atoms are modified during :class:`.ConstructedMolecule` construction. placer_ids: The ids of *placer* atoms. These are the atoms which should be used for calculating the position of the building block. Depending on the values passed to `placer_ids`, and the functional groups in the building block, different *placer* ids will be used by the building block. #. `placer_ids` is passed to the initializer: the passed *placer* ids will be used by the building block. #. `placer_ids` is ``None`` and the building block has functional groups: The *placer* ids of the functional groups will be used as the *placer* ids of the building block. #. `placer_ids` is ``None`` and `functional_groups` is empty. All atoms of the molecule will be used for *placer* ids. Returns: The building block. Raises: :class:`ValueError` If the file type cannot be used for initialization. """ _, extension = os.path.splitext(path) if extension == '.pdb': warnings.warn( 'Loading from .pdb files is deprecated and will be ' 'removed from stk versions released after 1st Nov ' '2022. Please use .mol files for loading molecules ' 'instead.', category=FutureWarning, ) if extension not in cls._init_funcs: raise ValueError( f'Unable to initialize from "{extension}" files.' ) # This remake needs to be here because molecules loaded # with rdkit often have issues, because rdkit tries to do # bits of structural analysis like stereocenters. Remake # gets rid of all this problematic metadata. molecule = remake(cls._init_funcs[extension](path)) return cls.init_from_rdkit_mol( molecule=molecule, functional_groups=functional_groups, placer_ids=placer_ids, ) @classmethod def init_from_rdkit_mol( cls, molecule: rdkit.Mol, functional_groups: typing.Iterable[ typing.Union[FunctionalGroup, FunctionalGroupFactory] ] = (), placer_ids: typing.Optional[tuple[int, ...]] = None, ) -> BuildingBlock: """ Initialize from an :mod:`rdkit` molecule. Warnings: For :mod:`rdkit` molecules with non-integer bond orders, such as 1.5, the molecule should be kekulized prior to calling this method. Otherwise, all bond orders will be set to an integer value in the building block. Parameters: molecule: The molecule. functional_groups: An :class:`iterable` of :class:`.FunctionalGroup` or :class:`.FunctionalGroupFactory` or both. :class:`.FunctionalGroup` instances are added to the building block and :class:`.FunctionalGroupFactory` instances are used to create :class:`.FunctionalGroup` instances the building block should hold. :class:`.FunctionalGroup` instances are used to identify which atoms are modified during :class:`.ConstructedMolecule` construction. placer_ids: The ids of *placer* atoms. These are the atoms which should be used for calculating the position of the building block. Depending on the values passed to `placer_ids`, and the functional groups in the building block, different *placer* ids will be used by the building block. #. `placer_ids` is passed to the initializer: the passed *placer* ids will be used by the building block. #. `placer_ids` is ``None`` and the building block has functional groups: The *placer* ids of the functional groups will be used as the *placer* ids of the building block. #. `placer_ids` is ``None`` and `functional_groups` is empty. All atoms of the molecule will be used for *placer* ids. Returns: The molecule. """ building_block = cls.__new__(cls) building_block._init_from_rdkit_mol( molecule=molecule, functional_groups=functional_groups, placer_ids=placer_ids, ) return building_block def _init_from_rdkit_mol( self, molecule: rdkit.Mol, functional_groups: typing.Iterable[ typing.Union[FunctionalGroup, FunctionalGroupFactory] ], placer_ids: typing.Optional[tuple[int, ...]], ) -> None: """ Initialize from an :mod:`rdkit` molecule. Parameters: molecule: The molecule. functional_groups: An :class:`iterable` of :class:`.FunctionalGroup` or :class:`.FunctionalGroupFactory` or both. :class:`.FunctionalGroup` instances are added to the building block and :class:`.FunctionalGroupFactory` instances are used to create :class:`.FunctionalGroup` instances the building block should hold. :class:`.FunctionalGroup` instances are used to identify which atoms are modified during :class:`.ConstructedMolecule` construction. placer_ids: The ids of *placer* atoms. These are the atoms which should be used for calculating the position of the building block. Depending on the values passed to `placer_ids`, and the functional groups in the building block, different *placer* ids will be used by the building block. #. `placer_ids` is passed to the initializer: the passed *placer* ids will be used by the building block. #. `placer_ids` is ``None`` and the building block has functional groups: The *placer* ids of the functional groups will be used as the *placer* ids of the building block. #. `placer_ids` is ``None`` and `functional_groups` is empty. All atoms of the molecule will be used for *placer* ids. """ atoms = tuple( Atom(a.GetIdx(), a.GetAtomicNum(), a.GetFormalCharge()) for a in molecule.GetAtoms() ) bonds = tuple( Bond( atom1=atoms[b.GetBeginAtomIdx()], atom2=atoms[b.GetEndAtomIdx()], order=( 9 if b.GetBondType() == rdkit.BondType.DATIVE else b.GetBondTypeAsDouble() ) ) for b in molecule.GetBonds() ) position_matrix = molecule.GetConformer().GetPositions() super().__init__(atoms, bonds, position_matrix) self._with_functional_groups(self._extract_functional_groups( functional_groups=functional_groups, )) self._placer_ids = self._normalize_placer_ids( placer_ids=placer_ids, functional_groups=self._functional_groups, ) self._core_ids = frozenset(self._get_core_ids( functional_groups=self._functional_groups, )) def _normalize_placer_ids( self, placer_ids: typing.Optional[tuple[int, ...]], functional_groups: Collection[FunctionalGroup], ) -> frozenset[int]: """ Get the final *placer* ids. Parameters: placer_ids: The ids of *placer* atoms or ``None``. functional_groups: The :class:`.FunctionalGroup` instances of the building block. Returns: Depending on the input values, this function will return different things. #. `placer_ids` is a :class:`tuple` of :class`int`: the `placer_ids` will be returned. #. `placer_ids` is ``None`` and `functional_groups` is not empty: The *placer* ids of the functional groups will be returned. #. `placer_ids` is ``None`` and `functional_groups` is empty. The ids of all atoms in the building block will be returned. """ if placer_ids is not None: return frozenset(placer_ids) if functional_groups: return frozenset(flatten( functional_group.get_placer_ids() for functional_group in functional_groups )) return frozenset(atom.get_id() for atom in self._atoms) def _get_core_ids( self, functional_groups: typing.Iterable[FunctionalGroup], ) -> typing.Iterator[int]: """ Get the final *core* ids. This method may return duplicate ids. Parameters: functional_groups: The :class:`.FunctionalGroup` instances of the building block. Yields: The id of an atom defining the core of the molecule. """ functional_group_atom_ids = { atom_id for functional_group in functional_groups for atom_id in functional_group.get_atom_ids() } for atom in self._atoms: atom_id = atom.get_id() if atom_id not in functional_group_atom_ids: yield atom_id for functional_group in functional_groups: for atom_id in functional_group.get_core_atom_ids(): yield atom_id def _extract_functional_groups( self, functional_groups: typing.Iterable[ typing.Union[FunctionalGroup, FunctionalGroupFactory] ], ) -> typing.Iterator[FunctionalGroup]: """ Yield functional groups. The input can be a mixture of :class:`.FunctionalGroup` and :class:`.FunctionalGroupFactory`. The output yields :class:`.FunctionalGroup` instances only. Either those held directly in `functional_groups` or created by the factories in `functional_groups`. Parameters: functional_groups: Can be an :class:`iterable` of both :class:`.FunctionalGroup` and :class:`.FunctionalGroupFactory`. Yields: A functional group from `functional_groups`, or created by a factory in `functional_groups`. """ for functional_group in functional_groups: if isinstance(functional_group, FunctionalGroup): yield functional_group else: # Else it's a factory. yield from functional_group.get_functional_groups(self) def _with_functional_groups( self, functional_groups: typing.Iterable[FunctionalGroup], ) -> BuildingBlock: """ Modify the molecule. """ self._functional_groups = tuple(functional_groups) return self def with_functional_groups( self, functional_groups: typing.Iterable[FunctionalGroup], ) -> BuildingBlock: """ Return a clone with specific functional groups. Parameters: functional_groups: :class:`.FunctionalGroup` instances which the clone should have. Returns: The clone. """ return self.clone()._with_functional_groups(functional_groups) def _with_canonical_atom_ordering(self) -> BuildingBlock: ordering = rdkit.CanonicalRankAtoms(self.to_rdkit_mol()) super()._with_canonical_atom_ordering() id_map = { old_id: new_id for old_id, new_id in enumerate(ordering) } self._functional_groups = tuple( functional_group.with_ids(id_map) for functional_group in self._functional_groups ) self._placer_ids = frozenset( id_map[placer_id] for placer_id in self._placer_ids ) self._core_ids = frozenset( id_map[core_id] for core_id in self._core_ids ) return self def get_num_functional_groups(self) -> int: """ Return the number of functional groups. Returns: The number of functional groups in the building block. """ return len(self._functional_groups) def get_functional_groups( self, fg_ids: typing.Optional[OneOrMany[int]] = None, ) -> typing.Iterator[FunctionalGroup]: """ Yield the functional groups, ordered by id. Parameters: fg_ids: The ids of functional groups yielded. If ``None``, then all functional groups are yielded. Can be a single :class:`int`, if a single functional group is desired. Yields: A functional group of the building block. """ if fg_ids is None: fg_ids = range(len(self._functional_groups)) elif isinstance(fg_ids, int): fg_ids = (fg_ids, ) for fg_id in fg_ids: yield self._functional_groups[fg_id] def clone(self) -> BuildingBlock: clone = self._clone() clone._functional_groups = self._functional_groups clone._placer_ids = self._placer_ids clone._core_ids = self._core_ids return clone def get_num_placers(self) -> int: """ Return the number of *placer* atoms in the building block. Returns: The number of *placer* atoms in the building block. """ return len(self._placer_ids) def get_placer_ids(self) -> typing.Iterator[int]: """ Yield the ids of *placer* atoms. *Placer* atoms are those, which should be used to calculate the position of the building block. See Also: :meth:`.FunctionalGroup.get_placer_ids` Yields: The id of a *placer* atom. """ yield from self._placer_ids def get_core_atom_ids(self) -> typing.Iterator[int]: """ Yield ids of atoms which form the core of the building block. This includes all atoms in the building block not part of a functional group, as well as any atoms in a functional group, specifically labelled as core atoms. See Also: :meth:`.FunctionalGroup.get_core_atom_ids` Yields: The id of a *core* atom. """ yield from self._core_ids def with_canonical_atom_ordering(self) -> BuildingBlock: return self.clone()._with_canonical_atom_ordering() def with_centroid( self, position: np.ndarray, atom_ids: typing.Optional[OneOrMany[int]] = None, ) -> BuildingBlock: return self.clone()._with_centroid(position, atom_ids) def with_displacement( self, displacement: np.ndarray, ) -> BuildingBlock: return self.clone()._with_displacement(displacement) def with_position_matrix( self, position_matrix: np.ndarray, ) -> BuildingBlock: return self.clone()._with_position_matrix(position_matrix) def with_rotation_about_axis( self, angle: float, axis: np.ndarray, origin: np.ndarray, ) -> BuildingBlock: return self.clone()._with_rotation_about_axis( angle=angle, axis=axis, origin=origin, ) def with_rotation_between_vectors( self, start: np.ndarray, target: np.ndarray, origin: np.ndarray, ) -> BuildingBlock: return self.clone()._with_rotation_between_vectors( start=start, target=target, origin=origin, ) def with_rotation_to_minimize_angle( self, start: np.ndarray, target: np.ndarray, axis: np.ndarray, origin: np.ndarray, ) -> BuildingBlock: return self.clone()._with_rotation_to_minimize_angle( start=start, target=target, axis=axis, origin=origin, ) def with_structure_from_file( self, path: str, extension: typing.Optional[str] = None, ) -> BuildingBlock: return typing.cast( BuildingBlock, super().with_structure_from_file( path=path, extension=extension, ) ) def write( self, path: str, atom_ids: typing.Optional[OneOrMany[int]] = None, ) -> BuildingBlock: return typing.cast( BuildingBlock, super().write(path, atom_ids) ) def __str__(self) -> str: if self._functional_groups: fg_repr = f', {self._functional_groups!r}' else: fg_repr = '' smiles = rdkit.MolToSmiles( mol=rdkit.RemoveHs(self.to_rdkit_mol()), ) return f'{self.__class__.__name__}({smiles!r}{fg_repr})' def __repr__(self) -> str: return str(self)
py
1a2f44021d4e117d94bf6f0a764f46971f807d64
#!/usr/bin/env python """ Entry point for starting Galaxy without starting as part of a web server. Example Usage: Start a job/workflow handler without a web server and with a given name using. galaxy-main --server-name handler0 Start as a daemon with (requires daemonized - install with 'pip install daemonize'): galaxy-main -d --daemon-log-file=handler0-daemon.log --pid-file handler0.pid --server-name handler0 In daemon mode logging of Galaxy (as opposed to this script) is configured via a loggers section in Galaxy's ini file - this can be overridden with sensible defaults logging to a single file with the following: galaxy-main -d --server-name handler0 --daemon-log-file=handler0-daemon.log --pid-file handler0.pid --log-file handler0.log This can also be used to start Galaxy as a uWSGI mule, e.g. for job handling: uwsgi ... --py-call-osafterfork --mule=lib/galaxy/main.py --mule=lib/galaxy/main.py --farm=job-handlers:1,2 The --py-call-osafterfork allows for proper shutdown on SIGTERM/SIGINT. """ import functools import logging import os import signal import sys import threading from argparse import ArgumentParser from configparser import ConfigParser from logging.config import fileConfig try: from daemonize import Daemonize except ImportError: Daemonize = None try: import uwsgi except ImportError: uwsgi = None log = logging.getLogger(__name__) real_file = os.path.realpath(__file__) GALAXY_ROOT_DIR_ = os.path.abspath(os.path.join(os.path.dirname(real_file), os.pardir)) if not os.path.exists(os.path.join(GALAXY_ROOT_DIR_, 'run.sh')): # Galaxy is installed GALAXY_ROOT_DIR = None else: GALAXY_ROOT_DIR = GALAXY_ROOT_DIR_ GALAXY_LIB_DIR = os.path.join(GALAXY_ROOT_DIR_, "lib") try: sys.path.insert(1, GALAXY_LIB_DIR) except Exception: log.exception("Failed to add Galaxy to sys.path") raise from galaxy.main_config import ( absolute_config_path, config_is_ini, DEFAULT_CONFIG_SECTION, DEFAULT_INI_APP, find_config, ) from galaxy.util import unicodify from galaxy.web_stack import get_app_kwds REQUIRES_DAEMONIZE_MESSAGE = "Attempted to use Galaxy in daemon mode, but daemonize is unavailable." DEFAULT_PID = "galaxy.pid" DEFAULT_VERBOSE = True DESCRIPTION = "Daemonized entry point for Galaxy." SHUTDOWN_MSG = '__SHUTDOWN__' UWSGI_FARMS_VAR = '_GALAXY_UWSGI_FARM_NAMES' exit = threading.Event() def load_galaxy_app( config_builder, config_env=False, log=None, attach_to_pools=None, **kwds ): # Allow specification of log so daemon can reuse properly configured one. if log is None: log = logging.getLogger(__name__) # If called in daemon mode, set the ROOT directory and ensure Galaxy is on # sys.path. if config_env: try: os.chdir(GALAXY_ROOT_DIR) except Exception: log.exception("Failed to chdir") raise config_builder.setup_logging() from galaxy.util.properties import load_app_properties kwds = config_builder.app_kwds() kwds = load_app_properties(**kwds) from galaxy.app import UniverseApplication app = UniverseApplication( global_conf=config_builder.global_conf(), attach_to_pools=attach_to_pools, **kwds ) app.database_heartbeat.start() app.application_stack.log_startup() return app def handle_signal(signum, frame): log.info('Received signal %d, exiting', signum) if uwsgi and 'mule_id' in dir(uwsgi) and uwsgi.mule_id() > 0: farms = os.environ.get(UWSGI_FARMS_VAR, None) if farms: for farm in farms.split(','): uwsgi.farm_msg(farm, SHUTDOWN_MSG) else: uwsgi.mule_msg(SHUTDOWN_MSG, uwsgi.mule_id()) exit.set() def register_signals(): for name in ('TERM', 'INT', 'HUP'): sig = getattr(signal, 'SIG%s' % name) signal.signal(sig, handle_signal) def app_loop(args, log): try: config_builder = GalaxyConfigBuilder(args) config_env = GALAXY_ROOT_DIR is not None galaxy_app = load_galaxy_app( config_builder, config_env=config_env, log=log, attach_to_pools=args.attach_to_pool, ) except BaseException: log.exception("Failed to initialize Galaxy application") raise try: # A timeout is required or the signals won't be handled while not exit.wait(20): pass except (KeyboardInterrupt, SystemExit): pass try: galaxy_app.shutdown() except Exception: log.exception("Failed to shutdown Galaxy application") raise class GalaxyConfigBuilder: """ Generate paste-like configuration from supplied command-line arguments. """ def __init__(self, args=None, **kwds): self.config_file = None self.config_section = None self.app_name = kwds.get("app") or (args and args.app) or DEFAULT_CONFIG_SECTION config_file = kwds.get("config_file", None) or (args and args.config_file) # If given app_conf_path - use that - else we need to ensure we have a # config file path. if not config_file and 'config_file' in self.app_kwds(): config_file = self.app_kwds()['config_file'] if not config_file: galaxy_root = kwds.get("galaxy_root", GALAXY_ROOT_DIR) config_file = find_config(config_file, galaxy_root) config_file = absolute_config_path(config_file, galaxy_root=galaxy_root) self.config_file = unicodify(config_file) # FIXME: this won't work for non-Paste ini configs if self.config_is_ini: self.config_section = "app:%s" % unicodify(kwds.get("app") or (args and args.app) or DEFAULT_INI_APP) else: self.config_section = self.app_name self.log_file = (args and args.log_file) @classmethod def populate_options(cls, arg_parser): arg_parser.add_argument("-c", "--config-file", default=None, help="Galaxy config file (defaults to config/galaxy.ini)") arg_parser.add_argument("--ini-path", default=None, help="DEPRECATED: use -c/--config-file") arg_parser.add_argument("--app", default=None, help="app section in config file (defaults to 'galaxy' for YAML/JSON, 'main' (w/ 'app:' prepended) for INI") arg_parser.add_argument("-d", "--daemonize", default=False, help="Daemonize process", action="store_true") arg_parser.add_argument("--daemon-log-file", default=None, help="log file for daemon script ") arg_parser.add_argument("--log-file", default=None, help="Galaxy log file (overrides log configuration in config_file if set)") arg_parser.add_argument("--pid-file", default=DEFAULT_PID, help="pid file (default is %s)" % DEFAULT_PID) arg_parser.add_argument("--server-name", default=None, help="set a galaxy server name") arg_parser.add_argument("--attach-to-pool", action="append", default=['job-handlers'], help="attach to asynchronous worker pool (specify multiple times for multiple pools)") @property def config_is_ini(self): return config_is_ini(self.config_file) def app_kwds(self): kwds = get_app_kwds(self.app_name, app_name=self.app_name) if 'config_file' not in kwds: kwds['config_file'] = self.config_file if 'config_section' not in kwds: kwds['config_section'] = self.config_section return kwds def global_conf(self): conf = {} if self.config_is_ini: conf["__file__"] = self.config_file return conf def setup_logging(self): # Galaxy will attempt to setup logging if loggers is not present in # ini config file - this handles that loggers block however if present # (the way paste normally would) if not self.config_file: return if self.config_is_ini: raw_config = ConfigParser() raw_config.read([self.config_file]) if raw_config.has_section('loggers'): config_file = os.path.abspath(self.config_file) fileConfig( config_file, dict(__file__=config_file, here=os.path.dirname(config_file)) ) def main(func=app_loop): arg_parser = ArgumentParser(description=DESCRIPTION) GalaxyConfigBuilder.populate_options(arg_parser) args = arg_parser.parse_args() if args.ini_path and not args.config_file: args.config_file = args.ini_path if args.log_file: os.environ["GALAXY_CONFIG_LOG_DESTINATION"] = os.path.abspath(args.log_file) if args.server_name: os.environ["GALAXY_CONFIG_SERVER_NAME"] = args.server_name pid_file = args.pid_file log.setLevel(logging.DEBUG) log.propagate = False register_signals() if args.daemonize: if Daemonize is None: raise ImportError(REQUIRES_DAEMONIZE_MESSAGE) keep_fds = [] if args.daemon_log_file: fh = logging.FileHandler(args.daemon_log_file, "w") fh.setLevel(logging.DEBUG) log.addHandler(fh) keep_fds.append(fh.stream.fileno()) else: fh = logging.StreamHandler(sys.stderr) fh.setLevel(logging.DEBUG) log.addHandler(fh) daemon = Daemonize( app="galaxy", pid=pid_file, action=functools.partial(func, args, log), verbose=DEFAULT_VERBOSE, logger=log, keep_fds=keep_fds, ) daemon.start() else: func(args, log) if __name__ == "__main__": main()
py
1a2f4535555f2da39e3616818a12b5a4c107d47b
# -*- coding: utf-8 -*- import sugartensor as tf import matplotlib.pyplot as plt # set log level to debug tf.sg_verbosity(10) # # hyper parameters # batch_size = 25 z_dim = 50 # # create generator # # random uniform seed z = tf.random_uniform((batch_size, z_dim)) with tf.sg_context(name='generator', size=5, stride=2, act='relu', bn=True): # generator network gen = (z.sg_dense(dim=1024) .sg_dense(dim=7*7*128) .sg_reshape(shape=(-1, 7, 7, 128)) .sg_upconv(dim=64) .sg_upconv(dim=1, act='sigmoid', bn=False) .sg_squeeze()) # # draw samples # with tf.Session() as sess: tf.sg_init(sess) # restore parameters saver = tf.train.Saver() saver.restore(sess, tf.train.latest_checkpoint('asset/train/ckpt')) # run generator imgs = sess.run(gen) # plot result _, ax = plt.subplots(5, 5, sharex=True, sharey=True) for i in range(5): for j in range(5): ax[i][j].imshow(imgs[i * 5 + j], 'gray') ax[i][j].set_axis_off() plt.savefig('asset/train/sample.png', dpi=600) tf.sg_info('Sample image saved to "asset/train/sample.png"') plt.close()
py
1a2f46441e61603778a4538a28f056da2e0c9228
# -*- coding: utf-8 -*- # Define here the models for your scraped items # # See documentation in: # https://docs.scrapy.org/en/latest/topics/items.html from scrapy import Item, Field # core thinker object """thinker item; represents one Wikipedia page, unless no page""" class Thinker(Item): name = Field() link = Field() # lists of names influences = Field() influenced = Field()
py
1a2f46b3cae6de88140e6ec40ea6d4d8cd467da9
from arm.logicnode.arm_nodes import * class RandomBooleanNode(ArmLogicTreeNode): """Generates a random boolean.""" bl_idname = 'LNRandomBooleanNode' bl_label = 'Random Boolean' arm_version = 1 def arm_init(self, context): self.add_output('ArmBoolSocket', 'Bool')
py
1a2f4a9003f7056f98e14377d576f3c434efd272
# Copyright 2019 Google LLC. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for tfx.components.example_gen.utils.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function # Standard Imports import tensorflow as tf from tfx.components.example_gen import utils class UtilsTest(tf.test.TestCase): def test_get_default_output_config(self): output_config = utils.get_default_output_config() self.assertEqual(2, len(output_config.split_config.splits)) if __name__ == '__main__': tf.test.main()
py
1a2f4ba51b1638be24e791d8458fa4801710132b
# -*- coding: utf-8 -*- from __future__ import unicode_literals import os from shutil import rmtree from tempfile import mkdtemp from pelican.generators import ArticlesGenerator from pelican.tests.support import unittest, get_settings import sub_parts CUR_DIR = os.path.dirname(__file__) class TestSubParts(unittest.TestCase): @classmethod def setUpClass(cls): cls.temp_path = mkdtemp(prefix='pelicantests.') settings = get_settings(filenames={}) settings['PATH'] = os.path.join(CUR_DIR, 'test_data') settings['AUTHOR'] = 'Me' settings['DEFAULT_DATE'] = (1970, 1, 1) settings['DEFAULT_CATEGORY'] = 'Default' settings['FILENAME_METADATA'] = '(?P<slug>[^.]+)' settings['PLUGINS'] = [sub_parts] settings['CACHE_CONTENT'] = False context = settings.copy() context['generated_content'] = dict() context['static_links'] = set() cls.generator = ArticlesGenerator( context=context, settings=settings, path=settings['PATH'], theme=settings['THEME'], output_path=cls.temp_path) cls.generator.generate_context() cls.all_articles = list(cls.generator.articles) sub_parts.patch_subparts(cls.generator) @classmethod def tearDownClass(cls): rmtree(cls.temp_path) def test_all_articles(self): self.assertEqual( sorted(['noparent', 'parent', 'parent--explicit', 'parent--implicit']), sorted([a.slug for a in self.all_articles])) def test_articles(self): self.assertEqual( sorted(['noparent', 'parent']), sorted([a.slug for a in self.generator.articles])) def test_dates(self): self.assertEqual( sorted(['noparent', 'parent']), sorted([a.slug for a in self.generator.dates])) def test_categories(self): self.assertEqual( sorted(['noparent', 'parent']), sorted([a.slug for a in self.generator.categories[0][1]])) def test_tags(self): self.assertEqual( sorted([a.slug for a in self.all_articles]), sorted([a.slug for a in self.generator.tags['atag']])) def test_authors(self): self.assertEqual( sorted([a.slug for a in self.all_articles]), sorted([a.slug for a in self.generator.authors[0][1]])) def test_subparts(self): for a in self.all_articles: if a.slug == 'parent': self.assertTrue(hasattr(a, 'subparts')) self.assertEqual( sorted(['parent--explicit', 'parent--implicit']), sorted([a.slug for a in a.subparts])) else: self.assertFalse(hasattr(a, 'subparts')) def test_subpart_of(self): for a in self.all_articles: if '--' in a.slug: self.assertTrue(hasattr(a, 'subpart_of')) self.assertEqual('parent', a.subpart_of.slug) else: self.assertFalse(hasattr(a, 'subpart_of')) def test_subtitle(self): for a in self.all_articles: if '--' in a.slug: self.assertTrue(hasattr(a, 'subtitle')) self.assertEqual(a.title, a.subtitle + ', ' + a.subpart_of.title) class TestSubPartsPhotos(unittest.TestCase): @classmethod def setUpClass(cls): cls.temp_path = mkdtemp(prefix='pelicantests.') settings = get_settings(filenames={}) settings['PATH'] = os.path.join(CUR_DIR, 'test_data') settings['AUTHOR'] = 'Me' settings['DEFAULT_DATE'] = (1970, 1, 1) settings['DEFAULT_CATEGORY'] = 'Default' settings['FILENAME_METADATA'] = '(?P<slug>[^.]+)' settings['PLUGINS'] = [sub_parts] settings['CACHE_CONTENT'] = False context = settings.copy() context['generated_content'] = dict() context['static_links'] = set() cls.generator = ArticlesGenerator( context=context, settings=settings, path=settings['PATH'], theme=settings['THEME'], output_path=cls.temp_path) cls.generator.generate_context() cls.all_articles = list(cls.generator.articles) for a in cls.all_articles: a.photo_gallery = [('i.jpg', 'i.jpg', 'it.jpg', '', '')] sub_parts.patch_subparts(cls.generator) @classmethod def tearDownClass(cls): rmtree(cls.temp_path) def test_subphotos(self): for a in self.all_articles: if a.slug == 'parent': self.assertTrue(hasattr(a, 'subphotos')) self.assertEqual(3, a.subphotos) else: self.assertFalse(hasattr(a, 'subphotos')) if __name__ == '__main__': unittest.main()
py
1a2f4bf38709bca0c900194842760c3304bedeff
import turtle import os t = turtle.Pen() #t.speed(0) t.shape("turtle") t.left(90) for i in range(6): t.forward(100) t.right(60) for j in range(3): t.forward(20) t.backward(20) t.left(60) t.right(120) t.backward(100) t.left(60) os.system("Pause")
py
1a2f4c2681d3490af76ce5c1f381f74258cb1d49
############################################################################## # Copyright (c) 2013-2018, Lawrence Livermore National Security, LLC. # Produced at the Lawrence Livermore National Laboratory. # # This file is part of Spack. # Created by Todd Gamblin, tgamblin@llnl.gov, All rights reserved. # LLNL-CODE-647188 # # For details, see https://github.com/spack/spack # Please also see the NOTICE and LICENSE files for our notice and the LGPL. # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU Lesser General Public License (as # published by the Free Software Foundation) version 2.1, February 1999. # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the IMPLIED WARRANTY OF # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the terms and # conditions of the GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA ############################################################################## from spack import * class Xbacklight(AutotoolsPackage): """Xbacklight is used to adjust the backlight brightness where supported. It uses the RandR extension to find all outputs on the X server supporting backlight brightness control and changes them all in the same way.""" homepage = "http://cgit.freedesktop.org/xorg/app/xbacklight" url = "https://www.x.org/archive/individual/app/xbacklight-1.2.1.tar.gz" version('1.2.1', 'e8e4c86b0f867e23aa3532618a697609') depends_on('libxcb') depends_on('xcb-util') depends_on('pkgconfig', type='build') depends_on('util-macros', type='build')
py
1a2f4c653b260f819dd0c1ba1773663bf9adac68
import os def align(step): path = os.getcwd() file_path = os.path.join(path, 'whales.txt') file1 = open(file_path, 'r') Lines = file1.readlines() initial_pos = [0] * 2000 final = 0 for line in Lines: input = line.strip() poss = input.split(",") prev_fuel = -1 for median in range(0, int(poss[len(poss)-1])+1): fuel = 0 for pos in poss: n = abs(int(pos) - int(median)) fuel += n * (n+1)/2 if step else n if fuel < prev_fuel or prev_fuel == -1: prev_fuel = fuel print(int(prev_fuel)) if __name__ == "__main__": align(False) align(True)
py
1a2f4e17d97df7bbbfca1934c1f3300bf31c9b4a
#!/usr/bin/env python # -*- coding: utf-8 -*- """Tests for the Zeitgeist event formatter.""" from __future__ import unicode_literals import unittest from plaso.formatters import zeitgeist from tests.formatters import test_lib class ZeitgeistFormatterTest(test_lib.EventFormatterTestCase): """Tests for the Zeitgeist activity database event formatter.""" def testInitialization(self): """Tests the initialization.""" event_formatter = zeitgeist.ZeitgeistFormatter() self.assertIsNotNone(event_formatter) def testGetFormatStringAttributeNames(self): """Tests the GetFormatStringAttributeNames function.""" event_formatter = zeitgeist.ZeitgeistFormatter() expected_attribute_names = ['subject_uri'] self._TestGetFormatStringAttributeNames( event_formatter, expected_attribute_names) # TODO: add test for GetMessages. if __name__ == '__main__': unittest.main()
py
1a2f4e17e44131e61bca54ff4dcb336000f8d1c2
#!/usr/bin/env python3 """ Convert $1 -- a markdown file to an HTML file in `/tmp/vim/<parent>/<basename>.html`. """ # Standard Library from os.path import realpath, basename, isfile, isdir, dirname, abspath import os import shutil import subprocess import sys import re import logging from logging import Logger from typing import Dict, Iterable, Optional, Set, Text, List from subprocess import DEVNULL, PIPE, run from meta import MATHJAX_CONFS, LongOpt, CodeStyle, FromFmt, Ext, ToFmt from utils import validate_executables, require_exists, vimdir_path, ensure_exists # initalise logging with sane configuration logging.basicConfig(level=logging.WARN, format="%(levelname)s:%(asctime)s %(message)s") log: Logger = logging.getLogger() validate_executables(["pandoc"]) # NOTE this will try to include a lua filter from `./task-list.lua`. class PandocCmd: def __init__( self, input_file: Text, stylesheet: Text = vimdir_path("css", "styles.css"), javascript: Text = vimdir_path("js", "script.js"), from_fmt: FromFmt = FromFmt.MARKDOWN, to_fmt: ToFmt = ToFmt.HTML5, exts: List[Ext] = [ Ext.ASCII_IDENTIFIERS, Ext.AUTOLINK_BARE_URIS, Ext.COMPACT_DEFINITION_LISTS, Ext.FENCED_DIVS, Ext.GFM_AUTO_IDENTIFIERS, Ext.INTRAWORD_UNDERSCORES, Ext.MARKDOWN_ATTRIBUTE, Ext.MMD_HEADER_IDENTIFIERS, Ext.MMD_LINK_ATTRIBUTES, Ext.SMART, Ext.TEX_MATH_DOLLARS, Ext.TEX_MATH_DOUBLE_BACKSLASH, Ext.TEX_MATH_SINGLE_BACKSLASH, ], no_exts: List[Ext] = [Ext.ALL_SYMBOLS_ESCAPABLE, Ext.ESCAPED_LINE_BREAKS], long_opts: Dict[LongOpt, Optional[Text]] = { LongOpt.ATX_HEADERS: None, LongOpt.REFERENCE_LOCATION: "document", LongOpt.SECTION_DIVS: None, LongOpt.EMAIL_OBFUSCATION: "javascript", }, code_style: CodeStyle = CodeStyle.TANGO, mathjax_version: Text = "2.7.5", mathjax_conf: Text = "TeX-AMS_HTML", width: int = 80, toc_depth: int = 3, ) -> None: log.debug("initalising a PandocCmd object") self.exts: List[Ext] = [] self.no_exts: List[Ext] = [] self.long_opts: Dict[LongOpt, Optional[Text]] = dict() self.set_opts(long_opts).set_exts(exts).set_no_exts(no_exts).set_input_file( input_file ).set_opt(LongOpt.STANDALONE).set_to_fmt(to_fmt).set_from_fmt( from_fmt ).set_mathjax( mathjax_version, mathjax_conf ).set_width( width ).set_stylesheet( stylesheet ).set_code_style( code_style ).set_javascript( javascript ) if toc_depth: self.set_toc_depth(toc_depth).set_opt(LongOpt.TOC) lua_filter: str = os.path.join(os.path.dirname( os.path.abspath(__file__)), 'task-list.lua') if isfile(lua_filter): self.set_opt(LongOpt.LUA_FILTER, lua_filter) else: log.error(f'failed to find lua filter ./{lua_filter}') def set_from_fmt(self, fmt: FromFmt) -> object: assert fmt in FromFmt, f"from format '{fmt}' invalid" self.from_fmt: FromFmt = fmt return self def set_opt(self, key: LongOpt, val: Optional[Text] = None) -> object: self.long_opts[key] = val return self def set_opts(self, pairs: Dict[LongOpt, Optional[Text]]) -> object: for (k, v) in pairs.items(): self.set_opt(k, v) return self def set_to_fmt(self, fmt: ToFmt) -> object: self.to_fmt = fmt return self def set_input_file(self, file_path: Text) -> object: require_exists(file_path) self.input_file = file_path return self def set_width(self, n: int) -> object: assert n and n >= 0, f"invalid value {str(n)}" return self.set_opt(LongOpt.COLUMNS, str(n)) def set_stylesheet(self, css_path: Text) -> object: require_exists(css_path) return self.set_opt(LongOpt.CSS, css_path) def set_javascript(self, js_path: Text) -> object: require_exists(js_path) self.javascript = js_path return self def set_toc_depth(self, n: int) -> object: assert n and n >= 0, f"invalid value {n}" return self.set_opt(LongOpt.TOC_DEPTH, str(n)) def set_code_style(self, style: CodeStyle) -> object: return self.set_opt(LongOpt.HIGHLIGHT_STYLE, style._name_.lower()) def set_mathjax(self, version: Text, cfg: Text) -> object: assert cfg and cfg in MATHJAX_CONFS, f"unreconginsed MathJax config {cfg}" assert ( version and len(version) >= 3 and version[0] == "2" and version[1] == "." and str.isdigit(version[2]) ), f"unrecognised MathJax version {version}" return self.set_opt( LongOpt.MATHJAX, f"https://cdnjs.cloudflare.com/ajax/libs/mathjax/{version}/MathJax.js?config={cfg}", ) def set_exts(self, exts: Iterable[Ext]) -> object: for ext in exts: self.set_ext(ext) return self def set_ext(self, ext: Ext) -> object: self.exts.append(ext) return self def set_no_ext(self, ext: Ext) -> object: self.no_exts.append(ext) return self def set_no_exts(self, exts: Iterable[Ext]) -> object: for ext in exts: self.set_no_ext(ext) return self @property def args(self) -> List[Text]: arguments = ["pandoc", "--from"] from_fmt = self.from_fmt._name_.lower() if len(self.exts) > 0: for ext in self.exts: from_fmt += f"+{ext._name_.lower()}" if len(self.no_exts) > 0: for ext in self.no_exts: from_fmt += f"-{ext._name_.lower()}" arguments.append(from_fmt) arguments.extend(["--to", self.to_fmt._name_.lower()]) for opt in self.long_opts.keys(): maybe_val: Optional[Text] = self.long_opts[opt] opt_name: Text = opt._name_.lower().replace("_", "-") if maybe_val: arguments.append(f"--{opt_name}={maybe_val}") else: arguments.append(f"--{opt_name}") log.debug(f"args: {arguments}") return arguments def before(self, text: Text) -> Text: """Fix badly formatted markdown where heading marker `#` is not followed by space. NOTE: This method is pure. :param text: input text before transformations :return: output text after transformations """ log.debug("preprocessing [before hook]") return re.sub(r"(#+)([A-Z])", "\1 \2", text, re.MULTILINE) def after(self, text: Text) -> Text: """Transform relative links and references, Inject JavaScript. NOTE: This method is pure. Match on either src or href e.g.: `src="./script.js"` and `href="styles.css"` skip over whitespace e.g.: `src=" address/file.png"` match if relative e.g.: `./` or match if not an external link with a protocol e.g.: `https://stackoverflow.com` or match if not a valid directory reference e.g.: `/srv/log/log.txt` and `~/file.txt` :param text: input text :return: output after transformations """ pat = re.compile(r'(href|src)="\s*(\./|(?![a-z]{2,10}://|~|\#|/))') d: Text = dirname(self.input_file) with open(self.javascript) as f: log.debug("postprocessing [after hook]") return re.sub(pat, f'\\1="{d}/', text).replace( "</body>", f"<script>{f.read()}</script></body>" ) def execute(self) -> Text: log.debug("executing") with open(self.input_file, encoding="utf-8") as input: return self.after( run( self.args, encoding="utf-8", input=self.before(input.read()), stderr=DEVNULL, stdout=PIPE, ).stdout ) if __name__ == "__main__": INPUT_FILE = sys.argv[1] log.debug(f"input file: {INPUT_FILE}") OUTPUT_FILE: Text = os.path.join( "/tmp/vim", basename(dirname(INPUT_FILE)), re.sub( r"^(.*)\.(?:r?md|m(?:ark)?down)$", r"\1.html", basename(INPUT_FILE), re.IGNORECASE | re.MULTILINE, ), ) log.debug(f"output file: {OUTPUT_FILE}") ensure_exists(OUTPUT_FILE) with open(OUTPUT_FILE, "w", encoding="utf-8") as output: cmd = PandocCmd(INPUT_FILE) output.write(cmd.execute()) print(f"Cmd: {' '.join(cmd.args)}") print(f'Output: {output.name}') # vim:foldmethod=manual:
py
1a2f4e20d52fd7abc28470acbfa72929fc69b781
import logging from typing import Any, Dict, List, Optional from starlette.applications import Starlette from starlette.exceptions import HTTPException from starlette.requests import Request from starlette.responses import HTMLResponse, JSONResponse from starlette.routing import Mount, Route from pait.api_doc.html import get_redoc_html as _get_redoc_html from pait.api_doc.html import get_swagger_ui_html as _get_swagger_ui_html from pait.api_doc.open_api import PaitOpenApi from pait.field import Depends, Query from pait.g import config from pait.model.core import PaitCoreModel from pait.model.status import PaitStatus from ._load_app import load_app from ._pait import Pait def add_doc_route( app: Starlette, scheme: Optional[str] = None, prefix: str = "/", pin_code: str = "", title: str = "Pait Doc", open_api_tag_list: Optional[List[Dict[str, Any]]] = None, ) -> None: if pin_code: logging.info(f"doc route start pin code:{pin_code}") doc_pait: Pait = Pait( author=config.author or ("so1n",), status=config.status or PaitStatus.release, tag=("pait_doc",), group="pait_doc", ) def _get_request_pin_code(r_pin_code: str = Query.i("", alias="pin_code")) -> Optional[str]: if pin_code: if r_pin_code != pin_code: raise HTTPException( status_code=404, detail=( "The requested URL was not found on the server. If you entered" " the URL manually please check your spelling and try again." ), ) return r_pin_code def _get_open_json_url(request: Request, r_pin_code: str) -> str: openapi_json_url: str = f"{'/'.join(request.url.path.split('/')[:-1])}/openapi.json" if r_pin_code: openapi_json_url += f"?pin_code={r_pin_code}" return openapi_json_url @doc_pait() def get_redoc_html(request: Request, r_pin_code: str = Depends.i(_get_request_pin_code)) -> HTMLResponse: return HTMLResponse(_get_redoc_html(_get_open_json_url(request, r_pin_code), title)) @doc_pait() def get_swagger_ui_html(request: Request, r_pin_code: str = Depends.i(_get_request_pin_code)) -> HTMLResponse: return HTMLResponse(_get_swagger_ui_html(_get_open_json_url(request, r_pin_code), title)) @doc_pait(pre_depend_list=[_get_request_pin_code]) def openapi_route(request: Request) -> JSONResponse: pait_dict: Dict[str, PaitCoreModel] = load_app(request.app) _scheme: str = scheme or request.url.scheme pait_openapi: PaitOpenApi = PaitOpenApi( pait_dict, title=title, open_api_server_list=[{"url": f"{_scheme}://{request.url.hostname}:{request.url.port}", "description": ""}], open_api_tag_list=open_api_tag_list, ) return JSONResponse(pait_openapi.open_api_dict) route: Mount = Mount( prefix, name="api doc", routes=[ Route("/redoc", get_redoc_html, methods=["GET"]), Route("/swagger", get_swagger_ui_html, methods=["GET"]), Route("/openapi.json", openapi_route, methods=["GET"]), ], ) app.routes.append(route)
py
1a2f4e390d7cb44e171aa0da52f9bab0974dc7d5
import hashlib import io import json import tempfile from random import randint from unittest.mock import Mock import pytest import requests from .shared import ( guess_mimetype, HashWrap, MogileFile, encode_int, future_waiter, make_bucket_map, maybe_update_max, md5_fileobj_b64, md5_fileobj_hex, sha1_fileobj_b64, sha1_fileobj_hex, ) # pylint: disable=C0115,C0116,R0201 class TestMigrate: MD5_HEX = "5eb63bbbe01eeed093cb22bb8f5acdc3" SHA1_HEX = "2aae6c35c94fcfb415dbe95f408b9ce91ee846ed" MD5_B64 = "XrY7u+Ae7tCTyyK7j1rNww==" SHA1_B64 = "Kq5sNclPz7QV2+lfQIuc6R7oRu0=" @pytest.fixture def mogile_client(self): mogile_client = Mock() paths = Mock() mogile_client.get_paths.return_value = paths paths.data = {"paths": {1: "http://example.org/1", 2: "http://example.org/2"}} return mogile_client @pytest.fixture def gcs_client(self): gcs_client = Mock() bucket = Mock() blob = Mock() gcs_client.get_bucket.return_value = bucket bucket.get_blob.return_value = blob bucket.blob.return_value = blob return gcs_client @pytest.fixture def row(self): # hello world sha1sum: 2aae6c35c94fcfb415dbe95f408b9ce91ee846ed # (fid, dmid, dkey, length, classid, devcount) return [1, 1, f"{self.SHA1_HEX}-mogilefs-prod-repo", 1593790, 0, 2] @pytest.yield_fixture def my_tempfile(self): with tempfile.TemporaryFile() as tmp: tmp.write(b"hello world") yield tmp @pytest.fixture def mogile_file(self): return MogileFile(dkey=f"{self.SHA1_HEX}-repo", fid=564879786, length=1) def test_parse_row(self, row): file = MogileFile.parse_row(row) assert file.sha1sum == self.SHA1_HEX assert file.fid == 1 assert file.length == 1593790 assert file.dkey == f"{self.SHA1_HEX}-mogilefs-prod-repo" assert file.skip is False assert file.mogile_bucket == "mogilefs-prod-repo" def test_parse_bad_dmid(self, row): row[1] = 2 with pytest.raises(AssertionError, match="Bad domain"): MogileFile.parse_row(row) def test_parse_temp_file(self, row): row[2] = "8d26b4da-bd3e-47eb-888a-13bb3579c7e9.tmp" file = MogileFile.parse_row(row) assert file.skip is True assert file.mogile_bucket is None assert file.sha1sum is None def test_parse_bad_class(self, row): row[4] = 1 with pytest.raises(AssertionError, match="Bad class"): MogileFile.parse_row(row) def test_make_intermediary_key(self, mogile_file): assert mogile_file.make_intermediary_key() == "0/564/879/0564879786.fid" def test_make_contentrepo_key(self, mogile_file): assert mogile_file.make_contentrepo_key() == self.SHA1_HEX def test_exists_in_bucket(self, mogile_file, gcs_client): blob = gcs_client.get_bucket().get_blob() blob.size = 1 blob.md5_hash = self.MD5_HEX assert ( mogile_file.exists_in_bucket(gcs_client, "my-bucket", "my-key") == self.MD5_HEX ) def test_does_not_exist_in_bucket(self, gcs_client, mogile_file): gcs_client.get_bucket().get_blob.return_value = None assert mogile_file.exists_in_bucket(gcs_client, "my-bucket", "my-file") is False def test_mogile_file_to_json(self, mogile_file): assert mogile_file == MogileFile.from_json(json.loads(mogile_file.to_json())) def test_md5_fileobj(self, my_tempfile): assert md5_fileobj_hex(my_tempfile) == self.MD5_HEX assert md5_fileobj_b64(my_tempfile) == self.MD5_B64 def test_sha1_fileobj(self, my_tempfile): assert sha1_fileobj_hex(my_tempfile) == self.SHA1_HEX assert sha1_fileobj_b64(my_tempfile) == self.SHA1_B64 def test_put( self, mogile_file: MogileFile, mogile_client, gcs_client, requests_mock ): requests_mock.get("http://example.org/1", content=b"hello world") requests_mock.get("http://example.org/2", content=b"hello world") blob = gcs_client.get_bucket().get_blob() def upload_from_file(fileobj, *args, **kwargs): fileobj.read() blob.upload_from_file.side_effect = upload_from_file blob.upload_from_file.return_value = "xyz" blob.md5_hash = self.MD5_B64 md5 = mogile_file.put(mogile_client, gcs_client, "my-bucket") assert md5 == self.MD5_B64 def test_make_bucket_map(self): assert make_bucket_map("a:b,c:d") == {"a": "b", "c": "d"} def test_future_waiter(self): futures_list = [] called_times = {} for i in range(0, 100): future = Mock() future.exception.return_value = None def mk_done(counter): called_times[counter] = 0 return_after = counter % 3 def done(): called_times[counter] += 1 return called_times[counter] > return_after return done future.done = mk_done(i) future.result.return_value = None futures_list.append(future) passthrough = future_waiter((f for f in futures_list), 10) leftovers = list(iter(passthrough)) assert leftovers == [None] * 100 def test_future_waiter_exception(self): with pytest.raises(Exception, match="huh"): futures_list = [] for i in range(0, 10): future = Mock() future.exception.return_value = None future.done.return_value = True future.result.return_value = None if i == 5: future.exception.return_value = Exception("huh") futures_list.append(future) passthrough = future_waiter((f for f in futures_list), 10) leftovers = list(iter(passthrough)) def test_hash_wrap(self): bio = io.BytesIO(b"hello world") md5 = hashlib.md5() with HashWrap(bio, md5) as pipe: assert pipe.read() == b"hello world" assert md5.hexdigest() == "5eb63bbbe01eeed093cb22bb8f5acdc3" def test_hash_wrap_requests(self, requests_mock): requests_mock.get("http://example.org/1", content=b"hello world") with requests.get("http://example.org/1", stream=True) as req: req.raise_for_status() sha1 = hashlib.sha1() req.raw.decode_content = True with HashWrap(req.raw, sha1) as pipe: assert "hello world" == pipe.read().decode("utf-8") assert sha1.hexdigest() == "2aae6c35c94fcfb415dbe95f408b9ce91ee846ed" def test_encode_int(self): assert b"100" == encode_int(100) i = randint(0, 100000000000) assert i == int(encode_int(i)) def test_maybe_update_max(self): db = {} maybe_update_max(db, "last", 1) assert db["last"] == b"1" db = {"last": b"1"} maybe_update_max(db, "last", 2) assert db["last"] == b"2" db = {"last": b"2"} maybe_update_max(db, "last", 1) assert db["last"] == b"2" def test_guess_mimetype(self): assert "image/png" == guess_mimetype( "corpus-dev-0242ac130003/10.1371/image.pbio.v01.i01/1/image.pbio.v01.i01.g001.PNG_I" ) assert "image/png" == guess_mimetype( "gs:///corpus-dev-0242ac130003/10.1371/image.pbio.v01.i01/1/image.pbio.v01.i01.g001.PNG_I" ) assert "image/png" == guess_mimetype("image.pbio.v01.i01.g001.PNG_I") assert "image/png" == guess_mimetype("image.pbio.v01.i01.g001.PNG") assert "image/png" == guess_mimetype("image.pbio.v01.i01.g001.PNG_I") assert "application/octet-stream" == guess_mimetype("foo") assert "text/csv" == guess_mimetype("foo.csv") assert "text/html" == guess_mimetype("foo.html")
py
1a2f4ea64c703fedcd44794e9fafec9ff20acd56
# coding: utf-8 """ YNAB API Endpoints Our API uses a REST based design, leverages the JSON data format, and relies upon HTTPS for transport. We respond with meaningful HTTP response codes and if an error occurs, we include error details in the response body. API Documentation is at https://api.youneedabudget.com # noqa: E501 OpenAPI spec version: 1.0.0 Generated by: https://github.com/swagger-api/swagger-codegen.git """ from __future__ import absolute_import import unittest import ynab_moka from ynab_moka.api.scheduled_transactions_api import ScheduledTransactionsApi # noqa: E501 from ynab_moka.rest import ApiException class TestScheduledTransactionsApi(unittest.TestCase): """ScheduledTransactionsApi unit test stubs""" def setUp(self): self.api = ynab_moka.api.scheduled_transactions_api.ScheduledTransactionsApi() # noqa: E501 def tearDown(self): pass def test_get_scheduled_transaction_by_id(self): """Test case for get_scheduled_transaction_by_id Single scheduled transaction # noqa: E501 """ pass def test_get_scheduled_transactions(self): """Test case for get_scheduled_transactions List scheduled transactions # noqa: E501 """ pass if __name__ == '__main__': unittest.main()
py
1a2f4f23daa5ac8ae414785f7aaf8ca3a86222bb
from ..boot import DATA_ROOT from ..utils import appid, on_production_server from .creation import DatabaseCreation from .stubs import stub_manager from django.db.backends.util import format_number from djangotoolbox.db.base import NonrelDatabaseFeatures, \ NonrelDatabaseOperations, NonrelDatabaseWrapper, NonrelDatabaseClient, \ NonrelDatabaseValidation, NonrelDatabaseIntrospection from google.appengine.ext.db.metadata import get_kinds, get_namespaces from google.appengine.api.datastore import Query, Delete from google.appengine.api.namespace_manager import set_namespace import logging import os DATASTORE_PATHS = { 'datastore_path': os.path.join(DATA_ROOT, 'datastore'), 'blobstore_path': os.path.join(DATA_ROOT, 'blobstore'), 'rdbms_sqlite_path': os.path.join(DATA_ROOT, 'rdbms'), 'prospective_search_path': os.path.join(DATA_ROOT, 'prospective-search'), } def get_datastore_paths(options): paths = {} for key, path in DATASTORE_PATHS.items(): paths[key] = options.get(key, path) return paths def destroy_datastore(paths): """Destroys the appengine datastore at the specified paths.""" for path in paths.values(): if not path: continue try: os.remove(path) except OSError, error: if error.errno != 2: logging.error("Failed to clear datastore: %s" % error) class DatabaseFeatures(NonrelDatabaseFeatures): allows_primary_key_0 = True supports_dicts = True class DatabaseOperations(NonrelDatabaseOperations): compiler_module = __name__.rsplit('.', 1)[0] + '.compiler' DEFAULT_MAX_DIGITS = 16 def value_to_db_decimal(self, value, max_digits, decimal_places): if value is None: return None sign = value < 0 and u'-' or u'' if sign: value = abs(value) if max_digits is None: max_digits = self.DEFAULT_MAX_DIGITS if decimal_places is None: value = unicode(value) else: value = format_number(value, max_digits, decimal_places) decimal_places = decimal_places or 0 n = value.find('.') if n < 0: n = len(value) if n < max_digits - decimal_places: value = u"0" * (max_digits - decimal_places - n) + value return sign + value def sql_flush(self, style, tables, sequences): self.connection.flush() return [] class DatabaseClient(NonrelDatabaseClient): pass class DatabaseValidation(NonrelDatabaseValidation): pass class DatabaseIntrospection(NonrelDatabaseIntrospection): def table_names(self): """Returns a list of names of all tables that exist in the database.""" return [kind.key().name() for kind in Query(kind='__kind__').Run()] class DatabaseWrapper(NonrelDatabaseWrapper): def __init__(self, *args, **kwds): super(DatabaseWrapper, self).__init__(*args, **kwds) self.features = DatabaseFeatures(self) self.ops = DatabaseOperations(self) self.client = DatabaseClient(self) self.creation = DatabaseCreation(self) self.validation = DatabaseValidation(self) self.introspection = DatabaseIntrospection(self) options = self.settings_dict self.remote_app_id = options.get('REMOTE_APP_ID', appid) self.domain = options.get('DOMAIN', 'appspot.com') self.remote_api_path = options.get('REMOTE_API_PATH', None) self.secure_remote_api = options.get('SECURE_REMOTE_API', True) remote = options.get('REMOTE', False) if on_production_server: remote = False if remote: stub_manager.setup_remote_stubs(self) else: stub_manager.setup_stubs(self) def flush(self): """Helper function to remove the current datastore and re-open the stubs""" if stub_manager.active_stubs == 'remote': import random import string code = ''.join([random.choice(string.ascii_letters) for x in range(4)]) print '\n\n!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!' print '!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!' print "Warning! You're about to delete the *production* datastore!" print 'Only models defined in your INSTALLED_APPS can be removed!' print 'If you want to clear the whole datastore you have to use the ' \ 'datastore viewer in the dashboard. Also, in order to delete all '\ 'unneeded indexes you have to run appcfg.py vacuum_indexes.' print 'In order to proceed you have to enter the following code:' print code response = raw_input('Repeat: ') if code == response: print 'Deleting...' delete_all_entities() print "Datastore flushed! Please check your dashboard's " \ 'datastore viewer for any remaining entities and remove ' \ 'all unneeded indexes with manage.py vacuum_indexes.' else: print 'Aborting' exit() elif stub_manager.active_stubs == 'test': stub_manager.deactivate_test_stubs() stub_manager.activate_test_stubs() else: destroy_datastore(get_datastore_paths(self.settings_dict)) stub_manager.setup_local_stubs(self) def delete_all_entities(): for namespace in get_namespaces(): set_namespace(namespace) for kind in get_kinds(): if kind.startswith('__'): continue while True: data = Query(kind=kind, keys_only=True).Get(200) if not data: break Delete(data)
py
1a2f4f50819b315f2c0f7883438652987d3e892e
# Copyright 2020 Pants project contributors (see CONTRIBUTORS.md). # Licensed under the Apache License, Version 2.0 (see LICENSE). from __future__ import annotations import itertools from pants.backend.python.goals import lockfile from pants.backend.python.goals.lockfile import GeneratePythonLockfile from pants.backend.python.subsystems.python_tool_base import PythonToolBase from pants.backend.python.subsystems.setup import PythonSetup from pants.backend.python.target_types import ConsoleScript, InterpreterConstraintsField from pants.backend.python.util_rules.interpreter_constraints import InterpreterConstraints from pants.core.goals.generate_lockfiles import GenerateToolLockfileSentinel from pants.engine.rules import Get, collect_rules, rule from pants.engine.target import AllTargets, AllTargetsRequest from pants.engine.unions import UnionRule from pants.util.docutil import git_url from pants.util.logging import LogLevel class IPython(PythonToolBase): options_scope = "ipython" help = "The IPython enhanced REPL (https://ipython.org/)." default_version = "ipython==7.16.1" # The last version to support Python 3.6. default_main = ConsoleScript("ipython") register_lockfile = True default_lockfile_resource = ("pants.backend.python.subsystems", "ipython_lockfile.txt") default_lockfile_path = "src/python/pants/backend/python/subsystems/ipython_lockfile.txt" default_lockfile_url = git_url(default_lockfile_path) @classmethod def register_options(cls, register): super().register_options(register) register( "--ignore-cwd", type=bool, advanced=True, default=True, help="Whether to tell IPython not to put the CWD on the import path.\n\n" "Normally you want this to be True, so that imports come from the hermetic " "environment Pants creates.\n\nHowever IPython<7.13.0 doesn't support this option, " "so if you're using an earlier version (e.g., because you have Python 2.7 code) " "then you will need to set this to False, and you may have issues with imports " "from your CWD shading the hermetic environment.", ) class IPythonLockfileSentinel(GenerateToolLockfileSentinel): resolve_name = IPython.options_scope @rule( desc=( "Determine all Python interpreter versions used by iPython in your project (for lockfile " "usage)" ), level=LogLevel.DEBUG, ) async def setup_ipython_lockfile( _: IPythonLockfileSentinel, ipython: IPython, python_setup: PythonSetup ) -> GeneratePythonLockfile: if not ipython.uses_lockfile: return GeneratePythonLockfile.from_tool(ipython) # IPython is often run against the whole repo (`./pants repl ::`), but it is possible to run # on subsets of the codebase with disjoint interpreter constraints, such as # `./pants repl py2::` and then `./pants repl py3::`. Still, even with those subsets possible, # we need a single lockfile that works with all possible Python interpreters in use. # # This ORs all unique interpreter constraints. The net effect is that every possible Python # interpreter used will be covered. all_tgts = await Get(AllTargets, AllTargetsRequest()) unique_constraints = { InterpreterConstraints.create_from_compatibility_fields( [tgt[InterpreterConstraintsField]], python_setup ) for tgt in all_tgts if tgt.has_field(InterpreterConstraintsField) } constraints = InterpreterConstraints(itertools.chain.from_iterable(unique_constraints)) return GeneratePythonLockfile.from_tool( ipython, constraints or InterpreterConstraints(python_setup.interpreter_constraints) ) def rules(): return ( *collect_rules(), *lockfile.rules(), UnionRule(GenerateToolLockfileSentinel, IPythonLockfileSentinel), )
py
1a2f4f774438953a0317b8d2432ecb4bd54a4d42
import os from acres.model import topic_list def test_parse(): topics = topic_list.parse("tests/resources/test_topics.tsv") types = topic_list.unique_types(topics) assert 'EKG' in types def test_create_topic_list(ngramstat): filename = "tests/resources/ngram_topics.tsv" topic_list.create(filename, 1.0, 3) assert os.path.exists(filename) assert os.path.getsize(filename) > 10
py
1a2f4f86bc4e5affbf717e120dd623459b963145
# Copyright 2011 10gen # # Modified by Antonin Amand <antonin.amand@gmail.com> to work with gevent. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os import pymongo.connection from gevent.hub import getcurrent import gevent.queue import gevent.greenlet import gevent.local import gevent.coros from gevent import socket import weakref class Pool(object): """ A greenlet safe connection pool for gevent (non-thread safe). """ DEFAULT_TIMEOUT = 3.0 def __init__(self, pool_size, network_timeout=None): self.network_timeout = network_timeout or self.DEFAULT_TIMEOUT self.pool_size = pool_size self._bootstrap(os.getpid()) self._lock = gevent.coros.RLock() def _bootstrap(self, pid): self._count = 0 self._pid = pid self._used = {} self._queue = gevent.queue.Queue(self.pool_size) def connect(self, host, port): """Connect to Mongo and return a new (connected) socket. """ try: # Prefer IPv4. If there is demand for an option # to specify one or the other we can add it later. s = socket.socket(socket.AF_INET) s.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1) s.settimeout(self.network_timeout or pymongo.connection._CONNECT_TIMEOUT) s.connect((host, port)) s.settimeout(self.network_timeout) return s except socket.gaierror: # If that fails try IPv6 s = socket.socket(socket.AF_INET6) s.setsockopt(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1) s.settimeout(self.network_timeout or pymongo.connection._CONNECT_TIMEOUT) s.connect((host, port)) s.settimeout(self.network_timeout) return s def get_socket(self, host, port): pid = os.getpid() if pid != self._pid: self._bootstrap(pid) greenlet = getcurrent() from_pool = True sock = self._used.get(greenlet) if sock is None: with self._lock: if self._count < self.pool_size: self._count += 1 from_pool = False sock = self.connect(host, port) if sock is None: from_pool = True sock = self._queue.get(timeout=self.network_timeout) if isinstance(greenlet, gevent.Greenlet): greenlet.link(self._return) self._used[greenlet] = sock else: ref = weakref.ref(greenlet, self._return) self._used[ref] = sock return sock, from_pool def return_socket(self): greenlet = getcurrent() self._return(greenlet) def _return(self, greenlet): try: sock = self._used.get(greenlet) if sock is not None: del self._used[greenlet] self._queue.put(sock) except: with self._lock: self._count -= 1 def patch(): import pymongo.connection pymongo.connection._Pool = Pool
py
1a2f4f92a113636a896106383cfcf97573818685
from buidl._libsec import ffi, lib GLOBAL_CTX = ffi.gc( lib.secp256k1_context_create( lib.SECP256K1_CONTEXT_SIGN | lib.SECP256K1_CONTEXT_VERIFY ), lib.secp256k1_context_destroy, ) def tagged_hash(tag, msg): result = ffi.new("unsigned char [32]") tag_length = len(tag) msg_length = len(msg) if not lib.secp256k1_tagged_sha256( GLOBAL_CTX, result, tag, tag_length, msg, msg_length, ): raise RuntimeError("libsecp256k1 tagged hash problem") return bytes(ffi.buffer(result, 32)) def hash_aux(msg): return tagged_hash(b"BIP0340/aux", msg) def hash_challenge(msg): return tagged_hash(b"BIP0340/challenge", msg) def hash_keyaggcoef(msg): return tagged_hash(b"KeyAgg coefficient", msg) def hash_keyagglist(msg): return tagged_hash(b"KeyAgg list", msg) def hash_musignonce(msg): return tagged_hash(b"MuSig/noncecoef", msg) def hash_nonce(msg): return tagged_hash(b"BIP0340/nonce", msg) def hash_tapbranch(msg): return tagged_hash(b"TapBranch", msg) def hash_tapleaf(msg): return tagged_hash(b"TapLeaf", msg) def hash_tapsighash(msg): return tagged_hash(b"TapSighash", msg) def hash_taptweak(msg): return tagged_hash(b"TapTweak", msg)
py
1a2f4f94f3c75c8f9e9d1fa2053fef78511e680e
from PyQt5 import QtWidgets from PyQt5.QtCore import QObject, pyqtSlot, pyqtSignal, QLocale, Qt from PyQt5.QtWidgets import QPushButton, QLabel from PyQt5.QtGui import QIcon, QPixmap import sys from pymodaq.daq_utils.parameter import utils as putils from pymodaq.daq_measurement.daq_measurement_main import DAQ_Measurement from collections import OrderedDict from pymodaq.daq_utils.plotting.crosshair import Crosshair import pyqtgraph as pg import numpy as np from pymodaq.daq_utils import daq_utils as utils from pymodaq.daq_utils.gui_utils import QAction from pymodaq.daq_utils.plotting.viewer1D.viewer1Dbasic import Viewer1DBasic from pymodaq.daq_utils.managers.roi_manager import ROIManager import datetime logger = utils.set_logger(utils.get_module_name(__file__)) class Viewer1D(QtWidgets.QWidget, QObject): """this plots 1D data on a plotwidget. Math and measurement can be done on it. Datas and measurements are then exported with the signal data_to_export_signal """ data_to_export_signal = pyqtSignal(OrderedDict) # self.data_to_export=edict(data0D=None,data1D=None,data2D=None) math_signal = pyqtSignal(OrderedDict) # OrderedDict:=[x_axis=...,data=...,ROI_bounds=...,operation=] ROI_changed = pyqtSignal() ROI_changed_finished = pyqtSignal() def __init__(self, parent=None): QLocale.setDefault(QLocale(QLocale.English, QLocale.UnitedStates)) super(Viewer1D, self).__init__() self.viewer_type = 'Data1D' self.title = 'viewer1D' if parent is None: parent = QtWidgets.QWidget() self.parent = parent self.roi_manager = ROIManager('1D') self.roi_manager.new_ROI_signal.connect(self.add_lineout) self.roi_manager.remove_ROI_signal.connect(self.remove_ROI) # self.roi_manager.ROI_changed_finished.connect(self.update_lineouts) self.setupUI() self.wait_time = 3000 self.measurement_module = None self.math_module = Viewer1D_math() if DAQ_Measurement is None: # pragma: no cover self.ui.do_measurements_pb.setVisible(False) self._labels = [] self.plot_channels = None self.plot_colors = utils.plot_colors self.color_list = ROIManager.color_list self.lo_items = OrderedDict([]) self.lo_data = OrderedDict([]) self.ROI_bounds = [] self._x_axis = None self.datas = [] # datas on each channel. list of 1D arrays self.data_to_export = None self.measurement_dict = OrderedDict(x_axis=None, datas=[], ROI_bounds=[], operations=[], channels=[]) # OrderedDict to be send to the daq_measurement module self.measure_data_dict = OrderedDict() # dictionnary with data to be put in the table on the form: key="Meas.{}:".format(ind) # and value is the result of a given lineout or measurement def setupUI(self): self.ui = QObject() self.parent.setLayout(QtWidgets.QVBoxLayout()) splitter_hor = QtWidgets.QSplitter(Qt.Horizontal) # self.ui.statusbar = QtWidgets.QStatusBar() # self.ui.statusbar.setMaximumHeight(15) self.parent.layout().addWidget(splitter_hor) #self.parent.layout().addWidget(self.ui.statusbar) splitter_ver = QtWidgets.QSplitter(Qt.Vertical) splitter_hor.addWidget(splitter_ver) splitter_hor.addWidget(self.roi_manager.roiwidget) self.roi_manager.roiwidget.hide() self.ui.button_widget = QtWidgets.QToolBar() splitter_ver.addWidget(self.ui.button_widget) self.ui.Graph_Lineouts = pg.PlotWidget() widg = QtWidgets.QWidget() self.viewer = Viewer1DBasic(widg) splitter_ver.addWidget(widg) splitter_ver.addWidget(self.ui.Graph_Lineouts) self.ui.Graph1D = self.viewer # for backcompatibility self.roi_manager.viewer_widget = self.viewer.plotwidget self.setup_buttons(self.ui.button_widget) self.setup_zoom() self.legend = None self.axis_settings = dict(orientation='bottom', label='x axis', units='pxls') self.ui.xaxis_item = self.viewer.plotwidget.plotItem.getAxis('bottom') self.ui.Graph_Lineouts.hide() self.ui.aspect_ratio_pb.clicked.connect(self.lock_aspect_ratio) # #crosshair self.ui.crosshair = Crosshair(self.viewer.plotwidget.plotItem, orientation='vertical') self.ui.crosshair.crosshair_dragged.connect(self.update_crosshair_data) self.ui.crosshair_pb.clicked.connect(self.crosshairClicked) self.crosshairClicked() # self.ui.Measurement_widget=Dock("Measurement Module", size=(300, 100), closable=True) # self.dockarea.addDock(self.ui.Measurement_widget) self.ui.Measurement_widget = QtWidgets.QWidget() self.ui.Measurement_widget.setVisible(False) # #Connecting buttons: self.ui.Do_math_pb.clicked.connect(self.do_math_fun) self.ui.do_measurements_pb.clicked.connect(self.open_measurement_module) self.ui.zoom_pb.clicked.connect(self.enable_zoom) self.ui.scatter.clicked.connect(self.do_scatter) self.ui.xyplot_action.triggered.connect(self.do_xy) def setup_buttons(self, button_widget): self.ui.zoom_pb = QAction(QIcon(QPixmap(":/icons/Icon_Library/Zoom_to_Selection.png")), 'Zoom Widget') self.ui.zoom_pb.setCheckable(True) button_widget.addAction(self.ui.zoom_pb) self.ui.Do_math_pb = QAction(QIcon(QPixmap(":/icons/Icon_Library/Calculator.png")), 'Do Math using ROI') self.ui.Do_math_pb.setCheckable(True) button_widget.addAction(self.ui.Do_math_pb) self.ui.do_measurements_pb = QAction(QIcon(QPixmap(":/icons/Icon_Library/MeasurementStudio_32.png")), 'Do Advanced measurements (fits,...)') self.ui.do_measurements_pb.setCheckable(True) button_widget.addAction(self.ui.do_measurements_pb) self.ui.crosshair_pb = QAction(QIcon(QPixmap(":/icons/Icon_Library/reset.png")), 'Show data cursor') self.ui.crosshair_pb.setCheckable(True) button_widget.addAction(self.ui.crosshair_pb) self.ui.aspect_ratio_pb = QAction(QIcon(QPixmap(":/icons/Icon_Library/zoomReset.png")), 'Fix the aspect ratio') self.ui.aspect_ratio_pb.setCheckable(True) button_widget.addAction(self.ui.aspect_ratio_pb) self.ui.scatter = QAction(QIcon(QPixmap(":/icons/Icon_Library/Marker.png")), 'Switch between line or scatter plots') self.ui.scatter.setCheckable(True) button_widget.addAction(self.ui.scatter) self.ui.xyplot_action = QAction(QIcon(QPixmap(":/icons/Icon_Library/2d.png")), 'Switch between normal or XY representation (valid for 2 channels)') self.ui.xyplot_action.setCheckable(True) button_widget.addAction(self.ui.xyplot_action) self.ui.xyplot_action.setVisible(False) self.ui.x_label = QAction('x:') button_widget.addAction(self.ui.x_label) self.ui.y_label = QAction('y:') button_widget.addAction(self.ui.y_label) def setup_zoom(self): # create and set the zoom widget # self.ui.zoom_widget=Dock("1DViewer zoom", size=(300, 100), closable=True) self.ui.zoom_widget = QtWidgets.QWidget() layout = QtWidgets.QHBoxLayout() self.ui.Graph_zoom = pg.PlotWidget() layout.addWidget(self.ui.Graph_zoom) self.ui.zoom_widget.setLayout(layout) self.ui.zoom_region = pg.LinearRegionItem() self.ui.zoom_region.setZValue(-10) self.ui.zoom_region.setBrush('r') self.ui.zoom_region.setOpacity(0.2) self.ui.Graph_zoom.addItem(self.ui.zoom_region) self.zoom_plot = [] # self.dockarea.addDock(self.ui.zoom_widget) self.ui.zoom_widget.setVisible(False) def do_scatter(self): self.update_graph1D(self.datas) def do_xy(self): if self.ui.xyplot_action.isChecked(): axis = self.viewer.plotwidget.plotItem.getAxis('bottom') axis.setLabel(text=self.labels[0], units='') axis = self.viewer.plotwidget.plotItem.getAxis('left') axis.setLabel(text=self.labels[1], units='') self.legend.setVisible(False) else: self.set_axis_label(dict(orientation='bottom', label=self.axis_settings['label'], units=self.axis_settings['units'])) axis = self.viewer.plotwidget.plotItem.getAxis('left') axis.setLabel(text='', units='') self.legend.setVisible(True) self.update_graph1D(self.datas) def update_lineouts(self): try: operations = [] channels = [] for ind, key in enumerate(self.roi_manager.ROIs): operations.append(self.roi_manager.settings.child('ROIs', key, 'math_function').value()) channels.append( self.roi_manager.settings.child('ROIs', key, 'use_channel').opts['limits'].index( self.roi_manager.settings.child('ROIs', key, 'use_channel').value())) self.lo_items[key].setPen(self.roi_manager.settings.child('ROIs', key, 'Color').value()) self.measurement_dict['datas'] = self.datas self.measurement_dict['ROI_bounds'] = [self.roi_manager.ROIs[item].getRegion() for item in self.roi_manager.ROIs] self.measurement_dict['channels'] = channels self.measurement_dict['operations'] = operations data_lo = self.math_module.update_math(self.measurement_dict) self.show_math(data_lo) except Exception as e: pass @pyqtSlot(str) def remove_ROI(self, roi_name): item = self.lo_items.pop(roi_name) self.ui.Graph_Lineouts.plotItem.removeItem(item) self.measure_data_dict.pop("Lineout_{:s}:".format(roi_name)) self.update_lineouts() @pyqtSlot(int, str) def add_lineout(self, index, roi_type=''): try: item = self.roi_manager.ROIs['ROI_{:02d}'.format(index)] item_param = self.roi_manager.settings.child('ROIs', 'ROI_{:02d}'.format(index)) item_param.child(('use_channel')).setOpts(limits=self.labels) if len(self.labels) == 0: # pragma: no cover lab = '' else: lab = self.labels[0] item_param.child(('use_channel')).setValue(lab) item.sigRegionChanged.connect(self.update_lineouts) item.sigRegionChangeFinished.connect(self.ROI_changed_finished.emit) for child in putils.iter_children_params(item_param, childlist=[]): if child.type() != 'group': child.sigValueChanged.connect(self.update_lineouts) item_lo = self.ui.Graph_Lineouts.plot() item_lo.setPen(item_param.child(('Color')).value()) self.lo_items['ROI_{:02d}'.format(index)] = item_lo self.lo_data = OrderedDict([]) for k in self.lo_items: self.lo_data[k] = np.zeros((1,)) self.update_lineouts() except Exception as e: logger.exception(str(e)) def clear_lo(self): self.lo_data = [[] for ind in range(len(self.lo_data))] self.update_lineouts() def crosshairClicked(self): if self.ui.crosshair_pb.isChecked(): self.ui.crosshair.setVisible(True) self.ui.x_label.setVisible(True) self.ui.y_label.setVisible(True) range = self.viewer.plotwidget.plotItem.vb.viewRange() self.ui.crosshair.set_crosshair_position(xpos=np.mean(np.array(range[0]))) else: self.ui.crosshair.setVisible(False) self.ui.x_label.setVisible(False) self.ui.y_label.setVisible(False) def do_math_fun(self): try: if self.ui.Do_math_pb.isChecked(): self.roi_manager.roiwidget.show() self.ui.Graph_Lineouts.show() else: self.ui.Graph_Lineouts.hide() self.roi_manager.roiwidget.hide() except Exception as e: logger.exception(str(e)) def do_zoom(self): bounds = self.ui.zoom_region.getRegion() self.viewer.plotwidget.setXRange(bounds[0], bounds[1]) def enable_zoom(self): try: if not (self.ui.zoom_pb.isChecked()): if self.zoom_plot != []: for plot in self.zoom_plot: self.ui.Graph_zoom.removeItem(plot) self.ui.zoom_widget.hide() self.ui.zoom_region.sigRegionChanged.disconnect(self.do_zoom) else: self.zoom_plot = [] for ind, data in enumerate(self.datas): channel = self.ui.Graph_zoom.plot() channel.setPen(self.plot_colors[ind]) self.zoom_plot.append(channel) self.update_graph1D(self.datas) self.ui.zoom_region.setRegion([np.min(self._x_axis), np.max(self._x_axis)]) self.ui.zoom_widget.show() self.ui.zoom_region.sigRegionChanged.connect(self.do_zoom) except Exception as e: logger.exception(str(e)) def ini_data_plots(self, Nplots): try: self.plot_channels = [] # if self.legend is not None: # self.viewer.plotwidget.plotItem.removeItem(self.legend) self.legend = self.viewer.plotwidget.plotItem.legend flag = True while flag: items = [item[1].text for item in self.legend.items] if len(items) == 0: flag = False else: self.legend.removeItem(items[0]) channels = [] for ind in range(Nplots): channel = self.viewer.plotwidget.plot() channel.setPen(self.plot_colors[ind]) self.legend.addItem(channel, self._labels[ind]) channels.append(ind) self.plot_channels.append(channel) except Exception as e: logger.exception(str(e)) def update_labels(self, labels=[]): try: labels_tmp = labels[:] if self.labels == labels: if self.labels == [] or len(self.labels) < len(self.datas): self._labels = ["CH{}".format(ind) for ind in range(len(self.datas))] else: if self.legend is not None: flag = True while flag: items = [item[1].text for item in self.legend.items] if len(items) == 0: flag = False else: self.legend.removeItem(items[0]) if len(labels) < len(self.plot_channels): for ind in range(len(labels), len(self.plot_channels)): labels_tmp.append('CH{:02d}'.format(ind)) if len(labels_tmp) == len(self.plot_channels): for ind, channel in enumerate(self.plot_channels): self.legend.addItem(channel, labels_tmp[ind]) self._labels = labels_tmp if self.labels != labels: for ind in range(len(self.roi_manager.ROIs)): val = self.roi_manager.settings.child('ROIs', 'ROI_{:02d}'.format(ind), 'use_channel').value() self.roi_manager.settings.child('ROIs', 'ROI_{:02d}'.format(ind), 'use_channel').setOpts( limits=self.labels) if val not in self.labels: self.roi_manager.settings.child('ROIs', 'ROI_{:02d}'.format(ind), 'use_channel').setValue( self.labels[0]) self.ui.xyplot_action.setVisible(len(self.labels) == 2) except Exception as e: logger.exception(str(e)) @property def labels(self): return self._labels @labels.setter def labels(self, labels): self.update_labels(labels) self._labels = labels def lock_aspect_ratio(self): if self.ui.aspect_ratio_pb.isChecked(): self.viewer.plotwidget.plotItem.vb.setAspectLocked(lock=True, ratio=1) else: self.viewer.plotwidget.plotItem.vb.setAspectLocked(lock=False) def open_measurement_module(self): if not (self.ui.Do_math_pb.isChecked()): self.ui.Do_math_pb.setChecked(True) QtWidgets.QApplication.processEvents() self.ui.Do_math_pb.clicked.emit() QtWidgets.QApplication.processEvents() self.ui.Measurement_widget.setVisible(True) if self.ui.do_measurements_pb.isChecked(): Form = self.ui.Measurement_widget self.measurement_module = DAQ_Measurement(Form) # self.ui.Measurement_widget.addWidget(Form) self.measurement_module.measurement_signal[list].connect(self.show_measurement) self.update_measurement_module() elif self.measurement_module is not None: self.measurement_module.Quit_fun() def remove_plots(self): if self.plot_channels is not None: for channel in self.plot_channels: self.viewer.plotwidget.removeItem(channel) self.plot_channels = None if self.legend is not None: self.viewer.plotwidget.removeItem(self.legend) def set_axis_label(self, axis_settings=dict(orientation='bottom', label='x axis', units='pxls')): axis = self.viewer.plotwidget.plotItem.getAxis(axis_settings['orientation']) axis.setLabel(text=axis_settings['label'], units=axis_settings['units']) self.axis_settings = axis_settings @pyqtSlot(list) def show_data(self, datas, labels=None, x_axis=None): try: self.datas = datas self.update_labels(self.labels) self.data_to_export = OrderedDict(name=self.title, data0D=OrderedDict(), data1D=OrderedDict(), data2D=None) for ind, data in enumerate(datas): self.data_to_export['data1D']['CH{:03d}'.format(ind)] = utils.DataToExport() if self.plot_channels == [] or self.plot_channels is None: # initialize data and plots self.ini_data_plots(len(datas)) elif len(self.plot_channels) != len(datas): self.remove_plots() self.ini_data_plots(len(datas)) self.update_graph1D(datas) if x_axis is not None: self.x_axis = x_axis if labels is not None: self.update_labels(labels) if self.ui.do_measurements_pb.isChecked(): self.update_measurement_module() except Exception as e: logger.exception(str(e)) @pyqtSlot(list) def show_data_temp(self, datas): """f to plot temporary data, for instance when all pixels are not yet populated... """ try: self.update_labels(self.labels) self.datas = datas if self.plot_channels is None: # initialize data and plots self.ini_data_plots(len(datas)) elif len(self.plot_channels) != len(datas): self.remove_plots() self.ini_data_plots(len(datas)) for ind_plot, data in enumerate(datas): if self.x_axis is None: self.x_axis = np.linspace(0, len(data), len(data), endpoint=False) x_axis = self.x_axis elif len(self.x_axis) != len(data): x_axis = np.linspace(0, len(data), len(data), endpoint=False) else: x_axis = self.x_axis self.plot_channels[ind_plot].setData(x=x_axis, y=data) except Exception as e: logger.exception(str(e)) @pyqtSlot(list) def show_math(self, data_lo): # self.data_to_export=OrderedDict(x_axis=None,y_axis=None,z_axis=None,data0D=None,data1D=None,data2D=None) if len(data_lo) != 0: for ind, key in enumerate(self.lo_items): self.measure_data_dict["Lineout_{:s}:".format(key)] = data_lo[ind] self.data_to_export['data0D']['Measure_{:03d}'.format(ind)] = utils.DataToExport(name=self.title, data=data_lo[ind], source='roi') self.roi_manager.settings.child(('measurements')).setValue(self.measure_data_dict) for ind, key in enumerate(self.lo_items): self.lo_data[key] = np.append(self.lo_data[key], data_lo[ind]) self.lo_items[key].setData(y=self.lo_data[key]) if not (self.ui.do_measurements_pb.isChecked()): # otherwise you export data from measurement self.data_to_export['acq_time_s'] = datetime.datetime.now().timestamp() self.data_to_export_signal.emit(self.data_to_export) @pyqtSlot(list) def show_measurement(self, data_meas): ind_offset = len(self.data_to_export['data0D']) for ind, res in enumerate(data_meas): self.measure_data_dict["Meas.{}:".format(ind)] = res self.data_to_export['data0D']['Measure_{:03d}'.format(ind + ind_offset)] = \ utils.DataToExport(name=self.title, data=res, source='roi') self.roi_manager.settings.child('measurements').setValue(self.measure_data_dict) self.data_to_export['acq_time_s'] = datetime.datetime.now().timestamp() self.data_to_export_signal.emit(self.data_to_export) def update_crosshair_data(self, posx, posy, name=""): try: indx = utils.find_index(self._x_axis, posx)[0][0] string = "y=" for data in self.datas: string += "{:.6e} / ".format(data[indx]) self.ui.y_label.setText(string) self.ui.x_label.setText("x={:.6e} ".format(posx)) except Exception as e: pass def update_graph1D(self, datas): # self.data_to_export=OrderedDict(data0D=OrderedDict(),data1D=OrderedDict(),data2D=None) try: pens = [] symbolBrushs = [] symbolSize = 5 for ind, ch in enumerate(self.plot_channels): if self.ui.scatter.isChecked(): pens.append(None) symbol = 'o' symbolBrushs.append(self.plot_colors[ind]) else: pens.append(self.plot_colors[ind]) symbol = None symbolBrushs.append(None) if self.x_axis is None: self._x_axis = np.linspace(0, len(datas[0]), len(datas[0]), endpoint=False) elif len(self.x_axis) != len(datas[0]): self._x_axis = np.linspace(0, len(datas[0]), len(datas[0]), endpoint=False) for ind_plot, data in enumerate(datas): if not self.ui.xyplot_action.isChecked() or len(datas) == 0: self.plot_channels[ind_plot].setData(x=self.x_axis, y=data, pen=pens[ind_plot], symbol=symbol, symbolBrush=symbolBrushs[ind_plot], symbolSize=symbolSize, pxMode=True) else: self.plot_channels[ind_plot].setData(x=np.array([]), y=np.array([]), pen=pens[ind_plot], symbol=symbol, symbolBrush=symbolBrushs[ind_plot], symbolSize=symbolSize, pxMode=True) if self.ui.zoom_pb.isChecked(): self.zoom_plot[ind_plot].setData(x=self.x_axis, y=data) x_axis = utils.Axis(data=self.x_axis, units=self.axis_settings['units'], label=self.axis_settings['label']) self.data_to_export['data1D']['CH{:03d}'.format(ind_plot)].update( OrderedDict(name=self.title, data=data, x_axis=x_axis, source='raw')) # to be saved or exported if self.ui.xyplot_action.isChecked() and len(datas) > 1: self.plot_channels[0].setData(x=datas[0], y=datas[1], pen=pens[0], symbol=symbol, symbolBrush=symbolBrushs[0], symbolSize=symbolSize, pxMode=True) if not self.ui.Do_math_pb.isChecked(): # otherwise math is done and then data is exported self.data_to_export['acq_time_s'] = datetime.datetime.now().timestamp() self.data_to_export_signal.emit(self.data_to_export) else: self.measurement_dict['datas'] = datas if self.measurement_dict['x_axis'] is None: self.measurement_dict['x_axis'] = self._x_axis data_lo = self.math_module.update_math(self.measurement_dict) self.show_math(data_lo) except Exception as e: logger.exception(str(e)) def update_measurement_module(self): xdata = self.measurement_dict['x_axis'] ydata = self.measurement_dict['datas'][0] if xdata is None: self.measurement_module.update_data(ydata=ydata) else: self.measurement_module.update_data(xdata=xdata, ydata=ydata) def update_status(self, txt): logger.info(txt) @property def x_axis(self): return self._x_axis @x_axis.setter def x_axis(self, x_axis): label = 'Pxls' units = '' if isinstance(x_axis, dict): if 'data' in x_axis: xdata = x_axis['data'] if 'label' in x_axis: label = x_axis['label'] if 'units' in x_axis: units = x_axis['units'] else: xdata = x_axis self._x_axis = xdata self.measurement_dict['x_axis'] = self._x_axis if self.datas != []: self.show_data_temp(self.datas) self.set_axis_label(dict(orientation='bottom', label=label, units=units)) class Viewer1D_math(QObject): status_sig = pyqtSignal(list) def __init__(self): super(QObject, self).__init__() self.datas = [] self.ROI_bounds = [] self.x_axis = None self.operations = [] self.channels = [] def update_math(self, measurement_dict): try: if 'datas' in measurement_dict: self.datas = measurement_dict['datas'] if 'ROI_bounds' in measurement_dict: self.ROI_bounds = measurement_dict['ROI_bounds'] if 'x_axis' in measurement_dict: self.x_axis = measurement_dict['x_axis'] if 'operations' in measurement_dict: self.operations = measurement_dict['operations'] if 'channels' in measurement_dict: self.channels = measurement_dict['channels'] # self.status_sig.emit(["Update_Status","doing math"]) data_lo = [] for ind_meas in range(len(self.operations)): indexes = utils.find_index(self.x_axis, self.ROI_bounds[ind_meas]) ind1 = indexes[0][0] ind2 = indexes[1][0] sub_data = self.datas[self.channels[ind_meas]][ind1:ind2] sub_xaxis = self.x_axis[ind1:ind2] if self.operations[ind_meas] == "Mean": data_lo.append(float(np.mean(sub_data))) elif self.operations[ind_meas] == "Sum": data_lo.append(float(np.sum(sub_data))) elif self.operations[ind_meas] == 'half-life' or self.operations[ind_meas] == 'expotime': ind_x0 = utils.find_index(sub_data, np.max(sub_data))[0][0] x0 = sub_xaxis[ind_x0] sub_xaxis = sub_xaxis[ind_x0:] sub_data = sub_data[ind_x0:] offset = sub_data[-1] N0 = np.max(sub_data) - offset if self.operations[ind_meas] == 'half-life': time = sub_xaxis[utils.find_index(sub_data - offset, 0.5 * N0)[0][0]] - x0 elif self.operations[ind_meas] == 'expotime': time = sub_xaxis[utils.find_index(sub_data - offset, 0.37 * N0)[0][0]] - x0 data_lo.append(time) return data_lo except Exception as e: logger.exception(str(e)) return [] if __name__ == '__main__': # pragma: no cover app = QtWidgets.QApplication(sys.argv) Form = QtWidgets.QWidget() prog = Viewer1D(Form) from pymodaq.daq_utils.daq_utils import gauss1D x = np.linspace(0, 200, 201) y1 = gauss1D(x, 75, 25) y2 = gauss1D(x, 120, 50, 2) tau_half = 27 tau2 = 100 x0 = 50 dx = 20 ydata_expodec = np.zeros((len(x))) ydata_expodec[:50] = 1 * gauss1D(x[:50], x0, dx, 2) ydata_expodec[50:] = 1 * np.exp(-(x[50:] - x0) / (tau_half / np.log(2))) # +1*np.exp(-(x[50:]-x0)/tau2) ydata_expodec += 0.1 * np.random.rand(len(x)) # x = np.sin(np.linspace(0,6*np.pi,201)) # y = np.sin(np.linspace(0, 6*np.pi, 201)+np.pi/2) Form.show() prog.ui.Do_math_pb.click() QtWidgets.QApplication.processEvents() prog.x_axis = x # prog.show_data([y, y+2]) prog.show_data([y1, y2, ydata_expodec]) QtWidgets.QApplication.processEvents() prog.update_labels(['coucou', 'label2']) sys.exit(app.exec_())
py
1a2f4fa11d88a0f3f192fd7453d835cbb54f7831
# Author: Phyllipe Bezerra (https://github.com/pmba) clothes = { 0: "underwear", 1: "pants", 2: "belt", 3: "suit", 4: "shoe", 5: "socks", 6: "shirt", 7: "tie", 8: "clock", } graph = [[1, 4], [2, 4], [3], [], [], [4], [2, 7], [3], []] visited = [0 for x in range(len(graph))] stack = [] def print_stack(stack, clothes): order = 1 while stack: cur_clothe = stack.pop() print(order, clothes[cur_clothe]) order += 1 def dfs(u, visited, graph): visited[u] = 1 for v in graph[u]: if not visited[v]: dfs(v, visited, graph) stack.append(u) def top_sort(graph, visited): for v in range(len(graph)): if not visited[v]: dfs(v, visited, graph) if __name__ == "__main__": top_sort(graph, visited) print(stack) print_stack(stack, clothes)
bzl
1a2f4fd9aca8b324e74562a7cd5cb686149c0cd9
# Copyright 2018 The Bazel Authors. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Actions related to codesigning.""" load( "@build_bazel_rules_apple//apple/internal/utils:defines.bzl", "defines", ) load( "@build_bazel_rules_apple//apple/internal/utils:legacy_actions.bzl", "legacy_actions", ) load( "@build_bazel_rules_apple//apple/internal:intermediates.bzl", "intermediates", ) load( "@build_bazel_rules_apple//apple/internal:platform_support.bzl", "platform_support", ) load( "@build_bazel_rules_apple//apple/internal:rule_support.bzl", "rule_support", ) load( "@bazel_skylib//lib:paths.bzl", "paths", ) load( "@bazel_skylib//lib:shell.bzl", "shell", ) def _double_quote(raw_string): """Add double quotes around the string and preserve existing quote characters. Args: raw_string: A string that might have shell-syntaxed environment variables. Returns: The string with double quotes. """ return "\"" + raw_string.replace("\"", "\\\"") + "\"" def _no_op(x): """Helper that does not nothing be return the result.""" return x def _codesign_args_for_path( ctx, path_to_sign, provisioning_profile, entitlements_file, shell_quote = True): """Returns a command line for the codesigning tool wrapper script. Args: ctx: The Starlark context. path_to_sign: A struct indicating the path that should be signed and its optionality (see `_path_to_sign`). provisioning_profile: The provisioning profile file. May be `None`. entitlements_file: The entitlements file to pass to codesign. May be `None` for non-app binaries (e.g. test bundles). shell_quote: Sanitizes the arguments to be evaluated in a shell. Returns: The codesign command invocation for the given directory as a list. """ if not path_to_sign.is_directory and path_to_sign.signed_frameworks: fail("Internal Error: Received a list of signed frameworks as exceptions " + "for code signing, but path to sign is not a directory.") for x in path_to_sign.signed_frameworks: if not x.startswith(path_to_sign.path): fail("Internal Error: Signed framework does not have the current path " + "to sign (%s) as its prefix (%s)." % (path_to_sign.path, x)) cmd_codesigning = [ "--codesign", "/usr/bin/codesign", ] is_device = platform_support.is_device_build(ctx) # Add quotes for sanitizing inputs when they're invoked directly from a shell script, for # instance when using this string to assemble the output of codesigning_command. maybe_quote = shell.quote if shell_quote else _no_op maybe_double_quote = _double_quote if shell_quote else _no_op # First, try to use the identity passed on the command line, if any. If it's a simulator build, # use an ad hoc identity. identity = ctx.fragments.objc.signing_certificate_name if is_device else "-" if not identity: if provisioning_profile: cmd_codesigning.extend([ "--mobileprovision", maybe_quote(provisioning_profile.path), ]) else: identity = "-" if identity: cmd_codesigning.extend([ "--identity", maybe_quote(identity), ]) # The entitlements rule ensures that entitlements_file is None or a file # containing only "com.apple.security.get-task-allow" when building for the # simulator. if path_to_sign.use_entitlements and entitlements_file: cmd_codesigning.extend([ "--entitlements", maybe_quote(entitlements_file.path), ]) if is_device: cmd_codesigning.append("--force") else: cmd_codesigning.extend([ "--force", "--disable_timestamp", ]) if path_to_sign.is_directory: cmd_codesigning.append("--directory_to_sign") else: cmd_codesigning.append("--target_to_sign") # Because the path does include environment variables which need to be expanded, path has to be # quoted using double quotes, this means that path can't be quoted using shell.quote. cmd_codesigning.append(maybe_double_quote(path_to_sign.path)) if path_to_sign.signed_frameworks: for signed_framework in path_to_sign.signed_frameworks: # Signed frameworks must also be double quoted, as they too have an environment # variable to be expanded. cmd_codesigning.extend([ "--signed_path", maybe_double_quote(signed_framework), ]) return cmd_codesigning def _path_to_sign(path, is_directory = False, signed_frameworks = [], use_entitlements = True): """Returns a "path to sign" value to be passed to `_signing_command_lines`. Args: path: The path to sign, relative to wherever the code signing command lines are being executed. is_directory: If `True`, the path is a directory and not a bundle, indicating that the contents of each item in the directory should be code signed except for the invisible files prefixed with a period. signed_frameworks: If provided, a list of frameworks that have already been signed. use_entitlements: If provided, indicates if the entitlements on the bundling target should be used for signing this path (useful to disabled the use when signing frameworks within an iOS app). Returns: A `struct` that can be passed to `_signing_command_lines`. """ return struct( path = path, is_directory = is_directory, signed_frameworks = signed_frameworks, use_entitlements = use_entitlements, ) def _provisioning_profile(ctx): # Verify that a provisioning profile was provided for device builds on # platforms that require it. is_device = platform_support.is_device_build(ctx) provisioning_profile = getattr(ctx.file, "provisioning_profile", None) rule_descriptor = rule_support.rule_descriptor(ctx) if (is_device and rule_descriptor.requires_signing_for_device and not provisioning_profile): fail("The provisioning_profile attribute must be set for device " + "builds on this platform (%s)." % platform_support.platform_type(ctx)) return provisioning_profile def _signing_command_lines( ctx, paths_to_sign, entitlements_file): """Returns a multi-line string with codesign invocations for the bundle. For any signing identity other than ad hoc, the identity is verified as being valid in the keychain and an error will be emitted if the identity cannot be used for signing for any reason. Args: ctx: The Starlark context. paths_to_sign: A list of values returned from `path_to_sign` that indicate paths that should be code-signed. entitlements_file: The entitlements file to pass to codesign. Returns: A multi-line string with codesign invocations for the bundle. """ provisioning_profile = _provisioning_profile(ctx) commands = [] # Use of the entitlements file is not recommended for the signing of frameworks. As long as # this remains the case, we do have to split the "paths to sign" between multiple invocations # of codesign. for path_to_sign in paths_to_sign: codesign_command = [ctx.executable._codesigningtool.path] codesign_command.extend(_codesign_args_for_path( ctx, path_to_sign, provisioning_profile, entitlements_file, )) commands.append(" ".join(codesign_command)) return "\n".join(commands) def _should_sign_simulator_bundles(ctx): """Check if a main bundle should be codesigned. The Frameworks/* bundles should *always* be signed, this is just for the other bundles. Args: ctx: The Starlark context. Returns: True/False for if the bundle should be signed. """ rule_descriptor = rule_support.rule_descriptor(ctx) if not rule_descriptor.skip_simulator_signing_allowed: return True # Default is to sign. return defines.bool_value( ctx, "apple.codesign_simulator_bundles", True, ) def _should_sign_bundles(ctx): should_sign_bundles = True rule_descriptor = rule_support.rule_descriptor(ctx) codesigning_exceptions = rule_descriptor.codesigning_exceptions if (codesigning_exceptions == rule_support.codesigning_exceptions.sign_with_provisioning_profile): # If the rule doesn't have a provisioning profile, do not sign the binary or its # frameworks. provisioning_profile = getattr(ctx.file, "provisioning_profile", None) if not provisioning_profile: should_sign_bundles = False elif codesigning_exceptions == rule_support.codesigning_exceptions.skip_signing: should_sign_bundles = False elif codesigning_exceptions != rule_support.codesigning_exceptions.none: fail("Internal Error: Encountered unsupported state for codesigning_exceptions.") return should_sign_bundles def _codesigning_args( ctx, entitlements, full_archive_path, is_framework = False): """Returns a set of codesigning arguments to be passed to the codesigning tool. Args: ctx: The rule context. entitlements: The entitlements file to sign with. Can be None. full_archive_path: The full path to the codesigning target. is_framework: If the target is a framework. False by default. Returns: A list containing the arguments to pass to the codesigning tool. """ if not _should_sign_bundles(ctx): return [] is_device = platform_support.is_device_build(ctx) if not is_framework and not is_device and not _should_sign_simulator_bundles(ctx): return [] return _codesign_args_for_path( ctx, _path_to_sign(full_archive_path), provisioning_profile = _provisioning_profile(ctx), entitlements_file = entitlements, shell_quote = False, ) def _codesigning_command( ctx, entitlements, frameworks_path, signed_frameworks, bundle_path = ""): """Returns a codesigning command that includes framework embedded bundles. Args: ctx: The rule context. entitlements: The entitlements file to sign with. Can be None. frameworks_path: The location of the Frameworks directory, relative to the archive. signed_frameworks: A depset containing each framework that has already been signed. bundle_path: The location of the bundle, relative to the archive. Returns: A string containing the codesigning commands. """ if not _should_sign_bundles(ctx): return "" paths_to_sign = [] # The command returned by this function is executed as part of a bundling shell script. # Each directory to be signed must be prefixed by $WORK_DIR, which is the variable in that # script that contains the path to the directory where the bundle is being built. if frameworks_path: framework_root = paths.join("$WORK_DIR", frameworks_path) + "/" full_signed_frameworks = [] for signed_framework in signed_frameworks.to_list(): full_signed_frameworks.append(paths.join(framework_root, signed_framework)) paths_to_sign.append( _path_to_sign( framework_root, is_directory = True, signed_frameworks = full_signed_frameworks, use_entitlements = False, ), ) is_device = platform_support.is_device_build(ctx) if is_device or _should_sign_simulator_bundles(ctx): path_to_sign = paths.join("$WORK_DIR", bundle_path) paths_to_sign.append( _path_to_sign(path_to_sign), ) return _signing_command_lines( ctx, paths_to_sign = paths_to_sign, entitlements_file = entitlements, ) def _post_process_and_sign_archive_action( ctx, archive_codesigning_path, frameworks_path, input_archive, output_archive, output_archive_root_path, signed_frameworks, entitlements = None): """Post-processes and signs an archived bundle. Args: ctx: The target's rule context. archive_codesigning_path: The codesigning path relative to the archive. frameworks_path: The Frameworks path relative to the archive. input_archive: The `File` representing the archive containing the bundle that has not yet been processed or signed. output_archive: The `File` representing the processed and signed archive. output_archive_root_path: The `string` path to where the processed, uncompressed archive should be located. signed_frameworks: Depset containing each framework that has already been signed. entitlements: Optional file representing the entitlements to sign with. """ input_files = [input_archive] processing_tools = [] signing_command_lines = _codesigning_command( ctx, entitlements, frameworks_path, signed_frameworks, bundle_path = archive_codesigning_path, ) if signing_command_lines: processing_tools.append(ctx.executable._codesigningtool) if entitlements: input_files.append(entitlements) provisioning_profile = getattr(ctx.file, "provisioning_profile", None) if provisioning_profile: input_files.append(provisioning_profile) ipa_post_processor = ctx.executable.ipa_post_processor ipa_post_processor_path = "" if ipa_post_processor: processing_tools.append(ipa_post_processor) ipa_post_processor_path = ipa_post_processor.path # Only compress the IPA for optimized (release) builds or when requested. # For debug builds, zip without compression, which will speed up the build. compression_requested = defines.bool_value(ctx, "apple.compress_ipa", False) should_compress = (ctx.var["COMPILATION_MODE"] == "opt") or compression_requested # TODO(b/163217926): These are kept the same for the three different actions # that could be run to ensure anything keying off these values continues to # work. After some data is collected, the values likely can be revisited and # changed. mnemonic = "ProcessAndSign" progress_message = "Processing and signing %s" % ctx.label.name # If there is no work to be done, skip the processing/signing action, just # copy the file over. has_work = any([signing_command_lines, ipa_post_processor_path, should_compress]) if not has_work: ctx.actions.run_shell( inputs = [input_archive], outputs = [output_archive], mnemonic = mnemonic, progress_message = progress_message, command = "cp -p '%s' '%s'" % (input_archive.path, output_archive.path), ) return process_and_sign_template = intermediates.file( ctx.actions, ctx.label.name, "process-and-sign-%s.sh" % hash(output_archive.path), ) ctx.actions.expand_template( template = ctx.file._process_and_sign_template, output = process_and_sign_template, is_executable = True, substitutions = { "%ipa_post_processor%": ipa_post_processor_path or "", "%output_path%": output_archive.path, "%should_compress%": "1" if should_compress else "", "%signing_command_lines%": signing_command_lines, "%unprocessed_archive_path%": input_archive.path, "%work_dir%": output_archive_root_path, }, ) # Build up some arguments for the script to allow logging to tell what work # is being done within the action's script. arguments = [] if signing_command_lines: arguments.append("should_sign") if ipa_post_processor_path: arguments.append("should_process") if should_compress: arguments.append("should_compress") run_on_darwin = any([signing_command_lines, ipa_post_processor_path]) if run_on_darwin: legacy_actions.run( ctx, inputs = input_files, outputs = [output_archive], executable = process_and_sign_template, arguments = arguments, mnemonic = mnemonic, progress_message = progress_message, execution_requirements = { # Added so that the output of this action is not cached remotely, in case multiple # developers sign the same artifact with different identities. "no-cache": "1", # Unsure, but may be needed for keychain access, especially for files that live in # $HOME. "no-sandbox": "1", }, tools = processing_tools, ) else: ctx.actions.run( inputs = input_files, outputs = [output_archive], executable = process_and_sign_template, arguments = arguments, mnemonic = mnemonic, progress_message = progress_message, ) def _sign_binary_action(ctx, input_binary, output_binary): """Signs the input binary file, copying it into the given output binary file. Args: ctx: The target's rule context. input_binary: The `File` representing the binary to be signed. output_binary: The `File` representing signed binary. """ # It's not hermetic to sign the binary that was built by the apple_binary # target that this rule takes as an input, so we copy it and then execute the # code signing commands on that copy in the same action. path_to_sign = _path_to_sign(output_binary.path) signing_commands = _signing_command_lines( ctx, [path_to_sign], None, ) legacy_actions.run_shell( ctx, inputs = [input_binary], outputs = [output_binary], command = [ "/bin/bash", "-c", "cp {input_binary} {output_binary}".format( input_binary = input_binary.path, output_binary = output_binary.path, ) + "\n" + signing_commands, ], mnemonic = "SignBinary", execution_requirements = { # Added so that the output of this action is not cached remotely, in case multiple # developers sign the same artifact with different identities. "no-cache": "1", # Unsure, but may be needed for keychain access, especially for files that live in # $HOME. "no-sandbox": "1", }, tools = [ ctx.executable._codesigningtool, ], ) codesigning_support = struct( codesigning_args = _codesigning_args, codesigning_command = _codesigning_command, post_process_and_sign_archive_action = _post_process_and_sign_archive_action, provisioning_profile = _provisioning_profile, sign_binary_action = _sign_binary_action, )
py
1a2f50f2434ae8e043b6001f11ce27ca488d5dc3
from panda3d.core import * from direct.gui.DirectGui import * from direct.task.Task import Task from direct.interval.IntervalGlobal import * import DistributedCloset, ClosetGlobals, TrunkGUI from toontown.toon import ToonDNA from toontown.toonbase import TTLocalizer from toontown.toonbase import ToontownGlobals N_A = 0 class DistributedTrunk(DistributedCloset.DistributedCloset): notify = directNotify.newCategory('DistributedTrunk') def __init__(self, cr): DistributedCloset.DistributedCloset.__init__(self, cr) self.hatList = [] self.glassesList = [] self.backpackList = [] self.shoesList = [] self.oldHatList = [] self.oldGlassesList = [] self.oldBackpackList = [] self.oldShoesList = [] self.swapHatEvent = '' self.swapGlassesEvent = '' self.swapBackpackEvent = '' self.swapShoesEvent = '' self.hatDeleted = 0 self.glassesDeleted = 0 self.backpackDeleted = 0 self.shoesDeleted = 0 self.isFreePlayer = 0 def printInfo(self): print 'avid: %s, gender: %s' % (self.av.doId, self.av.style.gender) print 'current hat = %s, glasses = %s, backpack = %s, shoes = %s' % (self.av.getHat(), self.av.getGlasses(), self.av.getBackpack(), self.av.getShoes()) print 'hatList = %s' % self.av.getHatList() print 'glassesList = %s' % self.av.getGlassesList() print 'backpackList = %s' % self.av.getBackpackList() print 'shoesList = %s' % self.av.getShoesList() def setState(self, mode, avId, ownerId, gender, hatList, glassesList, backpackList, shoesList): self.notify.debug('setState, mode=%s, avId=%s, ownerId=%d' % (mode, avId, ownerId)) self.isOwner = avId == ownerId self.ownerGender = gender if mode == ClosetGlobals.CLOSED: self.fsm.request('closed') return else: if mode == ClosetGlobals.OPEN: self.customerId = avId self.av = self.cr.doId2do.get(self.customerId, None) if self.av: if self.av.getGameAccess() != ToontownGlobals.AccessFull: self.isOwner = 0 self.isFreePlayer = 1 else: self.isFreePlayer = 0 if base.localAvatar.getDoId() == self.customerId: self.gender = self.av.style.gender self.hatList = hatList self.glassesList = glassesList self.backpackList = backpackList self.shoesList = shoesList self.oldHatList = self.hatList[0:] self.oldGlassesList = self.glassesList[0:] self.oldBackpackList = self.backpackList[0:] self.oldShoesList = self.shoesList[0:] print '-----------Starting trunk interaction-----------' self.printInfo() print '-------------------------------------------------' if not self.isOwner: self.__popupNotOwnerPanel() else: taskMgr.doMethodLater(0.5, self.popupChangeClothesGUI, self.uniqueName('popupChangeClothesGUI')) self.fsm.request('open') return def load(self): lNode = self.find('**/lid_origin') lLid = self.find('**/lid') if lNode.isEmpty() or lLid.isEmpty(): self.lid = None else: lLid.wrtReparentTo(lNode) self.lid = lNode if not lNode.isEmpty(): self.scale = lLid.getScale()[0] * 0.6 return def popupChangeClothesGUI(self, task): self.notify.debug('popupChangeClothesGUI') self.purchaseDoneEvent = self.uniqueName('purchaseDone') self.swapHatEvent = self.uniqueName('swapHat') self.swapGlassesEvent = self.uniqueName('swapGlasses') self.swapBackpackEvent = self.uniqueName('swapBackpack') self.swapShoesEvent = self.uniqueName('swapShoes') self.cancelEvent = self.uniqueName('cancel') self.accept(self.purchaseDoneEvent, self.__proceedToCheckout) self.accept(self.swapHatEvent, self.__handleSwapHat) self.accept(self.swapGlassesEvent, self.__handleSwapGlasses) self.accept(self.swapBackpackEvent, self.__handleSwapBackpack) self.accept(self.swapShoesEvent, self.__handleSwapShoes) self.accept(self.cancelEvent, self._handleCancel) self.deleteEvent = self.uniqueName('delete') if self.isOwner: self.accept(self.deleteEvent, self.__handleDelete) if not self.closetGUI: self.closetGUI = TrunkGUI.TrunkGUI(self.isOwner, self.purchaseDoneEvent, self.cancelEvent, self.swapHatEvent, self.swapGlassesEvent, self.swapBackpackEvent, self.swapShoesEvent, self.deleteEvent, self.hatList, self.glassesList, self.backpackList, self.shoesList) self.closetGUI.load() if self.gender != self.ownerGender: self.closetGUI.setGender(self.ownerGender) self.closetGUI.enter(base.localAvatar) self.closetGUI.showButtons() oldHat = self.av.getHat() oldGlasses = self.av.getGlasses() oldBackpack = self.av.getBackpack() oldShoes = self.av.getShoes() self.oldStyle = {ToonDNA.HAT: oldHat, ToonDNA.GLASSES: oldGlasses, ToonDNA.BACKPACK: oldBackpack, ToonDNA.SHOES: oldShoes} return Task.done def resetCloset(self): self.ignoreAll() taskMgr.remove(self.uniqueName('popupChangeClothesGUI')) taskMgr.remove(self.uniqueName('lerpCamera')) taskMgr.remove(self.uniqueName('lerpToon')) if self.closetGUI: self.closetGUI.hideButtons() self.closetGUI.exit() self.closetGUI.unload() self.closetGUI = None del self.av self.av = base.localAvatar oldHat = self.av.getHat() oldGlasses = self.av.getGlasses() oldBackpack = self.av.getBackpack() oldShoes = self.av.getShoes() self.oldStyle = {ToonDNA.HAT: oldHat, ToonDNA.GLASSES: oldGlasses, ToonDNA.BACKPACK: oldBackpack, ToonDNA.SHOES: oldShoes} self.hatDeleted = 0 self.glassesDeleted = 0 self.backpackDeleted = 0 self.shoesDeleted = 0 return Task.done def _handleCancel(self): if self.oldStyle: oldHat = self.oldStyle[ToonDNA.HAT] oldGlasses = self.oldStyle[ToonDNA.GLASSES] oldBackpack = self.oldStyle[ToonDNA.BACKPACK] oldShoes = self.oldStyle[ToonDNA.SHOES] self.d_setDNA(oldHat[0], oldHat[1], oldHat[2], oldGlasses[0], oldGlasses[1], oldGlasses[2], oldBackpack[0], oldBackpack[1], oldBackpack[2], oldShoes[0], oldShoes[1], oldShoes[2], 1) else: self.notify.info('avoided crash in handleCancel') self._handlePurchaseDone() if self.closetGUI: self.closetGUI.resetClothes(self.oldStyle) if self.popupInfo != None: self.popupInfo.destroy() self.popupInfo = None return def __handleSwapHat(self): item = self.av.getHat() self.d_setDNA(item[0], item[1], item[2], N_A, N_A, N_A, N_A, N_A, N_A, N_A, N_A, N_A, 0, ToonDNA.HAT) if self.closetGUI: self.closetGUI.updateTrashButtons() def __handleSwapGlasses(self): item = self.av.getGlasses() self.d_setDNA(N_A, N_A, N_A, item[0], item[1], item[2], N_A, N_A, N_A, N_A, N_A, N_A, 0, ToonDNA.GLASSES) if self.closetGUI: self.closetGUI.updateTrashButtons() def __handleSwapBackpack(self): item = self.av.getBackpack() self.d_setDNA(N_A, N_A, N_A, N_A, N_A, N_A, item[0], item[1], item[2], N_A, N_A, N_A, 0, ToonDNA.BACKPACK) if self.closetGUI: self.closetGUI.updateTrashButtons() def __handleSwapShoes(self): item = self.av.getShoes() self.d_setDNA(N_A, N_A, N_A, N_A, N_A, N_A, N_A, N_A, N_A, item[0], item[1], item[2], 0, ToonDNA.SHOES) if self.closetGUI: self.closetGUI.updateTrashButtons() def __handleDelete(self, which): if which == ToonDNA.HAT: itemList = self.closetGUI.hats trashIndex = self.closetGUI.hatChoice swapFunc = self.closetGUI.swapHat removeFunc = self.closetGUI.removeHat trashItem = self.av.getHat() self.hatDeleted = self.hatDeleted | 1 elif which == ToonDNA.GLASSES: itemList = self.closetGUI.glasses trashIndex = self.closetGUI.glassesChoice swapFunc = self.closetGUI.swapGlasses removeFunc = self.closetGUI.removeGlasses trashItem = self.av.getGlasses() self.glassesDeleted = self.glassesDeleted | 1 elif which == ToonDNA.BACKPACK: itemList = self.closetGUI.backpacks trashIndex = self.closetGUI.backpackChoice swapFunc = self.closetGUI.swapBackpack removeFunc = self.closetGUI.removeBackpack trashItem = self.av.getBackpack() self.backpackDeleted = self.backpackDeleted | 1 elif which == ToonDNA.SHOES: itemList = self.closetGUI.shoes trashIndex = self.closetGUI.shoesChoice swapFunc = self.closetGUI.swapShoes removeFunc = self.closetGUI.removeShoes trashItem = self.av.getShoes() self.shoesDeleted = self.shoesDeleted | 1 else: self.notify.warning("we don't know about this item(type = %s)" % which) return if len(itemList) > 1: if trashIndex == 0: swapFunc(1) else: swapFunc(-1) removeFunc(trashIndex) self.sendUpdate('removeItem', [trashItem[0], trashItem[1], trashItem[2], which]) swapFunc(0) self.closetGUI.updateTrashButtons() else: self.notify.warning("cant delete this item(type = %s), since we don't have a replacement" % which) def resetItemLists(self): self.hatList = self.oldHatList[0:] self.glassesList = self.oldGlassesList[0:] self.backpackList = self.oldBackpackList[0:] self.shoesList = self.oldShoesList[0:] self.closetGUI.hat = self.hatList self.closetGUI.glasses = self.glassesList self.closetGUI.backpack = self.backpackList self.closetGUI.shoes = self.shoesList self.hatDeleted = 0 self.glassesDeleted = 0 self.backpackDeleted = 0 self.shoesDeleted = 0 def __proceedToCheckout(self): if self.hatDeleted or self.glassesDeleted or self.backpackDeleted or self.shoesDeleted: self.__popupAreYouSurePanel() else: self._handlePurchaseDone() def _handlePurchaseDone(self, timeout=0): if timeout == 1: oldHat = self.oldStyle[ToonDNA.HAT] oldGlasses = self.oldStyle[ToonDNA.GLASSES] oldBackpack = self.oldStyle[ToonDNA.BACKPACK] oldShoes = self.oldStyle[ToonDNA.SHOES] self.d_setDNA(oldHat[0], oldHat[1], oldHat[2], oldGlasses[0], oldGlasses[1], oldGlasses[2], oldBackpack[0], oldBackpack[1], oldBackpack[2], oldShoes[0], oldShoes[1], oldShoes[2], 1) else: which = 0 if hasattr(self.closetGUI, 'hatChoice') and hasattr(self.closetGUI, 'glassesChoice') and hasattr(self.closetGUI, 'backpackChoice') and hasattr(self.closetGUI, 'shoesChoice'): if self.closetGUI.hatChoice != 0 or self.hatDeleted: which = which | ToonDNA.HAT if self.closetGUI.glassesChoice != 0 or self.glassesDeleted: which = which | ToonDNA.GLASSES if self.closetGUI.backpackChoice != 0 or self.backpackDeleted: which = which | ToonDNA.BACKPACK if self.closetGUI.shoesChoice != 0 or self.shoesDeleted: which = which | ToonDNA.SHOES hat = self.av.getHat() glasses = self.av.getGlasses() backpack = self.av.getBackpack() shoes = self.av.getShoes() self.d_setDNA(hat[0], hat[1], hat[2], glasses[0], glasses[1], glasses[2], backpack[0], backpack[1], backpack[2], shoes[0], shoes[1], shoes[2], 2, which) def d_setDNA(self, hatIdx, hatTexture, hatColor, glassesIdx, glassesTexture, glassesColor, backpackIdx, backpackTexture, backpackColor, shoesIdx, shoesTexture, shoesColor, finished, which=ToonDNA.HAT | ToonDNA.GLASSES | ToonDNA.BACKPACK | ToonDNA.SHOES): self.sendUpdate('setDNA', [hatIdx, hatTexture, hatColor, glassesIdx, glassesTexture, glassesColor, backpackIdx, backpackTexture, backpackColor, shoesIdx, shoesTexture, shoesColor, finished, which]) def setCustomerDNA(self, avId, hatIdx, hatTexture, hatColor, glassesIdx, glassesTexture, glassesColor, backpackIdx, backpackTexture, backpackColor, shoesIdx, shoesTexture, shoesColor, which): if avId and avId != base.localAvatar.doId: av = base.cr.doId2do.get(avId, None) if av: if self.av == base.cr.doId2do[avId]: if which & ToonDNA.HAT: self.av.setHat(hatIdx, hatTexture, hatColor) if which & ToonDNA.GLASSES: self.av.setGlasses(glassesIdx, glassesTexture, glassesColor) if which & ToonDNA.BACKPACK: self.av.setBackpack(backpackIdx, backpackTexture, backpackColor) if which & ToonDNA.SHOES: self.av.setShoes(shoesIdx, shoesTexture, shoesColor) self.av.generateToonAccessories() return def __popupNotOwnerPanel(self): if self.popupInfo != None: self.popupInfo.destroy() self.popupInfo = None self.purchaseDoneEvent = self.uniqueName('purchaseDone') self.swapHatEvent = self.uniqueName('swapHat') self.swapGlassesEvent = self.uniqueName('swapGlasses') self.swapBackpackEvent = self.uniqueName('swapBackpack') self.swapShoesEvent = self.uniqueName('swapShoes') self.cancelEvent = self.uniqueName('cancel') self.accept(self.purchaseDoneEvent, self.__proceedToCheckout) self.accept(self.swapHatEvent, self.__handleSwapHat) self.accept(self.swapGlassesEvent, self.__handleSwapGlasses) self.accept(self.swapBackpackEvent, self.__handleSwapBackpack) self.accept(self.swapShoesEvent, self.__handleSwapShoes) self.accept(self.cancelEvent, self._handleCancel) self.deleteEvent = self.uniqueName('delete') if self.isOwner: self.accept(self.deleteEvent, self.__handleDelete) buttons = loader.loadModel('phase_3/models/gui/dialog_box_buttons_gui') okButtonImage = (buttons.find('**/ChtBx_OKBtn_UP'), buttons.find('**/ChtBx_OKBtn_DN'), buttons.find('**/ChtBx_OKBtn_Rllvr')) if self.isFreePlayer: textMsg = TTLocalizer.TrunkNotPaidMessage else: textMsg = TTLocalizer.TrunkNotOwnerMessage self.popupInfo = DirectFrame(parent=hidden, relief=None, state='normal', text=textMsg, frameSize=(-1, 1, -1, 1), text_wordwrap=10, geom=DGG.getDefaultDialogGeom(), geom_color=ToontownGlobals.GlobalDialogColor, geom_scale=(0.88, 1, 0.55), geom_pos=(0, 0, -0.08), text_scale=0.08, text_pos=(0, 0.06)) DirectButton(self.popupInfo, image=okButtonImage, relief=None, text=TTLocalizer.ClosetPopupOK, text_scale=0.05, text_pos=(0.0, -0.1), textMayChange=0, pos=(0.0, 0.0, -0.21), command=self._handleNotOwnerMessageOK) buttons.removeNode() self.popupInfo.reparentTo(aspect2d) return def __popupAreYouSurePanel(self): if self.popupInfo != None: self.popupInfo.destroy() self.popupInfo = None buttons = loader.loadModel('phase_3/models/gui/dialog_box_buttons_gui') okButtonImage = (buttons.find('**/ChtBx_OKBtn_UP'), buttons.find('**/ChtBx_OKBtn_DN'), buttons.find('**/ChtBx_OKBtn_Rllvr')) cancelButtonImage = (buttons.find('**/CloseBtn_UP'), buttons.find('**/CloseBtn_DN'), buttons.find('**/CloseBtn_Rllvr')) self.popupInfo = DirectFrame(parent=hidden, relief=None, state='normal', text=TTLocalizer.TrunkAreYouSureMessage, frameSize=(-1, 1, -1, 1), text_wordwrap=10, geom=DGG.getDefaultDialogGeom(), geom_color=ToontownGlobals.GlobalDialogColor, geom_scale=(0.88, 1, 0.55), geom_pos=(0, 0, -0.08), text_scale=0.08, text_pos=(0, 0.08)) DirectButton(self.popupInfo, image=okButtonImage, relief=None, text=TTLocalizer.ClosetPopupOK, text_scale=0.05, text_pos=(0.0, -0.1), textMayChange=0, pos=(-0.1, 0.0, -0.21), command=self._handleYesImSure) DirectButton(self.popupInfo, image=cancelButtonImage, relief=None, text=TTLocalizer.ClosetPopupCancel, text_scale=0.05, text_pos=(0.0, -0.1), textMayChange=0, pos=(0.1, 0.0, -0.21), command=self._handleNotSure) buttons.removeNode() self.popupInfo.reparentTo(aspect2d) return def _openDoors(self): if self.closetTrack: self.closetTrack.finish() openHpr = Vec3(0, -80, 0) if self.av: self.av.applyCheesyEffect(ToontownGlobals.CENormal) self.closetTrack = Parallel() if self.lid: self.closetTrack.append(self.lid.hprInterval(0.5, openHpr)) self.closetTrack.start() def _closeDoors(self): if self.closetTrack: self.closetTrack.finish() closeHpr = Vec3(0, 0, 0) if self.av: self.av.reconsiderCheesyEffect() self.closetTrack = Parallel() if self.lid: self.closetTrack.append(self.lid.hprInterval(0.5, closeHpr)) self.closetTrack.start()
py
1a2f5181a9936e57ec9cd7a2a19c2fae6dc7cd65
from django.test import TestCase from django.contrib.auth import get_user_model class ModelTests(TestCase): def test_create_user_with_email_successful(self): email = 'kamransadixov@mail.ru' password = 'kam125486' """ get_user_model() returns the default user model. If it has to return custom usermodel, create a new model and in settings.py Set AUTH_USER_MODEL pointing to custom model """ user = get_user_model().objects.create_user( email=email, password=password ) self.assertEqual(user.email, email) self.assertTrue(user.check_password(password)) def test_new_user_email_normalize(self): email = 'tets@SADASD.com' user = get_user_model().objects.create_user(email, 'kam125486') self.assertEqual(user.email, email.lower()) def test_new_user_invalid_email(self): """Test email user with no email raises error""" with self.assertRaises(ValueError): get_user_model().objects.create_user(None, 'kam125486') def test_create_new_superuser(self): user = get_user_model().objects.create_superuser( 'kamransadikhov@yandex.com', 'kam125486' ) self.assertTrue(user.is_superuser) self.assertTrue(user.is_staff)
py
1a2f51ef1dc8277bf84dd1bc84b6dd06be963ddc
import os.path as osp import time import joblib import numpy as np import tensorflow as tf from baselines import logger from collections import deque from baselines.common import set_global_seeds, explained_variance from baselines.common.runners import AbstractEnvRunner from baselines.common import tf_util from baselines.a2c.utils import discount_with_dones from baselines.a2c.utils import Scheduler, make_path, find_trainable_variables from baselines.a2c.utils import cat_entropy, mse class Model(object): def __init__(self, policy, ob_space, ac_space, nenvs, nsteps, ent_coef=0.01, vf_coef=0.5, max_grad_norm=0.5, lr=7e-4, alpha=0.99, epsilon=1e-5, total_timesteps=int(80e6), lrschedule='linear'): sess = tf_util.make_session() nbatch = nenvs*nsteps A = tf.placeholder(tf.int32, [nbatch]) ADV = tf.placeholder(tf.float32, [nbatch]) R = tf.placeholder(tf.float32, [nbatch]) LR = tf.placeholder(tf.float32, []) step_model = policy(sess, ob_space, ac_space, nenvs, 1, reuse=False) train_model = policy(sess, ob_space, ac_space, nenvs*nsteps, nsteps, reuse=True) neglogpac = tf.nn.sparse_softmax_cross_entropy_with_logits(logits=train_model.pi, labels=A) pg_loss = tf.reduce_mean(ADV * neglogpac) vf_loss = tf.reduce_mean(mse(tf.squeeze(train_model.vf), R)) entropy = tf.reduce_mean(cat_entropy(train_model.pi)) loss = pg_loss - entropy*ent_coef + vf_loss * vf_coef params = find_trainable_variables("model") grads = tf.gradients(loss, params) if max_grad_norm is not None: grads, grad_norm = tf.clip_by_global_norm(grads, max_grad_norm) grads = list(zip(grads, params)) trainer = tf.train.RMSPropOptimizer(learning_rate=LR, decay=alpha, epsilon=epsilon) _train = trainer.apply_gradients(grads) lr = Scheduler(v=lr, nvalues=total_timesteps, schedule=lrschedule) def train(obs, states, rewards, masks, actions, values): advs = rewards - values for step in range(len(obs)): cur_lr = lr.value() td_map = {train_model.X:obs, A:actions, ADV:advs, R:rewards, LR:cur_lr} if states is not None: td_map[train_model.S] = states td_map[train_model.M] = masks policy_loss, value_loss, policy_entropy, _ = sess.run( [pg_loss, vf_loss, entropy, _train], td_map ) return policy_loss, value_loss, policy_entropy def save(save_path): ps = sess.run(params) make_path(osp.dirname(save_path)) joblib.dump(ps, save_path) def load(load_path): loaded_params = joblib.load(load_path) restores = [] for p, loaded_p in zip(params, loaded_params): restores.append(p.assign(loaded_p)) sess.run(restores) self.train = train self.train_model = train_model self.step_model = step_model self.step = step_model.step self.value = step_model.value self.initial_state = step_model.initial_state self.save = save self.load = load tf.global_variables_initializer().run(session=sess) class Runner(AbstractEnvRunner): def __init__(self, env, model, nsteps=5, gamma=0.99): super().__init__(env=env, model=model, nsteps=nsteps) self.gamma = gamma self.episodes_count = 0 def run(self): mb_obs, mb_rewards, mb_actions, mb_values, mb_dones = [],[],[],[],[] mb_states = self.states epinfos = [] for n in range(self.nsteps): actions, values, states, _ = self.model.step(self.obs, self.states, self.dones) mb_obs.append(np.copy(self.obs)) mb_actions.append(actions) mb_values.append(values) mb_dones.append(self.dones) obs, rewards, dones, infos = self.env.step(actions) for info in infos: maybeepinfo = info.get('episode') if maybeepinfo: self.episodes_count += 1 epinfos.append(maybeepinfo) self.states = states self.dones = dones for n, done in enumerate(dones): if done: self.obs[n] = self.obs[n]*0 self.obs = obs mb_rewards.append(rewards) mb_dones.append(self.dones) #batch of steps to batch of rollouts mb_obs = np.asarray(mb_obs, dtype=np.uint8).swapaxes(1, 0).reshape(self.batch_ob_shape) mb_rewards = np.asarray(mb_rewards, dtype=np.float32).swapaxes(1, 0) mb_actions = np.asarray(mb_actions, dtype=np.int32).swapaxes(1, 0) mb_values = np.asarray(mb_values, dtype=np.float32).swapaxes(1, 0) mb_dones = np.asarray(mb_dones, dtype=np.bool).swapaxes(1, 0) mb_masks = mb_dones[:, :-1] mb_dones = mb_dones[:, 1:] last_values = self.model.value(self.obs, self.states, self.dones).tolist() #discount/bootstrap off value fn for n, (rewards, dones, value) in enumerate(zip(mb_rewards, mb_dones, last_values)): rewards = rewards.tolist() dones = dones.tolist() if dones[-1] == 0: rewards = discount_with_dones(rewards+[value], dones+[0], self.gamma)[:-1] else: rewards = discount_with_dones(rewards, dones, self.gamma) mb_rewards[n] = rewards mb_rewards = mb_rewards.flatten() mb_actions = mb_actions.flatten() mb_values = mb_values.flatten() mb_masks = mb_masks.flatten() return mb_obs, mb_states, mb_rewards, mb_masks, mb_actions, mb_values, epinfos def learn(policy, env, seed, nsteps=5, total_timesteps=int(80e6), vf_coef=0.5, ent_coef=0.01, max_grad_norm=0.5, lr=7e-4, lrschedule='linear', epsilon=1e-5, alpha=0.99, gamma=0.99, log_interval=100): set_global_seeds(seed) nenvs = env.num_envs ob_space = env.observation_space ac_space = env.action_space model = Model(policy=policy, ob_space=ob_space, ac_space=ac_space, nenvs=nenvs, nsteps=nsteps, ent_coef=ent_coef, vf_coef=vf_coef, max_grad_norm=max_grad_norm, lr=lr, alpha=alpha, epsilon=epsilon, total_timesteps=total_timesteps, lrschedule=lrschedule) runner = Runner(env, model, nsteps=nsteps, gamma=gamma) epinfobuf = deque(maxlen=100) nbatch = nenvs*nsteps tstart = time.time() for update in range(1, total_timesteps//nbatch+1): obs, states, rewards, masks, actions, values, epinfos = runner.run() epinfobuf.extend(epinfos) policy_loss, value_loss, policy_entropy = model.train(obs, states, rewards, masks, actions, values) nseconds = time.time()-tstart fps = int((update*nbatch)/nseconds) if update % log_interval == 0 or update == 1: ev = explained_variance(values, rewards) logger.record_tabular("nupdates", update) logger.record_tabular("total_timesteps", update*nbatch) logger.record_tabular("fps", fps) logger.logkv('eprewmean', safemean([epinfo['r'] for epinfo in epinfobuf])) logger.logkv('eplenmean', safemean([epinfo['l'] for epinfo in epinfobuf])) logger.record_tabular("policy_entropy", float(policy_entropy)) logger.record_tabular("value_loss", float(value_loss)) logger.record_tabular("explained_variance", float(ev)) logger.logkv('time_elapsed', nseconds) logger.dump_tabular() logger.logkv('episodes', runner.episodes_count) env.close() return model def safemean(xs): return np.nan if len(xs) == 0 else np.mean(xs)
py
1a2f526a0b6bfaf86b7b7d940e635a431f8a6643
# coding: utf-8 """ Octopus Server API No description provided (generated by Swagger Codegen https://github.com/swagger-api/swagger-codegen) # noqa: E501 OpenAPI spec version: 2019.6.7+Branch.tags-2019.6.7.Sha.aa18dc6809953218c66f57eff7d26481d9b23d6a Generated by: https://github.com/swagger-api/swagger-codegen.git """ from __future__ import absolute_import import unittest import octopus_deploy_swagger_client from octopus_deploy_swagger_client.models.display_info import DisplayInfo # noqa: E501 from octopus_deploy_swagger_client.rest import ApiException class TestDisplayInfo(unittest.TestCase): """DisplayInfo unit test stubs""" def setUp(self): pass def tearDown(self): pass def testDisplayInfo(self): """Test DisplayInfo""" # FIXME: construct object with mandatory attributes with example values # model = octopus_deploy_swagger_client.models.display_info.DisplayInfo() # noqa: E501 pass if __name__ == '__main__': unittest.main()
py
1a2f52e30a61dbd98a07f21b4a7c3f8256082840
# Copyright 2018 The Chromium OS Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. from cros.factory.probe import function class ActionFunction(function.Function): """The base class of action functions. While evaluation, an action function executes a side-effect action. If the action is successfully executed, it returns the input data. Otherwise it returns an empty list to notify the computation failed. """ def Apply(self, data): if self.Action(): return data return function.NOTHING def Action(self): """Execute an action and return the action is successfully or not.""" raise NotImplementedError
py
1a2f5306e653c319496dc7b1f573882ee48345f4
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HugginFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # Copyright (c) 2019, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """PyTorch BERT model.""" from __future__ import absolute_import, division, print_function, unicode_literals import os import copy import json import math import logging import tarfile import tempfile import shutil import torch from torch import nn import torch.nn.functional as F from torch.nn import CrossEntropyLoss from torch.utils.checkpoint import checkpoint from olfmlm.data_utils.file_utils import cached_path logger = logging.getLogger(__name__) PRETRAINED_MODEL_ARCHIVE_MAP = { 'bert-base-uncased': "https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-uncased.tar.gz", 'bert-large-uncased': "https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-uncased.tar.gz", 'bert-base-cased': "https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-cased.tar.gz", 'bert-large-cased': "https://s3.amazonaws.com/models.huggingface.co/bert/bert-large-cased.tar.gz", 'bert-base-multilingual-uncased': "https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-multilingual-uncased.tar.gz", 'bert-base-multilingual-cased': "https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-multilingual-cased.tar.gz", 'bert-base-chinese': "https://s3.amazonaws.com/models.huggingface.co/bert/bert-base-chinese.tar.gz", } CONFIG_NAME = 'bert_config.json' WEIGHTS_NAME = 'pytorch_model.bin' TF_WEIGHTS_NAME = 'model.ckpt' def load_tf_weights_in_bert(model, tf_checkpoint_path): """ Load tf checkpoints in a pytorch model """ try: import re import numpy as np import tensorflow as tf except ImportError: print("Loading a TensorFlow models in PyTorch, requires TensorFlow to be installed. Please see " "https://www.tensorflow.org/install/ for installation instructions.") raise tf_path = os.path.abspath(tf_checkpoint_path) print("Converting TensorFlow checkpoint from {}".format(tf_path)) # Load weights from TF model init_vars = tf.train.list_variables(tf_path) names = [] arrays = [] for name, shape in init_vars: print("Loading TF weight {} with shape {}".format(name, shape)) array = tf.train.load_variable(tf_path, name) names.append(name) arrays.append(array) for name, array in zip(names, arrays): name = name.split('/') # adam_v and adam_m are variables used in AdamWeightDecayOptimizer to calculated m and v # which are not required for using pretrained model if any(n in ["adam_v", "adam_m"] for n in name): print("Skipping {}".format("/".join(name))) continue pointer = model for m_name in name: if re.fullmatch(r'[A-Za-z]+_\d+', m_name): l = re.split(r'_(\d+)', m_name) else: l = [m_name] if l[0] == 'kernel' or l[0] == 'gamma': pointer = getattr(pointer, 'weight') elif l[0] == 'output_bias' or l[0] == 'beta': pointer = getattr(pointer, 'bias') elif l[0] == 'output_weights': pointer = getattr(pointer, 'weight') else: pointer = getattr(pointer, l[0]) if len(l) >= 2: num = int(l[1]) pointer = pointer[num] if m_name[-11:] == '_embeddings': pointer = getattr(pointer, 'weight') elif m_name == 'kernel': array = np.transpose(array) try: assert pointer.shape == array.shape except AssertionError as e: e.args += (pointer.shape, array.shape) raise print("Initialize PyTorch weight {}".format(name)) pointer.data = torch.from_numpy(array) return model def gelu(x): """Implementation of the gelu activation function. For information: OpenAI GPT's gelu is slightly different (and gives slightly different results): 0.5 * x * (1 + torch.tanh(math.sqrt(2 / math.pi) * (x + 0.044715 * torch.pow(x, 3)))) """ return x * 0.5 * (1.0 + torch.erf(x / math.sqrt(2.0))) def swish(x): return x * torch.sigmoid(x) ACT2FN = {"gelu": gelu, "relu": torch.nn.functional.relu, "swish": swish} class BertConfig(object): """Configuration class to store the configuration of a `BertModel`. """ def __init__(self, vocab_size_or_config_json_file, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072, hidden_act="gelu", hidden_dropout_prob=0.1, attention_probs_dropout_prob=0.1, max_position_embeddings=512, type_vocab_size=2, initializer_range=0.02, fp32_layernorm=True, fp32_embedding=True, fp32_tokentypes=False, layernorm_epsilon=1e-12): """Constructs BertConfig. Args: vocab_size_or_config_json_file: Vocabulary size of `inputs_ids` in `BertModel`. hidden_size: Size of the encoder layers and the pooler layer. num_hidden_layers: Number of hidden layers in the Transformer encoder. num_attention_heads: Number of attention heads for each attention layer in the Transformer encoder. intermediate_size: The size of the "intermediate" (i.e., feed-forward) layer in the Transformer encoder. hidden_act: The non-linear activation function (function or string) in the encoder and pooler. If string, "gelu", "relu" and "swish" are supported. hidden_dropout_prob: The dropout probabilitiy for all fully connected layers in the embeddings, encoder, and pooler. attention_probs_dropout_prob: The dropout ratio for the attention probabilities. max_position_embeddings: The maximum sequence length that this model might ever be used with. Typically set this to something large just in case (e.g., 512 or 1024 or 2048). type_vocab_size: The vocabulary size of the `token_type_ids` passed into `BertModel`. initializer_range: The sttdev of the truncated_normal_initializer for initializing all weight matrices. """ if isinstance(vocab_size_or_config_json_file, str): with open(vocab_size_or_config_json_file, "r", encoding='utf-8') as reader: json_config = json.loads(reader.read()) for key, value in json_config.items(): self.__dict__[key] = value elif isinstance(vocab_size_or_config_json_file, int): self.vocab_size = vocab_size_or_config_json_file self.hidden_size = hidden_size self.num_hidden_layers = num_hidden_layers self.num_attention_heads = num_attention_heads self.hidden_act = hidden_act self.intermediate_size = intermediate_size self.hidden_dropout_prob = hidden_dropout_prob self.attention_probs_dropout_prob = attention_probs_dropout_prob self.max_position_embeddings = max_position_embeddings self.type_vocab_size = type_vocab_size self.initializer_range = initializer_range self.fp32_layernorm = fp32_layernorm self.fp32_embedding = fp32_embedding self.layernorm_epsilon = layernorm_epsilon self.fp32_tokentypes = fp32_tokentypes else: raise ValueError("First argument must be either a vocabulary size (int)" "or the path to a pretrained model config file (str)") @classmethod def from_dict(cls, json_object): """Constructs a `BertConfig` from a Python dictionary of parameters.""" config = BertConfig(vocab_size_or_config_json_file=-1) for key, value in json_object.items(): config.__dict__[key] = value return config @classmethod def from_json_file(cls, json_file): """Constructs a `BertConfig` from a json file of parameters.""" with open(json_file, "r", encoding='utf-8') as reader: text = reader.read() return cls.from_dict(json.loads(text)) def __repr__(self): return str(self.to_json_string()) def to_dict(self): """Serializes this instance to a Python dictionary.""" output = copy.deepcopy(self.__dict__) return output def to_json_string(self): """Serializes this instance to a JSON string.""" return json.dumps(self.to_dict(), indent=2, sort_keys=True) + "\n" # try: # from apex.normalization.fused_layer_norm import FusedLayerNorm as BertLayerNorm # except ImportError: # print("Better speed can be achieved with apex installed from https://www.github.com/nvidia/apex.") # class BertLayerNorm(nn.Module): # def __init__(self, hidden_size, eps=1e-12): # """Construct a layernorm module in the TF style (epsilon inside the square root). # """ # super(BertLayerNorm, self).__init__() # self.weight = nn.Parameter(torch.ones(hidden_size)) # self.bias = nn.Parameter(torch.zeros(hidden_size)) # self.variance_epsilon = eps # def forward(self, x): # u = x.mean(-1, keepdim=True) # s = (x - u).pow(2).mean(-1, keepdim=True) # x = (x - u) / torch.sqrt(s + self.variance_epsilon) # return self.weight * x + self.bias class BertLayerNorm(nn.Module): def __init__(self, hidden_size, eps=1e-12): """Construct a layernorm module in the TF style (epsilon inside the square root). """ super(BertLayerNorm, self).__init__() self.weight = nn.Parameter(torch.ones(hidden_size)) self.bias = nn.Parameter(torch.zeros(hidden_size)) self.variance_epsilon = eps def forward(self, x): u = x.mean(-1, keepdim=True) s = (x - u).pow(2).mean(-1, keepdim=True) x = (x - u) / torch.sqrt(s + self.variance_epsilon) return self.weight * x + self.bias class BertEmbeddings(nn.Module): """Construct the embeddings from word, position and token_type embeddings. """ def __init__(self, config): super(BertEmbeddings, self).__init__() self.word_embeddings = nn.Embedding(config.vocab_size, config.hidden_size, padding_idx=0) self.position_embeddings = nn.Embedding(config.max_position_embeddings, config.hidden_size) self.token_type_embeddings = nn.Embedding(config.type_vocab_size, config.hidden_size) self.task_type_embeddings = nn.Embedding(config.num_tasks, config.hidden_size) # self.LayerNorm is not snake-cased to stick with TensorFlow model variable name and be able to load # any TensorFlow checkpoint file self.LayerNorm = BertLayerNorm(config.hidden_size, eps=config.layer_norm_eps) self.dropout = nn.Dropout(config.hidden_dropout_prob) def forward(self, input_ids, token_type_ids=None, task_ids=None): seq_length = input_ids.size(1) position_ids = torch.arange(seq_length, dtype=torch.long, device=input_ids.device) position_ids = position_ids.unsqueeze(0).expand_as(input_ids) if token_type_ids is None: token_type_ids = torch.zeros_like(input_ids) if task_ids is None: task_ids = torch.zeros_like(input_ids) words_embeddings = self.word_embeddings(input_ids) position_embeddings = self.position_embeddings(position_ids) token_type_embeddings = self.token_type_embeddings(token_type_ids) task_embeddings = self.task_type_embeddings(task_ids) embeddings = words_embeddings + position_embeddings + token_type_embeddings + task_embeddings embeddings = self.LayerNorm(embeddings) embeddings = self.dropout(embeddings) return embeddings class BertSelfAttention(nn.Module): def __init__(self, config): super(BertSelfAttention, self).__init__() if config.hidden_size % config.num_attention_heads != 0: raise ValueError( "The hidden size (%d) is not a multiple of the number of attention " "heads (%d)" % (config.hidden_size, config.num_attention_heads)) self.num_attention_heads = config.num_attention_heads self.attention_head_size = int(config.hidden_size / config.num_attention_heads) self.all_head_size = self.num_attention_heads * self.attention_head_size self.query = nn.Linear(config.hidden_size, self.all_head_size) self.key = nn.Linear(config.hidden_size, self.all_head_size) self.value = nn.Linear(config.hidden_size, self.all_head_size) self.dropout = nn.Dropout(config.attention_probs_dropout_prob) def transpose_for_scores(self, x): new_x_shape = x.size()[:-1] + (self.num_attention_heads, self.attention_head_size) x = x.view(*new_x_shape) return x.permute(0, 2, 1, 3) def forward(self, hidden_states, attention_mask): mixed_query_layer = self.query(hidden_states) mixed_key_layer = self.key(hidden_states) mixed_value_layer = self.value(hidden_states) query_layer = self.transpose_for_scores(mixed_query_layer) key_layer = self.transpose_for_scores(mixed_key_layer) value_layer = self.transpose_for_scores(mixed_value_layer) # Take the dot product between "query" and "key" to get the raw attention scores. attention_scores = torch.matmul(query_layer, key_layer.transpose(-1, -2)) attention_scores = attention_scores / math.sqrt(self.attention_head_size) # Apply the attention mask is (precomputed for all layers in BertModel forward() function) attention_scores = attention_scores + attention_mask # Normalize the attention scores to probabilities. attention_probs = nn.Softmax(dim=-1)(attention_scores) # This is actually dropping out entire tokens to attend to, which might # seem a bit unusual, but is taken from the original Transformer paper. attention_probs = self.dropout(attention_probs) previous_type = attention_probs.type() context_layer = torch.matmul(attention_probs, value_layer) context_layer = context_layer.permute(0, 2, 1, 3).contiguous() new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) context_layer = context_layer.view(*new_context_layer_shape) return context_layer class BertSelfOutput(nn.Module): def __init__(self, config): super(BertSelfOutput, self).__init__() self.dense = nn.Linear(config.hidden_size, config.hidden_size) self.fp32_layernorm = config.fp32_layernorm self.LayerNorm = BertLayerNorm(config.hidden_size, eps=config.layernorm_epsilon) self.dropout = nn.Dropout(config.hidden_dropout_prob) def forward(self, hidden_states, input_tensor): hidden_states = self.dense(hidden_states) hidden_states = self.dropout(hidden_states) ln_input = hidden_states + input_tensor previous_type = ln_input.type() if self.fp32_layernorm: ln_input = ln_input.float() hidden_states = self.LayerNorm(ln_input) if self.fp32_layernorm: hidden_states = hidden_states.type(previous_type) return hidden_states class BertAttention(nn.Module): def __init__(self, config): super(BertAttention, self).__init__() self.self = BertSelfAttention(config) self.output = BertSelfOutput(config) def forward(self, input_tensor, attention_mask): self_output = self.self(input_tensor, attention_mask) attention_output = self.output(self_output, input_tensor) return attention_output class BertIntermediate(nn.Module): def __init__(self, config): super(BertIntermediate, self).__init__() self.dense = nn.Linear(config.hidden_size, config.intermediate_size) self.intermediate_act_fn = ACT2FN[config.hidden_act] \ if isinstance(config.hidden_act, str) else config.hidden_act def forward(self, hidden_states): hidden_states = self.dense(hidden_states) hidden_states = self.intermediate_act_fn(hidden_states) return hidden_states class BertOutput(nn.Module): def __init__(self, config): super(BertOutput, self).__init__() self.dense = nn.Linear(config.intermediate_size, config.hidden_size) self.fp32_layernorm = config.fp32_layernorm self.LayerNorm = BertLayerNorm(config.hidden_size, eps=config.layernorm_epsilon) self.dropout = nn.Dropout(config.hidden_dropout_prob) def forward(self, hidden_states, input_tensor): hidden_states = self.dense(hidden_states) hidden_states = self.dropout(hidden_states) ln_input = hidden_states + input_tensor previous_type = ln_input.type() if self.fp32_layernorm: ln_input = ln_input.float() hidden_states = self.LayerNorm(ln_input) if self.fp32_layernorm: hidden_states = hidden_states.type(previous_type) return hidden_states class BertLayer(nn.Module): def __init__(self, config): super(BertLayer, self).__init__() self.attention = BertAttention(config) self.intermediate = BertIntermediate(config) self.output = BertOutput(config) def forward(self, hidden_states, attention_mask): attention_output = self.attention(hidden_states, attention_mask) intermediate_output = self.intermediate(attention_output) layer_output = self.output(intermediate_output, attention_output) return layer_output class BertEncoder(nn.Module): def __init__(self, config): super(BertEncoder, self).__init__() layer = BertLayer(config) self.layer = nn.ModuleList([copy.deepcopy(layer) for _ in range(config.num_hidden_layers)]) # def forward(self, hidden_states, attention_mask, output_all_encoded_layers=True): # all_encoder_layers = [] # for layer_module in self.layer: # hidden_states = layer_module(hidden_states, attention_mask) # if output_all_encoded_layers: # all_encoder_layers.append(hidden_states) # if not output_all_encoded_layers: # all_encoder_layers.append(hidden_states) # return all_encoder_layers def forward(self, hidden_states, attention_mask, output_all_encoded_layers=True, checkpoint_activations=False): all_encoder_layers = [] def custom(start, end): def custom_forward(*inputs): layers = self.layer[start:end] x_ = inputs[0] for layer in layers: x_ = layer(x_, inputs[1]) return x_ return custom_forward if checkpoint_activations: l = 0 num_layers = len(self.layer) chunk_length = math.ceil(math.sqrt(num_layers)) while l < num_layers: hidden_states = checkpoint(custom(l, l+chunk_length), hidden_states, attention_mask*1) l += chunk_length # decoder layers else: for i,layer_module in enumerate(self.layer): hidden_states = layer_module(hidden_states, attention_mask) if output_all_encoded_layers: all_encoder_layers.append(hidden_states) if not output_all_encoded_layers or checkpoint_activations: all_encoder_layers.append(hidden_states) return all_encoder_layers class BertPooler(nn.Module): def __init__(self, config): super(BertPooler, self).__init__() self.dense = nn.Linear(config.hidden_size, config.hidden_size) self.activation = nn.Tanh() def forward(self, hidden_states): # We "pool" the model by simply taking the hidden state corresponding # to the first token. first_token_tensor = hidden_states[:, 0] pooled_output = self.dense(first_token_tensor) pooled_output = self.activation(pooled_output) return pooled_output class BertPredictionHeadTransform(nn.Module): def __init__(self, config): super(BertPredictionHeadTransform, self).__init__() self.dense = nn.Linear(config.hidden_size, config.hidden_size) self.transform_act_fn = ACT2FN[config.hidden_act] \ if isinstance(config.hidden_act, str) else config.hidden_act self.LayerNorm = BertLayerNorm(config.hidden_size, eps=config.layernorm_epsilon) self.fp32_layernorm = config.fp32_layernorm def forward(self, hidden_states): hidden_states = self.dense(hidden_states) hidden_states = self.transform_act_fn(hidden_states) previous_type = hidden_states.type() if self.fp32_layernorm: hidden_states = hidden_states.float() hidden_states = self.LayerNorm(hidden_states) if self.fp32_layernorm: hidden_states = hidden_states.type(previous_type) return hidden_states class BertLMPredictionHead(nn.Module): def __init__(self, config, bert_model_embedding_weights): super(BertLMPredictionHead, self).__init__() self.transform = BertPredictionHeadTransform(config) # The output weights are the same as the input embeddings, but there is # an output-only bias for each token. self.decoder = nn.Linear(bert_model_embedding_weights.size(1), bert_model_embedding_weights.size(0), bias=False) self.decoder.weight = bert_model_embedding_weights self.bias = nn.Parameter(torch.zeros(bert_model_embedding_weights.size(0))) def forward(self, hidden_states): hidden_states = self.transform(hidden_states) hidden_states = self.decoder(hidden_states) + self.bias return hidden_states class BertOnlyMLMHead(nn.Module): def __init__(self, config, bert_model_embedding_weights): super(BertOnlyMLMHead, self).__init__() self.predictions = BertLMPredictionHead(config, bert_model_embedding_weights) def forward(self, sequence_output): prediction_scores = self.predictions(sequence_output) return prediction_scores class BertOnlyNSPHead(nn.Module): def __init__(self, config): super(BertOnlyNSPHead, self).__init__() self.seq_relationship = nn.Linear(config.hidden_size, 2) def forward(self, pooled_output): seq_relationship_score = self.seq_relationship(pooled_output) return seq_relationship_score class BertPreTrainingHeads(nn.Module): def __init__(self, config, bert_model_embedding_weights): super(BertPreTrainingHeads, self).__init__() self.predictions = BertLMPredictionHead(config, bert_model_embedding_weights) self.seq_relationship = nn.Linear(config.hidden_size, 2) def forward(self, sequence_output, pooled_output): prediction_scores = self.predictions(sequence_output) for p in self.seq_relationship.parameters(): if p is None: continue pooled_output = pooled_output.type_as(p) seq_relationship_score = self.seq_relationship(pooled_output) return prediction_scores, seq_relationship_score class PreTrainedBertModel(nn.Module): """ An abstract class to handle weights initialization and a simple interface for dowloading and loading pretrained models. """ def __init__(self, config, *inputs, **kwargs): super(PreTrainedBertModel, self).__init__() if not isinstance(config, BertConfig): raise ValueError( "Parameter config in `{}(config)` should be an instance of class `BertConfig`. " "To create a model from a Google pretrained model use " "`model = {}.from_pretrained(PRETRAINED_MODEL_NAME)`".format( self.__class__.__name__, self.__class__.__name__ )) self.config = config def init_bert_weights(self, module): """ Initialize the weights. """ if isinstance(module, (nn.Linear, nn.Embedding)): # Slightly different from the TF version which uses truncated_normal for initialization # cf https://github.com/pytorch/pytorch/pull/5617 module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) elif isinstance(module, BertLayerNorm): module.bias.data.zero_() module.weight.data.fill_(1.0) if isinstance(module, nn.Linear) and module.bias is not None: module.bias.data.zero_() @classmethod def from_pretrained(cls, pretrained_model_name, state_dict=None, cache_dir=None, fp32_layernorm=False, fp32_embedding=False, layernorm_epsilon=1e-12, fp32_tokentypes=False, *inputs, **kwargs): """ Instantiate a PreTrainedBertModel from a pre-trained model file or a pytorch state dict. Download and cache the pre-trained model file if needed. Params: pretrained_model_name: either: - a str with the name of a pre-trained model to load selected in the list of: . `bert-base-uncased` . `bert-large-uncased` . `bert-base-cased` . `bert-large-cased` . `bert-base-multilingual-uncased` . `bert-base-multilingual-cased` . `bert-base-chinese` - a path or url to a pretrained model archive containing: . `bert_config.json` a configuration file for the model . `pytorch_model.bin` a PyTorch dump of a BertForPreTraining instance cache_dir: an optional path to a folder in which the pre-trained models will be cached. state_dict: an optional state dictionnary (collections.OrderedDict object) to use instead of Google pre-trained models *inputs, **kwargs: additional input for the specific Bert class (ex: num_labels for BertForSequenceClassification) """ if pretrained_model_name in PRETRAINED_MODEL_ARCHIVE_MAP: archive_file = PRETRAINED_MODEL_ARCHIVE_MAP[pretrained_model_name] else: archive_file = pretrained_model_name # redirect to the cache, if necessary try: resolved_archive_file = cached_path(archive_file, cache_dir=cache_dir) except FileNotFoundError: logger.error( "Model name '{}' was not found in model name list ({}). " "We assumed '{}' was a path or url but couldn't find any file " "associated to this path or url.".format( pretrained_model_name, ', '.join(PRETRAINED_MODEL_ARCHIVE_MAP.keys()), archive_file)) return None if resolved_archive_file == archive_file: logger.info("loading archive file {}".format(archive_file)) else: logger.info("loading archive file {} from cache at {}".format( archive_file, resolved_archive_file)) tempdir = None if os.path.isdir(resolved_archive_file): serialization_dir = resolved_archive_file else: # Extract archive to temp dir tempdir = tempfile.mkdtemp() logger.info("extracting archive file {} to temp dir {}".format( resolved_archive_file, tempdir)) with tarfile.open(resolved_archive_file, 'r:gz') as archive: archive.extractall(tempdir) serialization_dir = tempdir # Load config config_file = os.path.join(serialization_dir, CONFIG_NAME) config = BertConfig.from_json_file(config_file) config.fp32_layernorm = fp32_layernorm config.fp32_embedding = fp32_embedding config.layernorm_epsilon = layernorm_epsilon config.fp32_tokentypes = fp32_tokentypes logger.info("Model config {}".format(config)) # Instantiate model. model = cls(config, *inputs, **kwargs) if state_dict is None: weights_path = os.path.join(serialization_dir, WEIGHTS_NAME) state_dict = torch.load(weights_path) old_keys = [] new_keys = [] for key in state_dict.keys(): new_key = None if 'gamma' in key: new_key = key.replace('gamma', 'weight') if 'beta' in key: new_key = key.replace('beta', 'bias') if new_key: old_keys.append(key) new_keys.append(new_key) for old_key, new_key in zip(old_keys, new_keys): state_dict[new_key] = state_dict.pop(old_key) missing_keys = [] unexpected_keys = [] error_msgs = [] # copy state_dict so _load_from_state_dict can modify it metadata = getattr(state_dict, '_metadata', None) state_dict = state_dict.copy() if metadata is not None: state_dict._metadata = metadata def load(module, prefix=''): local_metadata = {} if metadata is None else metadata.get(prefix[:-1], {}) module._load_from_state_dict( state_dict, prefix, local_metadata, True, missing_keys, unexpected_keys, error_msgs) for name, child in module._modules.items(): if child is not None: load(child, prefix + name + '.') load(model, prefix='' if hasattr(model, 'bert') else 'bert.') if len(missing_keys) > 0: logger.info("Weights of {} not initialized from pretrained model: {}".format( model.__class__.__name__, missing_keys)) if len(unexpected_keys) > 0: logger.info("Weights from pretrained model not used in {}: {}".format( model.__class__.__name__, unexpected_keys)) if tempdir: # Clean up temp dir shutil.rmtree(tempdir) return model class BertModel(PreTrainedBertModel): """BERT model ("Bidirectional Embedding Representations from a Transformer"). Params: config: a BertConfig class instance with the configuration to build a new model Inputs: `input_ids`: a torch.LongTensor of shape [batch_size, sequence_length] with the word token indices in the vocabulary(see the tokens preprocessing logic in the scripts `extract_features.py`, `run_classifier.py` and `run_squad.py`) `token_type_ids`: an optional torch.LongTensor of shape [batch_size, sequence_length] with the token types indices selected in [0, 1]. Type 0 corresponds to a `sentence A` and type 1 corresponds to a `sentence B` token (see BERT paper for more details). `attention_mask`: an optional torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [0, 1]. It's a mask to be used if the input sequence length is smaller than the max input sequence length in the current batch. It's the mask that we typically use for attention when a batch has varying length sentences. `output_all_encoded_layers`: boolean which controls the content of the `encoded_layers` output as described below. Default: `True`. Outputs: Tuple of (encoded_layers, pooled_output) `encoded_layers`: controled by `output_all_encoded_layers` argument: - `output_all_encoded_layers=True`: outputs a list of the full sequences of encoded-hidden-states at the end of each attention block (i.e. 12 full sequences for BERT-base, 24 for BERT-large), each encoded-hidden-state is a torch.FloatTensor of size [batch_size, sequence_length, hidden_size], - `output_all_encoded_layers=False`: outputs only the full sequence of hidden-states corresponding to the last attention block of shape [batch_size, sequence_length, hidden_size], `pooled_output`: a torch.FloatTensor of size [batch_size, hidden_size] which is the output of a classifier pretrained on top of the hidden state associated to the first character of the input (`CLF`) to train on the Next-Sentence task (see BERT's paper). Example usage: ```python # Already been converted into WordPiece token ids input_ids = torch.LongTensor([[31, 51, 99], [15, 5, 0]]) input_mask = torch.LongTensor([[1, 1, 1], [1, 1, 0]]) token_type_ids = torch.LongTensor([[0, 0, 1], [0, 1, 0]]) config = modeling.BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072) model = modeling.BertModel(config=config) all_encoder_layers, pooled_output = model(input_ids, token_type_ids, input_mask) ``` """ def __init__(self, config): super(BertModel, self).__init__(config) self.embeddings = BertEmbeddings(config) self.encoder = BertEncoder(config) self.pooler = BertPooler(config) self.apply(self.init_bert_weights) def forward(self, input_ids, token_type_ids=None, task_ids=None, attention_mask=None, output_all_encoded_layers=True, checkpoint_activations=False): if attention_mask is None: attention_mask = torch.ones_like(input_ids) if token_type_ids is None: token_type_ids = torch.zeros_like(input_ids) if task_ids is None: task_ids = torch.zeros_like(input_ids) # We create a 3D attention mask from a 2D tensor mask. # Sizes are [batch_size, 1, 1, to_seq_length] # So we can broadcast to [batch_size, num_heads, from_seq_length, to_seq_length] # this attention mask is more simple than the triangular masking of causal attention # used in OpenAI GPT, we just need to prepare the broadcast dimension here. extended_attention_mask = attention_mask.unsqueeze(1).unsqueeze(2) # Since attention_mask is 1.0 for positions we want to attend and 0.0 for # masked positions, this operation will create a tensor which is 0.0 for # positions we want to attend and -10000.0 for masked positions. # Since we are adding it to the raw scores before the softmax, this is # effectively the same as removing these entirely. extended_attention_mask = extended_attention_mask.to(dtype=next(self.encoder.parameters()).dtype) extended_attention_mask = (1.0 - extended_attention_mask) * -10000.0 embedding_output = self.embeddings(input_ids, token_type_ids, task_ids) encoded_layers = self.encoder(embedding_output, extended_attention_mask, output_all_encoded_layers=output_all_encoded_layers, checkpoint_activations=checkpoint_activations) sequence_output = encoded_layers[-1] for p in self.pooler.parameters(): if p is None: continue sequence_output = sequence_output.type_as(p) break pooled_output = self.pooler(sequence_output) if not output_all_encoded_layers or checkpoint_activations: encoded_layers = encoded_layers[-1] return encoded_layers, pooled_output class BertForPreTraining(PreTrainedBertModel): """BERT model with pre-training heads. This module comprises the BERT model followed by the two pre-training heads: - the masked language modeling head, and - the next sentence classification head. Params: config: a BertConfig class instance with the configuration to build a new model. Inputs: `input_ids`: a torch.LongTensor of shape [batch_size, sequence_length] with the word token indices in the vocabulary(see the tokens preprocessing logic in the scripts `extract_features.py`, `run_classifier.py` and `run_squad.py`) `token_type_ids`: an optional torch.LongTensor of shape [batch_size, sequence_length] with the token types indices selected in [0, 1]. Type 0 corresponds to a `sentence A` and type 1 corresponds to a `sentence B` token (see BERT paper for more details). `attention_mask`: an optional torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [0, 1]. It's a mask to be used if the input sequence length is smaller than the max input sequence length in the current batch. It's the mask that we typically use for attention when a batch has varying length sentences. `masked_lm_labels`: masked language modeling labels: torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [-1, 0, ..., vocab_size]. All labels set to -1 are ignored (masked), the loss is only computed for the labels set in [0, ..., vocab_size] `next_sentence_label`: next sentence classification loss: torch.LongTensor of shape [batch_size] with indices selected in [0, 1]. 0 => next sentence is the continuation, 1 => next sentence is a random sentence. Outputs: if `masked_lm_labels` and `next_sentence_label` are not `None`: Outputs the total_loss which is the sum of the masked language modeling loss and the next sentence classification loss. if `masked_lm_labels` or `next_sentence_label` is `None`: Outputs a tuple comprising - the masked language modeling logits of shape [batch_size, sequence_length, vocab_size], and - the next sentence classification logits of shape [batch_size, 2]. Example usage: ```python # Already been converted into WordPiece token ids input_ids = torch.LongTensor([[31, 51, 99], [15, 5, 0]]) input_mask = torch.LongTensor([[1, 1, 1], [1, 1, 0]]) token_type_ids = torch.LongTensor([[0, 0, 1], [0, 1, 0]]) config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072) model = BertForPreTraining(config) masked_lm_logits_scores, seq_relationship_logits = model(input_ids, token_type_ids, input_mask) ``` """ def __init__(self, config): super(BertForPreTraining, self).__init__(config) self.bert = BertModel(config) self.cls = BertPreTrainingHeads(config, self.bert.embeddings.word_embeddings.weight) self.apply(self.init_bert_weights) def forward(self, input_ids, token_type_ids=None, attention_mask=None, masked_lm_labels=None, next_sentence_label=None, checkpoint_activations=False): sequence_output, pooled_output = self.bert(input_ids, token_type_ids, attention_mask, output_all_encoded_layers=False, checkpoint_activations=checkpoint_activations) prediction_scores, seq_relationship_score = self.cls(sequence_output, pooled_output) if masked_lm_labels is not None and next_sentence_label is not None: loss_fct = CrossEntropyLoss(ignore_index=-1) masked_lm_loss = loss_fct(prediction_scores.view(-1, self.config.vocab_size), masked_lm_labels.view(-1)) next_sentence_loss = loss_fct(seq_relationship_score.view(-1, 2), next_sentence_label.view(-1)) total_loss = masked_lm_loss + next_sentence_loss return total_loss else: return prediction_scores, seq_relationship_score class BertForMaskedLM(PreTrainedBertModel): """BERT model with the masked language modeling head. This module comprises the BERT model followed by the masked language modeling head. Params: config: a BertConfig class instance with the configuration to build a new model. Inputs: `input_ids`: a torch.LongTensor of shape [batch_size, sequence_length] with the word token indices in the vocabulary(see the tokens preprocessing logic in the scripts `extract_features.py`, `run_classifier.py` and `run_squad.py`) `token_type_ids`: an optional torch.LongTensor of shape [batch_size, sequence_length] with the token types indices selected in [0, 1]. Type 0 corresponds to a `sentence A` and type 1 corresponds to a `sentence B` token (see BERT paper for more details). `attention_mask`: an optional torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [0, 1]. It's a mask to be used if the input sequence length is smaller than the max input sequence length in the current batch. It's the mask that we typically use for attention when a batch has varying length sentences. `masked_lm_labels`: masked language modeling labels: torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [-1, 0, ..., vocab_size]. All labels set to -1 are ignored (masked), the loss is only computed for the labels set in [0, ..., vocab_size] Outputs: if `masked_lm_labels` is not `None`: Outputs the masked language modeling loss. if `masked_lm_labels` is `None`: Outputs the masked language modeling logits of shape [batch_size, sequence_length, vocab_size]. Example usage: ```python # Already been converted into WordPiece token ids input_ids = torch.LongTensor([[31, 51, 99], [15, 5, 0]]) input_mask = torch.LongTensor([[1, 1, 1], [1, 1, 0]]) token_type_ids = torch.LongTensor([[0, 0, 1], [0, 1, 0]]) config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072) model = BertForMaskedLM(config) masked_lm_logits_scores = model(input_ids, token_type_ids, input_mask) ``` """ def __init__(self, config): super(BertForMaskedLM, self).__init__(config) self.bert = BertModel(config) self.cls = BertOnlyMLMHead(config, self.bert.embeddings.word_embeddings.weight) self.apply(self.init_bert_weights) def forward(self, input_ids, token_type_ids=None, attention_mask=None, masked_lm_labels=None, checkpoint_activations=False): sequence_output, _ = self.bert(input_ids, token_type_ids, attention_mask, output_all_encoded_layers=False, checkpoint_activations=checkpoint_activations) prediction_scores = self.cls(sequence_output) if masked_lm_labels is not None: loss_fct = CrossEntropyLoss(ignore_index=-1) masked_lm_loss = loss_fct(prediction_scores.view(-1, self.config.vocab_size), masked_lm_labels.view(-1)) return masked_lm_loss else: return prediction_scores class BertForNextSentencePrediction(PreTrainedBertModel): """BERT model with next sentence prediction head. This module comprises the BERT model followed by the next sentence classification head. Params: config: a BertConfig class instance with the configuration to build a new model. Inputs: `input_ids`: a torch.LongTensor of shape [batch_size, sequence_length] with the word token indices in the vocabulary(see the tokens preprocessing logic in the scripts `extract_features.py`, `run_classifier.py` and `run_squad.py`) `token_type_ids`: an optional torch.LongTensor of shape [batch_size, sequence_length] with the token types indices selected in [0, 1]. Type 0 corresponds to a `sentence A` and type 1 corresponds to a `sentence B` token (see BERT paper for more details). `attention_mask`: an optional torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [0, 1]. It's a mask to be used if the input sequence length is smaller than the max input sequence length in the current batch. It's the mask that we typically use for attention when a batch has varying length sentences. `next_sentence_label`: next sentence classification loss: torch.LongTensor of shape [batch_size] with indices selected in [0, 1]. 0 => next sentence is the continuation, 1 => next sentence is a random sentence. Outputs: if `next_sentence_label` is not `None`: Outputs the total_loss which is the sum of the masked language modeling loss and the next sentence classification loss. if `next_sentence_label` is `None`: Outputs the next sentence classification logits of shape [batch_size, 2]. Example usage: ```python # Already been converted into WordPiece token ids input_ids = torch.LongTensor([[31, 51, 99], [15, 5, 0]]) input_mask = torch.LongTensor([[1, 1, 1], [1, 1, 0]]) token_type_ids = torch.LongTensor([[0, 0, 1], [0, 1, 0]]) config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072) model = BertForNextSentencePrediction(config) seq_relationship_logits = model(input_ids, token_type_ids, input_mask) ``` """ def __init__(self, config): super(BertForNextSentencePrediction, self).__init__(config) self.bert = BertModel(config) self.cls = BertOnlyNSPHead(config) self.apply(self.init_bert_weights) def forward(self, input_ids, token_type_ids=None, attention_mask=None, next_sentence_label=None, checkpoint_activations=False): _, pooled_output = self.bert(input_ids, token_type_ids, attention_mask, output_all_encoded_layers=False, checkpoint_activations=checkpoint_activations) seq_relationship_score = self.cls( pooled_output) if next_sentence_label is not None: loss_fct = CrossEntropyLoss(ignore_index=-1) next_sentence_loss = loss_fct(seq_relationship_score.view(-1, 2), next_sentence_label.view(-1)) return next_sentence_loss else: return seq_relationship_score class BertForSequenceClassification(PreTrainedBertModel): """BERT model for classification. This module is composed of the BERT model with a linear layer on top of the pooled output. Params: `config`: a BertConfig class instance with the configuration to build a new model. `num_labels`: the number of classes for the classifier. Default = 2. Inputs: `input_ids`: a torch.LongTensor of shape [batch_size, sequence_length] with the word token indices in the vocabulary(see the tokens preprocessing logic in the scripts `extract_features.py`, `run_classifier.py` and `run_squad.py`) `token_type_ids`: an optional torch.LongTensor of shape [batch_size, sequence_length] with the token types indices selected in [0, 1]. Type 0 corresponds to a `sentence A` and type 1 corresponds to a `sentence B` token (see BERT paper for more details). `attention_mask`: an optional torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [0, 1]. It's a mask to be used if the input sequence length is smaller than the max input sequence length in the current batch. It's the mask that we typically use for attention when a batch has varying length sentences. `labels`: labels for the classification output: torch.LongTensor of shape [batch_size] with indices selected in [0, ..., num_labels]. Outputs: if `labels` is not `None`: Outputs the CrossEntropy classification loss of the output with the labels. if `labels` is `None`: Outputs the classification logits of shape [batch_size, num_labels]. Example usage: ```python # Already been converted into WordPiece token ids input_ids = torch.LongTensor([[31, 51, 99], [15, 5, 0]]) input_mask = torch.LongTensor([[1, 1, 1], [1, 1, 0]]) token_type_ids = torch.LongTensor([[0, 0, 1], [0, 1, 0]]) config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072) num_labels = 2 model = BertForSequenceClassification(config, num_labels) logits = model(input_ids, token_type_ids, input_mask) ``` """ def __init__(self, config, num_labels=2): super(BertForSequenceClassification, self).__init__(config) self.num_labels = num_labels self.bert = BertModel(config) self.dropout = nn.Dropout(config.hidden_dropout_prob) self.classifier = nn.Linear(config.hidden_size, num_labels) self.apply(self.init_bert_weights) def forward(self, input_ids, token_type_ids=None, attention_mask=None, labels=None, checkpoint_activations=False): _, pooled_output = self.bert(input_ids, token_type_ids, attention_mask, output_all_encoded_layers=False, checkpoint_activations=checkpoint_activations) pooled_output = self.dropout(pooled_output) logits = self.classifier(pooled_output) if labels is not None: loss_fct = CrossEntropyLoss() loss = loss_fct(logits.view(-1, self.num_labels), labels.view(-1)) return loss else: return logits class BertForMultipleChoice(PreTrainedBertModel): """BERT model for multiple choice tasks. This module is composed of the BERT model with a linear layer on top of the pooled output. Params: `config`: a BertConfig class instance with the configuration to build a new model. `num_choices`: the number of classes for the classifier. Default = 2. Inputs: `input_ids`: a torch.LongTensor of shape [batch_size, num_choices, sequence_length] with the word token indices in the vocabulary(see the tokens preprocessing logic in the scripts `extract_features.py`, `run_classifier.py` and `run_squad.py`) `token_type_ids`: an optional torch.LongTensor of shape [batch_size, num_choices, sequence_length] with the token types indices selected in [0, 1]. Type 0 corresponds to a `sentence A` and type 1 corresponds to a `sentence B` token (see BERT paper for more details). `attention_mask`: an optional torch.LongTensor of shape [batch_size, num_choices, sequence_length] with indices selected in [0, 1]. It's a mask to be used if the input sequence length is smaller than the max input sequence length in the current batch. It's the mask that we typically use for attention when a batch has varying length sentences. `labels`: labels for the classification output: torch.LongTensor of shape [batch_size] with indices selected in [0, ..., num_choices]. Outputs: if `labels` is not `None`: Outputs the CrossEntropy classification loss of the output with the labels. if `labels` is `None`: Outputs the classification logits of shape [batch_size, num_labels]. Example usage: ```python # Already been converted into WordPiece token ids input_ids = torch.LongTensor([[[31, 51, 99], [15, 5, 0]], [[12, 16, 42], [14, 28, 57]]]) input_mask = torch.LongTensor([[[1, 1, 1], [1, 1, 0]],[[1,1,0], [1, 0, 0]]]) token_type_ids = torch.LongTensor([[[0, 0, 1], [0, 1, 0]],[[0, 1, 1], [0, 0, 1]]]) config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072) num_choices = 2 model = BertForMultipleChoice(config, num_choices) logits = model(input_ids, token_type_ids, input_mask) ``` """ def __init__(self, config, num_choices=2): super(BertForMultipleChoice, self).__init__(config) self.num_choices = num_choices self.bert = BertModel(config) self.dropout = nn.Dropout(config.hidden_dropout_prob) self.classifier = nn.Linear(config.hidden_size, 1) self.apply(self.init_bert_weights) def forward(self, input_ids, token_type_ids=None, attention_mask=None, labels=None, checkpoint_activations=False): flat_input_ids = input_ids.view(-1, input_ids.size(-1)) flat_token_type_ids = token_type_ids.view(-1, token_type_ids.size(-1)) flat_attention_mask = attention_mask.view(-1, attention_mask.size(-1)) _, pooled_output = self.bert(flat_input_ids, flat_token_type_ids, flat_attention_mask, output_all_encoded_layers=False, checkpoint_activations=checkpoint_activations) pooled_output = self.dropout(pooled_output) logits = self.classifier(pooled_output) reshaped_logits = logits.view(-1, self.num_choices) if labels is not None: loss_fct = CrossEntropyLoss() loss = loss_fct(reshaped_logits, labels) return loss else: return reshaped_logits class BertForTokenClassification(PreTrainedBertModel): """BERT model for token-level classification. This module is composed of the BERT model with a linear layer on top of the full hidden state of the last layer. Params: `config`: a BertConfig class instance with the configuration to build a new model. `num_labels`: the number of classes for the classifier. Default = 2. Inputs: `input_ids`: a torch.LongTensor of shape [batch_size, sequence_length] with the word token indices in the vocabulary(see the tokens preprocessing logic in the scripts `extract_features.py`, `run_classifier.py` and `run_squad.py`) `token_type_ids`: an optional torch.LongTensor of shape [batch_size, sequence_length] with the token types indices selected in [0, 1]. Type 0 corresponds to a `sentence A` and type 1 corresponds to a `sentence B` token (see BERT paper for more details). `attention_mask`: an optional torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [0, 1]. It's a mask to be used if the input sequence length is smaller than the max input sequence length in the current batch. It's the mask that we typically use for attention when a batch has varying length sentences. `labels`: labels for the classification output: torch.LongTensor of shape [batch_size] with indices selected in [0, ..., num_labels]. Outputs: if `labels` is not `None`: Outputs the CrossEntropy classification loss of the output with the labels. if `labels` is `None`: Outputs the classification logits of shape [batch_size, sequence_length, num_labels]. Example usage: ```python # Already been converted into WordPiece token ids input_ids = torch.LongTensor([[31, 51, 99], [15, 5, 0]]) input_mask = torch.LongTensor([[1, 1, 1], [1, 1, 0]]) token_type_ids = torch.LongTensor([[0, 0, 1], [0, 1, 0]]) config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072) num_labels = 2 model = BertForTokenClassification(config, num_labels) logits = model(input_ids, token_type_ids, input_mask) ``` """ def __init__(self, config, num_labels=2): super(BertForTokenClassification, self).__init__(config) self.num_labels = num_labels self.bert = BertModel(config) self.dropout = nn.Dropout(config.hidden_dropout_prob) self.classifier = nn.Linear(config.hidden_size, num_labels) self.apply(self.init_bert_weights) def forward(self, input_ids, token_type_ids=None, attention_mask=None, labels=None, checkpoint_activations=False): sequence_output, _ = self.bert(input_ids, token_type_ids, attention_mask, output_all_encoded_layers=False, checkpoint_activations=checkpoint_activations) sequence_output = self.dropout(sequence_output) logits = self.classifier(sequence_output) if labels is not None: loss_fct = CrossEntropyLoss() loss = loss_fct(logits.view(-1, self.num_labels), labels.view(-1)) return loss else: return logits class BertForQuestionAnswering(PreTrainedBertModel): """BERT model for Question Answering (span extraction). This module is composed of the BERT model with a linear layer on top of the sequence output that computes start_logits and end_logits Params: `config`: a BertConfig class instance with the configuration to build a new model. Inputs: `input_ids`: a torch.LongTensor of shape [batch_size, sequence_length] with the word token indices in the vocabulary(see the tokens preprocessing logic in the scripts `extract_features.py`, `run_classifier.py` and `run_squad.py`) `token_type_ids`: an optional torch.LongTensor of shape [batch_size, sequence_length] with the token types indices selected in [0, 1]. Type 0 corresponds to a `sentence A` and type 1 corresponds to a `sentence B` token (see BERT paper for more details). `attention_mask`: an optional torch.LongTensor of shape [batch_size, sequence_length] with indices selected in [0, 1]. It's a mask to be used if the input sequence length is smaller than the max input sequence length in the current batch. It's the mask that we typically use for attention when a batch has varying length sentences. `start_positions`: position of the first token for the labeled span: torch.LongTensor of shape [batch_size]. Positions are clamped to the length of the sequence and position outside of the sequence are not taken into account for computing the loss. `end_positions`: position of the last token for the labeled span: torch.LongTensor of shape [batch_size]. Positions are clamped to the length of the sequence and position outside of the sequence are not taken into account for computing the loss. Outputs: if `start_positions` and `end_positions` are not `None`: Outputs the total_loss which is the sum of the CrossEntropy loss for the start and end token positions. if `start_positions` or `end_positions` is `None`: Outputs a tuple of start_logits, end_logits which are the logits respectively for the start and end position tokens of shape [batch_size, sequence_length]. Example usage: ```python # Already been converted into WordPiece token ids input_ids = torch.LongTensor([[31, 51, 99], [15, 5, 0]]) input_mask = torch.LongTensor([[1, 1, 1], [1, 1, 0]]) token_type_ids = torch.LongTensor([[0, 0, 1], [0, 1, 0]]) config = BertConfig(vocab_size_or_config_json_file=32000, hidden_size=768, num_hidden_layers=12, num_attention_heads=12, intermediate_size=3072) model = BertForQuestionAnswering(config) start_logits, end_logits = model(input_ids, token_type_ids, input_mask) ``` """ def __init__(self, config): super(BertForQuestionAnswering, self).__init__(config) self.bert = BertModel(config) # TODO check with Google if it's normal there is no dropout on the token classifier of SQuAD in the TF version # self.dropout = nn.Dropout(config.hidden_dropout_prob) self.qa_outputs = nn.Linear(config.hidden_size, 2) self.apply(self.init_bert_weights) def forward(self, input_ids, token_type_ids=None, attention_mask=None, start_positions=None, end_positions=None, checkpoint_activations=False): sequence_output, _ = self.bert(input_ids, token_type_ids, attention_mask, output_all_encoded_layers=False, checkpoint_activations=checkpoint_activations) logits = self.qa_outputs(sequence_output) start_logits, end_logits = logits.split(1, dim=-1) start_logits = start_logits.squeeze(-1) end_logits = end_logits.squeeze(-1) if start_positions is not None and end_positions is not None: # If we are on multi-GPU, split add a dimension if len(start_positions.size()) > 1: start_positions = start_positions.squeeze(-1) if len(end_positions.size()) > 1: end_positions = end_positions.squeeze(-1) # sometimes the start/end positions are outside our model inputs, we ignore these terms ignored_index = start_logits.size(1) start_positions.clamp_(0, ignored_index) end_positions.clamp_(0, ignored_index) loss_fct = CrossEntropyLoss(ignore_index=ignored_index) start_loss = loss_fct(start_logits, start_positions) end_loss = loss_fct(end_logits, end_positions) total_loss = (start_loss + end_loss) / 2 return total_loss else: return start_logits, end_logits
py
1a2f53363c723cd3945e5c889bccc764eabdd84d
""" objectives.py: Objective (or loss) functions. They require its derivatives with respect to the network prediction. Copyright 2017 Ramon Vinas Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ import math class Objective: def __init__(self): raise NotImplementedError def get_loss(self, y_true, y_pred): raise NotImplementedError def derivative(self, y_true, y_pred): raise NotImplementedError class CrossEntropy(Objective): def __init__(self): pass def get_loss(self, y_true, y_pred, eps=1e-6): return -y_true * math.log(y_pred + eps) - (1 - y_true) * math.log(1 - y_pred + eps) def derivative(self, y_true, y_pred): return (y_pred - y_true) / (y_pred * (1 - y_pred)) class SquaredError(Objective): def __init__(self): pass def get_loss(self, y_true, y_pred): return (y_true - y_pred) ** 2 def derivative(self, y_true, y_pred): return 2 * (y_true - y_pred)
py
1a2f537443b67da5f567a524f1a86305586022a2
#!/usr/bin/env python3 # # linearize-hashes.py: List blocks in a linear, no-fork version of the chain. # # Copyright (c) 2013-2014 The Bitcoin Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. # from __future__ import print_function try: # Python 3 import http.client as httplib except ImportError: # Python 2 import httplib import json import re import base64 import sys import os import os.path settings = {} ##### Switch endian-ness ##### def hex_switchEndian(s): """ Switches the endianness of a hex string (in pairs of hex chars) """ pairList = [s[i:i+2].encode() for i in range(0, len(s), 2)] return b''.join(pairList[::-1]).decode() class BitcoinRPC: def __init__(self, host, port, username, password): authpair = "%s:%s" % (username, password) authpair = authpair.encode('utf-8') self.authhdr = b"Basic " + base64.b64encode(authpair) self.conn = httplib.HTTPConnection(host, port=port, timeout=30) def execute(self, obj): try: self.conn.request('POST', '/', json.dumps(obj), { 'Authorization' : self.authhdr, 'Content-type' : 'application/json' }) except ConnectionRefusedError: print('RPC connection refused. Check RPC settings and the server status.', file=sys.stderr) return None resp = self.conn.getresponse() if resp is None: print("JSON-RPC: no response", file=sys.stderr) return None body = resp.read().decode('utf-8') resp_obj = json.loads(body) return resp_obj @staticmethod def build_request(idx, method, params): obj = { 'version' : '1.1', 'method' : method, 'id' : idx } if params is None: obj['params'] = [] else: obj['params'] = params return obj @staticmethod def response_is_error(resp_obj): return 'error' in resp_obj and resp_obj['error'] is not None def get_block_hashes(settings, max_blocks_per_call=10000): rpc = BitcoinRPC(settings['host'], settings['port'], settings['rpcuser'], settings['rpcpassword']) height = settings['min_height'] while height < settings['max_height']+1: num_blocks = min(settings['max_height']+1-height, max_blocks_per_call) batch = [] for x in range(num_blocks): batch.append(rpc.build_request(x, 'getblockhash', [height + x])) reply = rpc.execute(batch) if reply is None: print('Cannot continue. Program will halt.') return None for x,resp_obj in enumerate(reply): if rpc.response_is_error(resp_obj): print('JSON-RPC: error at height', height+x, ': ', resp_obj['error'], file=sys.stderr) exit(1) assert(resp_obj['id'] == x) # assume replies are in-sequence if settings['rev_hash_bytes'] == 'true': resp_obj['result'] = hex_switchEndian(resp_obj['result']) print(resp_obj['result']) height += num_blocks def get_rpc_cookie(): # Open the cookie file with open(os.path.join(os.path.expanduser(settings['datadir']), '.cookie'), 'r') as f: combined = f.readline() combined_split = combined.split(":") settings['rpcuser'] = combined_split[0] settings['rpcpassword'] = combined_split[1] if __name__ == '__main__': if len(sys.argv) != 2: print("Usage: linearize-hashes.py CONFIG-FILE") sys.exit(1) f = open(sys.argv[1]) for line in f: # skip comment lines m = re.search('^\s*#', line) if m: continue # parse key=value lines m = re.search('^(\w+)\s*=\s*(\S.*)$', line) if m is None: continue settings[m.group(1)] = m.group(2) f.close() if 'host' not in settings: settings['host'] = '127.0.0.1' if 'port' not in settings: settings['port'] = 9998 if 'min_height' not in settings: settings['min_height'] = 0 if 'max_height' not in settings: settings['max_height'] = 313000 if 'rev_hash_bytes' not in settings: settings['rev_hash_bytes'] = 'false' use_userpass = True use_datadir = False if 'rpcuser' not in settings or 'rpcpassword' not in settings: use_userpass = False if 'datadir' in settings and not use_userpass: use_datadir = True if not use_userpass and not use_datadir: print("Missing datadir or username and/or password in cfg file", file=stderr) sys.exit(1) settings['port'] = int(settings['port']) settings['min_height'] = int(settings['min_height']) settings['max_height'] = int(settings['max_height']) # Force hash byte format setting to be lowercase to make comparisons easier. settings['rev_hash_bytes'] = settings['rev_hash_bytes'].lower() # Get the rpc user and pass from the cookie if the datadir is set if use_datadir: get_rpc_cookie() get_block_hashes(settings)
py
1a2f5695c7fdc5d6cc79a8dfa8b419505f24e4d5
""" Django settings for tests project. Generated by 'django-admin startproject' using Django 1.9.1. For more information on this file, see https://docs.djangoproject.com/en/1.9/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/1.9/ref/settings/ """ import os # Build paths inside the project like this: os.path.join(BASE_DIR, ...) BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/1.9/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = '5bg%^f37a=%mh8(qkq1#)a$e*d-pt*dzox0_39-ywqh=@m(_ii' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True ALLOWED_HOSTS = [u'testserver'] # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'tests', ] MIDDLEWARE_CLASSES = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.auth.middleware.SessionAuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'project.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] WSGI_APPLICATION = 'project.wsgi.application' # Database # https://docs.djangoproject.com/en/1.9/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(BASE_DIR, 'tests', 'db.sqlite3'), } } # Password validation # https://docs.djangoproject.com/en/1.9/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/1.9/topics/i18n/ LANGUAGE_CODE = 'en' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = False USE_TZ = False # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.9/howto/static-files/ STATIC_URL = '/static/' # Caching CACHE_DIR = os.path.join(BASE_DIR, '.cache') CACHES = { 'default': { 'BACKEND': 'diskcache.DjangoCache', 'LOCATION': CACHE_DIR, }, }
py
1a2f56ae602d3bfd4f34db75bcdb1e42f9cd14d0
#!/usr/bin/env python """ Easy Install ------------ A tool for doing automatic download/extract/build of distutils-based Python packages. For detailed documentation, see the accompanying EasyInstall.txt file, or visit the `EasyInstall home page`__. __ https://pythonhosted.org/setuptools/easy_install.html """ from glob import glob from distutils.util import get_platform from distutils.util import convert_path, subst_vars from distutils.errors import DistutilsArgError, DistutilsOptionError, \ DistutilsError, DistutilsPlatformError from distutils.command.install import INSTALL_SCHEMES, SCHEME_KEYS from distutils import log, dir_util from distutils.command.build_scripts import first_line_re import sys import os import zipimport import shutil import tempfile import zipfile import re import stat import random import platform import textwrap import warnings import site import struct import contextlib from setuptools import Command from setuptools.sandbox import run_setup from setuptools.py31compat import get_path, get_config_vars from setuptools.command import setopt from setuptools.archive_util import unpack_archive from setuptools.package_index import PackageIndex from setuptools.package_index import URL_SCHEME from setuptools.command import bdist_egg, egg_info from setuptools.compat import (iteritems, maxsize, basestring, unicode, reraise, PY2, PY3) from pkg_resources import ( yield_lines, normalize_path, resource_string, ensure_directory, get_distribution, find_distributions, Environment, Requirement, Distribution, PathMetadata, EggMetadata, WorkingSet, DistributionNotFound, VersionConflict, DEVELOP_DIST, ) import pkg_resources # Turn on PEP440Warnings warnings.filterwarnings("default", category=pkg_resources.PEP440Warning) sys_executable = os.environ.get('__PYVENV_LAUNCHER__', os.path.normpath(sys.executable)) __all__ = [ 'samefile', 'easy_install', 'PthDistributions', 'extract_wininst_cfg', 'main', 'get_exe_prefixes', ] def is_64bit(): return struct.calcsize("P") == 8 def samefile(p1, p2): both_exist = os.path.exists(p1) and os.path.exists(p2) use_samefile = hasattr(os.path, 'samefile') and both_exist if use_samefile: return os.path.samefile(p1, p2) norm_p1 = os.path.normpath(os.path.normcase(p1)) norm_p2 = os.path.normpath(os.path.normcase(p2)) return norm_p1 == norm_p2 if PY2: def _to_ascii(s): return s def isascii(s): try: unicode(s, 'ascii') return True except UnicodeError: return False else: def _to_ascii(s): return s.encode('ascii') def isascii(s): try: s.encode('ascii') return True except UnicodeError: return False class easy_install(Command): """Manage a download/build/install process""" description = "Find/get/install Python packages" command_consumes_arguments = True user_options = [ ('prefix=', None, "installation prefix"), ("zip-ok", "z", "install package as a zipfile"), ("multi-version", "m", "make apps have to require() a version"), ("upgrade", "U", "force upgrade (searches PyPI for latest versions)"), ("install-dir=", "d", "install package to DIR"), ("script-dir=", "s", "install scripts to DIR"), ("exclude-scripts", "x", "Don't install scripts"), ("always-copy", "a", "Copy all needed packages to install dir"), ("index-url=", "i", "base URL of Python Package Index"), ("find-links=", "f", "additional URL(s) to search for packages"), ("build-directory=", "b", "download/extract/build in DIR; keep the results"), ('optimize=', 'O', "also compile with optimization: -O1 for \"python -O\", " "-O2 for \"python -OO\", and -O0 to disable [default: -O0]"), ('record=', None, "filename in which to record list of installed files"), ('always-unzip', 'Z', "don't install as a zipfile, no matter what"), ('site-dirs=', 'S', "list of directories where .pth files work"), ('editable', 'e', "Install specified packages in editable form"), ('no-deps', 'N', "don't install dependencies"), ('allow-hosts=', 'H', "pattern(s) that hostnames must match"), ('local-snapshots-ok', 'l', "allow building eggs from local checkouts"), ('version', None, "print version information and exit"), ('no-find-links', None, "Don't load find-links defined in packages being installed") ] boolean_options = [ 'zip-ok', 'multi-version', 'exclude-scripts', 'upgrade', 'always-copy', 'editable', 'no-deps', 'local-snapshots-ok', 'version' ] if site.ENABLE_USER_SITE: help_msg = "install in user site-package '%s'" % site.USER_SITE user_options.append(('user', None, help_msg)) boolean_options.append('user') negative_opt = {'always-unzip': 'zip-ok'} create_index = PackageIndex def initialize_options(self): if site.ENABLE_USER_SITE: whereami = os.path.abspath(__file__) self.user = whereami.startswith(site.USER_SITE) else: self.user = 0 self.zip_ok = self.local_snapshots_ok = None self.install_dir = self.script_dir = self.exclude_scripts = None self.index_url = None self.find_links = None self.build_directory = None self.args = None self.optimize = self.record = None self.upgrade = self.always_copy = self.multi_version = None self.editable = self.no_deps = self.allow_hosts = None self.root = self.prefix = self.no_report = None self.version = None self.install_purelib = None # for pure module distributions self.install_platlib = None # non-pure (dists w/ extensions) self.install_headers = None # for C/C++ headers self.install_lib = None # set to either purelib or platlib self.install_scripts = None self.install_data = None self.install_base = None self.install_platbase = None if site.ENABLE_USER_SITE: self.install_userbase = site.USER_BASE self.install_usersite = site.USER_SITE else: self.install_userbase = None self.install_usersite = None self.no_find_links = None # Options not specifiable via command line self.package_index = None self.pth_file = self.always_copy_from = None self.site_dirs = None self.installed_projects = {} self.sitepy_installed = False # Always read easy_install options, even if we are subclassed, or have # an independent instance created. This ensures that defaults will # always come from the standard configuration file(s)' "easy_install" # section, even if this is a "develop" or "install" command, or some # other embedding. self._dry_run = None self.verbose = self.distribution.verbose self.distribution._set_command_options( self, self.distribution.get_option_dict('easy_install') ) def delete_blockers(self, blockers): for filename in blockers: if os.path.exists(filename) or os.path.islink(filename): log.info("Deleting %s", filename) if not self.dry_run: if (os.path.isdir(filename) and not os.path.islink(filename)): rmtree(filename) else: os.unlink(filename) def finalize_options(self): if self.version: print('setuptools %s' % get_distribution('setuptools').version) sys.exit() py_version = sys.version.split()[0] prefix, exec_prefix = get_config_vars('prefix', 'exec_prefix') self.config_vars = { 'dist_name': self.distribution.get_name(), 'dist_version': self.distribution.get_version(), 'dist_fullname': self.distribution.get_fullname(), 'py_version': py_version, 'py_version_short': py_version[0:3], 'py_version_nodot': py_version[0] + py_version[2], 'sys_prefix': prefix, 'prefix': prefix, 'sys_exec_prefix': exec_prefix, 'exec_prefix': exec_prefix, # Only python 3.2+ has abiflags 'abiflags': getattr(sys, 'abiflags', ''), } if site.ENABLE_USER_SITE: self.config_vars['userbase'] = self.install_userbase self.config_vars['usersite'] = self.install_usersite # fix the install_dir if "--user" was used # XXX: duplicate of the code in the setup command if self.user and site.ENABLE_USER_SITE: self.create_home_path() if self.install_userbase is None: raise DistutilsPlatformError( "User base directory is not specified") self.install_base = self.install_platbase = self.install_userbase if os.name == 'posix': self.select_scheme("unix_user") else: self.select_scheme(os.name + "_user") self.expand_basedirs() self.expand_dirs() self._expand('install_dir', 'script_dir', 'build_directory', 'site_dirs') # If a non-default installation directory was specified, default the # script directory to match it. if self.script_dir is None: self.script_dir = self.install_dir if self.no_find_links is None: self.no_find_links = False # Let install_dir get set by install_lib command, which in turn # gets its info from the install command, and takes into account # --prefix and --home and all that other crud. self.set_undefined_options( 'install_lib', ('install_dir', 'install_dir') ) # Likewise, set default script_dir from 'install_scripts.install_dir' self.set_undefined_options( 'install_scripts', ('install_dir', 'script_dir') ) if self.user and self.install_purelib: self.install_dir = self.install_purelib self.script_dir = self.install_scripts # default --record from the install command self.set_undefined_options('install', ('record', 'record')) # Should this be moved to the if statement below? It's not used # elsewhere normpath = map(normalize_path, sys.path) self.all_site_dirs = get_site_dirs() if self.site_dirs is not None: site_dirs = [ os.path.expanduser(s.strip()) for s in self.site_dirs.split(',') ] for d in site_dirs: if not os.path.isdir(d): log.warn("%s (in --site-dirs) does not exist", d) elif normalize_path(d) not in normpath: raise DistutilsOptionError( d + " (in --site-dirs) is not on sys.path" ) else: self.all_site_dirs.append(normalize_path(d)) if not self.editable: self.check_site_dir() self.index_url = self.index_url or "https://pypi.python.org/simple" self.shadow_path = self.all_site_dirs[:] for path_item in self.install_dir, normalize_path(self.script_dir): if path_item not in self.shadow_path: self.shadow_path.insert(0, path_item) if self.allow_hosts is not None: hosts = [s.strip() for s in self.allow_hosts.split(',')] else: hosts = ['*'] if self.package_index is None: self.package_index = self.create_index( self.index_url, search_path=self.shadow_path, hosts=hosts, ) self.local_index = Environment(self.shadow_path + sys.path) if self.find_links is not None: if isinstance(self.find_links, basestring): self.find_links = self.find_links.split() else: self.find_links = [] if self.local_snapshots_ok: self.package_index.scan_egg_links(self.shadow_path + sys.path) if not self.no_find_links: self.package_index.add_find_links(self.find_links) self.set_undefined_options('install_lib', ('optimize', 'optimize')) if not isinstance(self.optimize, int): try: self.optimize = int(self.optimize) if not (0 <= self.optimize <= 2): raise ValueError except ValueError: raise DistutilsOptionError("--optimize must be 0, 1, or 2") if self.editable and not self.build_directory: raise DistutilsArgError( "Must specify a build directory (-b) when using --editable" ) if not self.args: raise DistutilsArgError( "No urls, filenames, or requirements specified (see --help)") self.outputs = [] def _expand_attrs(self, attrs): for attr in attrs: val = getattr(self, attr) if val is not None: if os.name == 'posix' or os.name == 'nt': val = os.path.expanduser(val) val = subst_vars(val, self.config_vars) setattr(self, attr, val) def expand_basedirs(self): """Calls `os.path.expanduser` on install_base, install_platbase and root.""" self._expand_attrs(['install_base', 'install_platbase', 'root']) def expand_dirs(self): """Calls `os.path.expanduser` on install dirs.""" self._expand_attrs(['install_purelib', 'install_platlib', 'install_lib', 'install_headers', 'install_scripts', 'install_data', ]) def run(self): if self.verbose != self.distribution.verbose: log.set_verbosity(self.verbose) try: for spec in self.args: self.easy_install(spec, not self.no_deps) if self.record: outputs = self.outputs if self.root: # strip any package prefix root_len = len(self.root) for counter in range(len(outputs)): outputs[counter] = outputs[counter][root_len:] from distutils import file_util self.execute( file_util.write_file, (self.record, outputs), "writing list of installed files to '%s'" % self.record ) self.warn_deprecated_options() finally: log.set_verbosity(self.distribution.verbose) def pseudo_tempname(self): """Return a pseudo-tempname base in the install directory. This code is intentionally naive; if a malicious party can write to the target directory you're already in deep doodoo. """ try: pid = os.getpid() except: pid = random.randint(0, maxsize) return os.path.join(self.install_dir, "test-easy-install-%s" % pid) def warn_deprecated_options(self): pass def check_site_dir(self): """Verify that self.install_dir is .pth-capable dir, if needed""" instdir = normalize_path(self.install_dir) pth_file = os.path.join(instdir, 'easy-install.pth') # Is it a configured, PYTHONPATH, implicit, or explicit site dir? is_site_dir = instdir in self.all_site_dirs if not is_site_dir and not self.multi_version: # No? Then directly test whether it does .pth file processing is_site_dir = self.check_pth_processing() else: # make sure we can write to target dir testfile = self.pseudo_tempname() + '.write-test' test_exists = os.path.exists(testfile) try: if test_exists: os.unlink(testfile) open(testfile, 'w').close() os.unlink(testfile) except (OSError, IOError): self.cant_write_to_target() if not is_site_dir and not self.multi_version: # Can't install non-multi to non-site dir raise DistutilsError(self.no_default_version_msg()) if is_site_dir: if self.pth_file is None: self.pth_file = PthDistributions(pth_file, self.all_site_dirs) else: self.pth_file = None PYTHONPATH = os.environ.get('PYTHONPATH', '').split(os.pathsep) if instdir not in map(normalize_path, [_f for _f in PYTHONPATH if _f]): # only PYTHONPATH dirs need a site.py, so pretend it's there self.sitepy_installed = True elif self.multi_version and not os.path.exists(pth_file): self.sitepy_installed = True # don't need site.py in this case self.pth_file = None # and don't create a .pth file self.install_dir = instdir def cant_write_to_target(self): template = """can't create or remove files in install directory The following error occurred while trying to add or remove files in the installation directory: %s The installation directory you specified (via --install-dir, --prefix, or the distutils default setting) was: %s """ msg = template % (sys.exc_info()[1], self.install_dir,) if not os.path.exists(self.install_dir): msg += """ This directory does not currently exist. Please create it and try again, or choose a different installation directory (using the -d or --install-dir option). """ else: msg += """ Perhaps your account does not have write access to this directory? If the installation directory is a system-owned directory, you may need to sign in as the administrator or "root" account. If you do not have administrative access to this machine, you may wish to choose a different installation directory, preferably one that is listed in your PYTHONPATH environment variable. For information on other options, you may wish to consult the documentation at: https://pythonhosted.org/setuptools/easy_install.html Please make the appropriate changes for your system and try again. """ raise DistutilsError(msg) def check_pth_processing(self): """Empirically verify whether .pth files are supported in inst. dir""" instdir = self.install_dir log.info("Checking .pth file support in %s", instdir) pth_file = self.pseudo_tempname() + ".pth" ok_file = pth_file + '.ok' ok_exists = os.path.exists(ok_file) try: if ok_exists: os.unlink(ok_file) dirname = os.path.dirname(ok_file) if not os.path.exists(dirname): os.makedirs(dirname) f = open(pth_file, 'w') except (OSError, IOError): self.cant_write_to_target() else: try: f.write("import os; f = open(%r, 'w'); f.write('OK'); " "f.close()\n" % (ok_file,)) f.close() f = None executable = sys.executable if os.name == 'nt': dirname, basename = os.path.split(executable) alt = os.path.join(dirname, 'pythonw.exe') if (basename.lower() == 'python.exe' and os.path.exists(alt)): # use pythonw.exe to avoid opening a console window executable = alt from distutils.spawn import spawn spawn([executable, '-E', '-c', 'pass'], 0) if os.path.exists(ok_file): log.info( "TEST PASSED: %s appears to support .pth files", instdir ) return True finally: if f: f.close() if os.path.exists(ok_file): os.unlink(ok_file) if os.path.exists(pth_file): os.unlink(pth_file) if not self.multi_version: log.warn("TEST FAILED: %s does NOT support .pth files", instdir) return False def install_egg_scripts(self, dist): """Write all the scripts for `dist`, unless scripts are excluded""" if not self.exclude_scripts and dist.metadata_isdir('scripts'): for script_name in dist.metadata_listdir('scripts'): if dist.metadata_isdir('scripts/' + script_name): # The "script" is a directory, likely a Python 3 # __pycache__ directory, so skip it. continue self.install_script( dist, script_name, dist.get_metadata('scripts/' + script_name) ) self.install_wrapper_scripts(dist) def add_output(self, path): if os.path.isdir(path): for base, dirs, files in os.walk(path): for filename in files: self.outputs.append(os.path.join(base, filename)) else: self.outputs.append(path) def not_editable(self, spec): if self.editable: raise DistutilsArgError( "Invalid argument %r: you can't use filenames or URLs " "with --editable (except via the --find-links option)." % (spec,) ) def check_editable(self, spec): if not self.editable: return if os.path.exists(os.path.join(self.build_directory, spec.key)): raise DistutilsArgError( "%r already exists in %s; can't do a checkout there" % (spec.key, self.build_directory) ) def easy_install(self, spec, deps=False): tmpdir = tempfile.mkdtemp(prefix="easy_install-") download = None if not self.editable: self.install_site_py() try: if not isinstance(spec, Requirement): if URL_SCHEME(spec): # It's a url, download it to tmpdir and process self.not_editable(spec) download = self.package_index.download(spec, tmpdir) return self.install_item(None, download, tmpdir, deps, True) elif os.path.exists(spec): # Existing file or directory, just process it directly self.not_editable(spec) return self.install_item(None, spec, tmpdir, deps, True) else: spec = parse_requirement_arg(spec) self.check_editable(spec) dist = self.package_index.fetch_distribution( spec, tmpdir, self.upgrade, self.editable, not self.always_copy, self.local_index ) if dist is None: msg = "Could not find suitable distribution for %r" % spec if self.always_copy: msg += " (--always-copy skips system and development eggs)" raise DistutilsError(msg) elif dist.precedence == DEVELOP_DIST: # .egg-info dists don't need installing, just process deps self.process_distribution(spec, dist, deps, "Using") return dist else: return self.install_item(spec, dist.location, tmpdir, deps) finally: if os.path.exists(tmpdir): rmtree(tmpdir) def install_item(self, spec, download, tmpdir, deps, install_needed=False): # Installation is also needed if file in tmpdir or is not an egg install_needed = install_needed or self.always_copy install_needed = install_needed or os.path.dirname(download) == tmpdir install_needed = install_needed or not download.endswith('.egg') install_needed = install_needed or ( self.always_copy_from is not None and os.path.dirname(normalize_path(download)) == normalize_path(self.always_copy_from) ) if spec and not install_needed: # at this point, we know it's a local .egg, we just don't know if # it's already installed. for dist in self.local_index[spec.project_name]: if dist.location == download: break else: install_needed = True # it's not in the local index log.info("Processing %s", os.path.basename(download)) if install_needed: dists = self.install_eggs(spec, download, tmpdir) for dist in dists: self.process_distribution(spec, dist, deps) else: dists = [self.egg_distribution(download)] self.process_distribution(spec, dists[0], deps, "Using") if spec is not None: for dist in dists: if dist in spec: return dist def select_scheme(self, name): """Sets the install directories by applying the install schemes.""" # it's the caller's problem if they supply a bad name! scheme = INSTALL_SCHEMES[name] for key in SCHEME_KEYS: attrname = 'install_' + key if getattr(self, attrname) is None: setattr(self, attrname, scheme[key]) def process_distribution(self, requirement, dist, deps=True, *info): self.update_pth(dist) self.package_index.add(dist) if dist in self.local_index[dist.key]: self.local_index.remove(dist) self.local_index.add(dist) self.install_egg_scripts(dist) self.installed_projects[dist.key] = dist log.info(self.installation_report(requirement, dist, *info)) if (dist.has_metadata('dependency_links.txt') and not self.no_find_links): self.package_index.add_find_links( dist.get_metadata_lines('dependency_links.txt') ) if not deps and not self.always_copy: return elif requirement is not None and dist.key != requirement.key: log.warn("Skipping dependencies for %s", dist) return # XXX this is not the distribution we were looking for elif requirement is None or dist not in requirement: # if we wound up with a different version, resolve what we've got distreq = dist.as_requirement() requirement = requirement or distreq requirement = Requirement( distreq.project_name, distreq.specs, requirement.extras ) log.info("Processing dependencies for %s", requirement) try: distros = WorkingSet([]).resolve( [requirement], self.local_index, self.easy_install ) except DistributionNotFound: e = sys.exc_info()[1] raise DistutilsError( "Could not find required distribution %s" % e.args ) except VersionConflict: e = sys.exc_info()[1] raise DistutilsError( "Installed distribution %s conflicts with requirement %s" % e.args ) if self.always_copy or self.always_copy_from: # Force all the relevant distros to be copied or activated for dist in distros: if dist.key not in self.installed_projects: self.easy_install(dist.as_requirement()) log.info("Finished processing dependencies for %s", requirement) def should_unzip(self, dist): if self.zip_ok is not None: return not self.zip_ok if dist.has_metadata('not-zip-safe'): return True if not dist.has_metadata('zip-safe'): return True return False def maybe_move(self, spec, dist_filename, setup_base): dst = os.path.join(self.build_directory, spec.key) if os.path.exists(dst): msg = ("%r already exists in %s; build directory %s will not be " "kept") log.warn(msg, spec.key, self.build_directory, setup_base) return setup_base if os.path.isdir(dist_filename): setup_base = dist_filename else: if os.path.dirname(dist_filename) == setup_base: os.unlink(dist_filename) # get it out of the tmp dir contents = os.listdir(setup_base) if len(contents) == 1: dist_filename = os.path.join(setup_base, contents[0]) if os.path.isdir(dist_filename): # if the only thing there is a directory, move it instead setup_base = dist_filename ensure_directory(dst) shutil.move(setup_base, dst) return dst def install_wrapper_scripts(self, dist): if not self.exclude_scripts: for args in get_script_args(dist): self.write_script(*args) def install_script(self, dist, script_name, script_text, dev_path=None): """Generate a legacy script wrapper and install it""" spec = str(dist.as_requirement()) is_script = is_python_script(script_text, script_name) if is_script: script_text = (get_script_header(script_text) + self._load_template(dev_path) % locals()) self.write_script(script_name, _to_ascii(script_text), 'b') @staticmethod def _load_template(dev_path): """ There are a couple of template scripts in the package. This function loads one of them and prepares it for use. """ # See https://bitbucket.org/pypa/setuptools/issue/134 for info # on script file naming and downstream issues with SVR4 name = 'script.tmpl' if dev_path: name = name.replace('.tmpl', ' (dev).tmpl') raw_bytes = resource_string('setuptools', name) return raw_bytes.decode('utf-8') def write_script(self, script_name, contents, mode="t", blockers=()): """Write an executable file to the scripts directory""" self.delete_blockers( # clean up old .py/.pyw w/o a script [os.path.join(self.script_dir, x) for x in blockers] ) log.info("Installing %s script to %s", script_name, self.script_dir) target = os.path.join(self.script_dir, script_name) self.add_output(target) mask = current_umask() if not self.dry_run: ensure_directory(target) if os.path.exists(target): os.unlink(target) f = open(target, "w" + mode) f.write(contents) f.close() chmod(target, 0o777 - mask) def install_eggs(self, spec, dist_filename, tmpdir): # .egg dirs or files are already built, so just return them if dist_filename.lower().endswith('.egg'): return [self.install_egg(dist_filename, tmpdir)] elif dist_filename.lower().endswith('.exe'): return [self.install_exe(dist_filename, tmpdir)] # Anything else, try to extract and build setup_base = tmpdir if os.path.isfile(dist_filename) and not dist_filename.endswith('.py'): unpack_archive(dist_filename, tmpdir, self.unpack_progress) elif os.path.isdir(dist_filename): setup_base = os.path.abspath(dist_filename) if (setup_base.startswith(tmpdir) # something we downloaded and self.build_directory and spec is not None): setup_base = self.maybe_move(spec, dist_filename, setup_base) # Find the setup.py file setup_script = os.path.join(setup_base, 'setup.py') if not os.path.exists(setup_script): setups = glob(os.path.join(setup_base, '*', 'setup.py')) if not setups: raise DistutilsError( "Couldn't find a setup script in %s" % os.path.abspath(dist_filename) ) if len(setups) > 1: raise DistutilsError( "Multiple setup scripts in %s" % os.path.abspath(dist_filename) ) setup_script = setups[0] # Now run it, and return the result if self.editable: log.info(self.report_editable(spec, setup_script)) return [] else: return self.build_and_install(setup_script, setup_base) def egg_distribution(self, egg_path): if os.path.isdir(egg_path): metadata = PathMetadata(egg_path, os.path.join(egg_path, 'EGG-INFO')) else: metadata = EggMetadata(zipimport.zipimporter(egg_path)) return Distribution.from_filename(egg_path, metadata=metadata) def install_egg(self, egg_path, tmpdir): destination = os.path.join(self.install_dir, os.path.basename(egg_path)) destination = os.path.abspath(destination) if not self.dry_run: ensure_directory(destination) dist = self.egg_distribution(egg_path) if not samefile(egg_path, destination): if os.path.isdir(destination) and not os.path.islink(destination): dir_util.remove_tree(destination, dry_run=self.dry_run) elif os.path.exists(destination): self.execute(os.unlink, (destination,), "Removing " + destination) try: new_dist_is_zipped = False if os.path.isdir(egg_path): if egg_path.startswith(tmpdir): f, m = shutil.move, "Moving" else: f, m = shutil.copytree, "Copying" elif self.should_unzip(dist): self.mkpath(destination) f, m = self.unpack_and_compile, "Extracting" else: new_dist_is_zipped = True if egg_path.startswith(tmpdir): f, m = shutil.move, "Moving" else: f, m = shutil.copy2, "Copying" self.execute(f, (egg_path, destination), (m + " %s to %s") % (os.path.basename(egg_path), os.path.dirname(destination))) update_dist_caches(destination, fix_zipimporter_caches=new_dist_is_zipped) except: update_dist_caches(destination, fix_zipimporter_caches=False) raise self.add_output(destination) return self.egg_distribution(destination) def install_exe(self, dist_filename, tmpdir): # See if it's valid, get data cfg = extract_wininst_cfg(dist_filename) if cfg is None: raise DistutilsError( "%s is not a valid distutils Windows .exe" % dist_filename ) # Create a dummy distribution object until we build the real distro dist = Distribution( None, project_name=cfg.get('metadata', 'name'), version=cfg.get('metadata', 'version'), platform=get_platform(), ) # Convert the .exe to an unpacked egg egg_path = dist.location = os.path.join(tmpdir, dist.egg_name() + '.egg') egg_tmp = egg_path + '.tmp' _egg_info = os.path.join(egg_tmp, 'EGG-INFO') pkg_inf = os.path.join(_egg_info, 'PKG-INFO') ensure_directory(pkg_inf) # make sure EGG-INFO dir exists dist._provider = PathMetadata(egg_tmp, _egg_info) # XXX self.exe_to_egg(dist_filename, egg_tmp) # Write EGG-INFO/PKG-INFO if not os.path.exists(pkg_inf): f = open(pkg_inf, 'w') f.write('Metadata-Version: 1.0\n') for k, v in cfg.items('metadata'): if k != 'target_version': f.write('%s: %s\n' % (k.replace('_', '-').title(), v)) f.close() script_dir = os.path.join(_egg_info, 'scripts') self.delete_blockers( # delete entry-point scripts to avoid duping [os.path.join(script_dir, args[0]) for args in get_script_args(dist)] ) # Build .egg file from tmpdir bdist_egg.make_zipfile( egg_path, egg_tmp, verbose=self.verbose, dry_run=self.dry_run ) # install the .egg return self.install_egg(egg_path, tmpdir) def exe_to_egg(self, dist_filename, egg_tmp): """Extract a bdist_wininst to the directories an egg would use""" # Check for .pth file and set up prefix translations prefixes = get_exe_prefixes(dist_filename) to_compile = [] native_libs = [] top_level = {} def process(src, dst): s = src.lower() for old, new in prefixes: if s.startswith(old): src = new + src[len(old):] parts = src.split('/') dst = os.path.join(egg_tmp, *parts) dl = dst.lower() if dl.endswith('.pyd') or dl.endswith('.dll'): parts[-1] = bdist_egg.strip_module(parts[-1]) top_level[os.path.splitext(parts[0])[0]] = 1 native_libs.append(src) elif dl.endswith('.py') and old != 'SCRIPTS/': top_level[os.path.splitext(parts[0])[0]] = 1 to_compile.append(dst) return dst if not src.endswith('.pth'): log.warn("WARNING: can't process %s", src) return None # extract, tracking .pyd/.dll->native_libs and .py -> to_compile unpack_archive(dist_filename, egg_tmp, process) stubs = [] for res in native_libs: if res.lower().endswith('.pyd'): # create stubs for .pyd's parts = res.split('/') resource = parts[-1] parts[-1] = bdist_egg.strip_module(parts[-1]) + '.py' pyfile = os.path.join(egg_tmp, *parts) to_compile.append(pyfile) stubs.append(pyfile) bdist_egg.write_stub(resource, pyfile) self.byte_compile(to_compile) # compile .py's bdist_egg.write_safety_flag( os.path.join(egg_tmp, 'EGG-INFO'), bdist_egg.analyze_egg(egg_tmp, stubs)) # write zip-safety flag for name in 'top_level', 'native_libs': if locals()[name]: txt = os.path.join(egg_tmp, 'EGG-INFO', name + '.txt') if not os.path.exists(txt): f = open(txt, 'w') f.write('\n'.join(locals()[name]) + '\n') f.close() def installation_report(self, req, dist, what="Installed"): """Helpful installation message for display to package users""" msg = "\n%(what)s %(eggloc)s%(extras)s" if self.multi_version and not self.no_report: msg += """ Because this distribution was installed --multi-version, before you can import modules from this package in an application, you will need to 'import pkg_resources' and then use a 'require()' call similar to one of these examples, in order to select the desired version: pkg_resources.require("%(name)s") # latest installed version pkg_resources.require("%(name)s==%(version)s") # this exact version pkg_resources.require("%(name)s>=%(version)s") # this version or higher """ if self.install_dir not in map(normalize_path, sys.path): msg += """ Note also that the installation directory must be on sys.path at runtime for this to work. (e.g. by being the application's script directory, by being on PYTHONPATH, or by being added to sys.path by your code.) """ eggloc = dist.location name = dist.project_name version = dist.version extras = '' # TODO: self.report_extras(req, dist) return msg % locals() def report_editable(self, spec, setup_script): dirname = os.path.dirname(setup_script) python = sys.executable return """\nExtracted editable version of %(spec)s to %(dirname)s If it uses setuptools in its setup script, you can activate it in "development" mode by going to that directory and running:: %(python)s setup.py develop See the setuptools documentation for the "develop" command for more info. """ % locals() def run_setup(self, setup_script, setup_base, args): sys.modules.setdefault('distutils.command.bdist_egg', bdist_egg) sys.modules.setdefault('distutils.command.egg_info', egg_info) args = list(args) if self.verbose > 2: v = 'v' * (self.verbose - 1) args.insert(0, '-' + v) elif self.verbose < 2: args.insert(0, '-q') if self.dry_run: args.insert(0, '-n') log.info( "Running %s %s", setup_script[len(setup_base) + 1:], ' '.join(args) ) try: run_setup(setup_script, args) except SystemExit: v = sys.exc_info()[1] raise DistutilsError("Setup script exited with %s" % (v.args[0],)) def build_and_install(self, setup_script, setup_base): args = ['bdist_egg', '--dist-dir'] dist_dir = tempfile.mkdtemp( prefix='egg-dist-tmp-', dir=os.path.dirname(setup_script) ) try: self._set_fetcher_options(os.path.dirname(setup_script)) args.append(dist_dir) self.run_setup(setup_script, setup_base, args) all_eggs = Environment([dist_dir]) eggs = [] for key in all_eggs: for dist in all_eggs[key]: eggs.append(self.install_egg(dist.location, setup_base)) if not eggs and not self.dry_run: log.warn("No eggs found in %s (setup script problem?)", dist_dir) return eggs finally: rmtree(dist_dir) log.set_verbosity(self.verbose) # restore our log verbosity def _set_fetcher_options(self, base): """ When easy_install is about to run bdist_egg on a source dist, that source dist might have 'setup_requires' directives, requiring additional fetching. Ensure the fetcher options given to easy_install are available to that command as well. """ # find the fetch options from easy_install and write them out # to the setup.cfg file. ei_opts = self.distribution.get_option_dict('easy_install').copy() fetch_directives = ( 'find_links', 'site_dirs', 'index_url', 'optimize', 'site_dirs', 'allow_hosts', ) fetch_options = {} for key, val in ei_opts.items(): if key not in fetch_directives: continue fetch_options[key.replace('_', '-')] = val[1] # create a settings dictionary suitable for `edit_config` settings = dict(easy_install=fetch_options) cfg_filename = os.path.join(base, 'setup.cfg') setopt.edit_config(cfg_filename, settings) def update_pth(self, dist): if self.pth_file is None: return for d in self.pth_file[dist.key]: # drop old entries if self.multi_version or d.location != dist.location: log.info("Removing %s from easy-install.pth file", d) self.pth_file.remove(d) if d.location in self.shadow_path: self.shadow_path.remove(d.location) if not self.multi_version: if dist.location in self.pth_file.paths: log.info( "%s is already the active version in easy-install.pth", dist ) else: log.info("Adding %s to easy-install.pth file", dist) self.pth_file.add(dist) # add new entry if dist.location not in self.shadow_path: self.shadow_path.append(dist.location) if not self.dry_run: self.pth_file.save() if dist.key == 'setuptools': # Ensure that setuptools itself never becomes unavailable! # XXX should this check for latest version? filename = os.path.join(self.install_dir, 'setuptools.pth') if os.path.islink(filename): os.unlink(filename) f = open(filename, 'wt') f.write(self.pth_file.make_relative(dist.location) + '\n') f.close() def unpack_progress(self, src, dst): # Progress filter for unpacking log.debug("Unpacking %s to %s", src, dst) return dst # only unpack-and-compile skips files for dry run def unpack_and_compile(self, egg_path, destination): to_compile = [] to_chmod = [] def pf(src, dst): if dst.endswith('.py') and not src.startswith('EGG-INFO/'): to_compile.append(dst) elif dst.endswith('.dll') or dst.endswith('.so'): to_chmod.append(dst) self.unpack_progress(src, dst) return not self.dry_run and dst or None unpack_archive(egg_path, destination, pf) self.byte_compile(to_compile) if not self.dry_run: for f in to_chmod: mode = ((os.stat(f)[stat.ST_MODE]) | 0o555) & 0o7755 chmod(f, mode) def byte_compile(self, to_compile): if sys.dont_write_bytecode: self.warn('byte-compiling is disabled, skipping.') return from distutils.util import byte_compile try: # try to make the byte compile messages quieter log.set_verbosity(self.verbose - 1) byte_compile(to_compile, optimize=0, force=1, dry_run=self.dry_run) if self.optimize: byte_compile( to_compile, optimize=self.optimize, force=1, dry_run=self.dry_run ) finally: log.set_verbosity(self.verbose) # restore original verbosity def no_default_version_msg(self): template = """bad install directory or PYTHONPATH You are attempting to install a package to a directory that is not on PYTHONPATH and which Python does not read ".pth" files from. The installation directory you specified (via --install-dir, --prefix, or the distutils default setting) was: %s and your PYTHONPATH environment variable currently contains: %r Here are some of your options for correcting the problem: * You can choose a different installation directory, i.e., one that is on PYTHONPATH or supports .pth files * You can add the installation directory to the PYTHONPATH environment variable. (It must then also be on PYTHONPATH whenever you run Python and want to use the package(s) you are installing.) * You can set up the installation directory to support ".pth" files by using one of the approaches described here: https://pythonhosted.org/setuptools/easy_install.html#custom-installation-locations Please make the appropriate changes for your system and try again.""" return template % (self.install_dir, os.environ.get('PYTHONPATH', '')) def install_site_py(self): """Make sure there's a site.py in the target dir, if needed""" if self.sitepy_installed: return # already did it, or don't need to sitepy = os.path.join(self.install_dir, "site.py") source = resource_string("setuptools", "site-patch.py") current = "" if os.path.exists(sitepy): log.debug("Checking existing site.py in %s", self.install_dir) f = open(sitepy, 'rb') current = f.read() # we want str, not bytes if PY3: current = current.decode() f.close() if not current.startswith('def __boot():'): raise DistutilsError( "%s is not a setuptools-generated site.py; please" " remove it." % sitepy ) if current != source: log.info("Creating %s", sitepy) if not self.dry_run: ensure_directory(sitepy) f = open(sitepy, 'wb') f.write(source) f.close() self.byte_compile([sitepy]) self.sitepy_installed = True def create_home_path(self): """Create directories under ~.""" if not self.user: return home = convert_path(os.path.expanduser("~")) for name, path in iteritems(self.config_vars): if path.startswith(home) and not os.path.isdir(path): self.debug_print("os.makedirs('%s', 0o700)" % path) os.makedirs(path, 0o700) INSTALL_SCHEMES = dict( posix=dict( install_dir='$base/lib/python$py_version_short/site-packages', script_dir='$base/bin', ), ) DEFAULT_SCHEME = dict( install_dir='$base/Lib/site-packages', script_dir='$base/Scripts', ) def _expand(self, *attrs): config_vars = self.get_finalized_command('install').config_vars if self.prefix: # Set default install_dir/scripts from --prefix config_vars = config_vars.copy() config_vars['base'] = self.prefix scheme = self.INSTALL_SCHEMES.get(os.name, self.DEFAULT_SCHEME) for attr, val in scheme.items(): if getattr(self, attr, None) is None: setattr(self, attr, val) from distutils.util import subst_vars for attr in attrs: val = getattr(self, attr) if val is not None: val = subst_vars(val, config_vars) if os.name == 'posix': val = os.path.expanduser(val) setattr(self, attr, val) def get_site_dirs(): # return a list of 'site' dirs sitedirs = [_f for _f in os.environ.get('PYTHONPATH', '').split(os.pathsep) if _f] prefixes = [sys.prefix] if sys.exec_prefix != sys.prefix: prefixes.append(sys.exec_prefix) for prefix in prefixes: if prefix: if sys.platform in ('os2emx', 'riscos'): sitedirs.append(os.path.join(prefix, "Lib", "site-packages")) elif os.sep == '/': sitedirs.extend([os.path.join(prefix, "lib", "python" + sys.version[:3], "site-packages"), os.path.join(prefix, "lib", "site-python")]) else: sitedirs.extend( [prefix, os.path.join(prefix, "lib", "site-packages")] ) if sys.platform == 'darwin': # for framework builds *only* we add the standard Apple # locations. Currently only per-user, but /Library and # /Network/Library could be added too if 'Python.framework' in prefix: home = os.environ.get('HOME') if home: sitedirs.append( os.path.join(home, 'Library', 'Python', sys.version[:3], 'site-packages')) lib_paths = get_path('purelib'), get_path('platlib') for site_lib in lib_paths: if site_lib not in sitedirs: sitedirs.append(site_lib) if site.ENABLE_USER_SITE: sitedirs.append(site.USER_SITE) sitedirs = list(map(normalize_path, sitedirs)) return sitedirs def expand_paths(inputs): """Yield sys.path directories that might contain "old-style" packages""" seen = {} for dirname in inputs: dirname = normalize_path(dirname) if dirname in seen: continue seen[dirname] = 1 if not os.path.isdir(dirname): continue files = os.listdir(dirname) yield dirname, files for name in files: if not name.endswith('.pth'): # We only care about the .pth files continue if name in ('easy-install.pth', 'setuptools.pth'): # Ignore .pth files that we control continue # Read the .pth file f = open(os.path.join(dirname, name)) lines = list(yield_lines(f)) f.close() # Yield existing non-dupe, non-import directory lines from it for line in lines: if not line.startswith("import"): line = normalize_path(line.rstrip()) if line not in seen: seen[line] = 1 if not os.path.isdir(line): continue yield line, os.listdir(line) def extract_wininst_cfg(dist_filename): """Extract configuration data from a bdist_wininst .exe Returns a ConfigParser.RawConfigParser, or None """ f = open(dist_filename, 'rb') try: endrec = zipfile._EndRecData(f) if endrec is None: return None prepended = (endrec[9] - endrec[5]) - endrec[6] if prepended < 12: # no wininst data here return None f.seek(prepended - 12) from setuptools.compat import StringIO, ConfigParser import struct tag, cfglen, bmlen = struct.unpack("<iii", f.read(12)) if tag not in (0x1234567A, 0x1234567B): return None # not a valid tag f.seek(prepended - (12 + cfglen)) cfg = ConfigParser.RawConfigParser( {'version': '', 'target_version': ''}) try: part = f.read(cfglen) # part is in bytes, but we need to read up to the first null # byte. if sys.version_info >= (2, 6): null_byte = bytes([0]) else: null_byte = chr(0) config = part.split(null_byte, 1)[0] # Now the config is in bytes, but for RawConfigParser, it should # be text, so decode it. config = config.decode(sys.getfilesystemencoding()) cfg.readfp(StringIO(config)) except ConfigParser.Error: return None if not cfg.has_section('metadata') or not cfg.has_section('Setup'): return None return cfg finally: f.close() def get_exe_prefixes(exe_filename): """Get exe->egg path translations for a given .exe file""" prefixes = [ ('PURELIB/', ''), ('PLATLIB/pywin32_system32', ''), ('PLATLIB/', ''), ('SCRIPTS/', 'EGG-INFO/scripts/'), ('DATA/lib/site-packages', ''), ] z = zipfile.ZipFile(exe_filename) try: for info in z.infolist(): name = info.filename parts = name.split('/') if len(parts) == 3 and parts[2] == 'PKG-INFO': if parts[1].endswith('.egg-info'): prefixes.insert(0, ('/'.join(parts[:2]), 'EGG-INFO/')) break if len(parts) != 2 or not name.endswith('.pth'): continue if name.endswith('-nspkg.pth'): continue if parts[0].upper() in ('PURELIB', 'PLATLIB'): contents = z.read(name) if PY3: contents = contents.decode() for pth in yield_lines(contents): pth = pth.strip().replace('\\', '/') if not pth.startswith('import'): prefixes.append((('%s/%s/' % (parts[0], pth)), '')) finally: z.close() prefixes = [(x.lower(), y) for x, y in prefixes] prefixes.sort() prefixes.reverse() return prefixes def parse_requirement_arg(spec): try: return Requirement.parse(spec) except ValueError: raise DistutilsError( "Not a URL, existing file, or requirement spec: %r" % (spec,) ) class PthDistributions(Environment): """A .pth file with Distribution paths in it""" dirty = False def __init__(self, filename, sitedirs=()): self.filename = filename self.sitedirs = list(map(normalize_path, sitedirs)) self.basedir = normalize_path(os.path.dirname(self.filename)) self._load() Environment.__init__(self, [], None, None) for path in yield_lines(self.paths): list(map(self.add, find_distributions(path, True))) def _load(self): self.paths = [] saw_import = False seen = dict.fromkeys(self.sitedirs) if os.path.isfile(self.filename): f = open(self.filename, 'rt') for line in f: if line.startswith('import'): saw_import = True continue path = line.rstrip() self.paths.append(path) if not path.strip() or path.strip().startswith('#'): continue # skip non-existent paths, in case somebody deleted a package # manually, and duplicate paths as well path = self.paths[-1] = normalize_path( os.path.join(self.basedir, path) ) if not os.path.exists(path) or path in seen: self.paths.pop() # skip it self.dirty = True # we cleaned up, so we're dirty now :) continue seen[path] = 1 f.close() if self.paths and not saw_import: self.dirty = True # ensure anything we touch has import wrappers while self.paths and not self.paths[-1].strip(): self.paths.pop() def save(self): """Write changed .pth file back to disk""" if not self.dirty: return data = '\n'.join(map(self.make_relative, self.paths)) if data: log.debug("Saving %s", self.filename) data = ( "import sys; sys.__plen = len(sys.path)\n" "%s\n" "import sys; new=sys.path[sys.__plen:];" " del sys.path[sys.__plen:];" " p=getattr(sys,'__egginsert',0); sys.path[p:p]=new;" " sys.__egginsert = p+len(new)\n" ) % data if os.path.islink(self.filename): os.unlink(self.filename) f = open(self.filename, 'wt') f.write(data) f.close() elif os.path.exists(self.filename): log.debug("Deleting empty %s", self.filename) os.unlink(self.filename) self.dirty = False def add(self, dist): """Add `dist` to the distribution map""" new_path = ( dist.location not in self.paths and ( dist.location not in self.sitedirs or # account for '.' being in PYTHONPATH dist.location == os.getcwd() ) ) if new_path: self.paths.append(dist.location) self.dirty = True Environment.add(self, dist) def remove(self, dist): """Remove `dist` from the distribution map""" while dist.location in self.paths: self.paths.remove(dist.location) self.dirty = True Environment.remove(self, dist) def make_relative(self, path): npath, last = os.path.split(normalize_path(path)) baselen = len(self.basedir) parts = [last] sep = os.altsep == '/' and '/' or os.sep while len(npath) >= baselen: if npath == self.basedir: parts.append(os.curdir) parts.reverse() return sep.join(parts) npath, last = os.path.split(npath) parts.append(last) else: return path def _first_line_re(): """ Return a regular expression based on first_line_re suitable for matching strings. """ if isinstance(first_line_re.pattern, str): return first_line_re # first_line_re in Python >=3.1.4 and >=3.2.1 is a bytes pattern. return re.compile(first_line_re.pattern.decode()) def get_script_header(script_text, executable=sys_executable, wininst=False): """Create a #! line, getting options (if any) from script_text""" first = (script_text + '\n').splitlines()[0] match = _first_line_re().match(first) options = '' if match: options = match.group(1) or '' if options: options = ' ' + options if wininst: executable = "python.exe" else: executable = nt_quote_arg(executable) hdr = "#!%(executable)s%(options)s\n" % locals() if not isascii(hdr): # Non-ascii path to sys.executable, use -x to prevent warnings if options: if options.strip().startswith('-'): options = ' -x' + options.strip()[1:] # else: punt, we can't do it, let the warning happen anyway else: options = ' -x' executable = fix_jython_executable(executable, options) hdr = "#!%(executable)s%(options)s\n" % locals() return hdr def auto_chmod(func, arg, exc): if func is os.remove and os.name == 'nt': chmod(arg, stat.S_IWRITE) return func(arg) et, ev, _ = sys.exc_info() reraise(et, (ev[0], ev[1] + (" %s %s" % (func, arg)))) def update_dist_caches(dist_path, fix_zipimporter_caches): """ Fix any globally cached `dist_path` related data `dist_path` should be a path of a newly installed egg distribution (zipped or unzipped). sys.path_importer_cache contains finder objects that have been cached when importing data from the original distribution. Any such finders need to be cleared since the replacement distribution might be packaged differently, e.g. a zipped egg distribution might get replaced with an unzipped egg folder or vice versa. Having the old finders cached may then cause Python to attempt loading modules from the replacement distribution using an incorrect loader. zipimport.zipimporter objects are Python loaders charged with importing data packaged inside zip archives. If stale loaders referencing the original distribution, are left behind, they can fail to load modules from the replacement distribution. E.g. if an old zipimport.zipimporter instance is used to load data from a new zipped egg archive, it may cause the operation to attempt to locate the requested data in the wrong location - one indicated by the original distribution's zip archive directory information. Such an operation may then fail outright, e.g. report having read a 'bad local file header', or even worse, it may fail silently & return invalid data. zipimport._zip_directory_cache contains cached zip archive directory information for all existing zipimport.zipimporter instances and all such instances connected to the same archive share the same cached directory information. If asked, and the underlying Python implementation allows it, we can fix all existing zipimport.zipimporter instances instead of having to track them down and remove them one by one, by updating their shared cached zip archive directory information. This, of course, assumes that the replacement distribution is packaged as a zipped egg. If not asked to fix existing zipimport.zipimporter instances, we still do our best to clear any remaining zipimport.zipimporter related cached data that might somehow later get used when attempting to load data from the new distribution and thus cause such load operations to fail. Note that when tracking down such remaining stale data, we can not catch every conceivable usage from here, and we clear only those that we know of and have found to cause problems if left alive. Any remaining caches should be updated by whomever is in charge of maintaining them, i.e. they should be ready to handle us replacing their zip archives with new distributions at runtime. """ # There are several other known sources of stale zipimport.zipimporter # instances that we do not clear here, but might if ever given a reason to # do so: # * Global setuptools pkg_resources.working_set (a.k.a. 'master working # set') may contain distributions which may in turn contain their # zipimport.zipimporter loaders. # * Several zipimport.zipimporter loaders held by local variables further # up the function call stack when running the setuptools installation. # * Already loaded modules may have their __loader__ attribute set to the # exact loader instance used when importing them. Python 3.4 docs state # that this information is intended mostly for introspection and so is # not expected to cause us problems. normalized_path = normalize_path(dist_path) _uncache(normalized_path, sys.path_importer_cache) if fix_zipimporter_caches: _replace_zip_directory_cache_data(normalized_path) else: # Here, even though we do not want to fix existing and now stale # zipimporter cache information, we still want to remove it. Related to # Python's zip archive directory information cache, we clear each of # its stale entries in two phases: # 1. Clear the entry so attempting to access zip archive information # via any existing stale zipimport.zipimporter instances fails. # 2. Remove the entry from the cache so any newly constructed # zipimport.zipimporter instances do not end up using old stale # zip archive directory information. # This whole stale data removal step does not seem strictly necessary, # but has been left in because it was done before we started replacing # the zip archive directory information cache content if possible, and # there are no relevant unit tests that we can depend on to tell us if # this is really needed. _remove_and_clear_zip_directory_cache_data(normalized_path) def _collect_zipimporter_cache_entries(normalized_path, cache): """ Return zipimporter cache entry keys related to a given normalized path. Alternative path spellings (e.g. those using different character case or those using alternative path separators) related to the same path are included. Any sub-path entries are included as well, i.e. those corresponding to zip archives embedded in other zip archives. """ result = [] prefix_len = len(normalized_path) for p in cache: np = normalize_path(p) if (np.startswith(normalized_path) and np[prefix_len:prefix_len + 1] in (os.sep, '')): result.append(p) return result def _update_zipimporter_cache(normalized_path, cache, updater=None): """ Update zipimporter cache data for a given normalized path. Any sub-path entries are processed as well, i.e. those corresponding to zip archives embedded in other zip archives. Given updater is a callable taking a cache entry key and the original entry (after already removing the entry from the cache), and expected to update the entry and possibly return a new one to be inserted in its place. Returning None indicates that the entry should not be replaced with a new one. If no updater is given, the cache entries are simply removed without any additional processing, the same as if the updater simply returned None. """ for p in _collect_zipimporter_cache_entries(normalized_path, cache): # N.B. pypy's custom zipimport._zip_directory_cache implementation does # not support the complete dict interface: # * Does not support item assignment, thus not allowing this function # to be used only for removing existing cache entries. # * Does not support the dict.pop() method, forcing us to use the # get/del patterns instead. For more detailed information see the # following links: # https://bitbucket.org/pypa/setuptools/issue/202/more-robust-zipimporter-cache-invalidation#comment-10495960 # https://bitbucket.org/pypy/pypy/src/dd07756a34a41f674c0cacfbc8ae1d4cc9ea2ae4/pypy/module/zipimport/interp_zipimport.py#cl-99 old_entry = cache[p] del cache[p] new_entry = updater and updater(p, old_entry) if new_entry is not None: cache[p] = new_entry def _uncache(normalized_path, cache): _update_zipimporter_cache(normalized_path, cache) def _remove_and_clear_zip_directory_cache_data(normalized_path): def clear_and_remove_cached_zip_archive_directory_data(path, old_entry): old_entry.clear() _update_zipimporter_cache( normalized_path, zipimport._zip_directory_cache, updater=clear_and_remove_cached_zip_archive_directory_data) # PyPy Python implementation does not allow directly writing to the # zipimport._zip_directory_cache and so prevents us from attempting to correct # its content. The best we can do there is clear the problematic cache content # and have PyPy repopulate it as needed. The downside is that if there are any # stale zipimport.zipimporter instances laying around, attempting to use them # will fail due to not having its zip archive directory information available # instead of being automatically corrected to use the new correct zip archive # directory information. if '__pypy__' in sys.builtin_module_names: _replace_zip_directory_cache_data = \ _remove_and_clear_zip_directory_cache_data else: def _replace_zip_directory_cache_data(normalized_path): def replace_cached_zip_archive_directory_data(path, old_entry): # N.B. In theory, we could load the zip directory information just # once for all updated path spellings, and then copy it locally and # update its contained path strings to contain the correct # spelling, but that seems like a way too invasive move (this cache # structure is not officially documented anywhere and could in # theory change with new Python releases) for no significant # benefit. old_entry.clear() zipimport.zipimporter(path) old_entry.update(zipimport._zip_directory_cache[path]) return old_entry _update_zipimporter_cache( normalized_path, zipimport._zip_directory_cache, updater=replace_cached_zip_archive_directory_data) def is_python(text, filename='<string>'): "Is this string a valid Python script?" try: compile(text, filename, 'exec') except (SyntaxError, TypeError): return False else: return True def is_sh(executable): """Determine if the specified executable is a .sh (contains a #! line)""" try: fp = open(executable) magic = fp.read(2) fp.close() except (OSError, IOError): return executable return magic == '#!' def nt_quote_arg(arg): """Quote a command line argument according to Windows parsing rules""" result = [] needquote = False nb = 0 needquote = (" " in arg) or ("\t" in arg) if needquote: result.append('"') for c in arg: if c == '\\': nb += 1 elif c == '"': # double preceding backslashes, then add a \" result.append('\\' * (nb * 2) + '\\"') nb = 0 else: if nb: result.append('\\' * nb) nb = 0 result.append(c) if nb: result.append('\\' * nb) if needquote: result.append('\\' * nb) # double the trailing backslashes result.append('"') return ''.join(result) def is_python_script(script_text, filename): """Is this text, as a whole, a Python script? (as opposed to shell/bat/etc. """ if filename.endswith('.py') or filename.endswith('.pyw'): return True # extension says it's Python if is_python(script_text, filename): return True # it's syntactically valid Python if script_text.startswith('#!'): # It begins with a '#!' line, so check if 'python' is in it somewhere return 'python' in script_text.splitlines()[0].lower() return False # Not any Python I can recognize try: from os import chmod as _chmod except ImportError: # Jython compatibility def _chmod(*args): pass def chmod(path, mode): log.debug("changing mode of %s to %o", path, mode) try: _chmod(path, mode) except os.error: e = sys.exc_info()[1] log.debug("chmod failed: %s", e) def fix_jython_executable(executable, options): if sys.platform.startswith('java') and is_sh(executable): # Workaround for Jython is not needed on Linux systems. import java if java.lang.System.getProperty("os.name") == "Linux": return executable # Workaround Jython's sys.executable being a .sh (an invalid # shebang line interpreter) if options: # Can't apply the workaround, leave it broken log.warn( "WARNING: Unable to adapt shebang line for Jython," " the following script is NOT executable\n" " see http://bugs.jython.org/issue1112 for" " more information.") else: return '/usr/bin/env %s' % executable return executable class ScriptWriter(object): """ Encapsulates behavior around writing entry point scripts for console and gui apps. """ template = textwrap.dedent(""" # EASY-INSTALL-ENTRY-SCRIPT: %(spec)r,%(group)r,%(name)r __requires__ = %(spec)r import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.exit( load_entry_point(%(spec)r, %(group)r, %(name)r)() ) """).lstrip() @classmethod def get_script_args(cls, dist, executable=sys_executable, wininst=False): """ Yield write_script() argument tuples for a distribution's entrypoints """ gen_class = cls.get_writer(wininst) spec = str(dist.as_requirement()) header = get_script_header("", executable, wininst) for type_ in 'console', 'gui': group = type_ + '_scripts' for name, ep in dist.get_entry_map(group).items(): script_text = gen_class.template % locals() for res in gen_class._get_script_args(type_, name, header, script_text): yield res @classmethod def get_writer(cls, force_windows): if force_windows or sys.platform == 'win32': return WindowsScriptWriter.get_writer() return cls @classmethod def _get_script_args(cls, type_, name, header, script_text): # Simply write the stub with no extension. yield (name, header + script_text) class WindowsScriptWriter(ScriptWriter): @classmethod def get_writer(cls): """ Get a script writer suitable for Windows """ writer_lookup = dict( executable=WindowsExecutableLauncherWriter, natural=cls, ) # for compatibility, use the executable launcher by default launcher = os.environ.get('SETUPTOOLS_LAUNCHER', 'executable') return writer_lookup[launcher] @classmethod def _get_script_args(cls, type_, name, header, script_text): "For Windows, add a .py extension" ext = dict(console='.pya', gui='.pyw')[type_] if ext not in os.environ['PATHEXT'].lower().split(';'): warnings.warn("%s not listed in PATHEXT; scripts will not be " "recognized as executables." % ext, UserWarning) old = ['.pya', '.py', '-script.py', '.pyc', '.pyo', '.pyw', '.exe'] old.remove(ext) header = cls._adjust_header(type_, header) blockers = [name + x for x in old] yield name + ext, header + script_text, 't', blockers @staticmethod def _adjust_header(type_, orig_header): """ Make sure 'pythonw' is used for gui and and 'python' is used for console (regardless of what sys.executable is). """ pattern = 'pythonw.exe' repl = 'python.exe' if type_ == 'gui': pattern, repl = repl, pattern pattern_ob = re.compile(re.escape(pattern), re.IGNORECASE) new_header = pattern_ob.sub(string=orig_header, repl=repl) clean_header = new_header[2:-1].strip('"') if sys.platform == 'win32' and not os.path.exists(clean_header): # the adjusted version doesn't exist, so return the original return orig_header return new_header class WindowsExecutableLauncherWriter(WindowsScriptWriter): @classmethod def _get_script_args(cls, type_, name, header, script_text): """ For Windows, add a .py extension and an .exe launcher """ if type_ == 'gui': launcher_type = 'gui' ext = '-script.pyw' old = ['.pyw'] else: launcher_type = 'cli' ext = '-script.py' old = ['.py', '.pyc', '.pyo'] hdr = cls._adjust_header(type_, header) blockers = [name + x for x in old] yield (name + ext, hdr + script_text, 't', blockers) yield ( name + '.exe', get_win_launcher(launcher_type), 'b' # write in binary mode ) if not is_64bit(): # install a manifest for the launcher to prevent Windows # from detecting it as an installer (which it will for # launchers like easy_install.exe). Consider only # adding a manifest for launchers detected as installers. # See Distribute #143 for details. m_name = name + '.exe.manifest' yield (m_name, load_launcher_manifest(name), 't') # for backward-compatibility get_script_args = ScriptWriter.get_script_args def get_win_launcher(type): """ Load the Windows launcher (executable) suitable for launching a script. `type` should be either 'cli' or 'gui' Returns the executable as a byte string. """ launcher_fn = '%s.exe' % type if platform.machine().lower() == 'arm': launcher_fn = launcher_fn.replace(".", "-arm.") if is_64bit(): launcher_fn = launcher_fn.replace(".", "-64.") else: launcher_fn = launcher_fn.replace(".", "-32.") return resource_string('setuptools', launcher_fn) def load_launcher_manifest(name): manifest = pkg_resources.resource_string(__name__, 'launcher manifest.xml') if PY2: return manifest % vars() else: return manifest.decode('utf-8') % vars() def rmtree(path, ignore_errors=False, onerror=auto_chmod): """Recursively delete a directory tree. This code is taken from the Python 2.4 version of 'shutil', because the 2.3 version doesn't really work right. """ if ignore_errors: def onerror(*args): pass elif onerror is None: def onerror(*args): raise names = [] try: names = os.listdir(path) except os.error: onerror(os.listdir, path, sys.exc_info()) for name in names: fullname = os.path.join(path, name) try: mode = os.lstat(fullname).st_mode except os.error: mode = 0 if stat.S_ISDIR(mode): rmtree(fullname, ignore_errors, onerror) else: try: os.remove(fullname) except os.error: onerror(os.remove, fullname, sys.exc_info()) try: os.rmdir(path) except os.error: onerror(os.rmdir, path, sys.exc_info()) def current_umask(): tmp = os.umask(0o022) os.umask(tmp) return tmp def bootstrap(): # This function is called when setuptools*.egg is run using /bin/sh import setuptools argv0 = os.path.dirname(setuptools.__path__[0]) sys.argv[0] = argv0 sys.argv.append(argv0) main() def main(argv=None, **kw): from setuptools import setup from setuptools.dist import Distribution class DistributionWithoutHelpCommands(Distribution): common_usage = "" def _show_help(self, *args, **kw): with _patch_usage(): Distribution._show_help(self, *args, **kw) if argv is None: argv = sys.argv[1:] with _patch_usage(): setup( script_args=['-q', 'easy_install', '-v'] + argv, script_name=sys.argv[0] or 'easy_install', distclass=DistributionWithoutHelpCommands, **kw ) @contextlib.contextmanager def _patch_usage(): import distutils.core USAGE = textwrap.dedent(""" usage: %(script)s [options] requirement_or_url ... or: %(script)s --help """).lstrip() def gen_usage(script_name): return USAGE % dict( script=os.path.basename(script_name), ) saved = distutils.core.gen_usage distutils.core.gen_usage = gen_usage try: yield finally: distutils.core.gen_usage = saved
py
1a2f56cf093b461b930e93a13c1778c2f4a7b13d
import matplotlib.pyplot as plt import math, cv2, glob import numpy as np class ImageVisualizer: # region vars # endregion def __init__(self, shape): self.shape = shape # to get a name of parameter variable to print a name in subplot def param_to_str(self, obj, namespace): return [name for name in namespace if namespace[name] is obj] # tool to draw multiple images on plot def plot_images(self, images, descriptions = None, invert_colors=False): w = self.shape[0] h = self.shape[1] fig = plt.figure(figsize=(8, 8)) total_images = len(images) rows = math.ceil(total_images/3) columns = math.ceil(total_images/rows) for i in range(1, total_images + 1): fig.add_subplot(rows, columns, i) if invert_colors and len(images[i-1].shape) > 2: plt.imshow(cv2.cvtColor(images[i-1], cv2.COLOR_BGR2RGB)) # need to invert colors else: plt.imshow(images[i-1]) # do not need to invert colors when showing binary images #plt.gca().set_title(self.param_to_str(images[i-1], globals())) if not descriptions is None: plt.gca().set_title(descriptions[i-1]) mng = plt.get_current_fig_manager() # to maximize window mng.window.state('zoomed') plt.show() def get_image_info(self, image, alias ): print(f'Image {alias} shape is: {image.shape}') # tests solid green area on blank image def draw_solid_area_on_blank_image(self, array_of_points): blank_image = np.zeros(self.shape, dtype=np.uint8) cv2.fillPoly(blank_image, pts = [array_of_points],color = (255,255,0)) return blank_image def overlay_image_with_solid_area(self, main_image, image2): return cv2.addWeighted(main_image, 0.8, image2, 0.2, 0)
py
1a2f570b022acc9b4789b5e3b9b14ecd5de70a22
# coding=utf-8 # 异常处理 try: file = open("hejeffery.txt", "r") except IOError: print "没有找到文件" else: print "找到文件了" try: file = open("hejeffery.txt", "r") except IOError: print "没有找到文件" finally: print "始终都会执行这句话"
py
1a2f5942ee004989816d672fa57d7dab25fe65a7
import torch from tqdm import tqdm from utils import Logger, AverageMeter, accuracy import numpy as np def train(trainloader, model, criterion, optimizer): # switch to train mode model.train() losses = AverageMeter() top1 = AverageMeter() top3 = AverageMeter() top5 = AverageMeter() criterion.reset() bar = tqdm(enumerate(trainloader), total=len(trainloader)) for batch_idx, (inputs, targets) in bar: inputs, targets = inputs.cuda(), targets.cuda() # compute output outputs = model(inputs) loss_dict = criterion(outputs, targets) loss = loss_dict[0]["loss"] # measure accuracy and record loss prec1, prec3, prec5 = accuracy(outputs.data, targets.data, topk=(1, 3, 5)) losses.update(loss.item(), inputs.size(0)) top1.update(prec1.item(), inputs.size(0)) top3.update(prec3.item(), inputs.size(0)) top5.update(prec5.item(), inputs.size(0)) # compute gradient and do SGD step optimizer.zero_grad() loss.backward() optimizer.step() # plot progress bar.set_postfix_str('({batch}/{size}) Loss: {loss:.8f} | top1: {top1: .4f} | top3: {top3: .4f} | top5: {top5: .4f}'.format( batch=batch_idx + 1, size=len(trainloader), loss=losses.avg, top1=top1.avg, top3=top3.avg, top5=top5.avg, )) return (losses.avg, top1.avg, top3.avg, top5.avg) @torch.no_grad() def test(testloader, model, criterion, ece_criterion, sce_criterion, T=1.0): criterion.reset() ece_criterion.reset() sce_criterion.reset() losses = AverageMeter() top1 = AverageMeter() top3 = AverageMeter() top5 = AverageMeter() # switch to evaluate mode model.eval() bar = tqdm(enumerate(testloader), total=len(testloader)) for batch_idx, (inputs, targets) in bar: inputs, targets = inputs.cuda(), targets.cuda() # compute output outputs = model(inputs) outputs /= T loss_dict = criterion(outputs, targets) loss = loss_dict[0]["loss"] ece_criterion.forward(outputs,targets) sce_criterion.forward(outputs,targets) prec1, prec3, prec5 = accuracy(outputs.data, targets.data, topk=(1, 3, 5)) losses.update(loss.item(), inputs.size(0)) top1.update(prec1.item(), inputs.size(0)) top3.update(prec3.item(), inputs.size(0)) top5.update(prec5.item(), inputs.size(0)) # plot progress bar.set_postfix_str('({batch}/{size}) Loss: {loss:.8f} | top1: {top1: .4f} | top3: {top3: .4f} | top5: {top5: .4f}'.format( batch=batch_idx + 1, size=len(testloader), loss=losses.avg, top1=top1.avg, top3=top3.avg, top5=top5.avg, )) eces = ece_criterion.get_overall_ECELoss() cces = sce_criterion.get_overall_CCELoss() return (losses.avg, top1.avg, top3.avg, top5.avg, cces.item(), eces.item()) @torch.no_grad() def get_logits_targets(testloader, model): # switch to evaluate mode model.eval() all_targets = None all_outputs = None bar = tqdm(testloader, total=len(testloader)) for inputs, targets in bar: inputs = inputs.cuda() # compute output outputs = model(inputs) # to numpy targets = targets.cpu().numpy() outputs = outputs.cpu().numpy() if all_targets is None: all_outputs = outputs all_targets = targets else: all_targets = np.concatenate([all_targets, targets], axis=0) all_outputs = np.concatenate([all_outputs, outputs], axis=0) return all_outputs, all_targets @torch.no_grad() def get_logits_targets_torch(testloader, model): # switch to evaluate mode model.eval() all_targets = None all_outputs = None bar = tqdm(testloader, total=len(testloader)) for inputs, targets in bar: inputs = inputs.cuda() targets= targets.cuda() # compute output outputs = model(inputs) if all_targets is None: all_outputs = outputs all_targets = targets else: all_targets = torch.cat([all_targets, targets], dim=0) all_outputs = torch.cat([all_outputs, outputs], dim=0) return all_outputs, all_targets def fine_tune(trainloader, model, criterion, optimizer): # switch to train mode model.train() losses = AverageMeter() top1 = AverageMeter() top3 = AverageMeter() top5 = AverageMeter() criterion.reset() for batch_idx, (inputs, targets) in enumerate(trainloader): inputs, targets = inputs.cuda(), targets.cuda() # compute output outputs = model(inputs) loss_dict = criterion(outputs, targets) loss = loss_dict[0]["loss"] # measure accuracy and record loss losses.update(loss.item(), inputs.size(0)) # compute gradient and do SGD step optimizer.zero_grad() loss.backward() optimizer.step() return losses.avg
py
1a2f596b243c27f687f209e4729403099f81c8f4
from typing import Optional from server.crud.base import CRUDBase from server.db.models import RolesTable from server.schemas.role import RoleCreate, RoleUpdate class CRUDRole(CRUDBase[RolesTable, RoleCreate, RoleUpdate]): def get_by_name(self, *, name: str) -> Optional[RolesTable]: return RolesTable.query.filter(RolesTable.name == name).first() role_crud = CRUDRole(RolesTable)
py
1a2f5a912c80f5b551c5ebad298ebd156262d6dd
# Copyright 2014, Solly Ross (see LICENSE.txt) # Portions Copyright Python Software Foundation # (see COPYRIGHT.txt and PYTHON-LICENSE.txt) # main part import code import contextlib import doctest import getpass import linecache import pdb import re import sys import traceback import warnings import six from yalpt import ansi_helper as ansi from yalpt import formatters from yalpt import parsers __all__ = ["LiterateInterpreter"] @contextlib.contextmanager def noop_mgr(writer): yield writer class LiterateInterpreter(code.InteractiveConsole): def __init__(self, text_formatter=formatters.NoopFormatter(), code_parser=parsers.DocTestParser(), use_ansi=True, use_readline=True, env_driver=None, *args, **kwargs): code.InteractiveConsole.__init__(self, *args, **kwargs) self._output_checker = doctest.OutputChecker() self._fakeout = doctest._SpoofOut() self.chunks = None self.exc_msg = None self.name = 'literate program' self.text_formatter = text_formatter self.code_parser = code_parser self.use_ansi = use_ansi self.pause = True self.interactive = True if use_readline: self._readline = __import__('readline') self._add_readline() else: self._readline = None self._env_driver = env_driver self._correct_path() def runfunction(self, func): self.runcode(six.get_function_code(func)) def _correct_path(self): def correct_path(): import sys if '' not in sys.path: sys.path.insert(0, '') del sys self.runfunction(correct_path) def _add_readline(self): self.locals['__console_locals__'] = self.locals self.locals['readline'] = self._readline def add_readline(): # add support for completion import atexit import os import readline import rlcompleter completer = rlcompleter.Completer(__console_locals__) # noqa readline.set_completer(completer.complete) # tab completion readline.parse_and_bind('Control-space: complete') # history file histfile = os.path.join(os.environ['HOME'], '.literate-python-history') try: readline.read_history_file(histfile) except IOError: pass atexit.register(readline.write_history_file, histfile) del os, histfile, readline, rlcompleter self.runfunction(add_readline) del self.locals['__console_locals__'] del self.locals['readline'] def _interact_once(self, more): try: if more: prompt = sys.ps2 else: prompt = sys.ps1 try: line = self.raw_input(prompt) # Can be None if sys.stdin was redefined encoding = getattr(sys.stdin, "encoding", None) if encoding and not isinstance(line, six.text_type): line = line.decode(encoding) except EOFError: self.write("\n") return (False, None) else: if len(line) == 0: blank = True else: blank = False more = self.push(line) except KeyboardInterrupt: self.write("\nKeyboardInterrupt\n") self.resetbuffer() more = False blank = False return (blank, more) # BEGIN FROM PYTHON STD LIB # (extracted from the doctest module, made slight changes to the # filename RE, and refactored _capture_output into its own method) __LINECACHE_FILENAME_RE = re.compile(r'<literate ' r'(?P<name>.+)' r'\[(?P<chunknum>\d+)\]>$') def _patched_linecache_getlines(self, filename, module_globals=None): m = self.__LINECACHE_FILENAME_RE.match(filename) if m and m.group('name') == self.name: chunk = self.chunks[int(m.group('chunknum'))] source = chunk.source if isinstance(source, six.text_type): source = source.encode('ascii', 'backslashreplace') return source.splitlines(True) else: return self.save_linecache_getlines(filename, module_globals) @contextlib.contextmanager def _capture_output(self): save_stdout = sys.stdout sys.stdout = self._fakeout # set up pdb to work properly save_set_trace = pdb.set_trace self.debugger = doctest._OutputRedirectingPdb(save_stdout) self.debugger.reset() pdb.set_trace = self.debugger.set_trace self.save_linecache_getlines = linecache.getlines linecache.getlines = self._patched_linecache_getlines save_displayhook = sys.displayhook sys.displayhook = sys.__displayhook__ try: yield sys.stdout finally: sys.stdout = save_stdout pdb.set_trace = save_set_trace linecache.getlines = self.save_linecache_getlines sys.displayhook = save_displayhook # END FROM PYTHON STD LIB def _process_code_line(self, line, res, more): if self._readline is not None: self._readline.add_history(line) if more: self.write(sys.ps2) else: self.write(sys.ps1) if self.use_ansi: self.write(ansi.with_codes(line, 38, 5, 0xdf)) else: self.write(line) self.write("\n") with self._capture_output() as output: more = self.push(line) res += output.getvalue() output.truncate(0) exc = self.exc_msg self.exc_msg = None return (res, exc, more) def _format_tb(self): try: extype, value, tb = sys.exc_info() sys.last_type = extype sys.last_value = value sys.last_traceback = tb res = traceback.format_exception_only(extype, value) finally: tb = None if res: return res[-1] def showtraceback(self): try: extype, value, tb = sys.exc_info() sys.last_type = extype sys.last_value = value sys.last_traceback = tb tblist = traceback.extract_tb(tb) del tblist[:1] lst = traceback.format_list(tblist) if lst: lst.insert(0, "Traceback (most recent call last):\n") exc_msg = traceback.format_exception_only(extype, value) lst[len(lst):] = exc_msg finally: tblist = tb = None if self.filename.startswith("<literate "): self._fakeout.write(''.join(lst)) else: if sys.excepthook is sys.__excepthook__: self.write(''.join(lst)) else: sys.excepthook(extype, value, tb) self.exc_msg = ''.join(exc_msg) def _run_code(self, chunk, chunk_ind, pause=True): self.filename = "<literate {name}[{num}]>".format(name=self.name, num=chunk_ind) more = False res = "" lines = chunk.source.split("\n") for line in lines[:-1]: res, exc, more = self._process_code_line(line, res, more) if more: res, exc, more = self._process_code_line(lines[-1], res, more) if self.use_ansi: mgr = ansi.BoxMaker(self) else: mgr = noop_mgr(self) if chunk.want is not None or chunk.exc_msg is not None: if len(res) > 0 or exc is not None: # compare to the expected output # TODO(sross): convert options to optionsflags optionsflags = 0 checker = self._output_checker if exc is None: same = self._output_checker.check_output(chunk.want, res, 0) if not same: self.write('\n') with mgr as maker: maker.write('Warning, output different from ' 'expected:\n') maker.write('================================' '========\n\n') diff = checker.output_difference(chunk, res, optionsflags) maker.write(diff) self.write('\n') else: self.write(res) elif chunk.exc_msg is None: self.write('\n') with mgr as maker: maker.write('Warning, unexpected exception:\n') maker.write('==============================\n\n') maker.write(exc) self.write('\n') else: same_ex = checker.check_output(chunk.exc_msg, exc, optionsflags) if not same_ex: self.write('\n') with mgr as maker: maker.write('Warning, exception different from ' 'expected:\n') maker.write('==================================' '=========\n\n') diff = checker.output_difference(chunk, res, optionsflags) maker.write(diff) self.write('\n') else: self.write(res) else: if exc is not None: self.write('\n') with mgr as maker: maker.write('Warning, unexpected exception:\n') maker.write('==============================\n\n') maker.write(exc) self.write('\n') else: self.write(res) if not self.pause: self.write(sys.ps1 + '\n') def no_echo_input(self, prompt): with warnings.catch_warnings(): res = getpass.getpass(prompt) return res def interact(self, lit_string, name, pause=True, interactive=True): self.name = name self.pause = pause self.interactive = interactive try: sys.ps1 except AttributeError: sys.ps1 = ">>> " try: sys.ps2 except AttributeError: sys.ps2 = "... " try: sys.ps3 except AttributeError: sys.ps3 = '>>> ' if self._env_driver is not None: extra_locals = self._env_driver.setup() self.locals.update(extra_locals) extra_banner = self._env_driver.banner driver_text = " ({0})".format(self._env_driver.DRIVER_NAME) else: extra_banner = "" driver_text = "" cprt = ('Type "help", "copyright", "credits" or "license" for ' 'more information about Python.') self.write("Literate Python Shell{driver_text}\nPython {ver} " "on {platform}\n{cprt}\n" "{extra_banner}\n\n".format(driver_text=driver_text, ver=sys.version, platform=sys.platform, cprt=cprt, extra_banner=extra_banner)) if not interactive and pause: self.write('Press enter to continue after a code block\n\n') try: parser = self.code_parser start = True self.chunks = list(parser.parse(lit_string, name)) for chunk_ind, chunk in enumerate(self.chunks): if isinstance(chunk, parsers.CodeChunk): self._run_code(chunk, chunk_ind) elif not chunk: continue else: if not start and pause and interactive: self.filename = "<stdin>" more = False blanks = 0 while blanks < 2: blank, more = self._interact_once(more) if blank: blanks += 1 else: blanks = 0 if more is None: return # reset exc_msg so it doesn't get # raised after the next code block self.exc_msg = None elif not start and pause: self.no_echo_input(sys.ps3) self.write(self.text_formatter.format(chunk)) start = False complete_msg = ("\n{file} complete! Continuing to interactive " "console...\n\n".format(file=self.name)) if self.use_ansi: self.write(ansi.with_codes(complete_msg, 1)) else: self.write(complete_msg) self.filename = "<stdin>" self.locals['con'] = self more = False while more is not None: blank, more = self._interact_once(more) finally: if self._env_driver is not None: self._env_driver.teardown()
py
1a2f5ae8d3c02f130f839a951dcfb420bc66083b
# print absolute value of an integer # python的语法比较简单,采用缩进方式 a = 100 if a>=0 : print(a) else : print(-a)
py
1a2f5baed8b7607321747e371d736b51f44b1f86
""" datauri, library for "data:" URIs as defined in RFC 2397. """ from .datauri import ( # noqa: F401 DataURIError, discover, parse)
py
1a2f5dbd40eb6ca1bdd7c250d3497c57722d9386
# Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from __future__ import print_function import sys sys.path.append("..") import unittest import numpy as np from op_test import OpTest import paddle import paddle.fluid as fluid from paddle.fluid import compiler, Program, program_guard #Situation 1: repeat_times is a list (without tensor) class TestTileOpRank1(OpTest): def setUp(self): self.op_type = "tile" self.place = paddle.device.MLUPlace(0) self.__class__.use_mlu = True self.init_data() self.inputs = {'X': np.random.random(self.ori_shape).astype("float32")} self.attrs = {'repeat_times': self.repeat_times} output = np.tile(self.inputs['X'], self.repeat_times) self.outputs = {'Out': output} def init_data(self): self.ori_shape = [100] self.repeat_times = [2] def test_check_output(self): self.check_output_with_place(self.place) def test_check_grad(self): self.check_grad(['X'], 'Out') # with dimension expanding class TestTileOpRank2Expanding(TestTileOpRank1): def init_data(self): self.ori_shape = [120] self.repeat_times = [2, 2] class TestTileOpRank2(TestTileOpRank1): def init_data(self): self.ori_shape = [12, 14] self.repeat_times = [2, 3] class TestTileOpRank3_Corner(TestTileOpRank1): def init_data(self): self.ori_shape = (2, 10, 5) self.repeat_times = (1, 1, 1) class TestTileOpRank3_Corner2(TestTileOpRank1): def init_data(self): self.ori_shape = (2, 10, 5) self.repeat_times = (2, 2) class TestTileOpRank3(TestTileOpRank1): def init_data(self): self.ori_shape = (2, 4, 15) self.repeat_times = (2, 1, 4) class TestTileOpRank4(TestTileOpRank1): def init_data(self): self.ori_shape = (2, 4, 5, 7) self.repeat_times = (3, 2, 1, 2) # Situation 2: repeat_times is a list (with tensor) class TestTileOpRank1_tensor_attr(OpTest): def setUp(self): self.op_type = "tile" self.place = paddle.device.MLUPlace(0) self.__class__.use_mlu = True self.init_data() repeat_times_tensor = [] for index, ele in enumerate(self.repeat_times): repeat_times_tensor.append(("x" + str(index), np.ones( (1)).astype('int32') * ele)) self.inputs = { 'X': np.random.random(self.ori_shape).astype("float32"), 'repeat_times_tensor': repeat_times_tensor, } self.attrs = {"repeat_times": self.infer_repeat_times} output = np.tile(self.inputs['X'], self.repeat_times) self.outputs = {'Out': output} def init_data(self): self.ori_shape = [100] self.repeat_times = [2] self.infer_repeat_times = [-1] def test_check_output(self): self.check_output_with_place(self.place) def test_check_grad(self): self.check_grad(['X'], 'Out') class TestTileOpRank2_Corner_tensor_attr(TestTileOpRank1_tensor_attr): def init_data(self): self.ori_shape = [12, 14] self.repeat_times = [1, 1] self.infer_repeat_times = [1, -1] class TestTileOpRank2_attr_tensor(TestTileOpRank1_tensor_attr): def init_data(self): self.ori_shape = [12, 14] self.repeat_times = [2, 3] self.infer_repeat_times = [-1, 3] # Situation 3: repeat_times is a tensor class TestTileOpRank1_tensor(OpTest): def setUp(self): self.op_type = "tile" self.place = paddle.device.MLUPlace(0) self.__class__.use_mlu = True self.init_data() self.inputs = { 'X': np.random.random(self.ori_shape).astype("float32"), 'RepeatTimes': np.array(self.repeat_times).astype("int32"), } self.attrs = {} output = np.tile(self.inputs['X'], self.repeat_times) self.outputs = {'Out': output} def init_data(self): self.ori_shape = [100] self.repeat_times = [2] def test_check_output(self): self.check_output_with_place(self.place) def test_check_grad(self): self.check_grad(['X'], 'Out') class TestTileOpRank2_tensor(TestTileOpRank1_tensor): def init_data(self): self.ori_shape = [12, 14] self.repeat_times = [2, 3] # Situation 4: input x is Integer class TestTileOpInteger(OpTest): def setUp(self): self.op_type = "tile" self.place = paddle.device.MLUPlace(0) self.__class__.use_mlu = True self.inputs = { 'X': np.random.randint(10, size=(4, 4, 5)).astype("int32") } self.attrs = {'repeat_times': [2, 1, 4]} output = np.tile(self.inputs['X'], (2, 1, 4)) self.outputs = {'Out': output} def test_check_output(self): self.check_output_with_place(self.place) # Situation 5: input x is Bool class TestTileOpBoolean(OpTest): def setUp(self): self.op_type = "tile" self.place = paddle.device.MLUPlace(0) self.__class__.use_mlu = True self.inputs = {'X': np.random.randint(2, size=(2, 4, 5)).astype("bool")} self.attrs = {'repeat_times': [2, 1, 4]} output = np.tile(self.inputs['X'], (2, 1, 4)) self.outputs = {'Out': output} def test_check_output(self): self.check_output_with_place(self.place) # Situation 56: input x is Integer class TestTileOpInt64_t(OpTest): def setUp(self): self.op_type = "tile" self.place = paddle.device.MLUPlace(0) self.__class__.use_mlu = True self.inputs = { 'X': np.random.randint(10, size=(2, 4, 5)).astype("int64") } self.attrs = {'repeat_times': [2, 1, 4]} output = np.tile(self.inputs['X'], (2, 1, 4)) self.outputs = {'Out': output} def test_check_output(self): self.check_output_with_place(self.place) class TestTileError(unittest.TestCase): def test_errors(self): with program_guard(Program(), Program()): x1 = fluid.create_lod_tensor(np.array([[-1]]), [[1]], fluid.CPUPlace()) repeat_times = [2, 2] self.assertRaises(TypeError, paddle.tile, x1, repeat_times) x2 = fluid.layers.data(name='x2', shape=[4], dtype="uint8") self.assertRaises(TypeError, paddle.tile, x2, repeat_times) x3 = fluid.layers.data(name='x3', shape=[4], dtype="bool") x3.stop_gradient = False self.assertRaises(ValueError, paddle.tile, x3, repeat_times) class TestTileAPIStatic(unittest.TestCase): def test_api(self): with program_guard(Program(), Program()): repeat_times = [2, 2] x1 = fluid.layers.data(name='x1', shape=[4], dtype="int32") out = paddle.tile(x1, repeat_times) positive_2 = fluid.layers.fill_constant([1], dtype="int32", value=2) out2 = paddle.tile(x1, repeat_times=[positive_2, 2]) # Test python API class TestTileAPI(unittest.TestCase): def test_api(self): with fluid.dygraph.guard(): np_x = np.random.random([12, 14]).astype("float32") x = paddle.to_tensor(np_x) positive_2 = np.array([2]).astype("int32") positive_2 = paddle.to_tensor(positive_2) repeat_times = np.array([2, 3]).astype("int32") repeat_times = paddle.to_tensor(repeat_times) out_1 = paddle.tile(x, repeat_times=[2, 3]) out_2 = paddle.tile(x, repeat_times=[positive_2, 3]) out_3 = paddle.tile(x, repeat_times=repeat_times) assert np.array_equal(out_1.numpy(), np.tile(np_x, (2, 3))) assert np.array_equal(out_2.numpy(), np.tile(np_x, (2, 3))) assert np.array_equal(out_3.numpy(), np.tile(np_x, (2, 3))) if __name__ == "__main__": paddle.enable_static() unittest.main()
py
1a2f5e8bab97eb229d96cd075b7cc632d8e8b611
# Copyright 2016-2019 Dirk Thomas # Licensed under the Apache License, Version 2.0 from pathlib import Path import pytest from scspell import Report from scspell import SCSPELL_BUILTIN_DICT from scspell import spell_check spell_check_words_path = Path(__file__).parent / 'spell_check.words' @pytest.fixture(scope='module') def known_words(): global spell_check_words_path return spell_check_words_path.read_text().splitlines() def test_spell_check(known_words): source_filenames = [Path(__file__).parents[1] / 'setup.py'] + \ list( (Path(__file__).parents[1] / 'colcon_metadata') .glob('**/*.py')) + \ list((Path(__file__).parents[1] / 'test').glob('**/*.py')) for source_filename in sorted(source_filenames): print('Spell checking:', source_filename) # check all files report = Report(known_words) spell_check( [str(p) for p in source_filenames], base_dicts=[SCSPELL_BUILTIN_DICT], report_only=report, additional_extensions=[('', 'Python')]) unknown_word_count = len(report.unknown_words) assert unknown_word_count == 0, \ 'Found {unknown_word_count} unknown words: '.format_map(locals()) + \ ', '.join(sorted(report.unknown_words)) unused_known_words = set(known_words) - report.found_known_words unused_known_word_count = len(unused_known_words) assert unused_known_word_count == 0, \ '{unused_known_word_count} words in the word list are not used: ' \ .format_map(locals()) + ', '.join(sorted(unused_known_words)) def test_spell_check_word_list_order(known_words): assert known_words == sorted(known_words), \ 'The word list should be ordered alphabetically' def test_spell_check_word_list_duplicates(known_words): assert len(known_words) == len(set(known_words)), \ 'The word list should not contain duplicates'
py
1a2f5f20c7a5fba766f0599d46bbdc5f002967b3
#!/usr/bin/env python import os import sys if __name__ == "__main__": os.environ.setdefault("DJANGO_SETTINGS_MODULE", "bootProj.settings") try: from django.core.management import execute_from_command_line except ImportError as exc: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) from exc execute_from_command_line(sys.argv)
py
1a2f5fcebf615e5a0064a633180aae48ca9e6486
# -*- coding: utf-8 -*- """ sphinx.builders.gettext ~~~~~~~~~~~~~~~~~~~~~~~ The MessageCatalogBuilder class. :copyright: Copyright 2007-2015 by the Sphinx team, see AUTHORS. :license: BSD, see LICENSE for details. """ from __future__ import unicode_literals from os import path, walk from codecs import open from time import time from datetime import datetime, tzinfo, timedelta from collections import defaultdict from uuid import uuid4 from six import iteritems from sphinx.builders import Builder from sphinx.util import split_index_msg from sphinx.util.nodes import extract_messages, traverse_translatable_index from sphinx.util.osutil import safe_relpath, ensuredir, find_catalog, SEP from sphinx.util.console import darkgreen, purple, bold from sphinx.locale import pairindextypes POHEADER = r""" # SOME DESCRIPTIVE TITLE. # Copyright (C) %(copyright)s # This file is distributed under the same license as the %(project)s package. # FIRST AUTHOR <EMAIL@ADDRESS>, YEAR. # #, fuzzy msgid "" msgstr "" "Project-Id-Version: %(project)s %(version)s\n" "Report-Msgid-Bugs-To: \n" "POT-Creation-Date: %(ctime)s\n" "PO-Revision-Date: YEAR-MO-DA HO:MI+ZONE\n" "Last-Translator: FULL NAME <EMAIL@ADDRESS>\n" "Language-Team: LANGUAGE <LL@li.org>\n" "MIME-Version: 1.0\n" "Content-Type: text/plain; charset=UTF-8\n" "Content-Transfer-Encoding: 8bit\n" """[1:] class Catalog(object): """Catalog of translatable messages.""" def __init__(self): self.messages = [] # retain insertion order, a la OrderedDict self.metadata = {} # msgid -> file, line, uid def add(self, msg, origin): if not hasattr(origin, 'uid'): # Nodes that are replicated like todo don't have a uid, # however i18n is also unnecessary. return if msg not in self.metadata: # faster lookup in hash self.messages.append(msg) self.metadata[msg] = [] self.metadata[msg].append((origin.source, origin.line, origin.uid)) class MsgOrigin(object): """ Origin holder for Catalog message origin. """ def __init__(self, source, line): self.source = source self.line = line self.uid = uuid4().hex class I18nBuilder(Builder): """ General i18n builder. """ name = 'i18n' versioning_method = 'text' versioning_compare = None # be set by `gettext_uuid` def __init__(self, app): self.versioning_compare = app.env.config.gettext_uuid super(I18nBuilder, self).__init__(app) def init(self): Builder.init(self) self.catalogs = defaultdict(Catalog) def get_target_uri(self, docname, typ=None): return '' def get_outdated_docs(self): return self.env.found_docs def prepare_writing(self, docnames): return def compile_catalogs(self, catalogs, message): return def write_doc(self, docname, doctree): catalog = self.catalogs[find_catalog(docname, self.config.gettext_compact)] for node, msg in extract_messages(doctree): catalog.add(msg, node) if 'index' in self.env.config.gettext_enables: # Extract translatable messages from index entries. for node, entries in traverse_translatable_index(doctree): for typ, msg, tid, main in entries: for m in split_index_msg(typ, msg): if typ == 'pair' and m in pairindextypes.values(): # avoid built-in translated message was incorporated # in 'sphinx.util.nodes.process_index_entry' continue catalog.add(m, node) # determine tzoffset once to remain unaffected by DST change during build timestamp = time() tzdelta = datetime.fromtimestamp(timestamp) - \ datetime.utcfromtimestamp(timestamp) class LocalTimeZone(tzinfo): def __init__(self, *args, **kw): super(LocalTimeZone, self).__init__(*args, **kw) self.tzdelta = tzdelta def utcoffset(self, dt): return self.tzdelta def dst(self, dt): return timedelta(0) ltz = LocalTimeZone() class MessageCatalogBuilder(I18nBuilder): """ Builds gettext-style message catalogs (.pot files). """ name = 'gettext' def init(self): I18nBuilder.init(self) self.create_template_bridge() self.templates.init(self) def _collect_templates(self): template_files = set() for template_path in self.config.templates_path: tmpl_abs_path = path.join(self.app.srcdir, template_path) for dirpath, dirs, files in walk(tmpl_abs_path): for fn in files: if fn.endswith('.html'): filename = path.join(dirpath, fn) filename = filename.replace(path.sep, SEP) template_files.add(filename) return template_files def _extract_from_template(self): files = self._collect_templates() self.info(bold('building [%s]: ' % self.name), nonl=1) self.info('targets for %d template files' % len(files)) extract_translations = self.templates.environment.extract_translations for template in self.app.status_iterator( files, 'reading templates... ', purple, len(files)): with open(template, 'r', encoding='utf-8') as f: context = f.read() for line, meth, msg in extract_translations(context): origin = MsgOrigin(template, line) self.catalogs['sphinx'].add(msg, origin) def build(self, docnames, summary=None, method='update'): self._extract_from_template() I18nBuilder.build(self, docnames, summary, method) def finish(self): I18nBuilder.finish(self) data = dict( version = self.config.version, copyright = self.config.copyright, project = self.config.project, ctime = datetime.fromtimestamp( timestamp, ltz).strftime('%Y-%m-%d %H:%M%z'), ) for textdomain, catalog in self.app.status_iterator( iteritems(self.catalogs), "writing message catalogs... ", darkgreen, len(self.catalogs), lambda textdomain__: textdomain__[0]): # noop if config.gettext_compact is set ensuredir(path.join(self.outdir, path.dirname(textdomain))) pofn = path.join(self.outdir, textdomain + '.pot') pofile = open(pofn, 'w', encoding='utf-8') try: pofile.write(POHEADER % data) for message in catalog.messages: positions = catalog.metadata[message] if self.config.gettext_location: # generate "#: file1:line1\n#: file2:line2 ..." pofile.write("#: %s\n" % "\n#: ".join("%s:%s" % (safe_relpath(source, self.outdir), line) for source, line, _ in positions)) if self.config.gettext_uuid: # generate "# uuid1\n# uuid2\n ..." pofile.write("# %s\n" % "\n# ".join( uid for _, _, uid in positions)) # message contains *one* line of text ready for translation message = message.replace('\\', r'\\'). \ replace('"', r'\"'). \ replace('\n', '\\n"\n"') pofile.write('msgid "%s"\nmsgstr ""\n\n' % message) finally: pofile.close()
py
1a2f5fe5fbbe5ca306f72443569c2a9f39d15f63
import json from pathlib import Path from typing import List, Optional class ConfJSON: def __init__(self, dir_path: Path, name: str, prefix: str): self._dir_path = dir_path self.name = f'{prefix.rstrip("-")}-{name}' self._file_path = dir_path / name self.path = str(self._file_path) self.write_json() def write_json( self, tests: Optional[List[str]] = None, skipped_tests: Optional[List[str]] = None, skipped_integrations: Optional[List[str]] = None, nightly_integrations: Optional[List[str]] = None, unmockable_integrations: Optional[List[str]] = None, docker_thresholds: Optional[dict] = None ): if tests is None: tests = [] if skipped_tests is None: skipped_tests = None if skipped_integrations is None: skipped_integrations = [] if nightly_integrations is None: nightly_integrations = [] if unmockable_integrations is None: unmockable_integrations = [] if docker_thresholds is None: docker_thresholds = {} self._file_path.write_text(json.dumps({ 'tests': tests, 'skipped_tests': skipped_tests, 'skipped_integrations': skipped_integrations, 'nightly_integrations': nightly_integrations, 'unmockable_integrations': unmockable_integrations, 'docker_thresholds': docker_thresholds }))
py
1a2f60261566d276f5c5ef5e24872b2879b4b4a5
#!/usr/bin/env python3 # Copyright (c) 2018-2020 The Wflscoin Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. """Verify that starting wflscoin with -h works as expected.""" from test_framework.test_framework import WflscoinTestFramework from test_framework.util import assert_equal class HelpTest(WflscoinTestFramework): def set_test_params(self): self.setup_clean_chain = True self.num_nodes = 1 def setup_network(self): self.add_nodes(self.num_nodes) # Don't start the node def get_node_output(self, *, ret_code_expected): ret_code = self.nodes[0].process.wait(timeout=60) assert_equal(ret_code, ret_code_expected) self.nodes[0].stdout.seek(0) self.nodes[0].stderr.seek(0) out = self.nodes[0].stdout.read() err = self.nodes[0].stderr.read() self.nodes[0].stdout.close() self.nodes[0].stderr.close() # Clean up TestNode state self.nodes[0].running = False self.nodes[0].process = None self.nodes[0].rpc_connected = False self.nodes[0].rpc = None return out, err def run_test(self): self.log.info("Start wflscoin with -h for help text") self.nodes[0].start(extra_args=['-h']) # Node should exit immediately and output help to stdout. output, _ = self.get_node_output(ret_code_expected=0) assert b'Options' in output self.log.info("Help text received: {} (...)".format(output[0:60])) self.log.info("Start wflscoin with -version for version information") self.nodes[0].start(extra_args=['-version']) # Node should exit immediately and output version to stdout. output, _ = self.get_node_output(ret_code_expected=0) assert b'version' in output self.log.info("Version text received: {} (...)".format(output[0:60])) # Test that arguments not in the help results in an error self.log.info("Start wflscoind with -fakearg to make sure it does not start") self.nodes[0].start(extra_args=['-fakearg']) # Node should exit immediately and output an error to stderr _, output = self.get_node_output(ret_code_expected=1) assert b'Error parsing command line arguments' in output self.log.info("Error message received: {} (...)".format(output[0:60])) if __name__ == '__main__': HelpTest().main()
py
1a2f605bcd11fe30ea2f2421b0d5b8ff8242ff8a
import sys import click import json import random from .bakefile import Bakefile, TaskFilter from .exceptions import NoBakefileFound from .clint import eng_join import pygments import pygments.lexers import pygments.formatters from .constants import SKIP_NEXT, SAFE_ENVIRONS def indent(line): return f'{" " * 4}{line}' def do_help(exit=0): with click.Context(entrypoint) as ctx: help = entrypoint.get_help(ctx) help = help.replace( " bake", str(click.style(" $ ", fg="green", bold=True)) + str(click.style("bake", fg="yellow", bold=True)), ) help = help.replace( "the strangely familiar task–runner", str( click.style("the strangely familiar task–runner", fg="white", bold=True) ), ) help = help.replace( "Options", str(click.style("Options", fg="white", bold=True)) ) help = help.replace( "--insecure", str(click.style("--insecure", fg="red", bold=True)) ) help = help.replace("--yes", str(click.style("--yes", fg="red", bold=True))) help = help.replace( "--allow", str(click.style("--allow", fg="green", bold=True)) ) help = help.replace( "--no-deps", str(click.style("--no-deps", fg="yellow", bold=True)) ) help = help.replace( "--continue", str(click.style("--continue", fg="red", bold=True)) ) help = help.replace( "--environ-json", str(click.style("--environ-json", fg="green", bold=True)) ) help = help.replace("-e,", str(click.style("-e", fg="green", bold=True) + ",")) click.echo(help, err=True) sys.exit(exit) def echo_json(obj): _json = json.dumps(obj, indent=2) if sys.stdin.isatty(): _json = pygments.highlight( _json, pygments.lexers.JsonLexer(), pygments.formatters.TerminalFormatter() ) click.echo(_json, err=False) @click.command(context_settings=dict(help_option_names=["-h", "--help"])) @click.argument( "task", type=click.STRING, default="__LIST_ALL__", envvar="BAKE_TASK", # required=False, ) @click.option( "--bakefile", "-b", default="__BAKEFILE__", envvar="BAKEFILE_PATH", nargs=1, type=click.Path(), help="The Bakefile to use.", ) @click.option( "--list", "_list", default=False, is_flag=True, help="Lists available tasks (and their dependencies).", ) @click.option( "--levels", "-l", default=None, nargs=1, type=click.INT, help="List only a given number of '/' levels of tasks.", ) @click.option( "--help", "-h", default=False, is_flag=True, help="Show this message and exit." ) @click.option("--debug", default=False, is_flag=True, hidden=True) @click.option("--source", default=False, nargs=1, hidden=True) @click.option( "--allow", default=False, nargs=1, multiple=True, hidden=False, help="Whitelist an environment variable for use.", ) @click.option( "--no-deps", default=False, is_flag=True, hidden=False, help="Do not run dependent tasks.", ) @click.option("--yes", is_flag=True, help="Set medium–security prompts to yes.") @click.option( "--continue", "_continue", is_flag=True, # type=click.BOOL, help="Continue, if a task fails.", ) @click.option( "--interactive", "-i", is_flag=True, # type=click.BOOL, help="Run in interactive mode.", ) @click.option( "--insecure", is_flag=True, # type=click.BOOL, help="Inherit parent shell's environment variables.", ) @click.argument("arguments", nargs=-1, type=click.STRING) @click.option( "--silent", "-s", is_flag=True, # type=click.BOOL, help="Reduce output.", envvar="BAKE_SILENT", ) @click.option( "--sort", is_flag=True, type=click.BOOL, help="Sort tasks, alphabetially." ) @click.option( "--environ-json", "-e", nargs=1, type=click.STRING, help="Provide environment variables via JSON.", ) @click.option( "--json", "-j", "_json", is_flag=True, # type=click.BOOL, help="Output in JSON format (stdout).", ) def entrypoint( *, task, bakefile, arguments, _list, levels, _continue, environ_json, debug, silent, sort, insecure, allow, _json, no_deps, interactive, yes, help, source, ): """bake — the strangely familiar task–runner.""" if help: do_help(0) # Default to list behavior, when no task is provided. if _json or source: silent = True # Allow explicitly–passed environment variables. SAFE_ENVIRONS.extend(allow) # Enable list functionality, by default. if task == "__LIST_ALL__": _list = True task = None # Establish the Bakefile. try: if bakefile == "__BAKEFILE__": bakefile = Bakefile.find(root=".", filename="Bakefile") else: bakefile = Bakefile(path=bakefile) except NoBakefileFound: click.echo(click.style("No Bakefile found!", fg="red"), err=True) do_help(1) sys.exit(0) if debug: click.echo(f" + Bakefile: {bakefile.path}", err=True) # --source (internal API) if source: def echo_generator(g): for g in g: click.echo(g) if source == "__init__": source = random.choice(list(bakefile.tasks.keys())) task = bakefile.tasks[source] source = task.gen_source( sources=[task.bashfile.funcs_source, task.bashfile.root_source] ) else: task = bakefile.tasks[source] source = task.gen_source( sources=[ task.bashfile.funcs_source, task.bashfile.root_source, task.source, ] ) for source_line in source: click.echo(source_line) sys.exit(0) if not insecure: for key in bakefile.environ: if key not in SAFE_ENVIRONS: del bakefile.environ[key] if environ_json: bakefile.add_environ_json(environ_json) argv = [] environ = [] for i, argument in enumerate(arguments[:]): if "=" in argument: key, value = argument.split("=", 1) environ.append((key, value)) else: argv.append(argument) if debug: click.echo(f" + argv: {argv!r}", err=True) click.echo(f" + environ: {environ!r}", err=True) click.echo(err=True) for env in environ: key, value = env if debug: click.echo( f" + Setting environ: {click.style(key, fg='red')} {click.style('=', fg='white')} {value}.", err=True, ) bakefile.add_environ(key, value) bakefile.add_args(*argv) if _list: __list_json = {"tasks": {}} # Enable level filtering. if levels is not None: task_list = [] for _task in bakefile.tasks: if len(_task.split("/")) <= levels: task_list.append(_task) else: task_list = bakefile.tasks if sort: task_list = sorted(task_list) for _task in task_list: depends_on = bakefile[_task].depends_on( include_filters=False, recursive=True ) if no_deps: depends_on = () if depends_on: deps = [str(a) for a in depends_on] deps = f"\n {click.style('+', fg='yellow', bold=True)} {eng_join(deps, conj='&')}." else: deps = "" colon = "" if not deps else "…" __list_json["tasks"].update( {_task: {"depends_on": [str(d) for d in depends_on]}} ) if not silent: click.echo( f" {click.style('-', fg='green', bold=True)} {click.style(_task, bold=True)}{colon}{deps}", err=False, ) if not silent: tasks_unechoed = len(bakefile.tasks) - len(task_list) if tasks_unechoed: bake_command = str(click.style(f"bake --levels {levels + 1}", fg="red")) click.echo( f"Note: {tasks_unechoed} more tasks are available. " f"Please use $ {bake_command} to see more.", err=True, ) if _json: echo_json(__list_json) sys.exit(0) if task: try: task = bakefile[task] except KeyError: click.echo(click.style(f"Task {task} does not exist!", fg="red")) sys.exit(1) def execute_task(task, *, silent=False): try: edges = list(bakefile.graph.out_edges(task))[0] except IndexError: edges = list() skips = [] for edge in edges: if edge.do_skip is not None: skips.append(edge.do_skip) if not all(skips or [False]): # TODO: fully implement this? if "@" in f"{task}": silent = True if not silent: click.echo( click.style(" + ", fg="white") + click.style(f"Executing {task}", fg="yellow") + click.style(":", fg="white"), err=True, ) usually_bash_task = task.execute( yes=yes, debug=debug, silent=silent, interactive=interactive ) if not _continue: if hasattr(usually_bash_task, "ok"): if usually_bash_task.return_code > 0: if not silent: click.echo( click.style(f"Task {task} failed!", fg="red"), err=True, ) sys.exit(usually_bash_task.return_code) # This happens when it's a task filter. elif isinstance(usually_bash_task, tuple): key, value = ( usually_bash_task ) # But, in this instance, clearly isn't. else: click.echo( click.style(" + ", fg="green") + click.style(f"Skipping {task}", fg="white") + click.style(".", fg="white"), err=True, ) if not no_deps: tasks = task.depends_on(recursive=True) + [task] else: tasks = [task] for task in tasks: execute_task(task, silent=silent) if not silent: click.echo( click.style(" + ", fg="white") + click.style("Done", fg="green") + click.style(".", fg="white"), err=True, ) sys.exit(0) if __name__ == "__main__": entrypoint()
py
1a2f608c401acef3ba3d11f0434bb677bdceb6bc
#!/usr/bin/python3 ARD_DEVICE_ID = "1" ADC_READ_INTERVAL = 0.1 ADC_KEEP_VALS = 20 LIGHT_KEEP_VALS = 20 LOGFILE = "./logs/smarthome.log" # --- Unicodes --- # HOME = ' ⌂' HOMEON = ' ☗' KEY = ' ⚿' CLOUD = ' ☁' STAR = ' ★' SUN = ' ☀' REFRESH = ' 🗘' CHECKED = ' ✔' POINT = ' ☛' MISSING = ' ✘' MODESSYMBOL = ' ❖' MENUSYMBOL = ' ☰' TERMOMETER = ' 🌡' MUSIC = ' ♫' EMAIL = ' ✉' NOTIFICATION = ' 🔔' DEGREES = ' °C' SMILE = ' ☺' MOON = ' ☾' QUATER_MOON = '☽' PEACE = ' ☮' # --- Modes --- # LIGHT_TRESHOLD = 400
py
1a2f60af2272be77ef975238d91156b2983a193a
# the required python libraries imported import pandas as pd import matplotlib import matplotlib.pyplot as plt import numpy as np from trcrpm import TRCRP_Mixture from collections import Counter print "fuck yeah" data = pd.read_csv("./data/anomaly0245.csv", index_col=0) data = data.iloc[156600:240000].reset_index(drop=True) # Setup the placekeeping and initilizing variables chain = 0 x, eng_val, states, num_states = [], [], [], [] i = 0 step = 30 print(i) rng = np.random.RandomState(1) model = TRCRP_Mixture(chains=1, lag=10, variables=data.columns, rng=rng) model.incorporate(data[i:i + step]) model.resample_all(seconds=10) model.resample_hyperparameters(seconds=10) s = model.get_temporal_regimes('anomaly')[chain] num_states = step * [len(sorted(set(s)))] states = list(s[i:i + step]) eng_val = data.iloc[i:i + step, 0].tolist() x = list(range(i, i + step )) for i in range(step, len(data) - step, step): model.incorporate(data[i:i + step]) model.resample_all(seconds=10) model.resample_hyperparameters(seconds=10) s = model.get_temporal_regimes("anomaly")[chain] num_states = step * [len(sorted(set(s)))] states = list(s[i:i + step]) eng_val = data.iloc[i:i + step, 0].tolist() x = list(range(i, i + step )) print(i)
py
1a2f60bade4f75181ad28d3e096599995bccdd46
import pytest import os import utils import io import numpy import json import pickle import gzip from utils import kfp_client_utils from utils import minio_utils from utils import sagemaker_utils def run_predict_mnist(boto3_session, endpoint_name, download_dir): """ https://github.com/awslabs/amazon-sagemaker-examples/blob/a8c20eeb72dc7d3e94aaaf28be5bf7d7cd5695cb /sagemaker-python-sdk/1P_kmeans_lowlevel/kmeans_mnist_lowlevel.ipynb """ # Download and load dataset region = boto3_session.region_name download_path = os.path.join(download_dir, "mnist.pkl.gz") boto3_session.resource("s3", region_name=region).Bucket( "sagemaker-sample-data-{}".format(region) ).download_file("algorithms/kmeans/mnist/mnist.pkl.gz", download_path) with gzip.open(download_path, "rb") as f: train_set, valid_set, test_set = pickle.load(f, encoding="latin1") # Function to create a csv from numpy array def np2csv(arr): csv = io.BytesIO() numpy.savetxt(csv, arr, delimiter=",", fmt="%g") return csv.getvalue().decode().rstrip() # Run prediction on an image runtime = boto3_session.client("sagemaker-runtime") payload = np2csv(train_set[0][30:31]) response = runtime.invoke_endpoint( EndpointName=endpoint_name, ContentType="text/csv", Body=payload, ) return json.loads(response["Body"].read().decode()) @pytest.mark.parametrize( "test_file_dir", [ pytest.param( "resources/config/kmeans-mnist-endpoint", marks=pytest.mark.canary_test ) ], ) def test_create_endpoint( kfp_client, experiment_id, boto3_session, sagemaker_client, test_file_dir ): download_dir = utils.mkdir(os.path.join(test_file_dir + "/generated")) test_params = utils.load_params( utils.replace_placeholders( os.path.join(test_file_dir, "config.yaml"), os.path.join(download_dir, "config.yaml"), ) ) # Generate random prefix for model, endpoint config and endpoint name # to avoid errors if resources with same name exists test_params["Arguments"]["model_name"] = test_params["Arguments"][ "endpoint_config_name" ] = test_params["Arguments"]["endpoint_name"] = input_endpoint_name = ( utils.generate_random_string(5) + "-" + test_params["Arguments"]["model_name"] ) print(f"running test with model/endpoint name: {input_endpoint_name}") _, _, workflow_json = kfp_client_utils.compile_run_monitor_pipeline( kfp_client, experiment_id, test_params["PipelineDefinition"], test_params["Arguments"], download_dir, test_params["TestName"], test_params["Timeout"], ) try: outputs = {"sagemaker-deploy-model": ["endpoint_name"]} output_files = minio_utils.artifact_download_iterator( workflow_json, outputs, download_dir ) output_endpoint_name = utils.read_from_file_in_tar( output_files["sagemaker-deploy-model"]["endpoint_name"] ) print(f"endpoint name: {output_endpoint_name}") # Verify output from pipeline is endpoint name assert output_endpoint_name == input_endpoint_name # Verify endpoint is running assert ( sagemaker_utils.describe_endpoint(sagemaker_client, input_endpoint_name)[ "EndpointStatus" ] == "InService" ) # Validate the model for use by running a prediction result = run_predict_mnist(boto3_session, input_endpoint_name, download_dir) print(f"prediction result: {result}") assert json.dumps(result, sort_keys=True) == json.dumps( test_params["ExpectedPrediction"], sort_keys=True ) utils.remove_dir(download_dir) finally: # delete endpoint sagemaker_utils.delete_endpoint(sagemaker_client, input_endpoint_name)
py
1a2f616e58fa30c08acda8dc2317540f88f15600
# proxy module from pyface.ui.wx.grid.trait_grid_cell_adapter import *
py
1a2f62375f668ee9479eedce51823352feb3e903
default_app_config = 'sandbox.order.apps.OrderConfig'
py
1a2f628c8110b8d4f0d62d7b9055720ae4baad17
import random class PriQ(object): '''Binary-Heap based Priority Queue with uniquely named elements, name may \ be any hashable type. Defaults to min-heap, set maxpq=True for max-heap. \ Public methods: put, get, remove, update, contains, front, get_priority.''' def __init__(self, maxpq = False): self.q =[] # The priority queue, contains (priority, name) tuples self.elements = {} # Dict of all elements currently in queue, with current index self.maxmod = -1 if maxpq else 1 # modify pq to max instead of min def __len__(self): return len(self.q) def __str__(self): return str(self.q) def _propUp(self, index): '''Propagate up element to proper place in binary heap''' current = index while current > 0: parent = (current - 1) // 2 # parent node # swap with parent until parent <= child (>= for maxPQ) if self.q[parent][0] * self.maxmod <= self.q[current][0] * self.maxmod: break self.elements[self.q[current][1]], self.elements[self.q[parent][1]] = parent, current self.q[parent], self.q[current] = self.q[current], self.q[parent] current = parent def _propDown(self, index): '''Propagate down element to proper place in binary heap''' if len(self.q) == 1: self.elements[self.q[0][1]] = 0 # update index of last element current = index while current * 2 + 1 < len(self.q): # node has a child left, right = current * 2 + 1, current * 2 + 2 if right == len(self.q): # left child only if self.q[current][0] * self.maxmod >= self.q[left][0] * self.maxmod: # swap with left child and update elements dict self.elements[self.q[current][1]], self.elements[self.q[left][1]] = left, current self.q[current], self.q[left] = self.q[left], self.q[current] break if self.maxmod == 1: # min PQ minChild = left if self.q[left] <= self.q[right] else right if self.q[current] <= self.q[minChild]: # swap with lowest priority child as needed break self.elements[self.q[current][1]], self.elements[self.q[minChild][1]] = minChild, current self.q[current], self.q[minChild] = self.q[minChild], self.q[current] current = minChild else: # max PQ maxChild = left if self.q[left] >= self.q[right] else right if self.q[current] >= self.q[maxChild]: # swap with highest priority child as needed break self.elements[self.q[current][1]], self.elements[self.q[maxChild][1]] = maxChild, current self.q[current], self.q[maxChild] = self.q[maxChild], self.q[current] current = maxChild def put(self, name, priority): '''Add named element to priorty queue and place in binary heap''' if self.contains(name): return ValueError(name) self.q.append((priority, name)) self.elements[name] = len(self.q) - 1 self._propUp(self.elements[name]) def front(self): '''Element at front of queue''' return self.q[0] def get_priority(self, name): '''Current priority of named element''' return self.q[self.elements[name]][0] def get(self): '''Return element at front of queue and re-heapify queue''' if not self.q: return False # empty queue result = self.q[0] del(self.elements[result[1]]) if len(self.q) > 1: self.q[0] = self.q.pop() self._propDown(0) else: self.q = [] self.elements = {} return result def update(self, name, priority): '''Change priority of named element and re-heapify''' if not self.contains(name): return ValueError(name) index = self.elements[name] old_priority = self.q[index][0] self.q[index] = (priority, name) if priority * self.maxmod < old_priority * self.maxmod: self._propUp(index) if priority * self.maxmod > old_priority * self.maxmod: self._propDown(index) def contains(self, name): '''True if name currently exists in the queue''' return (name in self.elements) def remove(self, name): '''Remove named element and re-heapify''' if not self.contains(name): return ValueError(name) index = self.elements[name] old_priority = self.q[index][0] del(self.elements[name]) if len(self.q) > 1: self.q[index] = self.q.pop() # replace with last item in queue self.elements[self.q[index][1]] = index # re-heapify if self.q[index][0] * self.maxmod < old_priority * self.maxmod: self._propUp(index) elif self.q[index][0] * self.maxmod > old_priority * self.maxmod: self._propDown(index) else: self.q = [] ############################################################################### # Following are examples, including Dijkstra's shortest path ############################################################################### if __name__ == "__main__": # Sequential letter names with random integer weights (demonstration) pq = PriQ() name = 'a' for _ in range(26): pq.put(name, random.randint(1,99)) name = chr(ord(name) + 1) print("Initial: ['a'-'z' = randint(1,99)]\n", pq, "\nLength: ", len(pq)) print("Priority of 'a':", pq.get_priority('a')) print("Front:", pq.front()) pq.put('a', 17) # testing for duplicate put pq.update('a', 1) pq.update('b', 99) print("After updating 'a' to 1 and 'b' to 99:\n", pq) pq.remove('c') print("After removing 'c':\n", pq) print("Contains 'c'?", pq.contains('c')) print("Get elements until empty:") while pq: print(pq.get(), "contains 'd'?", pq.contains('d')) ''' # Max-heap priority queue with random float weights and integer names pq = PriQ(True) for i in range(20): pq.put(i, random.random()) print("['1'-'20' = random()]:\n", pq) while pq: print(pq.get()) ''' # Dijkstra's shortest path algorithm example def dsp(G, s): '''Dijkstra's shortest path algorithm. Input: weighted graph G, an\ adjacency list mapping each named node to it's neighbors with a {name: weight}\ dict, and s, the name of the starting node. Outputs a mapping for each node\ to a tuple of (weight of shortest path, name of predecessor node). ''' pq = PriQ() result = {} predecessor = {s: None} for key in G.keys(): pq.put(key, 0 if key == s else float('inf')) while pq: cur_weight, cur_node = pq.get() result[cur_node] = (cur_weight, predecessor[cur_node]) for adj in G[cur_node]: # for neighboring nodes # update weight from starting node if less than previous paths if adj not in result: if pq.get_priority(adj) > cur_weight + G[cur_node][adj]: pq.update(adj, cur_weight + G[cur_node][adj]) predecessor[adj] = cur_node return result G = { 'a': {'b': 2, 'f': 6, 'c': 1}, 'b': {'a': 2, 'd': 6, 'f': 3, 'g': 4}, 'c': {'a': 1, 'f': 2, 'g': 3, 'e': 5}, 'd': {'b': 6, 'f': 2, 'g': 1, 'h': 2}, 'e': {'c': 5, 'f': 4, 'g': 2, 'h': 3}, 'f': {'a': 6, 'b': 3, 'd': 2, 'g': 7, 'e': 4, 'c': 2}, 'g': {'f': 7, 'b': 4, 'd': 1, 'h': 5, 'e': 2, 'c': 3}, 'h': {'d': 2, 'g': 5, 'e': 3} } print( ''' B 6 D 2 3 2 4 1 2 A 6 F 7 G 5 H 1 2 3 4 2 3 C 5 E ''') print("\nCalculating shortest paths with Dijkstra's algorithm...") print("Shortest paths from 'a':", dsp(G, 'a')) print("Shortest paths from 'd':", dsp(G, 'd'))
py
1a2f631f3f1e62117c3b1a341897500e1a37b9f4
# Lint as: python3 # Copyright 2019 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """An abstract layer for processing sequences step-by-step. E.g.:: def ProcessSeq(step, external_inputs, input_batch): prepared_inputs = step.PrepareExternalInputs( step.theta, external_inputs) batch_size, T = tf.shape(input_batch.paddings)[:2] state = step.ZeroState( step.theta, prepared_inputs, batch_size) for t in range(T): step_inputs = input_batch.Transform(lambda x: x[:, i, ...]) step_outputs, state = step.FProp( step.theta, prepared_inputs, step_inputs, state) (processing step_outputs...) """ import collections from lingvo import compat as tf from lingvo.core import base_layer from lingvo.core import builder_layers from lingvo.core import py_utils from lingvo.core import recurrent class Step(base_layer.BaseLayer): """A layer that processes input sequences step-by-step. This can be seen as an RNNCell extended with optional external inputs. """ def PrepareExternalInputs(self, theta, external_inputs): """Returns the prepared external inputs, e.g., packed_src for attention.""" if not external_inputs: external_inputs = py_utils.NestedMap() packed = external_inputs.DeepCopy() for name, child in self.children.items(): child_external_inputs = external_inputs.get(name, py_utils.NestedMap()) if isinstance(child, (tuple, list)): output = [] for i, sub in enumerate(child): if isinstance(sub, Step): output.append( sub.PrepareExternalInputs(theta[name][i], child_external_inputs)) if output: if len(output) != len(child): raise ValueError('Expecting child list to be instances of Step.') packed[name] = type(child)(output) elif isinstance(child, Step): packed[name] = child.PrepareExternalInputs(theta[name], child_external_inputs) return packed def ZeroState(self, theta, prepared_inputs, batch_size): """Returns the initial state given external inputs and batch size. Args: theta: A `.NestedMap` object containing weights' values of this layer and its children layers. prepared_inputs: External inputs returned by PrepareExternalInputs(). batch_size: An int scalar representing the batch size of per-step inputs. Returns: A `.NestedMap` representing the initial state, which can be passed to FProp() for processing the first time step. """ state0 = py_utils.NestedMap() for name, child in self.children.items(): if isinstance(child, (tuple, list)): output = [] for i, sub in enumerate(child): if isinstance(sub, Step): output.append( sub.ZeroState(theta[name][i], prepared_inputs[name][i], batch_size)) if output: if len(output) != len(child): raise ValueError('Expecting child list to be instances of Step.') state0[name] = type(child)(output) elif isinstance(child, Step): state0[name] = child.ZeroState(theta[name], prepared_inputs[name], batch_size) return state0 def FProp(self, theta, prepared_inputs, step_inputs, padding, state0): """Forward function. step_inputs, state0, step_outputs, and state1 should each be a `.NestedMap` of tensor values. Each tensor must be of shape [batch_size ...]. The structure of NestedMaps are determined by the implementation. state0 and state1 must have exactly the same structure and tensor shapes. Args: theta: A `.NestedMap` object containing weights' values of this layer and its children layers. prepared_inputs: External inputs returned by PrepareExternalInputs(). step_inputs: The inputs for this time step. padding: A 0/1 float tensor of shape [batch_size]; 1.0 means that this batch element is empty in this step. state0: The previous recurrent state. Returns: A tuple (step_outputs, state1). - outputs: The outputs of this step. - state1: The next recurrent state. """ raise NotImplementedError(type(self)) class StatelessLayerStep(Step): """Allows BaseLayer subclasses to be used as Steps. Layers used with this class should be stateless: they should not return anything that must be passed back in the next invocation. """ @classmethod def Params(cls): p = super().Params() p.Define('layer', None, 'Params for the layer that this step wraps.') return p def __init__(self, params): super().__init__(params) p = params self.CreateChild('layer', p.layer) def FProp(self, theta, prepared_inputs, step_inputs, padding, state0): """Perform inference on a stateless layer. Args: theta: A `.NestedMap` object containing weights' values of this layer and its children layers. prepared_inputs: unused. step_inputs: A NestedMap containing 'inputs', which are passed directly to the layer. padding: A 0/1 float tensor of shape [batch_size]; 1.0 means that this batch element is empty in this step. state0: unused. Returns: (output, state1), where output is the output of the layer, and state1 is an empty NestedMap. """ del state0 del prepared_inputs args = {} if padding is not None: args['padding'] = padding output = self.layer.FProp(theta.layer, step_inputs.inputs, **args) return output, py_utils.NestedMap() class StackStep(Step): """A stack of steps. Each sub-step is assumed to accept step_inputs of type NestedMap(inputs=[]) and return a primary output of type NestedMap(output=tensor). The output of layer n-1 is sent to input of layer n. Per-step context vectors and per-sequence context vectors can also be supplied; see FProp for more details. """ @classmethod def Params(cls): p = super().Params() p.Define( 'sub', [], 'A list of sub-stack params. Each layer is ' 'expected to accept its input as NestedMap(inputs=[]), and ' 'produce output as NestedMap(output=tensor). ' 'The external_inputs parameter is passed directly to the ' 'PrepareExternalInputs method of each sub-step. ') p.Define( 'residual_start', -1, 'An index of the layer where residual ' 'connections start. Setting this parameter to a negative value turns ' 'off residual connections.' 'More precisely, when i >= residual_start, the output of each step ' 'is defined as: ' 'output[i] = output[i - residual_stride] + sub[i](output[i - 1]) ' 'where output[-1] is the step input.') p.Define( 'residual_stride', 1, 'If residual connections are active, this ' 'is the number of layers that each connection skips. For ' 'instance, setting residual_stride = 2 means the output of layer ' 'n is added to layer n + 2') return p def __init__(self, params): super().__init__(params) p = params self.sub_steps = [] self.CreateChildren('sub', p.sub) def PrepareExternalInputs(self, theta, external_inputs): """Delegates external inputs preparation to sub-layers. Args: theta: A `.NestedMap` object containing weights' values of this layer and its children layers. external_inputs: A `.NestedMap` object. The structure of the internal fields is defined by the sub-steps. Returns: A `.NestedMap` containing a pre-processed version of the external_inputs, one per sub-step. """ packed = py_utils.NestedMap(sub=[]) for i in range(len(self.sub)): packed.sub.append(self.sub[i].PrepareExternalInputs( theta.sub[i], external_inputs)) return packed def ZeroState(self, theta, prepared_inputs, batch_size): """Computes a zero state for each sub-step. Args: theta: A `.NestedMap` object containing weights' values of this layer and its children layers. prepared_inputs: An output from PrepareExternalInputs. batch_size: The number of items in the batch that FProp will process. Returns: A `.NestedMap` containing a state0 object for each sub-step. """ state = py_utils.NestedMap(sub=[]) for i in range(len(self.sub)): state.sub.append(self.sub[i].ZeroState(theta.sub[i], prepared_inputs, batch_size)) return state def FProp(self, theta, prepared_inputs, step_inputs, padding, state0): """Performs inference on the stack of sub-steps. There are three possible ways to feed input to the stack: * step_inputs.inputs: These tensors are fed only to the lowest layer. * step_inputs.context: [Optional] This tensor is fed to every layer. * prepared_inputs: [Optional] This tensor is fed to every layer and is assumed to stay constant over all steps. Args: theta: A `.NestedMap` object containing weights' values of this layer and its children layers. prepared_inputs: An output from PrepareExternalInputs. step_inputs: A `.NestedMap` containing a list called 'inputs', an optionally a tensor called 'context'. padding: A 0/1 float tensor of shape [batch_size]; 1.0 means that this batch element is empty in this step. state0: The previous recurrent state. Returns: A tuple (output, state1): - output: A `.NestedMap` containing the output of the top-most step. - state1: The recurrent state to feed to next invocation of this graph. """ state1 = py_utils.NestedMap(sub=[]) inputs = list(step_inputs.inputs) # We pretend that the input is the output of layer -1 for the purposes # of residual connections. residual_inputs = [tf.concat(inputs, axis=1)] additional = [] if 'context' in step_inputs: additional.append(step_inputs.context) for i in range(len(self.sub)): sub_inputs = py_utils.NestedMap(inputs=inputs + additional) sub_output, state1_i = self.sub[i].FProp(theta.sub[i], prepared_inputs.sub[i], sub_inputs, padding, state0.sub[i]) state1.sub.append(state1_i) output = sub_output.output if i >= self.params.residual_start >= 0: # residual_inputs contains the step input at residual_inputs[0]. assert i + 1 - self.params.residual_stride < len(residual_inputs) output += residual_inputs[i + 1 - self.params.residual_stride] residual_inputs.append(output) inputs = [output] return py_utils.NestedMap(output=output), state1 class ParallelStep(Step): """Runs many steps on the same input and concatenates their outputs.""" @classmethod def Params(cls): p = super().Params() p.Define( 'sub', [], 'A list of step params. Each step is ' 'expected to accept its input as NestedMap(inputs=[]), and ' 'produce output as NestedMap(output=tensor). ' 'The external_inputs parameter is passed directly to the ' 'PrepareExternalInputs method of each sub-step. ') return p def __init__(self, params): super().__init__(params) p = params self.CreateChildren('sub', p.sub) def FProp(self, theta, prepared_inputs, step_inputs, padding, state0): """Performs inference on N steps at once and concatenates the result. Args: theta: A `.NestedMap` object containing weights' values of this layer and its children layers. prepared_inputs: An output from PrepareExternalInputs. step_inputs: A `.NestedMap` containing a list called 'inputs'. padding: A 0/1 float tensor of shape [batch_size]; 1.0 means that this batch element is empty in this step. state0: The previous recurrent state. Returns: A tuple (output, state1): - output: A `.NestedMap` containing the output of the top-most step. - state1: The recurrent state to feed to next invocation of this graph. """ state1 = py_utils.NestedMap(sub=[None] * len(self.sub)) outputs = [None] * len(self.sub) for i in range(len(self.sub)): outputs[i], state1.sub[i] = self.sub[i].FProp(theta.sub[i], prepared_inputs.sub[i], step_inputs, padding, state0.sub[i]) output = py_utils.NestedMap(output=tf.concat(outputs, axis=1)) return output, state1 # signature: A GraphSignature string defining the input and output parameters # of this step. For example, (inputs=[a,b])->c means that step_inputs # should be NestedMap(inputs=[a,b]), and the output of FProp should be # stored in c. # external_signature: A GraphSignature string defining the input to # PrepareExternalInputs. For example, 'external_inputs.foo' means that # the tensor external_inputs.foo should be the 'external_inputs' parameter # when calling PrepareExternalInputs on this sub-step. # params: The parameters to use when constructing the sub-step. SubStep = collections.namedtuple('SubStep', ['signature', 'external_signature', 'params']) class GraphStep(Step): r"""A step that connects sub-steps in a simple data flow graph. This is an adaptation of builder_layers.GraphLayer to support steps. Params.sub specifies a list of Specs that define each sub-step. A spec contains: * step_inputs: The signature describing how to assemble the input and output for this step. The input part describes the 'step_inputs' parameter, while the output part describes the name of the output. The state0 input and state1 output are handled automatically and should not be specified. * external_inputs: if this Step requires external_inputs, this is the signature describing how to find those inputs. This value can also be set to None. * params: the params used to construct the sub-step. The format of signature strings is defined in detail in the GraphSignature class documentation. All inputs to a layer must have been produced by some previous layer. No cycles are allowed. All outputs must be uniquely named; no overwriting of previous names is allowed. Example ('(act=[layer_0.output,step_inputs.context])->layer_1', 'external_inputs.extra', step_params) This constructs the step defined by step_params. Its FProp method will be called with {act=[layer_0.output,step_inputs.context]} as the step_inputs parameter. Its PrepareExternalInputs method will be called with 'external_inputs.extra' as the external_inputs parameter. The output of that method will be passed to ZeroState and FProp. """ @classmethod def Params(cls): p = super().Params() p.Define('output_signature', '', 'Signature of the step output.') p.Define('sub', [], 'A list of SubSteps (defined above).') p.Define('dict_type', py_utils.NestedMap, 'Type of nested dicts.') return p _seq = collections.namedtuple( '_Seq', ['name', 'signature', 'external_signature', 'step']) def __init__(self, params): super().__init__(params) p = self.params assert p.name self._seq = [] for i, (signature, external_signature, sub_params) in enumerate(p.sub): assert signature sig = builder_layers.GraphSignature(signature) assert len(sig.inputs) == 1 assert sig.outputs external_sig = None if external_signature: external_sig = builder_layers.GraphSignature(external_signature) assert len(external_sig.inputs) == 1 assert not external_sig.outputs name = sub_params.name if not name: name = '%s_%02d' % (sig.outputs[0], i) sub_params.name = name self.CreateChild(name, sub_params) self._seq.append( GraphStep._seq(name, sig, external_sig, self.children[name])) self.output_signature = builder_layers.GraphSignature(p.output_signature) def PrepareExternalInputs(self, theta, external_inputs): """Prepares external inputs for each sub-step. The external_inputs parameter of this method is processed by the external_inputs of each sub-step, then processed by the sub-step's PrepareExternalInputs method. Args: theta: variables used by sub-steps. external_inputs: A NestedMap of [n_batch, ...] tensors. Returns: A NestedMap of prepared inputs, where the keys are the names of each sub-step. """ graph_tensors = builder_layers.GraphTensors() graph_tensors.StoreTensor('external_inputs', external_inputs) prepared_inputs = py_utils.NestedMap() with tf.name_scope(self.params.name): for seq in self._seq: if seq.external_signature: template = py_utils.NestedMap(inputs=seq.external_signature.inputs) packed = template.Transform(graph_tensors.GetTensor) seq_external_inputs = packed.inputs[0] prepared_inputs[seq.name] = seq.step.PrepareExternalInputs( theta[seq.name], seq_external_inputs) else: prepared_inputs[seq.name] = py_utils.NestedMap() return prepared_inputs def ZeroState(self, theta, prepared_inputs, batch_size): """Creates a zero state NestedMap for this step. Args: theta: variables used by sub-steps. prepared_inputs: Output from a call to PrepareExternalInputs. batch_size: The number of items in the batch that FProp will process. Returns: A NestedMap of ZeroState results for each sub-step. """ state0 = py_utils.NestedMap() with tf.name_scope(self.params.name): for seq in self._seq: state0[seq.name] = seq.step.ZeroState(theta[seq.name], prepared_inputs[seq.name], batch_size) return state0 def FProp(self, theta, prepared_inputs, step_inputs, padding, state0): """A single inference step for this step graph. Args: theta: variables used by sub-steps. prepared_inputs: A NestedMap containing external_inputs that were pre-processed by the PrepareExternalInputs method of each sub-step. The keys are the names of the sub-steps. step_inputs: A NestedMap of [batch, ...] tensors. The structure of this depends on the graph implementation. padding: A 0/1 float tensor of shape [batch_size]; 1.0 means that this batch element is empty in this step. state0: A NestedMap of state variables produced by either ZeroState or a previous invocation of this FProp step. The keys are the names of the sub-steps. Returns: (output, state1), both of which are NestedMaps. output is implementation-dependent and is defined by the output_signature parameter. state1 is a NestedMap where the keys are names of sub-steps and the values are state outputs from their FProp methods. """ p = self.params graph_tensors = builder_layers.GraphTensors() graph_tensors.StoreTensor('prepared_inputs', prepared_inputs) graph_tensors.StoreTensor('step_inputs', step_inputs) state1 = py_utils.NestedMap() with tf.name_scope(p.name): for seq in self._seq: tf.logging.vlog(1, 'GraphStep: call %s', seq.name) external = None if seq.external_signature: external = prepared_inputs[seq.name] template = py_utils.NestedMap(inputs=seq.signature.inputs) packed = template.Transform(graph_tensors.GetTensor) input_args = packed.inputs[0] out, seq_state1 = seq.step.FProp(theta[seq.name], external, input_args, padding, state0[seq.name]) graph_tensors.StoreTensor(seq.signature.outputs[0], out) state1[seq.name] = seq_state1 template = py_utils.NestedMap(inputs=self.output_signature.inputs) output_tensors = template.Transform(graph_tensors.GetTensor).inputs[0] return output_tensors, state1 class IteratorStep(Step): """An iterator over the time dimension of some tensors. It's common to have a tensor of shape [batch, time, ...] or [time, batch, ...]. This object will step through the time dimension, producing tensors of shape [batch, ...] in succession. The input tensors are passed to PrepareExternalInputs. The step_inputs argument of FProp is unused. """ @classmethod def Params(cls): p = super().Params() p.Define('axis', 1, 'The time dimension of the tensors.') return p def PrepareExternalInputs(self, theta, external_inputs): """Prepares the input for iteration. Args: theta: unused. external_inputs: A NestedMap containing tensors. The time axis of each tensor should be params.axis. Returns: A prepared NestedMap (current the same as the input). """ return external_inputs def ZeroState(self, theta, prepared_inputs, batch_size): """Returns the initial iterator state. Args: theta: unused. prepared_inputs: Output from a call to PrepareExternalInputs. batch_size: The number of items in the batch that FProp will process. Returns: An initial state NestedMap. """ return py_utils.NestedMap(t=tf.constant(0, dtype=tf.int32)) def FProp(self, theta, prepared_inputs, step_inputs, padding, state0): """Returns a A single inference step for this step graph. Args: theta: unused. prepared_inputs: Output from a call to PrepareExternalInputs. step_inputs: unused. padding: unused. state0: A NestedMap of state variables produced by either ZeroState or a previous invocation of this FProp step. Returns: (output, state1), both of which are NestedMaps. output is implementation-dependent and is defined by the output_signature parameter. state1 is a NestedMap where the keys are names of sub-steps and the values are state outputs from their FProp methods. """ del theta del step_inputs del padding def _Slice(tensor): """Return a slice of this tensor at time=state0.t.""" shape = py_utils.GetShape(tensor) # All zeros except for t in the time dimension. # e.g. if params.axis=1, begin is [0, t, 0, 0, 0, ...] begin = tf.one_hot(self.params.axis, tf.rank(tensor), on_value=state0.t) # Same as shape, but with a 1 in the time dimension. # e.g. if params.axis=1, shape is [shape[0], 1, shape[2], shape[3], ...] size = tf.concat([ shape[0:self.params.axis], tf.constant([1], dtype=tf.int32), shape[self.params.axis + 1:] ], axis=0) # Make a slice where the time dimension is fixed at state0.t. time_slice = tf.slice(tensor, begin, size) # Remove the time dimension. return tf.squeeze(time_slice, axis=self.params.axis) output = prepared_inputs.Transform(_Slice) state1 = py_utils.NestedMap(t=state0.t + 1) return output, state1 class RecurrentStepWrapper(base_layer.BaseLayer): """A layer that wraps a step in a recurrent.Recurrent call.""" @classmethod def Params(cls): p = super().Params() p.Define('step', None, 'The step params that this class wraps.') return p def __init__(self, params): super().__init__(params) self.CreateChild('step', self.params.step) def _CreateChildrenVariables(self): # Backwards compatibility: manually call child.InstantiateVariables() # outside of tf.variable_scope(p.name). self.step.InstantiateVariables() super()._CreateChildrenVariables() def PrepareExternalInputs(self, theta, external_inputs): """See Step.PrepareExternalInputs.""" return self.step.PrepareExternalInputs(theta.step, external_inputs) def ZeroState(self, theta, prepared_inputs, batch_size): """See Step.ZeroState.""" return self.step.ZeroState(theta.step, prepared_inputs, batch_size) def FProp(self, theta, prepared_inputs, inputs, padding, state0, **kwargs): """Runs a Step layer over multiple timesteps using Recurrent. Args: theta: A NestedMap containing weights' values of this layer and its children layers. prepared_inputs: External inputs returned by Step.PrepareExternalInputs(). inputs: A NestedMap of inputs of shape [time, batch_size, dim]. padding: A 0/1 float tensor of shape [time, batch_size]; 1.0 means that this batch element is empty in this step. state0: A NestedMap containing the initial recurrent state. **kwargs: Additional kwargs to pass to Recurrent. Returns: A tuple (outputs, state1). - outputs: A NestedMap containing the accumulated outputs of all steps, containing Tensors shaped [time, batch_size, dim]. - state1: A NestedMap containing the accumulated recurrent states, containing Tensors shaped [time, batch_size, dim]. """ def RnnStep(recurrent_theta, recurrent_state0, recurrent_inputs): """Compute a single timestep.""" output, state1 = self.step.FProp( theta=recurrent_theta.theta, prepared_inputs=recurrent_theta.prepared_inputs, step_inputs=recurrent_inputs.inputs, padding=recurrent_inputs.padding, state0=recurrent_state0.state) recurrent_state1 = py_utils.NestedMap(output=output, state=state1) return recurrent_state1, py_utils.NestedMap() # In order to pass Step outputs through Recurrent, they need to be # included as part of state. output0, _ = self.step.FProp(theta.step, prepared_inputs, inputs.Transform(lambda x: x[0]), padding[0], state0) accumulated_states, _ = recurrent.Recurrent( theta=py_utils.NestedMap( theta=theta.step, prepared_inputs=prepared_inputs), state0=py_utils.NestedMap(output=output0, state=state0), inputs=py_utils.NestedMap(inputs=inputs, padding=padding), cell_fn=RnnStep, **kwargs) return accumulated_states.output, accumulated_states.state
py
1a2f639f66ea104a409626485e774ec7720d7bf6
# USAGE # python facial_landmarks.py --shape-predictor shape_predictor_68_face_landmarks.dat --image images/example_01.jpg # import the necessary packages from imutils import face_utils import numpy as np import argparse import imutils import dlib import cv2 # construct the argument parser and parse the arguments ap = argparse.ArgumentParser() ap.add_argument("-p", "--shape-predictor", required=True, help="path to facial landmark predictor") ap.add_argument("-i", "--image", required=True, help="path to input image") args = vars(ap.parse_args()) # initialize dlib's face detector (HOG-based) and then create # the facial landmark predictor detector = dlib.get_frontal_face_detector() predictor = dlib.shape_predictor(args["shape_predictor"]) # load the input image, resize it, and convert it to grayscale image = cv2.imread(args["image"]) image = imutils.resize(image, width=500) gray = cv2.cvtColor(image, cv2.COLOR_BGR2GRAY) # detect faces in the grayscale image rects = detector(gray, 1) # loop over the face detections for (i, rect) in enumerate(rects): # determine the facial landmarks for the face region, then # convert the facial landmark (x, y)-coordinates to a NumPy # array shape = predictor(gray, rect) shape = face_utils.shape_to_np(shape) # convert dlib's rectangle to a OpenCV-style bounding box # [i.e., (x, y, w, h)], then draw the face bounding box (x, y, w, h) = face_utils.rect_to_bb(rect) cv2.rectangle(image, (x, y), (x + w, y + h), (0, 255, 0), 2) # show the face number cv2.putText(image, "Face #{}".format(i + 1), (x - 10, y - 10), cv2.FONT_HERSHEY_SIMPLEX, 0.5, (0, 255, 0), 2) # loop over the (x, y)-coordinates for the facial landmarks # and draw them on the image for (x, y) in shape: cv2.circle(image, (x, y), 1, (0, 0, 255), -1) # show the output image with the face detections + facial landmarks cv2.imshow("Output", image) cv2.waitKey(0)
py
1a2f641584a08008ee532ab93c06ced2336e4710
# -*- coding: utf-8 -*- # ---------------------------------------------------------------------------- # Copyright © 2021, Spyder Bot # # Licensed under the terms of the MIT license # ---------------------------------------------------------------------------- """ Status bar widgets setup. """ from setuptools import find_packages from setuptools import setup from status_bar_widgets import __version__ setup( # See: https://setuptools.readthedocs.io/en/latest/setuptools.html name="status-bar-widgets", version=__version__, author="Spyder Bot", author_email="spyder.python@gmail.com", description="Example that shows how to add wigets to the status bar", license="MIT license", url="https://github.com/spyder-bot/status-bar-widgets", python_requires='>= 3.7', install_requires=[ "qtpy", "qtawesome", "spyder>=5.1.1", ], packages=find_packages(), entry_points={ "spyder.plugins": [ "status_bar_widgets = status_bar_widgets.spyder.plugin:StatusbarWidgets" ], }, classifiers=[ "Operating System :: MacOS", "Operating System :: Microsoft :: Windows", "Operating System :: POSIX :: Linux", "Programming Language :: Python :: 3", "Programming Language :: Python :: 3.7", "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Development Status :: 5 - Production/Stable", "Intended Audience :: Education", "Intended Audience :: Science/Research", "Intended Audience :: Developers", "Topic :: Scientific/Engineering", ], )
py
1a2f642f91ccc54a3e54b279bb474d03a514954f
"""Support for Homematic thermostats.""" import logging from homeassistant.components.climate import ClimateDevice from homeassistant.components.climate.const import ( HVAC_MODE_AUTO, HVAC_MODE_HEAT, HVAC_MODE_OFF, PRESET_BOOST, PRESET_COMFORT, PRESET_ECO, SUPPORT_PRESET_MODE, SUPPORT_TARGET_TEMPERATURE, ) from homeassistant.const import ATTR_TEMPERATURE, TEMP_CELSIUS from . import ATTR_DISCOVER_DEVICES, HM_ATTRIBUTE_SUPPORT, HMDevice _LOGGER = logging.getLogger(__name__) HM_TEMP_MAP = ["ACTUAL_TEMPERATURE", "TEMPERATURE"] HM_HUMI_MAP = ["ACTUAL_HUMIDITY", "HUMIDITY"] HM_PRESET_MAP = { "BOOST_MODE": PRESET_BOOST, "COMFORT_MODE": PRESET_COMFORT, "LOWERING_MODE": PRESET_ECO, } HM_CONTROL_MODE = "CONTROL_MODE" HMIP_CONTROL_MODE = "SET_POINT_MODE" SUPPORT_FLAGS = SUPPORT_TARGET_TEMPERATURE | SUPPORT_PRESET_MODE def setup_platform(hass, config, add_entities, discovery_info=None): """Set up the Homematic thermostat platform.""" if discovery_info is None: return devices = [] for conf in discovery_info[ATTR_DISCOVER_DEVICES]: new_device = HMThermostat(conf) devices.append(new_device) add_entities(devices) class HMThermostat(HMDevice, ClimateDevice): """Representation of a Homematic thermostat.""" @property def supported_features(self): """Return the list of supported features.""" return SUPPORT_FLAGS @property def temperature_unit(self): """Return the unit of measurement that is used.""" return TEMP_CELSIUS @property def hvac_mode(self): """Return hvac operation ie. heat, cool mode. Need to be one of HVAC_MODE_*. """ if self.target_temperature <= self._hmdevice.OFF_VALUE + 0.5: return HVAC_MODE_OFF if "MANU_MODE" in self._hmdevice.ACTIONNODE: if self._hm_control_mode == self._hmdevice.MANU_MODE: return HVAC_MODE_HEAT return HVAC_MODE_AUTO # Simple devices if self._data.get("BOOST_MODE"): return HVAC_MODE_AUTO return HVAC_MODE_HEAT @property def hvac_modes(self): """Return the list of available hvac operation modes. Need to be a subset of HVAC_MODES. """ if "AUTO_MODE" in self._hmdevice.ACTIONNODE: return [HVAC_MODE_AUTO, HVAC_MODE_HEAT, HVAC_MODE_OFF] return [HVAC_MODE_HEAT, HVAC_MODE_OFF] @property def preset_mode(self): """Return the current preset mode, e.g., home, away, temp.""" if self._data.get("BOOST_MODE", False): return "boost" # Get the name of the mode mode = HM_ATTRIBUTE_SUPPORT[HM_CONTROL_MODE][1][self._hm_control_mode] mode = mode.lower() # Filter HVAC states if mode not in (HVAC_MODE_AUTO, HVAC_MODE_HEAT): return None return mode @property def preset_modes(self): """Return a list of available preset modes.""" preset_modes = [] for mode in self._hmdevice.ACTIONNODE: if mode in HM_PRESET_MAP: preset_modes.append(HM_PRESET_MAP[mode]) return preset_modes @property def current_humidity(self): """Return the current humidity.""" for node in HM_HUMI_MAP: if node in self._data: return self._data[node] @property def current_temperature(self): """Return the current temperature.""" for node in HM_TEMP_MAP: if node in self._data: return self._data[node] @property def target_temperature(self): """Return the target temperature.""" return self._data.get(self._state) def set_temperature(self, **kwargs): """Set new target temperature.""" temperature = kwargs.get(ATTR_TEMPERATURE) if temperature is None: return None self._hmdevice.writeNodeData(self._state, float(temperature)) def set_hvac_mode(self, hvac_mode): """Set new target hvac mode.""" if hvac_mode == HVAC_MODE_AUTO: self._hmdevice.MODE = self._hmdevice.AUTO_MODE elif hvac_mode == HVAC_MODE_HEAT: self._hmdevice.MODE = self._hmdevice.MANU_MODE elif hvac_mode == HVAC_MODE_OFF: self._hmdevice.turnoff() def set_preset_mode(self, preset_mode: str) -> None: """Set new preset mode.""" if preset_mode == PRESET_BOOST: self._hmdevice.MODE = self._hmdevice.BOOST_MODE elif preset_mode == PRESET_COMFORT: self._hmdevice.MODE = self._hmdevice.COMFORT_MODE elif preset_mode == PRESET_ECO: self._hmdevice.MODE = self._hmdevice.LOWERING_MODE @property def min_temp(self): """Return the minimum temperature.""" return 4.5 @property def max_temp(self): """Return the maximum temperature.""" return 30.5 @property def target_temperature_step(self): """Return the supported step of target temperature.""" return 0.5 @property def _hm_control_mode(self): """Return Control mode.""" if HMIP_CONTROL_MODE in self._data: return self._data[HMIP_CONTROL_MODE] # Homematic return self._data["CONTROL_MODE"] def _init_data_struct(self): """Generate a data dict (self._data) from the Homematic metadata.""" self._state = next(iter(self._hmdevice.WRITENODE.keys())) self._data[self._state] = None if ( HM_CONTROL_MODE in self._hmdevice.ATTRIBUTENODE or HMIP_CONTROL_MODE in self._hmdevice.ATTRIBUTENODE ): self._data[HM_CONTROL_MODE] = None for node in self._hmdevice.SENSORNODE.keys(): self._data[node] = None
py
1a2f643387f6c7bd230bc3199ccef4cb9b4c0248
import os,re import requests import time import json class Connect(object): def __init__(self, name='Athus', icon="Zaika"): self.name = name self.icon = icon self.session = requests.session() def save_cookie(self, file_name): f = open(file_name, 'w+') f.write(str(self.session.cookies.get_dict())) f.close() def login(self): home = self.session.get('https://drrr.com',headers={'User-Agent': 'Bot'}) token = re.search('<input type="hidden" name="token" data-value=".*?">', home.text).group(0)[-34:-2] home.close() login_body = { 'name': self.name, 'login': 'ENTER', 'token': token, 'direct-join': '', 'language': 'en-US', 'icon': self.icon } li = self.session.post('https://drrr.com', login_body, headers={'User-Agent': 'Bot'}) li.close()
py
1a2f6502b0f0af6974f54b5c71da4c379fa3cecb
from GlobalConstants import N_ROWS, N_COLS, FEATURE_COLOUR, FEATURE_SHAPE, FEATURE_SIZE, FEATURE_TEXT import numpy as np from scipy.spatial import distance from DisplayGenerator import DisplayGenerator class ObservationModel(object): def sample(self, action, current_display): """ Samples a random observation from a given display. :param seed: :return: 2D array with colour and shape noisy observation """ x = action / N_COLS y = action % N_COLS obs_space_col = self.observe_feature(current_display, x, y, FEATURE_COLOUR) obs_space_shp = self.observe_feature(current_display, x, y, FEATURE_SHAPE) obs_space_size = self.observe_feature(current_display, x, y, FEATURE_SIZE) obs_space_text = self.observe_feature(current_display, x, y, FEATURE_TEXT) return obs_space_col, obs_space_shp, obs_space_size, obs_space_text def observe_feature(self, display, x, y, feature): observation = self.add_feature_noise(display, x, y, feature) #observation = self.add_spatial_noise(temp, x, y, global_variables) return observation def add_feature_noise(self, features, x, y, feature): obs_space = np.ones((N_ROWS, N_COLS)) * -1 #for COLOUR. if feature == FEATURE_COLOUR: for ext_x in range(0, N_ROWS, 1): for ext_y in range(0, N_COLS, 1): e = self.get_eccentricity(features.x[x][y], features.y[x][y], features.x[ext_x][ext_y], features.y[ext_x][ext_y]) mu = 0.05 + (0.2*e) + (0*e*e) + (0.0004*e*e*e) if features.objects[ext_x][ext_y].color != "BLANK": if features.objects[ext_x][ext_y].size > np.random.normal(mu,0.5,1)[0]: if features.objects[ext_x][ext_y].color == features.target.color: obs_space[ext_x][ext_y] = 1 else: obs_space[ext_x][ext_y] = 0 #for SHAPE. if feature == FEATURE_SHAPE: for ext_x in range(0, N_ROWS, 1): for ext_y in range(0, N_COLS, 1): e = self.get_eccentricity(features.x[x][y], features.y[x][y], features.x[ext_x][ext_y], features.y[ext_x][ext_y]) mu = 0.05 + (0.2*e) + (0*e*e) + (0.025*e*e*e) if features.objects[ext_x][ext_y].color != "BLANK": if features.objects[ext_x][ext_y].size > np.random.normal(mu,0.5,1)[0]: if features.objects[ext_x][ext_y].shape == features.target.shape: obs_space[ext_x][ext_y] = 1 else: obs_space[ext_x][ext_y] = 0 #for SIZE. if feature == FEATURE_SIZE: for ext_x in range(0, N_ROWS, 1): for ext_y in range(0, N_COLS, 1): e = self.get_eccentricity(features.x[x][y], features.y[x][y], features.x[ext_x][ext_y], features.y[ext_x][ext_y]) mu = 0.05 + (0.2*e) + (0*e*e) + (0.0004*e*e*e) if features.objects[ext_x][ext_y].color != "BLANK": if features.objects[ext_x][ext_y].size > np.random.normal(mu,0.5,1)[0]: if features.objects[ext_x][ext_y].size == features.target.size: obs_space[ext_x][ext_y] = 1 else: obs_space[ext_x][ext_y] = 0 #for TEXT. if feature == FEATURE_TEXT: for ext_x in range(0, N_ROWS, 1): for ext_y in range(0, N_COLS, 1): e = self.get_eccentricity(features.x[x][y], features.y[x][y], features.x[ext_x][ext_y], features.y[ext_x][ext_y]) mu = 0.05 + (0.1*e) + (0*e*e) + (0.05*e*e*e) if features.objects[ext_x][ext_y].color != "BLANK": if .26 > np.random.normal(mu,1.0,1)[0]: if features.objects[ext_x][ext_y].text == features.target.text: obs_space[ext_x][ext_y] = 1 else: obs_space[ext_x][ext_y] = 0 return obs_space def get_eccentricity(self, fix_x, fix_y, ext_x, ext_y): return distance.euclidean([fix_x, fix_y], [ext_x, ext_y]) #gen = DisplayGenerator() #env = gen.sample() #space = np.ones((N_ROWS, N_COLS)) * -1 #print env.target #for ext_x in range(0, N_ROWS, 1): # for ext_y in range(0, N_COLS, 1): # if env.objects[ext_x][ext_y].text == env.target.text: # space[ext_x][ext_y] = 1 #print space #print "------------" #model = ObservationModel() #col, shp, size, txt = model.sample(0, env) #print txt #print "------------" #col, shp, size, txt = model.sample(10, env) #print txt #print "------------" #col, shp, size, txt = model.sample(38, env) #print txt #print "------------" #col, shp, size, txt = model.sample(70, env) #print txt #print "------------" #col, shp, size, txt = model.sample(76, env) #print txt
py
1a2f6603f13d1021c92282b072a8ec83bbe50b05
# http://people.duke.edu/~ccc14/sta-663-2016/16A_MCMC.html import numpy as np import matplotlib.pyplot as plt from scipy import stats thetas = np.linspace(0, 1, 200) n = 100 h = 61 a = 10 b = 10 def target(lik, prior, n, h, theta): if theta < 0 or theta > 1: return 0 else: return lik(n, theta).pmf(h)*prior.pdf(theta) def mh_coin(niters, n, h, theta, lik, prior, sigma): samples = [theta] while len(samples) < niters: theta_p = theta + stats.norm(0, sigma).rvs() rho = min(1, target(lik, prior, n, h, theta_p)/target(lik, prior, n, h, theta )) u = np.random.uniform() if u < rho: theta = theta_p samples.append(theta) return samples lik = stats.binom prior = stats.beta(a, b) sigma = 0.05 niters = 100 sampless = [mh_coin(niters, n, h, theta, lik, prior, sigma) for theta in np.arange(0.1, 1, 0.2)] for samples in sampless: plt.plot(samples, '-o') plt.xlim([0, niters]) plt.ylim([0, 1]);
py
1a2f6622255bf45a2dae7055f355c247c3f4b237
import datetime as dt from unittest.mock import patch, call from model_bakery import baker from django.utils import timezone as djangotime from tacticalrmm.test import TacticalTestCase from .models import AutomatedTask from logs.models import PendingAction from .serializers import AutoTaskSerializer from .tasks import remove_orphaned_win_tasks, run_win_task, create_win_task_schedule class TestAutotaskViews(TacticalTestCase): def setUp(self): self.authenticate() self.setup_coresettings() @patch("automation.tasks.generate_agent_tasks_from_policies_task.delay") @patch("autotasks.tasks.create_win_task_schedule.delay") def test_add_autotask( self, create_win_task_schedule, generate_agent_tasks_from_policies_task ): url = "/tasks/automatedtasks/" # setup data script = baker.make_recipe("scripts.script") agent = baker.make_recipe("agents.agent") policy = baker.make("automation.Policy") check = baker.make_recipe("checks.diskspace_check", agent=agent) # test script set to invalid pk data = {"autotask": {"script": 500}} resp = self.client.post(url, data, format="json") self.assertEqual(resp.status_code, 404) # test invalid policy data = {"autotask": {"script": script.id}, "policy": 500} resp = self.client.post(url, data, format="json") self.assertEqual(resp.status_code, 404) # test invalid agent data = { "autotask": {"script": script.id}, "agent": 500, } resp = self.client.post(url, data, format="json") self.assertEqual(resp.status_code, 404) # test add task to agent data = { "autotask": { "name": "Test Task Scheduled with Assigned Check", "run_time_days": ["Sunday", "Monday", "Friday"], "run_time_minute": "10:00", "timeout": 120, "enabled": True, "script": script.id, "script_args": None, "task_type": "scheduled", "assigned_check": check.id, }, "agent": agent.id, } resp = self.client.post(url, data, format="json") self.assertEqual(resp.status_code, 200) create_win_task_schedule.assert_called() # test add task to policy data = { "autotask": { "name": "Test Task Manual", "run_time_days": [], "timeout": 120, "enabled": True, "script": script.id, "script_args": None, "task_type": "manual", "assigned_check": None, }, "policy": policy.id, } resp = self.client.post(url, data, format="json") self.assertEqual(resp.status_code, 200) generate_agent_tasks_from_policies_task.assert_called_with(policy.id) self.check_not_authenticated("post", url) def test_get_autotask(self): # setup data agent = baker.make_recipe("agents.agent") baker.make("autotasks.AutomatedTask", agent=agent, _quantity=3) url = f"/tasks/{agent.id}/automatedtasks/" resp = self.client.get(url, format="json") serializer = AutoTaskSerializer(agent) self.assertEqual(resp.status_code, 200) self.assertEqual(resp.data, serializer.data) self.check_not_authenticated("get", url) @patch("autotasks.tasks.enable_or_disable_win_task.delay") @patch("automation.tasks.update_policy_task_fields_task.delay") def test_update_autotask( self, update_policy_task_fields_task, enable_or_disable_win_task ): # setup data agent = baker.make_recipe("agents.agent") agent_task = baker.make("autotasks.AutomatedTask", agent=agent) policy = baker.make("automation.Policy") policy_task = baker.make("autotasks.AutomatedTask", policy=policy) # test invalid url resp = self.client.patch("/tasks/500/automatedtasks/", format="json") self.assertEqual(resp.status_code, 404) url = f"/tasks/{agent_task.id}/automatedtasks/" # test editing agent task data = {"enableordisable": False} resp = self.client.patch(url, data, format="json") self.assertEqual(resp.status_code, 200) enable_or_disable_win_task.assert_called_with(pk=agent_task.id, action=False) url = f"/tasks/{policy_task.id}/automatedtasks/" # test editing policy task data = {"enableordisable": True} resp = self.client.patch(url, data, format="json") self.assertEqual(resp.status_code, 200) update_policy_task_fields_task.assert_called_with(policy_task.id, True) self.check_not_authenticated("patch", url) @patch("autotasks.tasks.delete_win_task_schedule.delay") @patch("automation.tasks.delete_policy_autotask_task.delay") def test_delete_autotask( self, delete_policy_autotask_task, delete_win_task_schedule ): # setup data agent = baker.make_recipe("agents.agent") agent_task = baker.make("autotasks.AutomatedTask", agent=agent) policy = baker.make("automation.Policy") policy_task = baker.make("autotasks.AutomatedTask", policy=policy) # test invalid url resp = self.client.delete("/tasks/500/automatedtasks/", format="json") self.assertEqual(resp.status_code, 404) # test delete agent task url = f"/tasks/{agent_task.id}/automatedtasks/" resp = self.client.delete(url, format="json") self.assertEqual(resp.status_code, 200) delete_win_task_schedule.assert_called_with(pk=agent_task.id) # test delete policy task url = f"/tasks/{policy_task.id}/automatedtasks/" resp = self.client.delete(url, format="json") self.assertEqual(resp.status_code, 200) delete_policy_autotask_task.assert_called_with(policy_task.id) self.check_not_authenticated("delete", url) @patch("agents.models.Agent.nats_cmd") def test_run_autotask(self, nats_cmd): # setup data agent = baker.make_recipe("agents.agent", version="1.1.0") task = baker.make("autotasks.AutomatedTask", agent=agent) # test invalid url resp = self.client.get("/tasks/runwintask/500/", format="json") self.assertEqual(resp.status_code, 404) # test run agent task url = f"/tasks/runwintask/{task.id}/" resp = self.client.get(url, format="json") self.assertEqual(resp.status_code, 200) nats_cmd.assert_called_with({"func": "runtask", "taskpk": task.id}, wait=False) nats_cmd.reset_mock() old_agent = baker.make_recipe("agents.agent", version="1.0.2") task2 = baker.make("autotasks.AutomatedTask", agent=old_agent) url = f"/tasks/runwintask/{task2.id}/" resp = self.client.get(url, format="json") self.assertEqual(resp.status_code, 400) nats_cmd.assert_not_called() self.check_not_authenticated("get", url) class TestAutoTaskCeleryTasks(TacticalTestCase): def setUp(self): self.authenticate() self.setup_coresettings() @patch("agents.models.Agent.nats_cmd") def test_remove_orphaned_win_task(self, nats_cmd): self.agent = baker.make_recipe("agents.agent") self.task1 = AutomatedTask.objects.create( agent=self.agent, name="test task 1", win_task_name=AutomatedTask.generate_task_name(), ) # test removing an orphaned task win_tasks = [ "Adobe Acrobat Update Task", "AdobeGCInvoker-1.0", "GoogleUpdateTaskMachineCore", "GoogleUpdateTaskMachineUA", "OneDrive Standalone Update Task-S-1-5-21-717461175-241712648-1206041384-1001", self.task1.win_task_name, "TacticalRMM_fixmesh", "TacticalRMM_SchedReboot_jk324kajd", "TacticalRMM_iggrLcOaldIZnUzLuJWPLNwikiOoJJHHznb", # orphaned task ] self.calls = [ call({"func": "listschedtasks"}, timeout=10), call( { "func": "delschedtask", "schedtaskpayload": { "name": "TacticalRMM_iggrLcOaldIZnUzLuJWPLNwikiOoJJHHznb" }, }, timeout=10, ), ] nats_cmd.side_effect = [win_tasks, "ok"] ret = remove_orphaned_win_tasks.s(self.agent.pk).apply() self.assertEqual(nats_cmd.call_count, 2) nats_cmd.assert_has_calls(self.calls) self.assertEqual(ret.status, "SUCCESS") # test nats delete task fail nats_cmd.reset_mock() nats_cmd.side_effect = [win_tasks, "error deleting task"] ret = remove_orphaned_win_tasks.s(self.agent.pk).apply() nats_cmd.assert_has_calls(self.calls) self.assertEqual(nats_cmd.call_count, 2) self.assertEqual(ret.status, "SUCCESS") # no orphaned tasks nats_cmd.reset_mock() win_tasks.remove("TacticalRMM_iggrLcOaldIZnUzLuJWPLNwikiOoJJHHznb") nats_cmd.side_effect = [win_tasks, "ok"] ret = remove_orphaned_win_tasks.s(self.agent.pk).apply() self.assertEqual(nats_cmd.call_count, 1) self.assertEqual(ret.status, "SUCCESS") @patch("agents.models.Agent.nats_cmd") def test_run_win_task(self, nats_cmd): self.agent = baker.make_recipe("agents.agent") self.task1 = AutomatedTask.objects.create( agent=self.agent, name="test task 1", win_task_name=AutomatedTask.generate_task_name(), ) nats_cmd.return_value = "ok" ret = run_win_task.s(self.task1.pk).apply() self.assertEqual(ret.status, "SUCCESS") @patch("agents.models.Agent.nats_cmd") def test_create_win_task_schedule(self, nats_cmd): self.agent = baker.make_recipe("agents.agent") task_name = AutomatedTask.generate_task_name() # test scheduled task self.task1 = AutomatedTask.objects.create( agent=self.agent, name="test task 1", win_task_name=task_name, task_type="scheduled", run_time_bit_weekdays=127, run_time_minute="21:55", ) self.assertEqual(self.task1.sync_status, "notsynced") nats_cmd.return_value = "ok" ret = create_win_task_schedule.s(pk=self.task1.pk, pending_action=False).apply() self.assertEqual(nats_cmd.call_count, 1) nats_cmd.assert_called_with( { "func": "schedtask", "schedtaskpayload": { "type": "rmm", "trigger": "weekly", "weekdays": 127, "pk": self.task1.pk, "name": task_name, "hour": 21, "min": 55, }, }, timeout=10, ) self.task1 = AutomatedTask.objects.get(pk=self.task1.pk) self.assertEqual(self.task1.sync_status, "synced") nats_cmd.return_value = "timeout" ret = create_win_task_schedule.s(pk=self.task1.pk, pending_action=False).apply() self.assertEqual(ret.status, "SUCCESS") self.task1 = AutomatedTask.objects.get(pk=self.task1.pk) self.assertEqual(self.task1.sync_status, "notsynced") # test pending action self.pending_action = PendingAction.objects.create( agent=self.agent, action_type="taskaction" ) self.assertEqual(self.pending_action.status, "pending") nats_cmd.return_value = "ok" ret = create_win_task_schedule.s( pk=self.task1.pk, pending_action=self.pending_action.pk ).apply() self.assertEqual(ret.status, "SUCCESS") self.pending_action = PendingAction.objects.get(pk=self.pending_action.pk) self.assertEqual(self.pending_action.status, "completed") # test runonce with future date nats_cmd.reset_mock() task_name = AutomatedTask.generate_task_name() run_time_date = djangotime.now() + djangotime.timedelta(hours=22) self.task2 = AutomatedTask.objects.create( agent=self.agent, name="test task 2", win_task_name=task_name, task_type="runonce", run_time_date=run_time_date, ) nats_cmd.return_value = "ok" ret = create_win_task_schedule.s(pk=self.task2.pk, pending_action=False).apply() nats_cmd.assert_called_with( { "func": "schedtask", "schedtaskpayload": { "type": "rmm", "trigger": "once", "pk": self.task2.pk, "name": task_name, "year": int(dt.datetime.strftime(self.task2.run_time_date, "%Y")), "month": dt.datetime.strftime(self.task2.run_time_date, "%B"), "day": int(dt.datetime.strftime(self.task2.run_time_date, "%d")), "hour": int(dt.datetime.strftime(self.task2.run_time_date, "%H")), "min": int(dt.datetime.strftime(self.task2.run_time_date, "%M")), }, }, timeout=10, ) self.assertEqual(ret.status, "SUCCESS") # test runonce with date in the past nats_cmd.reset_mock() task_name = AutomatedTask.generate_task_name() run_time_date = djangotime.now() - djangotime.timedelta(days=13) self.task3 = AutomatedTask.objects.create( agent=self.agent, name="test task 3", win_task_name=task_name, task_type="runonce", run_time_date=run_time_date, ) nats_cmd.return_value = "ok" ret = create_win_task_schedule.s(pk=self.task3.pk, pending_action=False).apply() self.task3 = AutomatedTask.objects.get(pk=self.task3.pk) self.assertEqual(ret.status, "SUCCESS") # test checkfailure nats_cmd.reset_mock() self.check = baker.make_recipe("checks.diskspace_check", agent=self.agent) task_name = AutomatedTask.generate_task_name() self.task4 = AutomatedTask.objects.create( agent=self.agent, name="test task 4", win_task_name=task_name, task_type="checkfailure", assigned_check=self.check, ) nats_cmd.return_value = "ok" ret = create_win_task_schedule.s(pk=self.task4.pk, pending_action=False).apply() nats_cmd.assert_called_with( { "func": "schedtask", "schedtaskpayload": { "type": "rmm", "trigger": "manual", "pk": self.task4.pk, "name": task_name, }, }, timeout=10, ) self.assertEqual(ret.status, "SUCCESS") # test manual nats_cmd.reset_mock() task_name = AutomatedTask.generate_task_name() self.task5 = AutomatedTask.objects.create( agent=self.agent, name="test task 5", win_task_name=task_name, task_type="manual", ) nats_cmd.return_value = "ok" ret = create_win_task_schedule.s(pk=self.task5.pk, pending_action=False).apply() nats_cmd.assert_called_with( { "func": "schedtask", "schedtaskpayload": { "type": "rmm", "trigger": "manual", "pk": self.task5.pk, "name": task_name, }, }, timeout=10, ) self.assertEqual(ret.status, "SUCCESS")