commit
stringlengths
40
40
old_file
stringlengths
7
86
new_file
stringlengths
7
86
old_contents
stringlengths
63
1.64k
new_contents
stringlengths
243
2.71k
subject
stringlengths
24
138
message
stringlengths
25
394
lang
stringclasses
1 value
license
stringclasses
7 values
repos
stringlengths
12
20k
70d2f182c09583802da2860994a99fd2bc9e39d5
kimochiconsumer/views.py
kimochiconsumer/views.py
from pyramid.view import view_config from pyramid.httpexceptions import ( HTTPNotFound, ) @view_config(route_name='page', renderer='templates/page.mako') @view_config(route_name='page_view', renderer='templates/page.mako') def page_view(request): if 'page_id' in request.matchdict: data = request.kimochi.page(request.matchdict['page_id']) else: data = request.kimochi.page('1') return data @view_config(route_name='gallery_view', renderer='templates/gallery.mako') def gallery_view(request): data = request.kimochi.gallery(request.matchdict['gallery_id']) if 'gallery' not in data or not data['gallery']: raise HTTPNotFound return data @view_config(route_name='gallery_image_view', renderer='templates/gallery_image.mako') def gallery_image_view(request): data = request.kimochi.gallery_image(request.matchdict['gallery_id'], request.matchdict['image_id']) if 'gallery' not in data or not data['gallery']: raise HTTPNotFound return data
from pyramid.view import view_config from pyramid.httpexceptions import ( HTTPNotFound, ) @view_config(route_name='page', renderer='templates/page.mako') @view_config(route_name='page_view', renderer='templates/page.mako') def page_view(request): if 'page_id' in request.matchdict: data = request.kimochi.page(request.matchdict['page_id']) else: data = request.kimochi.page('index') return data @view_config(route_name='gallery_view', renderer='templates/gallery.mako') def gallery_view(request): data = request.kimochi.gallery(request.matchdict['gallery_id']) if 'gallery' not in data or not data['gallery']: raise HTTPNotFound return data @view_config(route_name='gallery_image_view', renderer='templates/gallery_image.mako') def gallery_image_view(request): data = request.kimochi.gallery_image(request.matchdict['gallery_id'], request.matchdict['image_id']) if 'gallery' not in data or not data['gallery']: raise HTTPNotFound return data
Use 'index' as the default page alias for lookups
Use 'index' as the default page alias for lookups
Python
mit
matslindh/kimochi-consumer
5d61b4904057acbe235b74fc1122d09aa365bdeb
edx_data_research/monitor/monitor_tracking.py
edx_data_research/monitor/monitor_tracking.py
import sys import time from watchdog.observers import Observer from watchdog.events import FileSystemEventHandler class TrackingEventHandler(FileSystemEventHandler): def on_created(self, event): pass def on_moved(self, event): pass if __name__ == "__main__": if len(sys.argv) > 1: args = sys.argv[1] else: raise ValueError('Missing path to directory to monitor!!!') event_handler = TrackingEventHandler() observer = Observer() observer.schedule(event_handler, path, recursive=True) observer.start() try: while True: time.sleep(1) except KeyboardInterrupt: observer.stop() observer.join()
import sys import time from watchdog.observers import Observer from watchdog.events import PatternMatchingEventHandler class TrackingLogHandler(PatternMatchingEventHandler): def on_created(self, event): print event.__repr__() print event.event_type, event.is_directory, event.src_path if __name__ == "__main__": if len(sys.argv) > 1: path = sys.argv[1] else: raise ValueError('Missing path to directory to monitor!!!') event_handler = TrackingLogHandler(['*.log'], ['*.log-errors'], case_sensitive=True) observer = Observer() observer.schedule(event_handler, path, recursive=True) observer.start() try: while True: time.sleep(1) except KeyboardInterrupt: observer.stop() observer.join()
Define handler for tracking log files
Define handler for tracking log files
Python
mit
McGillX/edx_data_research,McGillX/edx_data_research,McGillX/edx_data_research
91853432d2e57bd7c01403c943fff4c2dad1cf5a
openquake/__init__.py
openquake/__init__.py
# -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # # Copyright (C) 2010-2016 GEM Foundation # # OpenQuake is free software: you can redistribute it and/or modify it # under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # OpenQuake is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with OpenQuake. If not, see <http://www.gnu.org/licenses/>. __import__('pkg_resources').declare_namespace(__name__)
# -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # # Copyright (C) 2010-2016 GEM Foundation # # OpenQuake is free software: you can redistribute it and/or modify it # under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # OpenQuake is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with OpenQuake. If not, see <http://www.gnu.org/licenses/>. # Make the namespace compatible with old setuptools, like the one # provided by QGIS 2.1x on Windows try: __import__('pkg_resources').declare_namespace(__name__) except ImportError: __path__ = __import__('pkgutil').extend_path(__path__, __name__)
Make the openquake namespace compatible with old setuptools
Make the openquake namespace compatible with old setuptools Former-commit-id: b1323f4831645a19d5e927fc342abe4b319a76bb [formerly 529c98ec0a7c5a3fefa4da6cdf2f6a58b5487ebc] [formerly 529c98ec0a7c5a3fefa4da6cdf2f6a58b5487ebc [formerly e5f4dc01e94694bf9bfcae3ecd6eca34a33a24eb]] Former-commit-id: e01df405c03f37a89cdf889c45de410cb1ca9b00 Former-commit-id: f8d3b5d4c1d3d81dee1c22a4e2563e6b8d116c74
Python
agpl-3.0
gem/oq-engine,gem/oq-engine,gem/oq-engine,gem/oq-engine,gem/oq-engine
ca2789ad15cba31449e4946494122ab271a83c92
inspirationforge/settings/production.py
inspirationforge/settings/production.py
from .base import * DEBUG = False # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.8/howto/static-files/ # TODO: Add MEDIA_ROOT setting. #MEDIA_ROOT = get_secret("MEDIA_ROOT") # Security-related settings ALLOWED_HOSTS = ["*"] SECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTO', 'https') CSRF_COOKIE_HTTPONLY = True X_FRAME_OPTIONS = 'DENY' # Static asset configuration STATIC_URL = '/static/' STATIC_ROOT = 'staticfiles' STATICFILES_DIRS = (os.path.join(BASE_DIR, 'static'),)
from .base import * DEBUG = False # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.8/howto/static-files/ MEDIA_ROOT = os.path.join(BASE_DIR, 'media') # Security-related settings ALLOWED_HOSTS = ["*"] SECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTO', 'https') CSRF_COOKIE_HTTPONLY = True X_FRAME_OPTIONS = 'DENY' # Static asset configuration STATIC_URL = '/static/' STATIC_ROOT = 'staticfiles' STATICFILES_DIRS = (os.path.join(BASE_DIR, 'static'),)
Set MEDIA_ROOT setting for Production.
Set MEDIA_ROOT setting for Production.
Python
mit
FarmCodeGary/InspirationForge,FarmCodeGary/InspirationForge,FarmCodeGary/InspirationForge
2eba1fc80263c11a1b2b5ee1707b19e98a7b2980
apps/submission/tests/test_models.py
apps/submission/tests/test_models.py
null
from django.test import TestCase from django.urls import reverse from apps.core.factories import PIXELER_PASSWORD, PixelerFactory from .. import models class SubmissionProcessTestCase(TestCase): def test_can_create_submission_process(self): label = 'Candida datasest 0001' qs = models.SubmissionProcess.objects.all() self.assertEqual(qs.count(), 0) process = models.SubmissionProcess.objects.create( label=label, ) self.assertEqual(process.label, label) self.assertEqual(qs.count(), 1) def test_archive_upload_to(self): # Create process and activate tasks self.user = PixelerFactory( is_active=True, is_staff=True, is_superuser=True, ) self.client.login( username=self.user.username, password=PIXELER_PASSWORD, ) self.client.post( reverse('submission:start'), data={ 'label': 'Candida datasest 0001', '_viewflow_activation-started': '2000-01-01', }, follow=True, ) process = models.SubmissionProcess.objects.get() filename = 'archive.zip' upload_path = models.SubmissionProcess.archive_upload_to( process, filename ) expected = '{}/submissions/{}/{}'.format( process.created_by.id, process.id, filename ) self.assertEqual(upload_path, expected)
Add tests for the SubmissionProcess model
Add tests for the SubmissionProcess model
Python
bsd-3-clause
Candihub/pixel,Candihub/pixel,Candihub/pixel,Candihub/pixel,Candihub/pixel
347038c528b07f2553f09daab6915828ab2a6113
tests/__init__.py
tests/__init__.py
# -*- coding: utf-8 -*- # Copyright 2011-2012 Antoine Bertin <diaoulael@gmail.com> # # This file is part of subliminal. # # subliminal is free software; you can redistribute it and/or modify it under # the terms of the GNU Lesser General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # subliminal is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public License # along with subliminal. If not, see <http://www.gnu.org/licenses/>. from . import test_language, test_services, test_subliminal import unittest suite = unittest.TestSuite([test_language.suite(), test_services.suite(), test_subliminal.suite()]) if __name__ == '__main__': unittest.TextTestRunner().run(suite)
# -*- coding: utf-8 -*- # Copyright 2011-2012 Antoine Bertin <diaoulael@gmail.com> # # This file is part of subliminal. # # subliminal is free software; you can redistribute it and/or modify it under # the terms of the GNU Lesser General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # # subliminal is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public License # along with subliminal. If not, see <http://www.gnu.org/licenses/>. from . import test_language, test_services, test_subliminal, test_videos import unittest suite = unittest.TestSuite([test_language.suite(), test_services.suite(), test_subliminal.suite(), test_videos.suite()]) if __name__ == '__main__': unittest.TextTestRunner().run(suite)
Add test_videos to the main test suite
Add test_videos to the main test suite
Python
mit
oxan/subliminal,t4lwh/subliminal,h3llrais3r/subliminal,getzze/subliminal,bogdal/subliminal,hpsbranco/subliminal,ravselj/subliminal,nvbn/subliminal,fernandog/subliminal,goll/subliminal,ratoaq2/subliminal,Elettronik/subliminal,neo1691/subliminal,juanmhidalgo/subliminal,SickRage/subliminal,Diaoul/subliminal,kbkailashbagaria/subliminal,ofir123/subliminal,pums974/subliminal
4c84dafeca9977543824653e354f113b5142d259
jsonsempai.py
jsonsempai.py
import imp import json import os import sys class Dot(dict): def __init__(self, d): super(dict, self).__init__() for k, v in d.iteritems(): if isinstance(v, dict): self[k] = Dot(v) else: self[k] = v def __getattr__(self, attr): try: return self[attr] except KeyError: raise AttributeError("'{}'".format(attr)) __setattr__ = dict.__setitem__ __delattr__ = dict.__delitem__ class SempaiLoader(object): def find_module(self, name, path=None): for d in sys.path: self.json_path = os.path.join(d, '{}.json'.format(name)) if os.path.isfile(self.json_path): return self return None def load_module(self, name): mod = imp.new_module(name) mod.__file__ = self.json_path mod.__loader__ = self try: with open(self.json_path) as f: d = json.load(f) except ValueError: raise ImportError( '"{}" does not contain valid json.'.format(self.json_path)) except: raise ImportError( 'Could not open "{}".'.format(self.json_path)) mod.__dict__.update(d) for k, i in mod.__dict__.items(): if isinstance(i, dict): mod.__dict__[k] = Dot(i) return mod sys.meta_path.append(SempaiLoader())
import imp import json import os import sys class Dot(dict): def __init__(self, d): super(dict, self).__init__() for k, v in iter(d.items()): if isinstance(v, dict): self[k] = Dot(v) else: self[k] = v def __getattr__(self, attr): try: return self[attr] except KeyError: raise AttributeError("'{}'".format(attr)) __setattr__ = dict.__setitem__ __delattr__ = dict.__delitem__ class SempaiLoader(object): def find_module(self, name, path=None): for d in sys.path: self.json_path = os.path.join(d, '{}.json'.format(name)) if os.path.isfile(self.json_path): return self return None def load_module(self, name): mod = imp.new_module(name) mod.__file__ = self.json_path mod.__loader__ = self try: with open(self.json_path) as f: d = json.load(f) except ValueError: raise ImportError( '"{}" does not contain valid json.'.format(self.json_path)) except: raise ImportError( 'Could not open "{}".'.format(self.json_path)) mod.__dict__.update(d) for k, i in mod.__dict__.items(): if isinstance(i, dict): mod.__dict__[k] = Dot(i) return mod sys.meta_path.append(SempaiLoader())
Fix python 3 use of iteritems
Fix python 3 use of iteritems
Python
mit
kragniz/json-sempai
743f4affcd89aa3d9fd37774e2e5f8e05525cb04
api/sync_wallet.py
api/sync_wallet.py
import urlparse import os, sys import json tools_dir = os.environ.get('TOOLSDIR') lib_path = os.path.abspath(tools_dir) sys.path.append(lib_path) from msc_apps import * data_dir_root = os.environ.get('DATADIR') def sync_wallet_response(request_dict): if not request_dict.has_key('type'): return (None, 'No field type in response dict '+str(request_dict)) req_type = request_dict['type'][0].upper() if req_type == "SYNCWALLET": response_data = syncWallets(request_dict['masterWallets'][0]) else: return (None, req_type + ' is not supported') response = { 'status': 'OK', 'data': response_data } return (json.dumps(response), None) def syncWallets(master_wallets_json): master_wallets = json.loads(master_wallets_json) for wallet in master_wallets: uuid = wallet['uuid'] filename = data_dir_root + '/wallets/' + uuid + '.json' with open(filename, 'w') as f: json.dump(wallet, f) return "OK" def sync_wallet_handler(environ, start_response): return general_handler(environ, start_response, sync_wallet_response)
import urlparse import os, sys import json tools_dir = os.environ.get('TOOLSDIR') lib_path = os.path.abspath(tools_dir) sys.path.append(lib_path) from msc_apps import * data_dir_root = os.environ.get('DATADIR') def sync_wallet_response(request_dict): if not request_dict.has_key('type'): return (None, 'No field type in response dict '+str(request_dict)) req_type = request_dict['type'][0].upper() if req_type == "SYNCWALLET": syncWallets(request_dict['masterWallets'][0]) else: return (None, req_type + ' is not supported') response = { 'status': 'OK' } return (json.dumps(response), None) def syncWallets(master_wallets_json): master_wallets = json.loads(master_wallets_json) for wallet in master_wallets: uuid = wallet['uuid'] filename = data_dir_root + '/wallets/' + uuid + '.json' with open(filename, 'w') as f: json.dump(wallet, f) return "OK" def sync_wallet_handler(environ, start_response): return general_handler(environ, start_response, sync_wallet_response)
Clean up return value for API
Clean up return value for API
Python
agpl-3.0
ripper234/omniwallet,maran/omniwallet,maran/omniwallet,Nevtep/omniwallet,FuzzyBearBTC/omniwallet,FuzzyBearBTC/omniwallet,achamely/omniwallet,curtislacy/omniwallet,habibmasuro/omniwallet,OmniLayer/omniwallet,ripper234/omniwallet,habibmasuro/omniwallet,ripper234/omniwallet,Nevtep/omniwallet,habibmasuro/omniwallet,curtislacy/omniwallet,OmniLayer/omniwallet,dexX7/omniwallet,arowser/omniwallet,habibmasuro/omniwallet,dexX7/omniwallet,Nevtep/omniwallet,VukDukic/omniwallet,arowser/omniwallet,achamely/omniwallet,FuzzyBearBTC/omniwallet,maran/omniwallet,VukDukic/omniwallet,OmniLayer/omniwallet,Nevtep/omniwallet,achamely/omniwallet,arowser/omniwallet,VukDukic/omniwallet,OmniLayer/omniwallet,achamely/omniwallet,dexX7/omniwallet,curtislacy/omniwallet
c145b2cc08b3bbf0d2506afb58116e1a0c2dc4fc
tests/core_tests.py
tests/core_tests.py
null
from graffiti import core from graffiti import util def test_schema(): assert "fn" in core.schema(1) fn = lambda x: 1 assert core.schema(fn) == util.fninfo(fn) def t(): return 1 t._schema = { "schema": 1 } assert core.schema(t) == { "schema": 1 } def test_dependencies(): g = { "a": util.fninfo(lambda x: 1), "b": util.fninfo(lambda y, z: 2), "c": util.fninfo(lambda: 3), "d": util.fninfo(lambda o=1: o) } assert core.dependencies(g) == { "a": {"x"}, "b": {"y", "z"}, "c": set(), "d": set() } def test_transitive(): g = { "a": {"b"}, "b": {"c"}, "c": {"d"} } assert core.transitive(g) == { "a": {"b", "c", "d"}, "b": {"c", "d"}, "c": {"d"} } def test_topological(): g = { "a": {"b", "c", "d"}, "b": {"c", "d"}, "c": {"d"}, "d": {} } res = core.topological(g) assert res.index("d") > res.index("c") assert res.index("c") > res.index("b") assert res.index("b") > res.index("a")
Add tests for core graph functions
Add tests for core graph functions
Python
mit
SegFaultAX/graffiti
08a6dddb866ec53ff45a302d7c163d041bbefe71
protoplot-test/test_options_resolving.py
protoplot-test/test_options_resolving.py
null
import unittest from protoplot.engine.item import Item from protoplot.engine.item_container import ItemContainer class Series(Item): pass Series.options.register("color", True) Series.options.register("lineWidth", False) Series.options.register("lineStyle", False) class TestOptionsResolving(unittest.TestCase): def setUp(self): pass def tearDown(self): pass def testOptionsResolving(self): pass if __name__ == "__main__": #import sys;sys.argv = ['', 'Test.testName'] unittest.main()
Add stub unit test for options resolving
Add stub unit test for options resolving
Python
agpl-3.0
deffi/protoplot
516bebe37212e72362b416bd1d9c87a83726fa5f
changes/api/cluster_nodes.py
changes/api/cluster_nodes.py
from __future__ import absolute_import from datetime import datetime, timedelta from flask.ext.restful import reqparse from changes.api.base import APIView from changes.models import Cluster, JobStep, Node class ClusterNodesAPIView(APIView): parser = reqparse.RequestParser() parser.add_argument('since', type=int, location='args') def get(self, cluster_id): cluster = Cluster.query.get(cluster_id) if cluster is None: return '', 404 queryset = Node.query.filter( Node.clusters.contains(cluster), ) args = self.parser.parse_args() if args.since: cutoff = datetime.utcnow() - timedelta(days=args.since) queryset = queryset.join( JobStep, JobStep.node_id == Node.id, ).filter( JobStep.date_created > cutoff, ).group_by(Node) return self.paginate(queryset)
from __future__ import absolute_import from datetime import datetime, timedelta from flask.ext.restful import reqparse from changes.api.base import APIView from changes.models import Cluster, JobStep, Node class ClusterNodesAPIView(APIView): parser = reqparse.RequestParser() parser.add_argument('since', type=int, location='args') def get(self, cluster_id): cluster = Cluster.query.get(cluster_id) if cluster is None: return '', 404 queryset = Node.query.filter( Node.clusters.contains(cluster), ).order_by(Node.label.asc()) args = self.parser.parse_args() if args.since: cutoff = datetime.utcnow() - timedelta(days=args.since) queryset = queryset.join( JobStep, JobStep.node_id == Node.id, ).filter( JobStep.date_created > cutoff, ).group_by(Node) return self.paginate(queryset)
Enforce ordering on cluster nodes endpoint
Enforce ordering on cluster nodes endpoint
Python
apache-2.0
bowlofstew/changes,bowlofstew/changes,dropbox/changes,wfxiang08/changes,wfxiang08/changes,dropbox/changes,dropbox/changes,dropbox/changes,bowlofstew/changes,bowlofstew/changes,wfxiang08/changes,wfxiang08/changes
c5a2167a63516c23390263408fcd2c9a4f654fc8
webcomix/tests/test_comic_spider.py
webcomix/tests/test_comic_spider.py
null
from webcomix.comic_spider import ComicSpider def test_parse_yields_good_page(mocker): mock_response = mocker.patch('scrapy.http.Response') mock_response.urljoin.return_value = "http://xkcd.com/3/" mock_response.url = "http://xkcd.com/2/" mock_selector = mocker.patch('scrapy.selector.SelectorList') mock_response.xpath.return_value = mock_selector mock_selector.extract_first.side_effect = [ '//imgs.xkcd.com/comics/tree_cropped_(1).jpg', 'xkcd.com/3/' ] spider = ComicSpider() result = spider.parse(mock_response) results = list(result) assert len(results) == 2 assert results[0].get( 'url') == "http://imgs.xkcd.com/comics/tree_cropped_(1).jpg" assert results[1].url == "http://xkcd.com/3/" def test_parse_yields_bad_page(mocker): mock_response = mocker.patch('scrapy.http.Response') mock_response.urljoin.return_value = "http://xkcd.com/3/" mock_response.url = "http://xkcd.com/2/" mock_selector = mocker.patch('scrapy.selector.SelectorList') mock_response.xpath.return_value = mock_selector mock_selector.extract_first.side_effect = [None, 'xkcd.com/3/'] spider = ComicSpider() result = spider.parse(mock_response) results = list(result) assert len(results) == 1 assert results[0].url == "http://xkcd.com/3/"
Add tests for the parse method of the spider
Add tests for the parse method of the spider
Python
mit
J-CPelletier/webcomix,J-CPelletier/webcomix,J-CPelletier/WebComicToCBZ
8f2f89129d24cdaa6bc37e4fec885ac78aa30ce4
test/automl/test_models.py
test/automl/test_models.py
null
# -*- encoding: utf-8 -*- from __future__ import print_function import unittest import mock from autosklearn.automl import AutoML from autosklearn.util.backend import Backend class AutoMLStub(object): def __init__(self): self.__class__ = AutoML class AutoMlModelsTest(unittest.TestCase): def setUp(self): self.automl = AutoMLStub() self.automl._shared_mode = False self.automl._seed = 42 self.automl._backend = mock.Mock(spec=Backend) self.automl._delete_output_directories = lambda: 0 def test_only_loads_ensemble_models(self): identifiers = [(1, 2), (3, 4)] models = [ 42 ] self.automl._backend.load_ensemble.return_value.identifiers_ \ = identifiers self.automl._backend.load_models_by_identifiers.side_effect \ = lambda ids: models if ids is identifiers else None self.automl._load_models() self.assertEqual(models, self.automl.models_) def test_loads_all_models_if_no_ensemble(self): models = [ 42 ] self.automl._backend.load_ensemble.return_value = None self.automl._backend.load_all_models.return_value = models self.automl._load_models() self.assertEqual(models, self.automl.models_) def test_raises_if_no_models(self): self.automl._backend.load_ensemble.return_value = None self.automl._backend.load_all_models.return_value = [] self.assertRaises(ValueError, self.automl._load_models)
Test AutoML usage of Backend to load models
Test AutoML usage of Backend to load models
Python
bsd-3-clause
automl/auto-sklearn,automl/auto-sklearn
7127489fc85537722c6216ea6af0005604214bdc
txircd/modules/umode_i.py
txircd/modules/umode_i.py
from txircd.modbase import Mode class InvisibleMode(Mode): def namesListEntry(self, recipient, channel, user, representation): if channel not in recipient.channels and "i" in user.mode: return "" return representation class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "modes": { "uni": InvisibleMode() } } def cleanup(self): self.ircd.removeMode("uni")
from txircd.modbase import Mode class InvisibleMode(Mode): def namesListEntry(self, recipient, channel, user, representation): if channel.name not in recipient.channels and "i" in user.mode: return "" return representation class Spawner(object): def __init__(self, ircd): self.ircd = ircd def spawn(self): return { "modes": { "uni": InvisibleMode() } } def cleanup(self): self.ircd.removeMode("uni")
Fix interpretation of parameters for names list modification
Fix interpretation of parameters for names list modification
Python
bsd-3-clause
DesertBus/txircd,ElementalAlchemist/txircd,Heufneutje/txircd
8803f6058255237dff39549426ca6a513a25193c
website_product_supplier/__openerp__.py
website_product_supplier/__openerp__.py
# -*- coding: utf-8 -*- # (c) 2015 Antiun Ingeniería S.L. - Sergio Teruel # (c) 2015 Antiun Ingeniería S.L. - Carlos Dauden # License AGPL-3 - See http://www.gnu.org/licenses/agpl-3.0.html { 'name': "Website Product Supplier", 'category': 'Website', 'version': '8.0.1.0.0', 'depends': [ 'website_sale', 'website_portal_purchase', ], 'data': [ 'security/ir.model.access.csv', 'views/product_supplier_view.xml', 'views/website_portal.xml', 'views/assets.xml', 'security/website_product_supplier_security.xml', ], 'qweb': ['static/src/xml/*.xml'], 'author': 'Antiun Ingeniería S.L., ' 'Incaser Informatica S.L., ' 'Odoo Community Association (OCA)', 'website': 'http://www.antiun.com', 'license': 'AGPL-3', 'installable': True, }
# -*- coding: utf-8 -*- # (c) 2015 Antiun Ingeniería S.L. - Sergio Teruel # (c) 2015 Antiun Ingeniería S.L. - Carlos Dauden # License AGPL-3 - See http://www.gnu.org/licenses/agpl-3.0.html { 'name': "Website Product Supplier", 'category': 'Website', 'version': '8.0.1.0.0', 'depends': [ 'website_sale', 'website_portal_purchase', ], 'data': [ 'security/ir.model.access.csv', 'views/product_supplier_view.xml', 'views/website_portal.xml', 'views/assets.xml', 'security/website_product_supplier_security.xml', ], 'images': [], 'qweb': ['static/src/xml/*.xml'], 'author': 'Antiun Ingeniería S.L., ' 'Incaser Informatica S.L., ' 'Odoo Community Association (OCA)', 'website': 'http://www.antiun.com', 'license': 'AGPL-3', 'installable': True, }
Add images key in manifest file
[FIX] website_product_supplier: Add images key in manifest file
Python
agpl-3.0
nuobit/website,open-synergy/website,gfcapalbo/website,LasLabs/website,acsone/website,nuobit/website,LasLabs/website,Yajo/website,LasLabs/website,gfcapalbo/website,kaerdsar/website,Yajo/website,nuobit/website,nuobit/website,Yajo/website,gfcapalbo/website,acsone/website,kaerdsar/website,LasLabs/website,open-synergy/website,acsone/website,acsone/website,gfcapalbo/website,open-synergy/website,brain-tec/website,brain-tec/website,open-synergy/website,brain-tec/website,kaerdsar/website,brain-tec/website,Yajo/website
323d52928fa58196299e1dde0c5b3ef0d1451d9f
setup.py
setup.py
from setuptools import find_packages, setup import sys if 'install' in sys.argv: import webbrowser webbrowser.open('https://www.youtube.com/watch?v=NMZcwXh7HDA', new=2, autoraise=True) setup( name='rdalal', version='1.0', description='Install some sweet Rehan', author='Will Kahn-Greene', author_email='willkg@bluesock.org', url='https://github.com/willkg/rdalal', zip_safe=True, packages=find_packages(), entry_points=""" [console_scripts] rdalal=rdalal.cmdline:run """, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'License :: OSI Approved :: BSD License', 'Operating System :: MacOS :: MacOS X', 'Operating System :: POSIX', 'Operating System :: POSIX :: Linux', 'Operating System :: Unix', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', ], )
from setuptools import find_packages, setup import sys if 'install' in sys.argv: import webbrowser webbrowser.open('https://www.youtube.com/watch?v=NMZcwXh7HDA', new=2, autoraise=True) setup( name='rdalal', version='1.1', description='Install some sweet Rehan', author='Will Kahn-Greene', author_email='willkg@bluesock.org', url='https://github.com/willkg/rdalal', zip_safe=True, packages=find_packages(), entry_points=""" [console_scripts] rdalal=rdalal.cmdline:run """, classifiers=[ 'Development Status :: 4 - Beta', 'Environment :: Console', 'License :: OSI Approved :: BSD License', 'Operating System :: MacOS :: MacOS X', 'Operating System :: POSIX', 'Operating System :: POSIX :: Linux', 'Operating System :: Unix', 'Programming Language :: Python', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', ], )
Set to 1.1 for new release
Set to 1.1 for new release
Python
bsd-3-clause
willkg/rdalal
6e9e6c0fbba6b1f6e97c40181ec58c55e4980995
pyipmi/fw.py
pyipmi/fw.py
"""FW records """ class FWInfo(object): """Object to hold device-reported SPI flash table""" def __str__(self): return "%s | %s | %s | %s | %s" % (self.slot, self.type, self.offset, self.size, self.flags) class FWDownloadResult(object): """Object to hold firmware update results""" start_fw_download_failed = None class FWUploadResult(object): """Object to hold firmware retrieve results""" pass class FWActivateResult(object): """Object to hold firmware activate results""" pass class FWDeactivateResult(object): """Object to hold firmware deactivate results""" pass class FWFlagsResult(object): """Object to hold firmware flag command results""" pass class FWStatus(object): """Object to hold firmware operation status""" pass class FWCancelResult(object): """Object to hold firmware operation cancelation results""" pass class FWCheckResult(object): """Object to hold firmware CRC check results""" pass class FWBlowResult(object): """Object to hold firmware blow results""" pass
"""FW records """ class FWInfo(object): """Object to hold device-reported SPI flash table""" def __str__(self): return "%s | %s | %s | %s | %s" % (self.slot, self.type, self.offset, self.size, self.flags) def __eq__(self, other): if isinstance(other, self.__class__): return (self.slot == other.slot and \ self.type == other.type and \ self.offset == other.offset and \ self.size == other.size and \ self.flags == other.flags) else: return False class FWDownloadResult(object): """Object to hold firmware update results""" start_fw_download_failed = None class FWUploadResult(object): """Object to hold firmware retrieve results""" pass class FWActivateResult(object): """Object to hold firmware activate results""" pass class FWDeactivateResult(object): """Object to hold firmware deactivate results""" pass class FWFlagsResult(object): """Object to hold firmware flag command results""" pass class FWStatus(object): """Object to hold firmware operation status""" pass class FWCancelResult(object): """Object to hold firmware operation cancelation results""" pass class FWCheckResult(object): """Object to hold firmware CRC check results""" pass class FWBlowResult(object): """Object to hold firmware blow results""" pass
Add equality operator to FWInfo
Add equality operator to FWInfo
Python
bsd-3-clause
Cynerva/pyipmi,emaadmanzoor/pyipmi
56471d264671b652b4b40619f709dc6b8e02eac1
dragonflow/db/models/host_route.py
dragonflow/db/models/host_route.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import dragonflow.db.field_types as df_fields import dragonflow.db.model_framework as mf @mf.construct_nb_db_model class HostRoute(mf.ModelBase): id = None destination = df_fields.IpNetworkField(required=True) nexthop = df_fields.IpAddressField(required=True)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from jsonmodels import models import dragonflow.db.field_types as df_fields class HostRoute(models.Base): destination = df_fields.IpNetworkField(required=True) nexthop = df_fields.IpAddressField(required=True)
Change HostRoute to a plain model
Change HostRoute to a plain model Since HostRoute doesn't have id, store it as a plain db model. Change-Id: I3dbb9e5ffa42bf48f47b7010ee6baf470b55e85e Partially-Implements: bp refactor-nb-api
Python
apache-2.0
openstack/dragonflow,openstack/dragonflow,openstack/dragonflow
2dbc4e2aec98aba8a0e307f951b412464db8b078
della/user_manager/urls.py
della/user_manager/urls.py
from django.conf.urls import url from django.contrib.auth import views as auth_views from .views import (SignupView, UserProfileUpdateView, DrawNamesView, UserProfileDetailView, ActivateView) urlpatterns = [ url(r'^login/$', auth_views.login, name='login', kwargs={'template_name': 'user_manager/login.html'}), url(r'^logout/$', auth_views.logout, name='logout', kwargs={'next_page': '/'}), url(r'^create/$', SignupView.as_view()), url(r'^update/$', UserProfileUpdateView.as_view(), name='update'), url(r'^activate/(?P<username>[0-9A-Za-z_]+)-(?P<code>[0-9A-Za-z_:-]+)/$', ActivateView.as_view()), url(r'^draw-names/$', DrawNamesView.as_view()), url(r'^@(?P<username>[a-zA-Z0-9_]+)/$', UserProfileDetailView.as_view(), name='user-detail'), ]
from django.conf.urls import url from django.contrib.auth import views as auth_views from .views import (SignupView, UserProfileUpdateView, DrawNamesView, UserProfileDetailView, ActivateView) urlpatterns = [ url(r'^login/$', auth_views.login, name='login', kwargs={'template_name': 'user_manager/login.html'}), url(r'^logout/$', auth_views.logout, name='logout', kwargs={'next_page': '/'}), url(r'^create/$', SignupView.as_view()), url(r'^update/$', UserProfileUpdateView.as_view(), name='update'), url(r'^activate/(?P<username>[0-9A-Za-z_]+)-(?P<code>[0-9A-Za-z_:-]+)/$', ActivateView.as_view(), name='activate-user'), url(r'^draw-names/$', DrawNamesView.as_view()), url(r'^@(?P<username>[a-zA-Z0-9_]+)/$', UserProfileDetailView.as_view(), name='user-detail'), ]
Add name for url config of `ActivateView`
Add name for url config of `ActivateView`
Python
mit
avinassh/della,avinassh/della,avinassh/della
3279d68859d947f2e618e2770a9fd1b7ce3f26c9
tests/test_cardxml.py
tests/test_cardxml.py
null
from hearthstone.enums import GameTag, Rarity import utils def test_all_tags_known(): """ Iterate through the card database and check that all specified GameTags are known in hearthstone.enums.GameTag """ unknown_tags = set() known_tags = list(GameTag) known_rarities = list(Rarity) # Check the db loaded correctly assert utils.fireplace.cards.db for card in utils.fireplace.cards.db.values(): card_tags = [int(e.attrib["enumID"]) for e in card.xml.findall("./Tag")] for tag in card_tags: # We have fake tags in fireplace.enums which are always negative if tag not in known_tags and tag > 0: unknown_tags.add(tag) # Test rarities as well (cf. TB_BlingBrawl_Blade1e in 10956...) assert card.rarity in known_rarities assert not unknown_tags
Add a test to verify that all GameTags are known
Add a test to verify that all GameTags are known
Python
agpl-3.0
Ragowit/fireplace,smallnamespace/fireplace,jleclanche/fireplace,beheh/fireplace,amw2104/fireplace,amw2104/fireplace,smallnamespace/fireplace,NightKev/fireplace,Ragowit/fireplace
dcfa7bfa11bea86d831959a217b558d704ece078
ensemble/ctf/tests/test_manager.py
ensemble/ctf/tests/test_manager.py
null
from contextlib import contextmanager from os.path import isfile, join import shutil import tempfile from numpy.testing import assert_allclose from ensemble.ctf.editor import ALPHA_DEFAULT, COLOR_DEFAULT, create_function from ensemble.ctf.manager import CTF_EXTENSION, CtfManager @contextmanager def temp_directory(): tempdir = tempfile.mkdtemp(suffix='', prefix='tmp', dir=None) try: yield tempdir finally: shutil.rmtree(tempdir) def sample_function_parts(): return create_function(COLOR_DEFAULT), create_function(ALPHA_DEFAULT) def test_ctf_manager_add(): name = 'test_function' color_func, alpha_func = sample_function_parts() with temp_directory() as root_dir: manager = CtfManager.from_directory(root_dir) manager.add(name, color_func, alpha_func) assert isfile(join(root_dir, name + CTF_EXTENSION)) def test_ctf_manager_get(): color_func, alpha_func = sample_function_parts() with temp_directory() as root_dir: manager = CtfManager.from_directory(root_dir) manager.add('test', color_func, alpha_func) ret_color, ret_alpha = manager.get('test') assert_allclose(ret_color.values(), COLOR_DEFAULT) assert_allclose(ret_alpha.values(), ALPHA_DEFAULT) def test_ctf_manager_load(): color_func, alpha_func = sample_function_parts() with temp_directory() as root_dir: manager = CtfManager.from_directory(root_dir) manager.add('test', color_func, alpha_func) del manager manager = CtfManager.from_directory(root_dir) manager.get('test')
Add unit tests for CtfManager.
Add unit tests for CtfManager.
Python
bsd-3-clause
dmsurti/ensemble
330bdcb3bc50acf4c15299a15fb7fdf8b58e152c
setup.py
setup.py
#!/usr/bin/env python import setuptools setuptools.setup( name='bumper', version='0.1.1', author='Max Zheng', author_email='maxzheng.os @t gmail.com', description=open('README.rst').read(), entry_points={ 'console_scripts': [ 'bump = bumper:bump', ], }, install_requires=open('requirements.txt').read(), license='MIT', setup_requires=['setuptools-git'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Topic :: Software Development :: Development Utilities', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', ], keywords='bump pin requirements requirements.txt pinned.txt', )
#!/usr/bin/env python import setuptools setuptools.setup( name='bumper', version='0.1.1', author='Max Zheng', author_email='maxzheng.os @t gmail.com', description='Bump (pin/manage) your dependency requirements with ease', long_description=open('README.rst').read(), url='https://github.com/maxzheng/bumper', entry_points={ 'console_scripts': [ 'bump = bumper:bump', ], }, install_requires=open('requirements.txt').read(), license='MIT', setup_requires=['setuptools-git'], classifiers=[ 'Development Status :: 5 - Production/Stable', 'Intended Audience :: Developers', 'Topic :: Software Development', 'License :: OSI Approved :: MIT License', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', ], keywords='bump pin requirements requirements.txt pinned.txt', )
Add long description / url
Add long description / url
Python
mit
maxzheng/bumper
02c74c5235b8ad821786213a3bcf5f824162454d
flax/linen/combinators.py
flax/linen/combinators.py
"""Combinators of modules, such as a Sequential.""" from typing import Callable, Sequence from flax.linen.module import Module class Sequential(Module): """Applies a linear chain of Modules. Meant to be used only for the simple case of fusing together callables where the input of a particular module/op is the output of the previous one. Modules will be applied in the order that they are passed in the constructor. The apply() method of Sequential accepts any input and forwards it to the first module it contains. It chains the output sequentially to the input of the next module and returns the output of the final module. Example usage:: class Foo(nn.Module): feature_sizes: Sequence[int] @nn.compact def __call__(self, x): return nn.Sequential([nn.Dense(layer_size, name=f'layers_{idx}') for idx, layer_size in enumerate(self.feature_sizes)])(x) """ layers: Sequence[Callable] def __call__(self, *args, **kwargs): if not self.layers: raise ValueError(f'Empty Sequential module {self.name}.') outputs = self.layers[0](*args, **kwargs) for layer in self.layers[1:]: outputs = layer(outputs) return outputs
"""Combinators of modules, such as a Sequential.""" from typing import Callable, Sequence from flax.linen.module import Module class Sequential(Module): """Applies a linear chain of Modules. Meant to be used only for the simple case of fusing together callables where the input of a particular module/op is the output of the previous one. Modules will be applied in the order that they are passed in the constructor. The apply() method of Sequential accepts any input and forwards it to the first module it contains. It chains the output sequentially to the input of the next module and returns the output of the final module. Example usage:: class Foo(nn.Module): feature_sizes: Sequence[int] @nn.compact def __call__(self, x): return nn.Sequential([nn.Dense(4), nn.relu, nn.Dense(2), nn.log_softmax])(x) """ layers: Sequence[Callable] def __call__(self, *args, **kwargs): if not self.layers: raise ValueError(f'Empty Sequential module {self.name}.') outputs = self.layers[0](*args, **kwargs) for layer in self.layers[1:]: outputs = layer(outputs) return outputs
Include activations in Sequential example.
Include activations in Sequential example.
Python
apache-2.0
google/flax,google/flax
ff618ea57b8f3d71772bcef5f7fecf9eceae4e3d
scripts/upsrv_schema.py
scripts/upsrv_schema.py
#!/usr/bin/python # Copyright (c) 2006 rPath, Inc # All rights reserved import sys import os import pwd from conary.server import schema from conary.lib import cfgtypes from conary.repository.netrepos.netserver import ServerConfig from conary import dbstore cnrPath = '/srv/conary/repository.cnr' cfg = ServerConfig() try: cfg.read(cnrPath) except cfgtypes.CfgEnvironmentError: print "Error reading %s" % cnrPath sys.exit(1) db = dbstore.connect(cfg.repositoryDB[1], cfg.repositoryDB[0]) schema.loadSchema(db, doMigrate=True) if cfg.repositoryDB[0] == 'sqlite': os.chown(cfg.repositoryDB[1], pwd.getpwnam('apache')[2], pwd.getpwnam('apache')[3])
#!/usr/bin/python # Copyright (c) 2006 rPath, Inc # All rights reserved import sys import os import pwd from conary.server import schema from conary.lib import cfgtypes, tracelog from conary.repository.netrepos.netserver import ServerConfig from conary import dbstore cnrPath = '/srv/conary/repository.cnr' cfg = ServerConfig() tracelog.initLog(filename='stdout', level=2) try: cfg.read(cnrPath) except cfgtypes.CfgEnvironmentError: print "Error reading %s" % cnrPath sys.exit(1) db = dbstore.connect(cfg.repositoryDB[1], cfg.repositoryDB[0]) schema.loadSchema(db, doMigrate=True) if cfg.repositoryDB[0] == 'sqlite': os.chown(cfg.repositoryDB[1], pwd.getpwnam('apache')[2], pwd.getpwnam('apache')[3])
Set log level to 2 when migrating so there is some indication it is running
Set log level to 2 when migrating so there is some indication it is running
Python
apache-2.0
sassoftware/rbm,sassoftware/rbm,sassoftware/rbm
4f6fde8329b0873f3568ce7153dc64017f5bc0cb
boto/beanstalk/__init__.py
boto/beanstalk/__init__.py
null
# Copyright (c) 2013 Mitch Garnaat http://garnaat.org/ # Copyright (c) 2013 Amazon.com, Inc. or its affiliates. # All Rights Reserved # # Permission is hereby granted, free of charge, to any person obtaining a # copy of this software and associated documentation files (the # "Software"), to deal in the Software without restriction, including # without limitation the rights to use, copy, modify, merge, publish, dis- # tribute, sublicense, and/or sell copies of the Software, and to permit # persons to whom the Software is furnished to do so, subject to the fol- # lowing conditions: # # The above copyright notice and this permission notice shall be included # in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS # OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABIL- # ITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT # SHALL THE AUTHOR BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, # WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS # IN THE SOFTWARE. from layer1 import Layer1 from boto.regioninfo import RegionInfo RegionData = { 'us-east-1': 'elasticbeanstalk.us-east-1.amazonaws.com', 'us-west-1': 'elasticbeanstalk.us-west-1.amazonaws.com', 'us-west-2': 'elasticbeanstalk.us-west-2.amazonaws.com', 'sa-east-1': 'elasticbeanstalk.sa-east-1.amazonaws.com', 'eu-west-1': 'elasticbeanstalk.eu-west-1.amazonaws.com', 'ap-northeast-1': 'elasticbeanstalk.ap-northeast-1.amazonaws.com', 'ap-southeast-1': 'elasticbeanstalk.ap-southeast-1.amazonaws.com', 'ap-southeast-2': 'elasticbeanstalk.ap-southeast-2.amazonaws.com', } def regions(): """ Get all available regions for the Elastic Beanstalk service. :rtype: list :return: A list of :class:`boto.RegionInfo` instances """ regions = [] for region_name in RegionData: region = RegionInfo(name=region_name, endpoint=RegionData[region_name], connection_cls=Layer1) regions.append(region) return regions def connect_to_region(region_name, **kw_params): """ Given a valid region name, return a :class:`boto.beanstalk.Layer1`. :param str region_name: The name of the region to connect to. :rtype: :class:`boto.beanstalk.Layer1` or ``None`` :return: A connection to the given region, or None if an invalid region name is given """ for region in regions(): if region.name == region_name: return region.connect(**kw_params) return None
Add connect_to_region to beanstalk module.
Add connect_to_region to beanstalk module.
Python
mit
jamesls/boto,appneta/boto,rjschwei/boto,SaranyaKarthikeyan/boto,drbild/boto,lra/boto,weebygames/boto,revmischa/boto,s0enke/boto,pfhayes/boto,drbild/boto,khagler/boto,dimdung/boto,Timus1712/boto,trademob/boto,alex/boto,nishigori/boto,jindongh/boto,janslow/boto,garnaat/boto,alfredodeza/boto,jameslegg/boto,disruptek/boto,dablak/boto,bleib1dj/boto,israelbenatar/boto,kouk/boto,nikhilraog/boto,dablak/boto,darjus-amzn/boto,shaunbrady/boto,elainexmas/boto,TiVoMaker/boto,lochiiconnectivity/boto,shipci/boto,bryx-inc/boto,weka-io/boto,rjschwei/boto,lochiiconnectivity/boto,j-carl/boto,ryansb/boto,alex/boto,serviceagility/boto,stevenbrichards/boto,appneta/boto,FATruden/boto,jamesls/boto,yangchaogit/boto,ddzialak/boto,cyclecomputing/boto,campenberger/boto,tpodowd/boto,vishnugonela/boto,nexusz99/boto,Asana/boto,awatts/boto,varunarya10/boto,jotes/boto,vijaylbais/boto,andresriancho/boto,abridgett/boto,andresriancho/boto,ric03uec/boto,Pretio/boto,ramitsurana/boto,acourtney2015/boto,tpodowd/boto,rosmo/boto,jameslegg/boto,disruptek/boto,podhmo/boto,ekalosak/boto,ocadotechnology/boto,felix-d/boto,kouk/boto,zachmullen/boto,rayluo/boto,clouddocx/boto,zzzirk/boto
b9b095a2a66f79e36bbad1affaeb57b38e20803b
cwod_site/cwod/models.py
cwod_site/cwod/models.py
from django.db import models # Create your models here. class CongressionalRecordVolume(models.Model): congress = models.IntegerField(db_index=True) session = models.CharField(max_length=10, db_index=True) volume = models.IntegerField()
from django.db import models # Create your models here. class CongressionalRecordVolume(models.Model): congress = models.IntegerField(db_index=True) session = models.CharField(max_length=10, db_index=True) volume = models.IntegerField() class NgramDateCount(models.Model): """Storing the total number of ngrams per date allows us to show the percentage of a given ngram on a given date, mainly for graphing purposes. """ n = models.IntegerField(db_index=True) date = models.DateField(db_index=True) count = models.IntegerField() class Meta: unique_together = (('n', 'date', ), )
Create model for storing total n-gram counts by date
Create model for storing total n-gram counts by date
Python
bsd-3-clause
sunlightlabs/Capitol-Words,sunlightlabs/Capitol-Words,sunlightlabs/Capitol-Words,sunlightlabs/Capitol-Words,sunlightlabs/Capitol-Words,propublica/Capitol-Words,propublica/Capitol-Words,sunlightlabs/Capitol-Words,sunlightlabs/Capitol-Words,propublica/Capitol-Words,propublica/Capitol-Words
9e900eb16e92027cfe990a07c5703a6adbb41a09
drivers/python/wappalyzer.py
drivers/python/wappalyzer.py
null
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import sys import PyV8 import urllib from urlparse import urlparse try: import json except ImportError: import simplejson as json class Wappalyzer(object): def __init__(self, url): self.file_dir = os.path.dirname(__file__) f = open(os.path.join(self.file_dir, '../../share/apps.json')) data = json.loads(f.read()) f.close() self.categories = data['categories'] self.apps = data['apps'] self.url = url def analyze(self): ctxt = PyV8.JSContext() ctxt.enter() f1 = open(os.path.join(self.file_dir, '../php/js/wappalyzer.js')) f2 = open(os.path.join(self.file_dir, '../php/js/driver.js')) ctxt.eval(f1.read()) ctxt.eval(f2.read()) f1.close() f2.close() host = urlparse(self.url).hostname html = urllib.urlopen(self.url).read() data = {'host': host, 'url': self.url, 'html': html, 'headers': {}} apps = json.dumps(self.apps) categories = json.dumps(self.categories) return ctxt.eval("w.apps = %s; w.categories = %s; w.driver.data = %s; w.driver.init();" % (apps, categories, json.dumps(data))) if __name__ == '__main__': try: w = Wappalyzer(sys.argv[1]) print w.analyze() except IndexError: print ('Usage: python %s <url>' % sys.argv[0])
Add python driver (depend on PyV8)
Add python driver (depend on PyV8)
Python
mit
WPO-Foundation/Wappalyzer,WPO-Foundation/Wappalyzer,WPO-Foundation/Wappalyzer,AliasIO/wappalyzer,AliasIO/wappalyzer
b5477239d7b1ee9e73265b023355e8e83826ec49
scrapy_rss/items.py
scrapy_rss/items.py
# -*- coding: utf-8 -*- import scrapy from scrapy.item import BaseItem from scrapy_rss.elements import * from scrapy_rss import meta import six @six.add_metaclass(meta.ItemMeta) class RssItem: title = TitleElement() link = LinkElement() description = DescriptionElement() author = AuthorElement() category = meta.MultipleElements(CategoryElement) comments = CommentsElement() enclosure = EnclosureElement() guid = GuidElement() pubDate = PubDateElement() source = SourceElement() class ExtendableItem(scrapy.Item): def __setattr__(self, name, value): if name in self.fields: raise AttributeError("Use item[{!r}] = {!r} to set field value".format(name, value)) super(BaseItem, self).__setattr__(name, value) class RssedItem(ExtendableItem): def __init__(self, **kwargs): super(RssedItem, self).__init__(**kwargs) self.rss = RssItem()
# -*- coding: utf-8 -*- import scrapy from scrapy.item import BaseItem from scrapy_rss.elements import * from scrapy_rss import meta import six @six.add_metaclass(meta.ItemMeta) class RssItem(BaseItem): title = TitleElement() link = LinkElement() description = DescriptionElement() author = AuthorElement() category = meta.MultipleElements(CategoryElement) comments = CommentsElement() enclosure = EnclosureElement() guid = GuidElement() pubDate = PubDateElement() source = SourceElement() class ExtendableItem(scrapy.Item): def __setattr__(self, name, value): if name in self.fields: raise AttributeError("Use item[{!r}] = {!r} to set field value".format(name, value)) super(BaseItem, self).__setattr__(name, value) class RssedItem(ExtendableItem): def __init__(self, **kwargs): super(RssedItem, self).__init__(**kwargs) self.rss = RssItem()
Fix RssItem when each scraped item is instance of RssItem
Fix RssItem when each scraped item is instance of RssItem
Python
bsd-3-clause
woxcab/scrapy_rss
01732179f923394bee6e46ae1bbee0f316df8297
amostra/tests/test_revert.py
amostra/tests/test_revert.py
null
from hypothesis import given, strategies as st from hypothesis.strategies import text from hypothesis import settings import random alphabet_list = '' for i in range(26): alphabet_list = alphabet_list + chr(97 + i) @given(names = st.lists(st.text(alphabet=alphabet_list, min_size=1, max_size=4), min_size=3, max_size=4, unique=True)) @settings(max_examples = 10, deadline = 1000) def test_revert(client_conf, names): client, mongo_client = client_conf() n = len(names) s = client.samples.new(name = names[0]) for name in names[1:]: s.name = name num = random.randint(0, n-2) revert_target_cursor = mongo_client['tests-amostra'].samples_revisions.find({'revision': num}) s.revert(num) target = next(revert_target_cursor) for name, trait in s.traits().items(): if name is 'revision': continue else: assert getattr(s, name) == target[name]
Add a revert function test
TST: Add a revert function test
Python
bsd-3-clause
NSLS-II/amostra
ab4983e577b9831b91290976be00917edb9fad6f
mlox/modules/resources.py
mlox/modules/resources.py
"""Handle program wide resources (files, images, etc...)""" import os import sys import base64 import tempfile def unpack_resource(data): """Convert base64 encoded data into a file handle, and a temporary file name to access the data""" file_handle = tempfile.NamedTemporaryFile() file_handle.write(base64.b64decode(data)) file_handle.seek(0) return (file_handle,file_handle.name) #Paths to resource files program_path = os.path.realpath(sys.path[0]) resources_path = os.path.join(program_path,"Resources") translation_file = os.path.join(resources_path,"mlox.msg") gif_file = os.path.join(resources_path,"mlox.gif") base_file = os.path.join(program_path,"mlox_base.txt") user_file = os.path.join(program_path,"mlox_user.txt") #For the updater UPDATE_BASE = "mlox-data.7z" update_file = os.path.join(program_path,UPDATE_BASE) UPDATE_URL = 'https://sourceforge.net/projects/mlox/files/mlox/' + UPDATE_BASE
"""Handle program wide resources (files, images, etc...)""" import os import sys import base64 import tempfile def unpack_resource(data): """Convert base64 encoded data into a file handle, and a temporary file name to access the data""" file_handle = tempfile.NamedTemporaryFile() file_handle.write(base64.b64decode(data)) file_handle.seek(0) return (file_handle,file_handle.name) #Paths to resource files program_path = os.path.realpath(sys.path[0]) resources_path = os.path.join(program_path,"Resources") translation_file = os.path.join(resources_path,"mlox.msg") gif_file = os.path.join(resources_path,"mlox.gif") base_file = os.path.join(program_path,"mlox_base.txt") user_file = os.path.join(program_path,"mlox_user.txt") #For the updater UPDATE_BASE = "mlox-data.7z" update_file = os.path.join(program_path,UPDATE_BASE) UPDATE_URL = 'https://svn.code.sf.net/p/mlox/code/trunk/downloads/' + UPDATE_BASE
Switch back to using the old SVN update location.
Switch back to using the old SVN update location. While changing the download location would be nice, this keeps the option of putting a final data file that would force users to update.
Python
mit
EmperorArthur/mlox,EmperorArthur/mlox,EmperorArthur/mlox
f18cf3c17e450eb6f8db5288ecf146eff0968a47
xmt/select.py
xmt/select.py
from itertools import groupby from nltk.translate.gleu_score import sentence_gleu as gleu from nltk.tokenize.toktok import ToktokTokenizer _tokenize = ToktokTokenizer().tokenize def select_first(p): """ Return (hypothesis, reference) translation pairs using the first realization result per item. """ pairs = [] rows = p.join('item', 'g-result') for i_id, group in groupby(rows, key=lambda row: row['g-result:i-id']): row = next(group) pairs.append((row['g-result:surface'], row['item:i-translation'])) return pairs def select_oracle(p): """ Return (hypothesis, reference) translation pairs using the realization result per item with the highest GLEU score. """ pairs = [] rows = p.join('item', 'g-result') for i_id, group in groupby(rows, key=lambda row: row['g-result:i-id']): hrs = ((r['g-result:surface'], r['item:i-translation']) for r in group) ranked = [(gleu(_tokenize(r), _tokenize(h)), h, r) for h, r in hrs] _, hyp, ref = sorted(ranked, key=lambda r: r[0])[-1] pairs.append((hyp, ref)) return pairs
from itertools import groupby from nltk.translate import bleu_score from nltk.tokenize.toktok import ToktokTokenizer _tokenize = ToktokTokenizer().tokenize _smoother = bleu_score.SmoothingFunction().method3 bleu = bleu_score.sentence_bleu def select_first(p): """ Return (hypothesis, reference) translation pairs using the first realization result per item. """ pairs = [] rows = p.join('item', 'g-result') for i_id, group in groupby(rows, key=lambda row: row['g-result:i-id']): row = next(group) pairs.append((row['g-result:surface'], row['item:i-translation'])) return pairs def select_oracle(p): """ Return (hypothesis, reference) translation pairs using the realization result per item with the highest GLEU score. """ pairs = [] rows = p.join('item', 'g-result') for i_id, group in groupby(rows, key=lambda row: row['g-result:i-id']): scored = [] for res in group: ref = res['item:i-translation'] hyp = res['g-result:surface'] scored.append( (bleu([_tokenize(ref)], _tokenize(hyp), smoothing_function=_smoother), hyp, ref) ) _, hyp, ref = sorted(scored, key=lambda r: r[0])[-1] pairs.append((hyp, ref)) return pairs
Use NIST-BLEU instead of GLEU for oracle.
Use NIST-BLEU instead of GLEU for oracle.
Python
mit
goodmami/xmt,goodmami/xmt
01163ce7fc43b4ab2e5b9ab1c5f94556d0509004
examples/tornado/auth_demo.py
examples/tornado/auth_demo.py
from mongrel2.config import * main = Server( uuid="f400bf85-4538-4f7a-8908-67e313d515c2", access_log="/logs/access.log", error_log="/logs/error.log", chroot="./", default_host="localhost", name="test", pid_file="/run/mongrel2.pid", port=6767, hosts = [ Host(name="localhost", routes={ r'/(.*)': Proxy(addr='127.0.0.1', port=8888) }) ] ) commit([main])
from mongrel2.config import * main = Server( uuid="f400bf85-4538-4f7a-8908-67e313d515c2", access_log="/logs/access.log", error_log="/logs/error.log", chroot="./", default_host="localhost", name="test", pid_file="/run/mongrel2.pid", port=6767, hosts = [ Host(name="localhost", routes={ r'/(.*)': Proxy(addr='127.0.0.1', port=8888) }) ] ) commit([main], settings={'limits.buffer_size': 4 * 1024})
Add the settings to the authdemo.
Add the settings to the authdemo.
Python
bsd-3-clause
niedbalski/mongrel2,jdesgats/mongrel2,winks/mongrel2,griffordson/mongrel2,cpick/mongrel2,duaneg/mongrel2,jagguli/mongrel2,mongrel2/mongrel2,issuu/mongrel2,metadave/mongrel2,steamraven/mongrel2,cpick/mongrel2,niedbalski/mongrel2,AustinWise/mongrel2,nmandery/mongrel2,markokr/mongrel2,reshefm/mongrel2,sshirokov/mongrel2,fanout/mongrel2,nickdesaulniers/mongrel2,mongrel2/mongrel2,ralphbean/mongrel2,aidenkeating/mongrel2,moai/mongrel2,jablkopp/mongrel2,jubarajborgohain/mongrel2,dermoth/mongrel2,bashi-bazouk/mongrel2,ameuret/mongrel2,metadave/mongrel2,markokr/mongrel2,AvdN/mongrel2,elo80ka/mongrel2,nickdesaulniers/mongrel2,wayneeseguin/mongrel2,mongrel2/mongrel2,steamraven/mongrel2,rpeterson/mongrel2,reshefm/mongrel2,ralphbean/mongrel2,winks/mongrel2,AlexVPopov/mongrel2,jiffyjeff/mongrel2,markokr/mongrel2,msteinert/mongrel2,AlexVPopov/mongrel2,duaneg/mongrel2,xrl/mongrel2,dermoth/mongrel2,musl/mongrel2,ameuret/mongrel2,aidenkeating/mongrel2,musl/mongrel2,moai/mongrel2,minrk/mongrel2,markokr/mongrel2,jagguli/mongrel2,jdesgats/mongrel2,apjanke/mongrel2,griffordson/mongrel2,AlexVPopov/mongrel2,xrl/mongrel2,krakensden/mongrel2,dermoth/mongrel2,AustinWise/mongrel2,elo80ka/mongrel2,cpick/mongrel2,msteinert/mongrel2,xrl/mongrel2,wayneeseguin/mongrel2,jasom/mongrel2,Gibheer/mongrel2,jubarajborgohain/mongrel2,jiffyjeff/mongrel2,musl/mongrel2,wayneeseguin/mongrel2,mbj/mongrel2,bashi-bazouk/mongrel2,apjanke/mongrel2,krakensden/mongrel2,fanout/mongrel2,musl/mongrel2,jasom/mongrel2,jubarajborgohain/mongrel2,sshirokov/mongrel2,jagguli/mongrel2,ameuret/mongrel2,msteinert/mongrel2,nickdesaulniers/mongrel2,jagguli/mongrel2,jablkopp/mongrel2,jablkopp/mongrel2,xrl/mongrel2,jasom/mongrel2,ralphbean/mongrel2,musl/mongrel2,rpeterson/mongrel2,yoink00/mongrel2,pjkundert/mongrel2,issuu/mongrel2,ameuret/mongrel2,yoink00/mongrel2,Gibheer/mongrel2,jdesgats/mongrel2,chickenkiller/mongrel2,nmandery/mongrel2,steamraven/mongrel2,nmandery/mongrel2,fanout/mongrel2,pjkundert/mongrel2,ralphbean/mongrel2,mbj/mongrel2,jagguli/mongrel2,steamraven/mongrel2,jdesgats/mongrel2,nickdesaulniers/mongrel2,rpeterson/mongrel2,fanout/mongrel2,minrk/mongrel2,elo80ka/mongrel2,aidenkeating/mongrel2,apjanke/mongrel2,apjanke/mongrel2,jdesgats/mongrel2,xrl/mongrel2,elo80ka/mongrel2,metadave/mongrel2,bashi-bazouk/mongrel2,krakensden/mongrel2,reshefm/mongrel2,mbj/mongrel2,jasom/mongrel2,moai/mongrel2,minrk/mongrel2,yoink00/mongrel2,AustinWise/mongrel2,moai/mongrel2,metadave/mongrel2,bashi-bazouk/mongrel2,nickdesaulniers/mongrel2,rpeterson/mongrel2,pjkundert/mongrel2,Gibheer/mongrel2,krakensden/mongrel2,krakensden/mongrel2,jablkopp/mongrel2,jablkopp/mongrel2,AvdN/mongrel2,wayneeseguin/mongrel2,issuu/mongrel2,bashi-bazouk/mongrel2,cpick/mongrel2,pjkundert/mongrel2,sshirokov/mongrel2,jiffyjeff/mongrel2,jasom/mongrel2,pjkundert/mongrel2,cpick/mongrel2,steamraven/mongrel2,minrk/mongrel2,cpick/mongrel2,jiffyjeff/mongrel2,issuu/mongrel2,yoink00/mongrel2,AlexVPopov/mongrel2,xrl/mongrel2,AlexVPopov/mongrel2,Gibheer/mongrel2,aidenkeating/mongrel2,apjanke/mongrel2,markokr/mongrel2,ameuret/mongrel2,mbj/mongrel2,griffordson/mongrel2,AvdN/mongrel2,AvdN/mongrel2,nmandery/mongrel2,ameuret/mongrel2,aidenkeating/mongrel2,chickenkiller/mongrel2,chickenkiller/mongrel2,fanout/mongrel2,jubarajborgohain/mongrel2,duaneg/mongrel2,winks/mongrel2,dermoth/mongrel2,moai/mongrel2,sshirokov/mongrel2,mongrel2/mongrel2,elo80ka/mongrel2,griffordson/mongrel2,minrk/mongrel2,niedbalski/mongrel2,mongrel2/mongrel2,aidenkeating/mongrel2,yoink00/mongrel2,yoink00/mongrel2,duaneg/mongrel2,winks/mongrel2,mongrel2/mongrel2,msteinert/mongrel2,niedbalski/mongrel2,jiffyjeff/mongrel2,chickenkiller/mongrel2,issuu/mongrel2,chickenkiller/mongrel2,reshefm/mongrel2,chickenkiller/mongrel2,AustinWise/mongrel2
967cf8774a5033f310ca69e7ad86fc79b2628882
infrastructure/aws/trigger-provision.py
infrastructure/aws/trigger-provision.py
# trigger-provision.py <indexer-provision.sh | web-server-provision.sh> import boto3 from datetime import datetime, timedelta import sys import os.path provisioners = sys.argv[1:] ec2 = boto3.resource('ec2') client = boto3.client('ec2') script = '' for provisioner in provisioners: script += open(provisioner).read() + '\n' user_data = '''#!/usr/bin/env bash cat > ~ubuntu/provision.sh <<"FINAL" {script} FINAL chmod +x ~ubuntu/provision.sh sudo -i -u ubuntu ~ubuntu/provision.sh '''.format(script=script) # ubuntu/images/hvm-ssd/ubuntu-xenial-16.04-amd64-server-20160815 (ami-f701cb97) image_id = 'ami-f701cb97' launch_spec = { 'ImageId': image_id, 'KeyName': 'Main Key Pair', 'SecurityGroups': ['indexer'], 'UserData': user_data, 'InstanceType': 'c3.2xlarge', 'BlockDeviceMappings': [] } client.run_instances(MinCount=1, MaxCount=1, **launch_spec)
# trigger-provision.py <indexer-provision.sh | web-server-provision.sh> import boto3 from datetime import datetime, timedelta import sys import os.path provisioners = sys.argv[1:] ec2 = boto3.resource('ec2') client = boto3.client('ec2') script = '' for provisioner in provisioners: script += open(provisioner).read() + '\n' user_data = '''#!/usr/bin/env bash cat > ~ubuntu/provision.sh <<"FINAL" {script} FINAL chmod +x ~ubuntu/provision.sh sudo -i -u ubuntu ~ubuntu/provision.sh '''.format(script=script) # ubuntu/images/hvm-ssd/ubuntu-xenial-16.04-amd64-server-20160815 (ami-f701cb97) image_id = 'ami-f701cb97' launch_spec = { 'ImageId': image_id, 'KeyName': 'Main Key Pair', 'SecurityGroups': ['indexer'], 'UserData': user_data, 'InstanceType': 'c3.2xlarge', 'BlockDeviceMappings': [], 'TagSpecifications': [{ 'ResourceType': 'instance', 'Tags': [{ 'Key': 'provisioner', 'Value': sys.argv[1], }], }], } client.run_instances(MinCount=1, MaxCount=1, **launch_spec)
Tag provisioning instances to make them easier to identify
Tag provisioning instances to make them easier to identify
Python
mpl-2.0
bill-mccloskey/searchfox,bill-mccloskey/searchfox,bill-mccloskey/searchfox,bill-mccloskey/searchfox,bill-mccloskey/searchfox,bill-mccloskey/searchfox
32e70ee06be67cb9058b2da7dc1a714272c6a07a
pyQuantuccia/setup.py
pyQuantuccia/setup.py
import setuptools qu_ext = setuptools.Extension( 'quantuccia', include_dirs=['src/Quantuccia'], sources=['src/pyQuantuccia.cpp'] ) setuptools.setup( name='pyQuantuccia', author='Jack Grahl', author_email='jack.grahl@gmail.com', version='0.1.0', packages=['pyQuantuccia'], ext_modules=[qu_ext] )
import setuptools qu_ext = setuptools.Extension( 'quantuccia', include_dirs=['src/Quantuccia'], sources=['src/pyQuantuccia.cpp'] ) setuptools.setup( name='pyQuantuccia', author='Jack Grahl', author_email='jack.grahl@gmail.com', version='0.1.0', packages=['pyQuantuccia'], test_suite='tests', ext_modules=[qu_ext] )
Add the location of tests.
Add the location of tests.
Python
bsd-3-clause
jwg4/pyQuantuccia,jwg4/pyQuantuccia
324ce82f25c78bce7f92af52952f036ba48e72e7
astrobin_apps_notifications/utils.py
astrobin_apps_notifications/utils.py
# Python import simplejson import urllib2 # Django from django.apps import apps from django.conf import settings from django.core.exceptions import ImproperlyConfigured # Third party from notification import models as notification from persistent_messages.models import Message def push_notification(recipients, notice_type, data): data.update({'notices_url': settings.ASTROBIN_BASE_URL}) notification.send(recipients, notice_type, data) def get_recent_notifications(user, n = 10): if not user.is_authenticated(): return None notifications = Message.objects.filter(user = user).order_by('-created') if n >= 0: notifications = notifications[:n] return notifications def get_unseen_notifications(user, n = 10): if not user.is_authenticated(): return None notifications =\ Message.objects.filter(user = user, read = False).order_by('-created') if n >= 0: notifications = notifications[:n] return notifications def get_seen_notifications(user, n = 10): if not user.is_authenticated(): return None notifications =\ Message.objects.filter(user = user, read = True).order_by('-created') if n >= 0: notifications = notifications[:n] return notifications
# Python import simplejson import urllib2 # Django from django.apps import apps from django.conf import settings from django.core.exceptions import ImproperlyConfigured # Third party from notification import models as notification from persistent_messages.models import Message def push_notification(recipients, notice_type, data): data.update({'notices_url': settings.ASTROBIN_BASE_URL + '/'}) notification.send(recipients, notice_type, data) def get_recent_notifications(user, n = 10): if not user.is_authenticated(): return None notifications = Message.objects.filter(user = user).order_by('-created') if n >= 0: notifications = notifications[:n] return notifications def get_unseen_notifications(user, n = 10): if not user.is_authenticated(): return None notifications =\ Message.objects.filter(user = user, read = False).order_by('-created') if n >= 0: notifications = notifications[:n] return notifications def get_seen_notifications(user, n = 10): if not user.is_authenticated(): return None notifications =\ Message.objects.filter(user = user, read = True).order_by('-created') if n >= 0: notifications = notifications[:n] return notifications
Revert "Drop extra trailing slash from notices_url"
Revert "Drop extra trailing slash from notices_url" This reverts commit 1eb4d00e005f22ae452ce9d36b9fce69fa9b96f7.
Python
agpl-3.0
astrobin/astrobin,astrobin/astrobin,astrobin/astrobin,astrobin/astrobin
0cfe6707cf02bab74741433dbe7a91b8c5c57f38
cinder/tests/unit/test_fixtures.py
cinder/tests/unit/test_fixtures.py
null
# Copyright 2010 United States Government as represented by the # Administrator of the National Aeronautics and Space Administration. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import fixtures as fx from oslo_log import log as logging import testtools from cinder.tests import fixtures class TestLogging(testtools.TestCase): def test_default_logging(self): stdlog = self.useFixture(fixtures.StandardLogging()) root = logging.getLogger() # there should be a null handler as well at DEBUG self.assertEqual(2, len(root.handlers), root.handlers) log = logging.getLogger(__name__) log.info("at info") log.debug("at debug") self.assertIn("at info", stdlog.logger.output) self.assertNotIn("at debug", stdlog.logger.output) # broken debug messages should still explode, even though we # aren't logging them in the regular handler self.assertRaises(TypeError, log.debug, "this is broken %s %s", "foo") # and, ensure that one of the terrible log messages isn't # output at info warn_log = logging.getLogger('migrate.versioning.api') warn_log.info("warn_log at info, should be skipped") warn_log.error("warn_log at error") self.assertIn("warn_log at error", stdlog.logger.output) self.assertNotIn("warn_log at info", stdlog.logger.output) def test_debug_logging(self): self.useFixture(fx.EnvironmentVariable('OS_DEBUG', '1')) stdlog = self.useFixture(fixtures.StandardLogging()) root = logging.getLogger() # there should no longer be a null handler self.assertEqual(1, len(root.handlers), root.handlers) log = logging.getLogger(__name__) log.info("at info") log.debug("at debug") self.assertIn("at info", stdlog.logger.output) self.assertIn("at debug", stdlog.logger.output)
Copy unit tests for StandardLogging fixture from Nova
Copy unit tests for StandardLogging fixture from Nova This comes from commit f96ec4411ce89606cf52211061003c14306dcfa1 in Nova by Sean Dague <sean@dague.net>. The StandardLogging fixture was already merged into Cinder, this adds the unit tests that were missed when copying over the fixture. Change-Id: I2fbe25ec71138e4b96ff175af72a2a56c1c8f52a Related-Bug: #1551325
Python
apache-2.0
Nexenta/cinder,bswartz/cinder,NetApp/cinder,mahak/cinder,Nexenta/cinder,phenoxim/cinder,Datera/cinder,openstack/cinder,cloudbase/cinder,Hybrid-Cloud/cinder,phenoxim/cinder,NetApp/cinder,Datera/cinder,cloudbase/cinder,j-griffith/cinder,openstack/cinder,mahak/cinder,ge0rgi/cinder,Hybrid-Cloud/cinder,bswartz/cinder,j-griffith/cinder,eharney/cinder,eharney/cinder
9116776bd62e6b7ae7a018fa2c2c0b3964c3fa7d
py/maximum-binary-tree.py
py/maximum-binary-tree.py
null
# Definition for a binary tree node. # class TreeNode(object): # def __init__(self, x): # self.val = x # self.left = None # self.right = None class Solution(object): def findMax(self, start, end): bit_length = (end - start).bit_length() - 1 d = 1 << bit_length return max(self.SparseTable[bit_length][start], self.SparseTable[bit_length][end - d]) def do_constructMaximumBinaryTree(self, start, end): if start == end: return None v, i = self.findMax(start, end) ret = TreeNode(v) ret.left = self.do_constructMaximumBinaryTree(start, i) ret.right = self.do_constructMaximumBinaryTree(i + 1, end) return ret def constructMaximumBinaryTree(self, nums): """ :type nums: List[int] :rtype: TreeNode """ self.SparseTable = [[(v, i) for i, v in enumerate(nums)]] l = len(nums) t = 1 while t * 2 < l: prevTable = self.SparseTable[-1] self.SparseTable.append([max(prevTable[i], prevTable[i + t]) for i in xrange(l - t * 2 + 1)]) t *= 2 return self.do_constructMaximumBinaryTree(0, l)
Add py solution for 654. Maximum Binary Tree
Add py solution for 654. Maximum Binary Tree 654. Maximum Binary Tree: https://leetcode.com/problems/maximum-binary-tree/
Python
apache-2.0
ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode
cf52a7c83e1479a99e95ab2125958a67febfccf5
dataviews/__init__.py
dataviews/__init__.py
import sys, os # Add param submodule to sys.path cwd = os.path.abspath(os.path.split(__file__)[0]) sys.path.insert(0, os.path.join(cwd, '..', 'param')) from .views import * # pyflakes:ignore (API import) from .dataviews import * # pyflakes:ignore (API import) from .sheetviews import * # pyflakes:ignore (API import) from .ndmapping import * # pyflakes:ignore (API import) def public(obj): if not isinstance(obj, type): return False baseclasses = [NdMapping, View, Dimension] return any([issubclass(obj, bc) for bc in baseclasses]) _public = list(set([_k for _k, _v in locals().items() if public(_v)])) __all__ = _public + ["boundingregion", "ipython", "plots", "sheetcoords" ]
import sys, os # Add param submodule to sys.path cwd = os.path.abspath(os.path.split(__file__)[0]) sys.path.insert(0, os.path.join(cwd, '..', 'param')) import param __version__ = param.Version(release=(0,7), fpath=__file__) from .views import * # pyflakes:ignore (API import) from .dataviews import * # pyflakes:ignore (API import) from .sheetviews import * # pyflakes:ignore (API import) from .ndmapping import * # pyflakes:ignore (API import) def public(obj): if not isinstance(obj, type): return False baseclasses = [NdMapping, View, Dimension] return any([issubclass(obj, bc) for bc in baseclasses]) _public = list(set([_k for _k, _v in locals().items() if public(_v)])) __all__ = _public + ["boundingregion", "ipython", "plots", "sheetcoords" ]
Set __version__ using param.Version (commit tagged as 'v0.7')
Set __version__ using param.Version (commit tagged as 'v0.7')
Python
bsd-3-clause
mjabri/holoviews,basnijholt/holoviews,ioam/holoviews,mjabri/holoviews,ioam/holoviews,vascotenner/holoviews,vascotenner/holoviews,ioam/holoviews,basnijholt/holoviews,basnijholt/holoviews,vascotenner/holoviews,mjabri/holoviews
608dc0db688be1dabe3c6ba7647807f6697fcefe
tools/misc/python/test-data-in-out.py
tools/misc/python/test-data-in-out.py
# TOOL test-data-in-out.py: "Test data input and output in Python" (Data input output test.) # INPUT input TYPE GENERIC # OUTPUT output # OUTPUT OPTIONAL missing_output.txt import shutil shutil.copyfile('input', 'output')
# TOOL test-data-in-out.py: "Test data input and output in Python" (Data input output test.) # INPUT input TYPE GENERIC # OUTPUT output # OUTPUT OPTIONAL missing_output.txt # IMAGE chipster-tools-python import shutil shutil.copyfile('input', 'output')
Test image definition in SADL
Test image definition in SADL
Python
mit
chipster/chipster-tools,chipster/chipster-tools,chipster/chipster-tools,chipster/chipster-tools
da3599ac6ed29e750d28834a8c0c0f39e9b57702
src/acquisition/covid_hosp/state_daily/network.py
src/acquisition/covid_hosp/state_daily/network.py
# first party from delphi.epidata.acquisition.covid_hosp.common.network import Network as BaseNetwork class Network(BaseNetwork): DATASET_ID = '823dd0e-c8c4-4206-953e-c6d2f451d6ed' def fetch_metadata(*args, **kwags): """Download and return metadata. See `fetch_metadata_for_dataset`. """ return Network.fetch_metadata_for_dataset( *args, **kwags, dataset_id=Network.DATASET_ID)
# first party from delphi.epidata.acquisition.covid_hosp.common.network import Network as BaseNetwork class Network(BaseNetwork): DATASET_ID = '7823dd0e-c8c4-4206-953e-c6d2f451d6ed' def fetch_metadata(*args, **kwags): """Download and return metadata. See `fetch_metadata_for_dataset`. """ return Network.fetch_metadata_for_dataset( *args, **kwags, dataset_id=Network.DATASET_ID)
Update state daily dataset ID
Update state daily dataset ID
Python
mit
cmu-delphi/delphi-epidata,cmu-delphi/delphi-epidata,cmu-delphi/delphi-epidata,cmu-delphi/delphi-epidata,cmu-delphi/delphi-epidata,cmu-delphi/delphi-epidata
0043fe9c8de4d8341afbcea388f472a50017de2c
jiradoc/__main__.py
jiradoc/__main__.py
# ------------------------------------------------------------ # __main__.py # # The main program which expects a jiradoc formatted file to # be passed in as a cmdline option. It reads the file and # parses its content to Story objects. # ------------------------------------------------------------ import argparse import pkg_resources from jiradoc.parser.parser import parser def main(args=None): argparser = argparse.ArgumentParser(description='The JIRAdoc parser') test_file = pkg_resources.resource_filename(__name__, 'data/test.jiradoc') argparser.add_argument('-f', dest='file', default=test_file, help='The jiradoc formatted file') args = argparser.parse_args() with open(args.file) as f: content = f.read() stories = parser.parse(content) for story in stories: print story if __name__ == "__main__": main()
# ------------------------------------------------------------ # __main__.py # # The main program # ------------------------------------------------------------ import argparse import os import pkg_resources import sys from jiradoc.parser.parser import parser as jiradoc_parser def main(args=None): parser = argparse.ArgumentParser(description='A tool that parses a JIRAdoc formatted file and returns a list of ' 'story objects') test_file = pkg_resources.resource_filename(__name__, 'data/test.jiradoc') parser.add_argument('-f', dest='file', default=test_file, help='A .jiradoc file containing sub-tasks to JIRA stories') args = parser.parse_args() filename, ext = os.path.splitext(args.file) if ext != '.jiradoc': print 'Invalid file extension: ' + ext print 'The only valid extension is .jiradoc' sys.exit(1) with open(args.file) as f: content = f.read() stories = jiradoc_parser.parse(content) for story in stories: print story if __name__ == "__main__": main()
Validate that the input file ends with .jiradoc
Validate that the input file ends with .jiradoc
Python
mit
lucianovdveekens/jiradoc
e91079ecab2d324ae5428239bcf792ab149faa4a
jacquard/storage/tests/test_dummy.py
jacquard/storage/tests/test_dummy.py
null
import unittest from jacquard.storage.dummy import DummyStore from jacquard.storage.testing_utils import StorageGauntlet class DummyGauntletTest(StorageGauntlet, unittest.TestCase): def open_storage(self): return DummyStore('')
Cover the dummy storage engine with tests
Cover the dummy storage engine with tests
Python
mit
prophile/jacquard,prophile/jacquard
3ff6b8a2e8eecf48bfe74d5a0b0972e29ace15fd
imagetagger/imagetagger/annotations/admin.py
imagetagger/imagetagger/annotations/admin.py
from django.contrib import admin from .models import Annotation, AnnotationType, Export, Verification, ExportFormat admin.site.register(Annotation) admin.site.register(AnnotationType) admin.site.register(Export) admin.site.register(Verification) admin.site.register(ExportFormat)
from django.contrib import admin from .models import Annotation, AnnotationType, Export, Verification, ExportFormat @admin.register(Annotation) class AnnotationAdmin(admin.ModelAdmin): raw_id_fields = ( 'image', ) @admin.register(Verification) class VerificationAdmin(admin.ModelAdmin): raw_id_fields = ( 'annotation', ) admin.site.register(AnnotationType) admin.site.register(Export) admin.site.register(ExportFormat)
Use raw id fields for annotation and verification foreign keys
Use raw id fields for annotation and verification foreign keys
Python
mit
bit-bots/imagetagger,bit-bots/imagetagger,bit-bots/imagetagger,bit-bots/imagetagger
24cbbd24e6398aa11956ac48282bd907806284c3
genderbot.py
genderbot.py
import re from twitterbot import TwitterBot import wikipedia class Genderbot(TwitterBot): boring_article_regex = (r"municipality|village|town|football|genus|family|" "administrative|district|community|region|hamlet|" "school|actor|mountain|basketball|city|species|film|" "county|located|politician|professional|settlement") def tweet(self): article = self.__random_wikipedia_article() match = re.search(r"\bis [^.?]+", article.content, re.UNICODE) if match: status = self.__format_status(match.group(0), article.url) if self.__is_interesting(status): self.post_tweet(status) def __format_status(self, is_phrase, url): status = 'gender %s' % (is_phrase) if len(status) > 114: status = status[0:113] + '...' return status + ' %s' % (url) def __is_interesting(self, status): boring_match = re.search(Genderbot.boring_article_regex, status, re.UNICODE) return boring_match is None def __random_wikipedia_article(self): random_title = wikipedia.random(pages=1) return wikipedia.page(title=random_title) if __name__ == "__main__": try: Genderbot("CustomGender").tweet() except: pass
import re from twitterbot import TwitterBot import wikipedia class Genderbot(TwitterBot): boring_regex = (r"municipality|village|town|football|genus|family|" "administrative|district|community|region|hamlet|" "school|actor|mountain|basketball|city|species|film|" "county|located|politician|professional|settlement|" "river|lake|province|replaced|origin|band|park|song" "approximately|north|south|east|west|business") def tweet(self): article = self.__random_wikipedia_article() match = re.search(r"\bis [^.?]+", article.content, re.UNICODE) if match: status = self.__format_status(match.group(0), article.url) if self.__is_interesting(status): self.post_tweet(status) def __format_status(self, is_phrase, url): status = 'gender %s' % (is_phrase) if len(status) > 114: status = status[0:113] + '...' return status + ' %s' % (url) def __is_interesting(self, status): flags = re.UNICODE | re.IGNORECASE boring = re.search(Genderbot.boring_regex, status, flags) return boring is None def __random_wikipedia_article(self): random_title = wikipedia.random(pages=1) return wikipedia.page(title=random_title) if __name__ == "__main__": try: Genderbot("CustomGender").tweet() except: pass
Tweak boring regex to exclude more terms
Tweak boring regex to exclude more terms
Python
mit
DanielleSucher/genderbot
f24fe32329625ec037a9afc8d3bdeed5f41e69a0
scripts/diff_incar.py
scripts/diff_incar.py
null
#!/usr/bin/env python ''' Created on Nov 12, 2011 ''' __author__="Shyue Ping Ong" __copyright__ = "Copyright 2011, The Materials Project" __version__ = "0.1" __maintainer__ = "Shyue Ping Ong" __email__ = "shyue@mit.edu" __date__ = "Nov 12, 2011" import sys import itertools from pymatgen.io.vaspio import Incar from pymatgen.util.string_utils import str_aligned filepath1 = sys.argv[1] filepath2 = sys.argv[2] incar1 = Incar.from_file(filepath1) incar2 = Incar.from_file(filepath2) def format_lists(v): if isinstance(v, (tuple, list)): return " ".join([str(i) + "*" + str(len(tuple(group))) for (i,group) in itertools.groupby(v)]) return v d = incar1.diff(incar2) output = [['SAME PARAMS','', '']] output.append(['---------------','', '']) output.extend([(k,format_lists(v),format_lists(v)) for k,v in d['Same'].items() if k != "SYSTEM"]) output.append(['','', '']) output.append(['DIFFERENT PARAM','', '']) output.append(['---------------','', '']) output.extend([(k,format_lists(v['INCAR1']),format_lists(v['INCAR2'])) for k, v in d['Different'].items() if k != "SYSTEM"]) print str_aligned(output, ['', filepath1, filepath2])
Add a script for easy diffing of two Incars.
Add a script for easy diffing of two Incars. Former-commit-id: 998a47c0b96b3024abd82b196f431926cc50847d [formerly 927396038d147b633bee31988cf1e016258c5320] Former-commit-id: 4a8c6bb9cfef4a3a3f6cc211b7ef558a06f523c3
Python
mit
gpetretto/pymatgen,aykol/pymatgen,Bismarrck/pymatgen,czhengsci/pymatgen,johnson1228/pymatgen,Bismarrck/pymatgen,gVallverdu/pymatgen,richardtran415/pymatgen,blondegeek/pymatgen,czhengsci/pymatgen,dongsenfo/pymatgen,montoyjh/pymatgen,czhengsci/pymatgen,fraricci/pymatgen,nisse3000/pymatgen,nisse3000/pymatgen,blondegeek/pymatgen,nisse3000/pymatgen,richardtran415/pymatgen,tschaume/pymatgen,fraricci/pymatgen,matk86/pymatgen,mbkumar/pymatgen,czhengsci/pymatgen,dongsenfo/pymatgen,tallakahath/pymatgen,blondegeek/pymatgen,johnson1228/pymatgen,Bismarrck/pymatgen,dongsenfo/pymatgen,Bismarrck/pymatgen,vorwerkc/pymatgen,gpetretto/pymatgen,johnson1228/pymatgen,vorwerkc/pymatgen,montoyjh/pymatgen,xhqu1981/pymatgen,mbkumar/pymatgen,vorwerkc/pymatgen,davidwaroquiers/pymatgen,matk86/pymatgen,gVallverdu/pymatgen,fraricci/pymatgen,tschaume/pymatgen,matk86/pymatgen,gpetretto/pymatgen,setten/pymatgen,richardtran415/pymatgen,setten/pymatgen,tallakahath/pymatgen,ndardenne/pymatgen,gVallverdu/pymatgen,tschaume/pymatgen,ndardenne/pymatgen,montoyjh/pymatgen,mbkumar/pymatgen,gmatteo/pymatgen,vorwerkc/pymatgen,richardtran415/pymatgen,xhqu1981/pymatgen,montoyjh/pymatgen,davidwaroquiers/pymatgen,matk86/pymatgen,nisse3000/pymatgen,Bismarrck/pymatgen,fraricci/pymatgen,davidwaroquiers/pymatgen,tallakahath/pymatgen,tschaume/pymatgen,gpetretto/pymatgen,xhqu1981/pymatgen,ndardenne/pymatgen,setten/pymatgen,gmatteo/pymatgen,davidwaroquiers/pymatgen,gVallverdu/pymatgen,setten/pymatgen,aykol/pymatgen,dongsenfo/pymatgen,mbkumar/pymatgen,blondegeek/pymatgen,aykol/pymatgen,johnson1228/pymatgen,tschaume/pymatgen
eae949e483e1d30e8c11b662bb07e9d30dcf39c5
lc0049_group_anagrams.py
lc0049_group_anagrams.py
"""Leetcode 49. Group Anagrams Medium URL: https://leetcode.com/problems/group-anagrams/ Given an array of strings, group anagrams together. Example: Input: ["eat", "tea", "tan", "ate", "nat", "bat"], Output: [ ["ate","eat","tea"], ["nat","tan"], ["bat"] ] Note: - All inputs will be in lowercase. - The order of your output does not matter. """ class SolutionSortedDict(object): def groupAnagrams(self, strs): """ :type strs: List[str] :rtype: List[List[str]] Output Limit Exceede. Time complexity: O(n*klogk), where - n is the length of strs, - k is the lenght of the longest string. Space complexity: O(n). """ from collections import defaultdict # Store in a dict with sorted string->string list. anagrams_d = defaultdict(list) for s in strs: # Use sorted string as dict key. k = ''.join(sorted(s)) anagrams_d[k].append(s) return anagrams_d.values() def main(): # Output: # [ # ["ate","eat","tea"], # ["nat","tan"], # ["bat"] # ] strs = ["eat", "tea", "tan", "ate", "nat", "bat"] print SolutionSortedDict().groupAnagrams(strs) if __name__ == '__main__': main()
"""Leetcode 49. Group Anagrams Medium URL: https://leetcode.com/problems/group-anagrams/ Given an array of strings, group anagrams together. Example: Input: ["eat", "tea", "tan", "ate", "nat", "bat"], Output: [ ["ate","eat","tea"], ["nat","tan"], ["bat"] ] Note: - All inputs will be in lowercase. - The order of your output does not matter. """ class SolutionSortedAnagramDict(object): def groupAnagrams(self, strs): """ :type strs: List[str] :rtype: List[List[str]] Output Limit Exceede. Time complexity: O(n*klogk), where - n is the length of strs, - k is the lenght of the longest string. Space complexity: O(n). """ from collections import defaultdict # Store in a dict with sorted string->string list. anagram_lists = defaultdict(list) for s in strs: # Use sorted string as dict key. k = ''.join(sorted(s)) anagram_lists[k].append(s) return anagram_lists.values() def main(): # Output: # [ # ["ate","eat","tea"], # ["nat","tan"], # ["bat"] # ] strs = ["eat", "tea", "tan", "ate", "nat", "bat"] print SolutionSortedAnagramDict().groupAnagrams(strs) if __name__ == '__main__': main()
Revise to anagram_lists and rename to sorted anagram dict class
Revise to anagram_lists and rename to sorted anagram dict class
Python
bsd-2-clause
bowen0701/algorithms_data_structures
e76ca364ab979e309d34ff458ef2629145a52ce2
magnum/db/sqlalchemy/alembic/versions/a1136d335540_add_docker_storage_driver_column.py
magnum/db/sqlalchemy/alembic/versions/a1136d335540_add_docker_storage_driver_column.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """Add docker storage driver column Revision ID: a1136d335540 Revises: d072f58ab240 Create Date: 2016-03-07 19:00:28.738486 """ # revision identifiers, used by Alembic. revision = 'a1136d335540' down_revision = 'd072f58ab240' from alembic import op import sqlalchemy as sa docker_storage_driver_enum = sa.Enum('devicemapper', 'overlay', name='docker_storage_driver') def upgrade(): op.add_column('baymodel', sa.Column('docker_storage_driver', docker_storage_driver_enum, nullable=True))
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """Add docker storage driver column Revision ID: a1136d335540 Revises: d072f58ab240 Create Date: 2016-03-07 19:00:28.738486 """ # revision identifiers, used by Alembic. revision = 'a1136d335540' down_revision = 'd072f58ab240' from alembic import op import sqlalchemy as sa docker_storage_driver_enum = sa.Enum('devicemapper', 'overlay', name='docker_storage_driver') def upgrade(): docker_storage_driver_enum.create(op.get_bind(), checkfirst=True) op.add_column('baymodel', sa.Column('docker_storage_driver', docker_storage_driver_enum, nullable=True))
Fix for enum type docker_storage_driver
Fix for enum type docker_storage_driver Create enum type "docker_storage_driver" for migration This is fixing oslo_db.exception.DBError: (psycopg2.ProgrammingError) type "docker_storage_driver" does not exist Closes-Bug: #1609776 Change-Id: I92d427e90bd73b4114d8688d3761cabac450fc9d
Python
apache-2.0
openstack/magnum,openstack/magnum,ArchiFleKs/magnum,ArchiFleKs/magnum
379e99a672537776ac0e160999967b5efce29305
tweepy/media.py
tweepy/media.py
# Tweepy # Copyright 2009-2021 Joshua Roesslein # See LICENSE for details. from tweepy.mixins import DataMapping class Media(DataMapping): __slots__ = ( "data", "media_key", "type", "duration_ms", "height", "non_public_metrics", "organic_metrics", "preview_image_url", "promoted_metrics", "public_metrics", "width" ) def __init__(self, data): self.data = data self.media_key = data["media_key"] self.type = data["type"] self.duration_ms = data.get("duration_ms") self.height = data.get("height") self.non_public_metrics = data.get("non_public_metrics") self.organic_metrics = data.get("organic_metrics") self.preview_image_url = data.get("preview_image_url") self.promoted_metrics = data.get("promoted_metrics") self.public_metrics = data.get("public_metrics") self.width = data.get("width") def __eq__(self, other): if isinstance(other, self.__class__): return self.media_key == other.media_key return NotImplemented def __hash__(self): return hash(self.media_key) def __repr__(self): return f"<Media media_key={self.media_key} type={self.type}>"
# Tweepy # Copyright 2009-2021 Joshua Roesslein # See LICENSE for details. from tweepy.mixins import DataMapping class Media(DataMapping): __slots__ = ( "data", "media_key", "type", "duration_ms", "height", "non_public_metrics", "organic_metrics", "preview_image_url", "promoted_metrics", "public_metrics", "width", "alt_text" ) def __init__(self, data): self.data = data self.media_key = data["media_key"] self.type = data["type"] self.duration_ms = data.get("duration_ms") self.height = data.get("height") self.non_public_metrics = data.get("non_public_metrics") self.organic_metrics = data.get("organic_metrics") self.preview_image_url = data.get("preview_image_url") self.promoted_metrics = data.get("promoted_metrics") self.public_metrics = data.get("public_metrics") self.width = data.get("width") self.alt_text = data.get("alt_text") def __eq__(self, other): if isinstance(other, self.__class__): return self.media_key == other.media_key return NotImplemented def __hash__(self): return hash(self.media_key) def __repr__(self): return f"<Media media_key={self.media_key} type={self.type}>"
Add alt_text field for Media
Add alt_text field for Media
Python
mit
svven/tweepy,tweepy/tweepy
052392da7980c4f4e2e86cd8eb65da5b91d3547b
CodeFights/differentSymbolsNaive.py
CodeFights/differentSymbolsNaive.py
null
#!/usr/local/bin/python # Code Fights Different Symbols Naive Problem from collections import Counter def differentSymbolsNaive(s): return len(Counter(s)) def main(): tests = [ ["cabca", 3], ["aba", 2] ] for t in tests: res = differentSymbolsNaive(t[0]) ans = t[1] if ans == res: print("PASSED: differentSymbolsNaive({}) returned {}" .format(t[0], res)) else: print(("FAILED: differentSymbolsNaive({}) returned {}," "answer: {}").format(t[0], res, ans)) if __name__ == '__main__': main()
Solve Code Fights different symbols naive problem
Solve Code Fights different symbols naive problem
Python
mit
HKuz/Test_Code
921d95a432316b2b76aadabcc3fc90fbf90376aa
simphony/testing/abc_container_check.py
simphony/testing/abc_container_check.py
null
import abc from functools import partial from simphony.testing.utils import ( create_data_container, compare_data_containers) from simphony.core.data_container import DataContainer from simphony.core.cuba import CUBA class ContainerCheck(object): __metaclass__ = abc.ABCMeta def setUp(self): self.addTypeEqualityFunc( DataContainer, partial(compare_data_containers, testcase=self)) self.container = self.container_factory(u'foo') self.data = create_data_container() self.container.data = DataContainer(self.data) @abc.abstractmethod def container_factory(self, name): """ Create and return the container object """ def test_name(self): self.assertEqual(self.contaoner.name, u'foo') def test_rename(self): container = self.container container.name = u'bar' self.assertEqual(container.name, u'bar') def test_data(self): self.assertEqual(self.container.data, self.data) self.assertIsNot(self.container.data, self.data) def test_update_data(self): container = self.container data = container.data data[CUBA.TEMPERATURE] = 23.4 self.assertNotEqual(container.data, data) self.assertEqual(container.data, self.data) container.data = data self.assertEqual(container.data, data) self.assertIsNot(container.data, data)
Add a basic test template for high level CUDS containers
Add a basic test template for high level CUDS containers
Python
bsd-2-clause
simphony/simphony-common
90d1a40175675b2950cb41b85434b522d6e21c4d
mass/cli.py
mass/cli.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # vim: set hls is ai et sw=4 sts=4 ts=8 nu ft=python: # built-in modules # 3rd-party modules import click # local modules from mass.monitor.app import app from mass.scheduler.swf import utils from mass.scheduler.swf import SWFWorker @click.group() def cli(): pass @cli.command() def init(): utils.register_domain() utils.register_workflow_type() utils.register_activity_type() @cli.group() def worker(): pass @cli.group() def job(): pass @cli.group() def monitor(): pass @worker.command('start') def worker_start(): worker = SWFWorker() worker.start() @job.command('submit') @click.option('-j', '--json', help='Job Description in JSON.') @click.option('-a', '--alfscript', help='Job Description in alfscript.') def job_submit(json_script, alf_script): pass @monitor.command('start') def monitor_start(): monitor = app.run(debug=True) cli.add_command(init) cli.add_command(worker) cli.add_command(job) cli.add_command(monitor)
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # vim: set hls is ai et sw=4 sts=4 ts=8 nu ft=python: # built-in modules # 3rd-party modules import click # local modules from mass.monitor.app import app from mass.scheduler.swf import utils from mass.scheduler.swf import SWFWorker @click.group() def cli(): pass @cli.command() @click.option('-d', '--domain', help='Amazon SWF Domain.') @click.option('-r', '--region', help='Amazon Region.') def init(domain, region): utils.register_domain(domain, region) utils.register_workflow_type(domain, region) utils.register_activity_type(domain, region) @cli.group() def worker(): pass @cli.group() def job(): pass @cli.group() def monitor(): pass @worker.command('start') def worker_start(): worker = SWFWorker() worker.start() @job.command('submit') @click.option('-j', '--json', help='Job Description in JSON.') @click.option('-a', '--alfscript', help='Job Description in alfscript.') def job_submit(json_script, alf_script): pass @monitor.command('start') def monitor_start(): monitor = app.run(debug=True) cli.add_command(init) cli.add_command(worker) cli.add_command(job) cli.add_command(monitor)
Add arguments --domain and --region to mass init.
Add arguments --domain and --region to mass init.
Python
apache-2.0
badboy99tw/mass,KKBOX/mass,badboy99tw/mass,badboy99tw/mass,KKBOX/mass,KKBOX/mass
fee11dbff232216726516eea6c8bf7645fdef1a7
pyxif/__init__.py
pyxif/__init__.py
from ._remove import remove from ._load_and_dump import load, dump, ZerothIFD, ExifIFD, GPSIFD from ._transplant import transplant from ._insert import insert try: from ._thumbnail import thumbnail except ImportError: print("'thumbnail' function depends on PIL or Pillow.") VERSION = '0.4.7'
from ._remove import remove from ._load_and_dump import load, dump, ZerothIFD, ExifIFD, GPSIFD from ._transplant import transplant from ._insert import insert try: from ._thumbnail import thumbnail except ImportError: print("'thumbnail' function depends on PIL or Pillow.") VERSION = '0.4.6'
Revert "up version to 0.4.7."
Revert "up version to 0.4.7." This reverts commit 9b1177d4a56070092faa89778911d11c70efdc54.
Python
mit
hMatoba/Piexif
8e065cb164b4ec1a9aafd7f448555707a875fab9
aggregator/espn_fc.py
aggregator/espn_fc.py
null
import datetime from aggregator import exceptions from aggregator.base import Aggregator, Article, InvalidArticle, make_soup EXCLUDE_IF_IN_TITLE = ['LIVE:', 'WATCH:', 'LISTEN:'] class ESPNFC(Aggregator): base_url = 'http://www.espnfc.com/?country-view=www&lang-view=en' source = 'ESPN FC' def extract(self): soup = make_soup(ESPNFC.base_url) divs = soup.find('div', {'alt': ' TOP STORIES '}) divs = iter(divs.find_all('div', {'class': 'grid-item-content'})) articles = (self.crawl(div) for div in divs) return list(article for article in articles if article is not None) def crawl(self, tag): try: anchor = tag.find('a', {'class': 'common-link'}) url = self.get_url(anchor) title = self.get_title(anchor) if any(exclude in title for exclude in EXCLUDE_IF_IN_TITLE): return None date_published = self.get_date_published(tag) author = self.get_author(tag) return Article(ESPNFC.source, title, url, author, date_published) except exceptions.WebCrawlException as e: return InvalidArticle(ESPNFC.source, e) def get_author(self, tag): try: author = tag.find('span', {'class': 'author byline'}) return author.text.strip() except AttributeError as e: raise exceptions.AuthorNotFoundException def get_date_published(self, tag): try: date_published = tag.find('time')['datetime'] date_published = date_published.split('T')[0] date_published = datetime.datetime.strptime(date_published, '%Y-%m-%d').date() return date_published except (IndexError, AttributeError, ValueError, TypeError): raise exceptions.DatePublishedNotFoundException def get_title(self, tag): try: return tag.text.strip() except AttributeError as e: raise exceptions.TitleNotFoundException def get_url(self, tag): try: url = tag['href'] url = url.replace('.us', '.com') return url except (KeyError, AttributeError, TypeError): raise exceptions.UrlNotFoundException if __name__ == '__main__': espn_fc = ESPNFC() print(espn_fc.extract())
Implement web scraping functionality for ESPN FC
Implement web scraping functionality for ESPN FC
Python
apache-2.0
footynews/fn_backend
ed9a02a769a64edae6aa0e4a06acf7a8fe31cd89
scripts/create_windows_virtenv.py
scripts/create_windows_virtenv.py
null
#!/usr/bin/env python import platform import shutil import subprocess import sys from os.path import dirname, isdir, join as join_path file_path = dirname(__file__) requirements_path = '/'.join(file_path.split('/')[:-1]) + "/requirements" root_path = file_path.split('/')[:-2] root_path = '/'.join(root_path) class VirtualEnvironmentBuilder(object): def __init__(self, virt_env_name): self.virt_env_name = virt_env_name @property def virt_env_path(self): print(join_path(root_path, self.virt_env_name)) return join_path(root_path, self.virt_env_name) @property def virt_env_path(self): return root_path + "/" + self.virt_env_name def clean_build(self): self.delete_env() self.build() def build(self): # Create a fresh virtual environment if it doesn't exist self.create_venv() try: print(requirements_path) self.run_in_venv('pip', ['install', '-r', requirements_path]) except Exception: print("Erorrrr") self.delete_env() def create_venv(self): if isdir(self.virt_env_path): return print(self.virt_env_path) try: subprocess.check_call([sys.executable, '-m', 'virtualenv', self.virt_env_path, '--no-site-packages']) except Exception: print("Something is wrong!") self.delete_env() if isdir(self.virt_env_name): print("Environment {} created".format(self.virt_env_path)) def delete_env(self): print("Deleting env!") try: if isdir(self.virt_env_path): shutil.rmtree(self.virt_env_path) except Exception: print("Could not delete environment!") def run_in_venv(self, cmd, args): virtual_env_bin_path = self.virt_env_path if platform.system() == 'Windows': cmd += '.exe' virtual_env_bin_path += r'\Scripts' else: virtual_env_bin_path += r'/bin' print("here") print(virtual_env_bin_path) print(cmd) a = join_path[virtual_env_bin_path, cmd] print(a) subprocess.check_call(join_path[file_path, virtual_env_bin_path, cmd] + args) if __name__ == '__main__': builder = VirtualEnvironmentBuilder('hack33-virtenv') builder.build()
Create virtual builder for windows OS
Create virtual builder for windows OS
Python
mit
Rositsazz/hack33,Rositsazz/hack33,Rositsazz/hack33,Rositsazz/hack33
522edf619f1bebf855f24da6f84c90b10f866745
slack_to_habitica/views.py
slack_to_habitica/views.py
import os import json import requests from django.http import HttpResponse from django.views.decorators.csrf import csrf_exempt @csrf_exempt def process_slack_message(request): send_message( request.POST.get('user_name'), request.POST.get('text')) return HttpResponse() def send_message(user, text): api_user = os.environ['HABITICA_APIUSER'] api_key = os.environ['HABITICA_APIKEY'] group_id = os.environ['HABITICA_GROUPID'] habitica_url = 'https://habitica.com/api/v3/groups/%s/chat' % group_id headers = { 'x-api-user': api_user, 'x-api-key': api_key } data = { 'groupId': group_id, 'message': '<%s says> %s' % (user, text) } response = requests.post(habitica_url, headers=headers, data=data)
import os import json import requests from django.http import HttpResponse from django.views.decorators.csrf import csrf_exempt @csrf_exempt def process_slack_message(request): send_message( request.POST.get('user_name'), request.POST.get('text')) return HttpResponse() def send_message(user, text): api_user = os.environ['HABITICA_APIUSER'] api_key = os.environ['HABITICA_APIKEY'] group_id = os.environ['HABITICA_GROUPID'] habitica_url = 'https://habitica.com/api/v3/groups/%s/chat' % group_id headers = { 'x-api-user': api_user, 'x-api-key': api_key } data = { 'groupId': group_id, 'message': '[%s says] %s' % (user, text) } response = requests.post(habitica_url, headers=headers, data=data)
Use square brackets instead of angle brackets in messages, as the latter are getting stripped out
Use square brackets instead of angle brackets in messages, as the latter are getting stripped out
Python
mit
niteshpatel/habitica-slack
d52f59911929eda6b8c0c42837ae9c19b9e133e4
twokenize_py/align.py
twokenize_py/align.py
"""Aligner for texts and their segmentations. """ from __future__ import absolute_import from __future__ import print_function from __future__ import unicode_literals __all__ = ['AlignmentFailed', 'Aligner'] class AlignmentFailed(Exception): pass class Aligner(object): """Align a text with its tokenization. """ def align(self, text, tokens): """Align text with its tokeniation. Parameters ---------- text : str Text. tokens : list of str Tokenization of ``text``. Returns ------- spans : list of tuple List of (``onset``, ``offset``) pairs, where ``spans[i]`` gives the onseta and offset in characters of ``tokens[i]`` relative to the beginning of ``text`` (0-indexed). """ spans = [] bi = 0 for token in tokens: try: token_len = len(token) token_bi = bi + txt[bi:].index(token) token_ei = token_bi + token_len - 1 spans.append([token_bi, token_ei]) bi = token_ei + 1 except ValueError: raise AlignmentFailed(token) return spans
"""Aligner for texts and their segmentations. """ from __future__ import absolute_import from __future__ import print_function from __future__ import unicode_literals __all__ = ['AlignmentFailed', 'Aligner'] class AlignmentFailed(Exception): pass class Aligner(object): """Align a text with its tokenization. """ def align(self, text, tokens): """Align text with its tokeniation. Parameters ---------- text : str Text. tokens : list of str Tokenization of ``text``. Returns ------- spans : list of tuple List of (``onset``, ``offset``) pairs, where ``spans[i]`` gives the onseta and offset in characters of ``tokens[i]`` relative to the beginning of ``text`` (0-indexed). """ spans = [] bi = 0 for token in tokens: try: token_len = len(token) token_bi = bi + text[bi:].index(token) token_ei = token_bi + token_len - 1 spans.append([token_bi, token_ei]) bi = token_ei + 1 except ValueError: raise AlignmentFailed(token) return spans
Fix typo in variable name.
BUG: Fix typo in variable name.
Python
apache-2.0
nryant/twokenize_py
00cbac852e83eb1f3ddc03ed70ad32494f16fdbf
caslogging.py
caslogging.py
""" file: caslogging.py author: Ben Grawi <bjg1568@rit.edu> date: October 2013 description: Sets up the logging information for the CAS Reader """ from config import config import logging as root_logging # Set up the logger logger = root_logging.getLogger() logger.setLevel(root_logging.INFO) logger_format = root_logging.Formatter('%(asctime)s %(levelname)s: %(message)s') logging_file_handler = root_logging.FileHandler(config['logging']['filename']) logging_file_handler.setLevel(root_logging.INFO) logging_file_handler.setFormatter(logger_format) logger.addHandler(logging_file_handler) logging_stream_handler = root_logging.StreamHandler() logging_stream_handler.setLevel(root_logging.INFO) logging_stream_handler.setFormatter(logger_format) logger.addHandler(logging_stream_handler) logging = root_logging
""" file: caslogging.py author: Ben Grawi <bjg1568@rit.edu> date: October 2013 description: Sets up the logging information for the CAS Reader """ from config import config import logging as root_logging # Set up the logger logger = root_logging.getLogger() logger.setLevel(root_logging.INFO) logger_format = root_logging.Formatter('%(asctime)s %(levelname)s: %(message)s', '%Y-%m-%d %H:%M:%S') logging_file_handler = root_logging.FileHandler(config['logging_system']['filename']) logging_file_handler.setLevel(root_logging.INFO) logging_file_handler.setFormatter(logger_format) logger.addHandler(logging_file_handler) logging_stream_handler = root_logging.StreamHandler() logging_stream_handler.setLevel(root_logging.INFO) logging_stream_handler.setFormatter(logger_format) logger.addHandler(logging_stream_handler) logging = root_logging
Fix of the logging system exception
Fix of the logging system exception Added a format to the date for the logging system. '%Y-%m-%d %H:%M:%S’. Fixed an exception opening the logging file because the variable name was not written correctly.
Python
mit
bumper-app/bumper-bianca,bumper-app/bumper-bianca
3df4cc086bf6c85eebc12094cc3ca459bd2bcd3d
project/members/tests/test_application.py
project/members/tests/test_application.py
null
# -*- coding: utf-8 -*- import pytest from members.tests.fixtures.memberlikes import MembershipApplicationFactory from members.tests.fixtures.types import MemberTypeFactory from members.models import Member @pytest.mark.django_db def test_application_approve(): mtypes = [MemberTypeFactory(label='Normal member')] application = MembershipApplicationFactory() email = application.email application.approve(set_mtypes=mtypes) Member.objects.get(email=email)
Add unit test for programmatic application and approval
Add unit test for programmatic application and approval
Python
mit
HelsinkiHacklab/asylum,HelsinkiHacklab/asylum,hacklab-fi/asylum,jautero/asylum,rambo/asylum,jautero/asylum,hacklab-fi/asylum,rambo/asylum,HelsinkiHacklab/asylum,rambo/asylum,jautero/asylum,jautero/asylum,hacklab-fi/asylum,hacklab-fi/asylum,rambo/asylum,HelsinkiHacklab/asylum
41fd6e8aae4044520a2e44d590c005dd71150c0c
web/attempts/migrations/0008_add_submission_date.py
web/attempts/migrations/0008_add_submission_date.py
# -*- coding: utf-8 -*- # Generated by Django 1.9.5 on 2017-05-09 09:24 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('attempts', '0007_auto_20161004_0927'), ] operations = [ migrations.AddField( model_name='attempt', name='submission_date', field=models.DateTimeField(null=True), ), migrations.AddField( model_name='historicalattempt', name='submission_date', field=models.DateTimeField(null=True), ), migrations.RunSQL( 'UPDATE attempts_historicalattempt SET submission_date = history_date' ), migrations.RunSQL( '''UPDATE attempts_attempt SET submission_date = subquery.submission_date FROM ( SELECT user_id, part_id, max(history_date) AS submission_date FROM attempts_historicalattempt GROUP BY user_id, part_id ) AS subquery WHERE attempts_attempt.user_id = subquery.user_id AND attempts_attempt.part_id = subquery.part_id ''' ), migrations.AlterField( model_name='attempt', name='submission_date', field=models.DateTimeField(auto_now=True), ), migrations.AlterField( model_name='historicalattempt', name='submission_date', field=models.DateTimeField(blank=True, editable=False), ), ]
# -*- coding: utf-8 -*- # Generated by Django 1.9.5 on 2017-05-09 09:24 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('attempts', '0007_auto_20161004_0927'), ] operations = [ migrations.AddField( model_name='attempt', name='submission_date', field=models.DateTimeField(null=True), ), migrations.AddField( model_name='historicalattempt', name='submission_date', field=models.DateTimeField(null=True), ), migrations.RunSQL( 'UPDATE attempts_historicalattempt SET submission_date = history_date' ), migrations.RunSQL( '''UPDATE attempts_attempt SET submission_date = ( SELECT max(history_date) FROM attempts_historicalattempt WHERE attempts_attempt.user_id = user_id AND attempts_attempt.part_id = part_id ) ''' ), migrations.AlterField( model_name='attempt', name='submission_date', field=models.DateTimeField(auto_now=True), ), migrations.AlterField( model_name='historicalattempt', name='submission_date', field=models.DateTimeField(blank=True, editable=False), ), ]
Revert "Revert "Make migration SQLite compatible""
Revert "Revert "Make migration SQLite compatible"" This reverts commit b16016994f20945a8a2bbb63b9cb920d856ab66f.
Python
agpl-3.0
ul-fmf/projekt-tomo,ul-fmf/projekt-tomo,matijapretnar/projekt-tomo,ul-fmf/projekt-tomo,matijapretnar/projekt-tomo,matijapretnar/projekt-tomo,matijapretnar/projekt-tomo,ul-fmf/projekt-tomo,ul-fmf/projekt-tomo,matijapretnar/projekt-tomo,ul-fmf/projekt-tomo
15cb279724a646368066591e81467e1b26d61938
examples/charts/file/steps.py
examples/charts/file/steps.py
from bokeh.charts import Step, show, output_file # build a dataset where multiple columns measure the same thing data = dict(python=[2, 3, 7, 5, 26, 221, 44, 233, 254, 265, 266, 267, 120, 111], pypy=[12, 33, 47, 15, 126, 121, 144, 233, 254, 225, 226, 267, 110, 130], jython=[22, 43, 10, 25, 26, 101, 114, 203, 194, 215, 201, 227, 139, 160], test=['foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar', 'foo', 'bar'] ) # create a line chart where each column of measures receives a unique color and dash style line = Step(data, y=['python', 'pypy', 'jython'], dash=['python', 'pypy', 'jython'], color=['python', 'pypy', 'jython'], title="Interpreter Sample Data", ylabel='Duration', legend=True) output_file("steps.html") show(line)
""" This example uses the U.S. postage rate per ounce for stamps and postcards. Source: https://en.wikipedia.org/wiki/History_of_United_States_postage_rates """ from bokeh.charts import Step, show, output_file # build a dataset where multiple columns measure the same thing data = dict(stamp=[ .33, .33, .34, .37, .37, .37, .37, .39, .41, .42, .44, .44, .44, .45, .46, .49, .49], postcard=[ .20, .20, .21, .23, .23, .23, .23, .24, .26, .27, .28, .28, .29, .32, .33, .34, .35], ) # create a line chart where each column of measures receives a unique color and dash style line = Step(data, y=['stamp', 'postcard'], dash=['stamp', 'postcard'], color=['stamp', 'postcard'], title="U.S. Postage Rates (1999-2015)", ylabel='Rate per ounce', legend=True) output_file("steps.html") show(line)
Change step example to plot US postage rates
Change step example to plot US postage rates
Python
bsd-3-clause
ptitjano/bokeh,timsnyder/bokeh,draperjames/bokeh,percyfal/bokeh,justacec/bokeh,clairetang6/bokeh,philippjfr/bokeh,ericmjl/bokeh,rs2/bokeh,azjps/bokeh,DuCorey/bokeh,clairetang6/bokeh,draperjames/bokeh,clairetang6/bokeh,DuCorey/bokeh,aavanian/bokeh,KasperPRasmussen/bokeh,justacec/bokeh,bokeh/bokeh,aiguofer/bokeh,rs2/bokeh,aavanian/bokeh,stonebig/bokeh,schoolie/bokeh,msarahan/bokeh,msarahan/bokeh,jakirkham/bokeh,schoolie/bokeh,Karel-van-de-Plassche/bokeh,Karel-van-de-Plassche/bokeh,phobson/bokeh,stonebig/bokeh,azjps/bokeh,aavanian/bokeh,jakirkham/bokeh,ptitjano/bokeh,bokeh/bokeh,azjps/bokeh,ptitjano/bokeh,msarahan/bokeh,Karel-van-de-Plassche/bokeh,schoolie/bokeh,KasperPRasmussen/bokeh,draperjames/bokeh,phobson/bokeh,schoolie/bokeh,timsnyder/bokeh,azjps/bokeh,quasiben/bokeh,percyfal/bokeh,ericmjl/bokeh,jakirkham/bokeh,timsnyder/bokeh,phobson/bokeh,justacec/bokeh,mindriot101/bokeh,azjps/bokeh,Karel-van-de-Plassche/bokeh,ericmjl/bokeh,quasiben/bokeh,dennisobrien/bokeh,philippjfr/bokeh,clairetang6/bokeh,ptitjano/bokeh,ericmjl/bokeh,quasiben/bokeh,aiguofer/bokeh,timsnyder/bokeh,Karel-van-de-Plassche/bokeh,draperjames/bokeh,dennisobrien/bokeh,phobson/bokeh,dennisobrien/bokeh,aavanian/bokeh,stonebig/bokeh,percyfal/bokeh,aiguofer/bokeh,aiguofer/bokeh,jakirkham/bokeh,rs2/bokeh,draperjames/bokeh,ericmjl/bokeh,bokeh/bokeh,dennisobrien/bokeh,schoolie/bokeh,bokeh/bokeh,KasperPRasmussen/bokeh,percyfal/bokeh,aavanian/bokeh,phobson/bokeh,justacec/bokeh,mindriot101/bokeh,DuCorey/bokeh,stonebig/bokeh,bokeh/bokeh,philippjfr/bokeh,philippjfr/bokeh,DuCorey/bokeh,rs2/bokeh,mindriot101/bokeh,timsnyder/bokeh,DuCorey/bokeh,dennisobrien/bokeh,ptitjano/bokeh,percyfal/bokeh,msarahan/bokeh,rs2/bokeh,KasperPRasmussen/bokeh,jakirkham/bokeh,KasperPRasmussen/bokeh,mindriot101/bokeh,aiguofer/bokeh,philippjfr/bokeh
7ac7583d714b15bb3e24bba1c5972bd1d0c235f2
python/receive.py
python/receive.py
null
#!/usr/bin/env python import pika # Host in which RabbitMQ is running. HOST = 'localhost' # Name of the queue. QUEUE = 'pages' # Specifies what to do after consuming messages from RabbitMQ. def handler(channel, method, properties, body): print '-> Handled: [%s]' % body # Getting the connection using pika. # Creating the channel. connection = pika.BlockingConnection(pika.ConnectionParameters(host=HOST)) channel = connection.channel() print '* Handling messages...' # The consumption is defined. # Notice the 'handler' as first argument. channel.basic_consume(handler, queue=QUEUE, no_ack=True) # Starting the consumption. channel.start_consuming()
Add consumer written in Python
Add consumer written in Python
Python
apache-2.0
jovannypcg/rabbitmq_usage,jovannypcg/rabbitmq_usage
6fc0c3884c38448956273d99a57e0c758ecbc658
crmapp/marketing/views.py
crmapp/marketing/views.py
from django.shortcuts import render # Create your views here.
from django.views.generic.base import TemplateView class HomePage(TemplateView): """ Because our needs are so simple, all we have to do is assign one value; template_name. The home.html file will be created in the next lesson. """ template_name = 'marketing/home.html'
Create the Home Page > Create the Home Page View
Create the Home Page > Create the Home Page View
Python
mit
tabdon/crmeasyapp,tabdon/crmeasyapp,deenaariff/Django
a0f030cd03d28d97924a3277722d7a51cf3a3e92
cms/test_utils/project/extensionapp/models.py
cms/test_utils/project/extensionapp/models.py
# -*- coding: utf-8 -*- from cms.extensions import PageExtension, TitleExtension from cms.extensions.extension_pool import extension_pool from django.db import models class MyPageExtension(PageExtension): extra = models.CharField(blank=True, default='', max_length=255) extension_pool.register(MyPageExtension) class MyTitleExtension(TitleExtension): extra_title = models.CharField(blank=True, default='', max_length=255) extension_pool.register(MyTitleExtension)
# -*- coding: utf-8 -*- from cms.extensions import PageExtension, TitleExtension from cms.extensions.extension_pool import extension_pool from django.contrib.auth.models import User from django.db import models class MyPageExtension(PageExtension): extra = models.CharField(blank=True, default='', max_length=255) favorite_users = models.ManyToManyField(User, blank=True, null=True) def copy_relations(self, other, language): for favorite_user in other.favorite_users.all(): favorite_user.pk = None favorite_user.mypageextension = self favorite_user.save() extension_pool.register(MyPageExtension) class MyTitleExtension(TitleExtension): extra_title = models.CharField(blank=True, default='', max_length=255) extension_pool.register(MyTitleExtension)
Update extension app to include a M2M
Update extension app to include a M2M
Python
bsd-3-clause
kk9599/django-cms,jrclaramunt/django-cms,farhaadila/django-cms,FinalAngel/django-cms,leture/django-cms,yakky/django-cms,wuzhihui1123/django-cms,czpython/django-cms,jproffitt/django-cms,astagi/django-cms,DylannCordel/django-cms,evildmp/django-cms,jrclaramunt/django-cms,SachaMPS/django-cms,netzkolchose/django-cms,donce/django-cms,bittner/django-cms,jeffreylu9/django-cms,cyberintruder/django-cms,takeshineshiro/django-cms,Vegasvikk/django-cms,nostalgiaz/django-cms,kk9599/django-cms,rryan/django-cms,rscnt/django-cms,SmithsonianEnterprises/django-cms,jsma/django-cms,sephii/django-cms,selecsosi/django-cms,jsma/django-cms,SmithsonianEnterprises/django-cms,donce/django-cms,sznekol/django-cms,robmagee/django-cms,rsalmaso/django-cms,Livefyre/django-cms,divio/django-cms,owers19856/django-cms,isotoma/django-cms,intip/django-cms,qnub/django-cms,divio/django-cms,farhaadila/django-cms,iddqd1/django-cms,josjevv/django-cms,stefanfoulis/django-cms,farhaadila/django-cms,SofiaReis/django-cms,wuzhihui1123/django-cms,owers19856/django-cms,MagicSolutions/django-cms,jproffitt/django-cms,FinalAngel/django-cms,benzkji/django-cms,360youlun/django-cms,bittner/django-cms,netzkolchose/django-cms,jeffreylu9/django-cms,vstoykov/django-cms,stefanw/django-cms,jeffreylu9/django-cms,chkir/django-cms,nimbis/django-cms,vxsx/django-cms,selecsosi/django-cms,chkir/django-cms,qnub/django-cms,Jaccorot/django-cms,evildmp/django-cms,bittner/django-cms,wuzhihui1123/django-cms,iddqd1/django-cms,datakortet/django-cms,Vegasvikk/django-cms,benzkji/django-cms,wyg3958/django-cms,andyzsf/django-cms,MagicSolutions/django-cms,vstoykov/django-cms,intip/django-cms,intip/django-cms,memnonila/django-cms,takeshineshiro/django-cms,philippze/django-cms,vxsx/django-cms,jproffitt/django-cms,Livefyre/django-cms,SachaMPS/django-cms,stefanfoulis/django-cms,rryan/django-cms,AlexProfi/django-cms,petecummings/django-cms,vxsx/django-cms,rscnt/django-cms,dhorelik/django-cms,rsalmaso/django-cms,Vegasvikk/django-cms,liuyisiyisi/django-cms,youprofit/django-cms,wyg3958/django-cms,FinalAngel/django-cms,sznekol/django-cms,360youlun/django-cms,jrief/django-cms,andyzsf/django-cms,stefanw/django-cms,nostalgiaz/django-cms,selecsosi/django-cms,jsma/django-cms,donce/django-cms,360youlun/django-cms,rryan/django-cms,benzkji/django-cms,petecummings/django-cms,memnonila/django-cms,DylannCordel/django-cms,intgr/django-cms,Jaccorot/django-cms,rscnt/django-cms,frnhr/django-cms,astagi/django-cms,rsalmaso/django-cms,irudayarajisawa/django-cms,andyzsf/django-cms,chmberl/django-cms,saintbird/django-cms,evildmp/django-cms,frnhr/django-cms,MagicSolutions/django-cms,evildmp/django-cms,mkoistinen/django-cms,liuyisiyisi/django-cms,datakortet/django-cms,jeffreylu9/django-cms,intip/django-cms,vad/django-cms,isotoma/django-cms,divio/django-cms,mkoistinen/django-cms,intgr/django-cms,stefanw/django-cms,AlexProfi/django-cms,rryan/django-cms,stefanfoulis/django-cms,chmberl/django-cms,dhorelik/django-cms,nimbis/django-cms,mkoistinen/django-cms,Livefyre/django-cms,jrclaramunt/django-cms,saintbird/django-cms,yakky/django-cms,datakortet/django-cms,irudayarajisawa/django-cms,vstoykov/django-cms,jsma/django-cms,irudayarajisawa/django-cms,astagi/django-cms,FinalAngel/django-cms,wyg3958/django-cms,sephii/django-cms,kk9599/django-cms,saintbird/django-cms,divio/django-cms,chmberl/django-cms,josjevv/django-cms,intgr/django-cms,jrief/django-cms,wuzhihui1123/django-cms,webu/django-cms,frnhr/django-cms,sznekol/django-cms,SofiaReis/django-cms,philippze/django-cms,czpython/django-cms,frnhr/django-cms,vxsx/django-cms,cyberintruder/django-cms,cyberintruder/django-cms,rsalmaso/django-cms,timgraham/django-cms,yakky/django-cms,isotoma/django-cms,benzkji/django-cms,Livefyre/django-cms,nimbis/django-cms,AlexProfi/django-cms,robmagee/django-cms,jrief/django-cms,ScholzVolkmer/django-cms,robmagee/django-cms,webu/django-cms,netzkolchose/django-cms,intgr/django-cms,keimlink/django-cms,memnonila/django-cms,timgraham/django-cms,yakky/django-cms,datakortet/django-cms,mkoistinen/django-cms,philippze/django-cms,youprofit/django-cms,SmithsonianEnterprises/django-cms,SofiaReis/django-cms,chkir/django-cms,vad/django-cms,ScholzVolkmer/django-cms,takeshineshiro/django-cms,DylannCordel/django-cms,jrief/django-cms,liuyisiyisi/django-cms,stefanfoulis/django-cms,czpython/django-cms,owers19856/django-cms,petecummings/django-cms,keimlink/django-cms,nimbis/django-cms,ScholzVolkmer/django-cms,selecsosi/django-cms,leture/django-cms,jproffitt/django-cms,iddqd1/django-cms,keimlink/django-cms,qnub/django-cms,timgraham/django-cms,andyzsf/django-cms,SachaMPS/django-cms,czpython/django-cms,vad/django-cms,dhorelik/django-cms,vad/django-cms,youprofit/django-cms,netzkolchose/django-cms,Jaccorot/django-cms,sephii/django-cms,bittner/django-cms,isotoma/django-cms,josjevv/django-cms,nostalgiaz/django-cms,webu/django-cms,stefanw/django-cms,nostalgiaz/django-cms,sephii/django-cms,leture/django-cms
6854f889e38f565acb80c52a74df09730e0f7e45
uitools/notifications/linux.py
uitools/notifications/linux.py
from gi.repository import GLib, Notify as LibNotify class Notification(object): def __init__(self, title, message, subtitle=None, sticky=False): self.title = title self.subtitle = subtitle self.message = message self.sticky = sticky self._sent = False def send(self): # see: https://developer.gnome.org/libnotify/0.7/libnotify-notify.html # see: https://developer.gnome.org/libnotify/0.7/NotifyNotification.html # Can check LibNotify.get_server_caps() for a list of capabilities. print 'capabilities', LibNotify.get_server_caps() self._notification = notification = LibNotify.Notification.new( self.title, self.message, 'folder-new' ) # If this is "default", then it is the default action for clicking the notification. notification.add_action('default', 'Default Action', _on_action, 'on_action_payload') notification.add_action('not_default', 'Another Action', _on_action, 'another_payload') notification.connect('closed', _on_action) notification.set_timeout(5000) # 5s notification.show() # NOTE: This object NEEDS to be held onto for the callback to work.
from gi.repository import GLib, Notify as LibNotify DEV = False class Notification(object): def __init__(self, title, message, subtitle=None, sticky=False): self.title = title self.subtitle = subtitle self.message = message self.sticky = sticky self._sent = False def send(self): # see: https://developer.gnome.org/libnotify/0.7/libnotify-notify.html # see: https://developer.gnome.org/libnotify/0.7/NotifyNotification.html LibNotify.init('com.westernx.uitools.notifications') self._notification = notification = LibNotify.Notification.new( self.title, self.message, 'dialog-information' ) if DEV: # Can check LibNotify.get_server_caps() for a list of capabilities. print 'capabilities', LibNotify.get_server_caps() # If this is "default", then it is the default action for clicking the notification. notification.add_action('default', 'Default Action', _on_action, 'on_action_payload') notification.add_action('not_default', 'Another Action', _on_action, 'another_payload') notification.connect('closed', _on_action) notification.set_timeout(5000) # 5s notification.show() # NOTE: This object NEEDS to be held onto for the callback to work.
Clean up for production Linux use
Clean up for production Linux use
Python
bsd-3-clause
westernx/uitools
cb5591228408ccd22cfb3df9106d4ecdeac295a9
nose2/plugins/collect.py
nose2/plugins/collect.py
""" This plugin implements :func:`startTestRun`, setting a test executor (``event.executeTests``) that just collects tests without executing them. To do so it calls result.startTest, result.addSuccess and result.stopTest for ech test, without calling the test itself. """ from nose2.events import Plugin from nose2.compat import unittest __unittest = True class CollectOnly(Plugin): """Collect but don't run tests""" configSection = 'collect-only' commandLineSwitch = (None, 'collect-only', 'Collect and output test names; do not run any tests') _mpmode = False def registerInSubprocess(self, event): event.pluginClasses.append(self.__class__) self._mpmode = True def startTestRun(self, event): """Replace ``event.executeTests``""" if self._mpmode: return event.executeTests = self.collectTests def startSubprocess(self, event): event.executeTests = self.collectTests def collectTests(self, suite, result): """Collect tests, but don't run them""" for test in suite: if isinstance(test, unittest.BaseTestSuite): self.collectTests(test, result) continue result.startTest(test) result.addSuccess(test) result.stopTest(test)
""" This plugin implements :func:`startTestRun`, setting a test executor (``event.executeTests``) that just collects tests without executing them. To do so it calls result.startTest, result.addSuccess and result.stopTest for each test, without calling the test itself. """ from nose2.events import Plugin from nose2.compat import unittest __unittest = True class CollectOnly(Plugin): """Collect but don't run tests""" configSection = 'collect-only' commandLineSwitch = (None, 'collect-only', 'Collect and output test names; do not run any tests') _mpmode = False def registerInSubprocess(self, event): event.pluginClasses.append(self.__class__) self._mpmode = True def startTestRun(self, event): """Replace ``event.executeTests``""" if self._mpmode: return event.executeTests = self.collectTests def startSubprocess(self, event): event.executeTests = self.collectTests def collectTests(self, suite, result): """Collect tests, but don't run them""" for test in suite: if isinstance(test, unittest.BaseTestSuite): self.collectTests(test, result) continue result.startTest(test) result.addSuccess(test) result.stopTest(test)
Fix typo "ech" -> "each"
Fix typo "ech" -> "each"
Python
bsd-2-clause
ptthiem/nose2,ptthiem/nose2
cc89c5222ec7f6d6f95b5efdce3958b3ca33814e
mica/archive/tests/test_aca_dark_cal.py
mica/archive/tests/test_aca_dark_cal.py
null
""" Basic functionality and regression tests for ACA dark cal module. """ import numpy as np from ..aca_dark import dark_cal def test_date_to_dark_id(): assert dark_cal.date_to_dark_id('2011-01-15T12:00:00') == '2011015' def test_dark_id_to_date(): assert dark_cal.dark_id_to_date('2011015') == '2011:015' def test_dark_temp_scale(): scale = dark_cal.dark_temp_scale(-10., -14) assert np.allclose(scale, 0.70) def test_get_dark_cal_id(): assert dark_cal.get_dark_cal_id('2007:008', 'nearest') == '2007006' assert dark_cal.get_dark_cal_id('2007:008', 'before') == '2007006' assert dark_cal.get_dark_cal_id('2007:008', 'after') == '2007069' def test_get_dark_cal_image(): image = dark_cal.get_dark_cal_image('2007:008') assert image.shape == (1024, 1024) def test_get_dark_cal_props(): props = dark_cal.get_dark_cal_props('2007:008') assert len(props['replicas']) == 5 assert props['start'] == '2007:006:01:56:46.817' props = dark_cal.get_dark_cal_props('2007:008', include_image=True) assert len(props['replicas']) == 5 assert props['start'] == '2007:006:01:56:46.817' assert props['image'].shape == (1024, 1024) def test_get_dark_cal_props_table(): props = dark_cal.get_dark_cal_props_table('2007:001', '2008:001') assert np.allclose(props['eb'], [24.6, 25.89, 51.13, 1.9])
Add basic functionality and regression tests for ACA dark cal module
Add basic functionality and regression tests for ACA dark cal module
Python
bsd-3-clause
sot/mica,sot/mica
eaed2b0b37c042f1dbaf5c60163022e2e2605486
setup.py
setup.py
#!/usr/bin/env python # encoding: utf-8 """ setup.py Setup the Keyring Lib for Python. """ import sys from distutils.core import setup, Extension from extensions import get_extensions setup(name = 'keyring', version = "0.5.1", description = "Store and access your passwords safely.", url = "http://home.python-keyring.org/", keywords = "keyring Keychain GnomeKeyring Kwallet password storage", maintainer = "Kang Zhang", maintainer_email = "jobo.zh@gmail.com", license="PSF", long_description = open('README.txt').read() + open('CHANGES.txt').read(), platforms = ["Many"], packages = ['keyring', 'keyring.tests', 'keyring.util', 'keyring.backends'], ext_modules = get_extensions() )
#!/usr/bin/env python # encoding: utf-8 """ setup.py Setup the Keyring Lib for Python. """ import sys from distutils.core import setup, Extension from extensions import get_extensions setup(name = 'keyring', version = "0.5.1", description = "Store and access your passwords safely.", url = "http://home.python-keyring.org/", keywords = "keyring Keychain GnomeKeyring Kwallet password storage", maintainer = "Kang Zhang", maintainer_email = "jobo.zh@gmail.com", license="PSF", long_description = open('README').read() + open('CHANGES.txt').read(), platforms = ["Many"], packages = ['keyring', 'keyring.tests', 'keyring.util', 'keyring.backends'], ext_modules = get_extensions() )
Use the new README file, which has been renamed in 4ec717c11604
Use the new README file, which has been renamed in 4ec717c11604
Python
mit
jaraco/keyring
3fb4d7b630fb7a4b34dcc4e1b72947e61f73a80f
TestData/download_test_data.py
TestData/download_test_data.py
null
def set_test_db(): from sys import path path.insert(0, "..") from MyEdgarDb import get_list_sec_filings, get_cik_ticker_lookup_db, lookup_cik_ticker get_list_sec_filings (7, 'test_idx.db') get_cik_ticker_lookup_db ('test_idx.db') def download_test_data(): import sqlite3 from datetime import datetime import pandas as pd testDir = "..\\TestData\\" testTickers = { "AAPL": [datetime(2014, 8, 1), datetime(2018, 8, 1)], "ACLS": [datetime(2014, 8, 31), datetime(2018, 8, 31)], "ADSK": [datetime(2014, 4, 15), datetime(2018, 4, 15)], "ALEX": [datetime(2015, 12, 31), datetime(2019, 12, 31)], "MMM": [datetime(2015, 7, 1), datetime(2019, 7, 1)], "NRP": [datetime(2015, 12, 31), datetime(2019, 12, 31)], "NVDA": [datetime(2015, 12, 31), datetime(2019, 12, 31)] } conn3 = sqlite3.connect('test_idx.db') cursor = conn3.cursor() for ticker in testTickers: #cursor.execute('''SELECT * FROM idx WHERE Symbol=?;''', ("ABBV",)) cursor.execute('''SELECT * FROM cik_ticker_name WHERE ticker=?;''',(ticker,)) res = cursor.fetchall() print(res) cursor.execute('''SELECT * FROM idx WHERE cik=?;''', (res[0][0],)) recs = cursor.fetchall() print(len(recs)) names = list(map(lambda x: x[0], cursor.description)) #print(names) df = pd.DataFrame(data=recs, columns=names) df['date'] = pd.to_datetime(df['date']) beginDate = testTickers[ticker][0] endDate = testTickers[ticker][1] df1 = df[(df.date >= beginDate) & (df.date <= endDate)] ## Sort by date in descending order (most recent is first) df1.sort_values(by=['date'], inplace=True, ascending=False) df1[df1.type == "10-Q"].to_csv(testDir+ticker.lower()+"_all_10qs.csv", index=None) df1[df1.type == "10-K"].to_csv(testDir+ticker.lower()+"_all_10ks.csv", index=None) conn3.close() if __name__ == "__main__": #set_test_db() download_test_data()
Create script to dowload requisite test urls.
Create script to dowload requisite test urls.
Python
agpl-3.0
cielling/jupyternbs
ca97a29dded7278b40785fe88b5e8c9ceb542d86
urllib3/util/wait.py
urllib3/util/wait.py
from .selectors import ( HAS_SELECT, DefaultSelector, EVENT_READ, EVENT_WRITE ) def _wait_for_io_events(socks, events, timeout=None): """ Waits for IO events to be available from a list of sockets or optionally a single socket if passed in. Returns a list of sockets that can be interacted with immediately. """ if not HAS_SELECT: raise ValueError('Platform does not have a selector') if not isinstance(socks, list): # Probably just a single socket. if hasattr(socks, "fileno"): socks = [socks] # Otherwise it might be a non-list iterable. else: socks = list(socks) selector = DefaultSelector() for sock in socks: selector.register(sock, events) return [key[0].fileobj for key in selector.select(timeout) if key[1] & events] def wait_for_read(socks, timeout=None): """ Waits for reading to be available from a list of sockets or optionally a single socket if passed in. Returns a list of sockets that can be read from immediately. """ return _wait_for_io_events(socks, EVENT_READ, timeout) def wait_for_write(socks, timeout=None): """ Waits for writing to be available from a list of sockets or optionally a single socket if passed in. Returns a list of sockets that can be written to immediately. """ return _wait_for_io_events(socks, EVENT_WRITE, timeout)
from .selectors import ( HAS_SELECT, DefaultSelector, EVENT_READ, EVENT_WRITE ) def _wait_for_io_events(socks, events, timeout=None): """ Waits for IO events to be available from a list of sockets or optionally a single socket if passed in. Returns a list of sockets that can be interacted with immediately. """ if not HAS_SELECT: raise ValueError('Platform does not have a selector') if not isinstance(socks, list): # Probably just a single socket. if hasattr(socks, "fileno"): socks = [socks] # Otherwise it might be a non-list iterable. else: socks = list(socks) with DefaultSelector() as selector: for sock in socks: selector.register(sock, events) return [key[0].fileobj for key in selector.select(timeout) if key[1] & events] def wait_for_read(socks, timeout=None): """ Waits for reading to be available from a list of sockets or optionally a single socket if passed in. Returns a list of sockets that can be read from immediately. """ return _wait_for_io_events(socks, EVENT_READ, timeout) def wait_for_write(socks, timeout=None): """ Waits for writing to be available from a list of sockets or optionally a single socket if passed in. Returns a list of sockets that can be written to immediately. """ return _wait_for_io_events(socks, EVENT_WRITE, timeout)
Use DefaultSelector as context manager.
Use DefaultSelector as context manager.
Python
mit
sigmavirus24/urllib3,Lukasa/urllib3,Disassem/urllib3,Lukasa/urllib3,Disassem/urllib3,urllib3/urllib3,urllib3/urllib3,sigmavirus24/urllib3
5bb90727efb62525995caad3b52fd588d8b08298
pregnancy/urls.py
pregnancy/urls.py
from django.conf.urls import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import contractions.views urlpatterns = patterns('', # Examples: # url(r'^$', 'pregnancy.views.home', name='home'), # url(r'^pregnancy/', include('pregnancy.foo.urls')), url(r'^contractions/$', contractions.views.ContractionList.as_view(), name='ContractionList'), url(r'^update_intensity/(?P<pk>\d+)/$', contractions.views.UpdateIntensity.as_view(), name='UpdateIntensity'), url(r'^update_intensity2/(?P<pk>\d+)/$', contractions.views.UpdateIntensity2.as_view(), name='UpdateIntensity2'), url(r'^ContractionListTable/$', contractions.views.ContractionListTable.as_view(), name='ContractionListTable'), url(r'^StartContraction/$', contractions.views.StartContraction.as_view(), name='StartContraction'), url(r'^StopContraction/(?P<pk>\d+)/$', contractions.views.StopContraction.as_view(), name='StopContraction'), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
from django.conf.urls import patterns, include, url # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import contractions.views urlpatterns = patterns('', # Examples: # url(r'^$', 'pregnancy.views.home', name='home'), # url(r'^pregnancy/', include('pregnancy.foo.urls')), url(r'^$', contractions.views.ContractionList.as_view(), name='ContractionList'), url(r'^contractions/$', contractions.views.ContractionList.as_view(), name='ContractionList'), url(r'^update_intensity/(?P<pk>\d+)/$', contractions.views.UpdateIntensity.as_view(), name='UpdateIntensity'), url(r'^update_intensity2/(?P<pk>\d+)/$', contractions.views.UpdateIntensity2.as_view(), name='UpdateIntensity2'), url(r'^ContractionListTable/$', contractions.views.ContractionListTable.as_view(), name='ContractionListTable'), url(r'^StartContraction/$', contractions.views.StartContraction.as_view(), name='StartContraction'), url(r'^StopContraction/(?P<pk>\d+)/$', contractions.views.StopContraction.as_view(), name='StopContraction'), # Uncomment the admin/doc line below to enable admin documentation: # url(r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: url(r'^admin/', include(admin.site.urls)), )
Update url to point / to the contractions app
Update url to point / to the contractions app
Python
bsd-2-clause
dreinhold/pregnancy,dreinhold/pregnancy,dreinhold/pregnancy
20b0e705fe6eedb05a94a3e9cb978b65a525fe91
conanfile.py
conanfile.py
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.2" class SanitizeTargetCMakeConan(ConanFile): name = "sanitize-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "cmake-multi-targets/master@smspillaz/cmake-multi-targets", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit", "sanitizers-cmake/0.0.1@smspillaz/sanitizers-cmake") url = "http://github.com/polysquare/sanitize-target-cmake" license = "MIT" def source(self): zip_name = "sanitize-target-cmake.zip" download("https://github.com/polysquare/" "sanitize-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/sanitize-target-cmake", src="sanitize-target-cmake-" + VERSION, keep_path=True)
from conans import ConanFile from conans.tools import download, unzip import os VERSION = "0.0.3" class SanitizeTargetCMakeConan(ConanFile): name = "sanitize-target-cmake" version = os.environ.get("CONAN_VERSION_OVERRIDE", VERSION) generators = "cmake" requires = ("cmake-include-guard/master@smspillaz/cmake-include-guard", "cmake-multi-targets/master@smspillaz/cmake-multi-targets", "tooling-cmake-util/master@smspillaz/tooling-cmake-util", "cmake-unit/master@smspillaz/cmake-unit", "sanitizers-cmake/0.0.1@smspillaz/sanitizers-cmake") url = "http://github.com/polysquare/sanitize-target-cmake" license = "MIT" def source(self): zip_name = "sanitize-target-cmake.zip" download("https://github.com/polysquare/" "sanitize-target-cmake/archive/{version}.zip" "".format(version="v" + VERSION), zip_name) unzip(zip_name) os.unlink(zip_name) def package(self): self.copy(pattern="*.cmake", dst="cmake/sanitize-target-cmake", src="sanitize-target-cmake-" + VERSION, keep_path=True)
Bump version: 0.0.2 -> 0.0.3
Bump version: 0.0.2 -> 0.0.3 [ci skip]
Python
mit
polysquare/sanitize-target-cmake
4130c082ae3008365c854ad65c4510cb04dfbf27
webcrawler.py
webcrawler.py
import requests from bs4 import BeautifulSoup def parse_html(html, **kwargs): parsed_html = BeautifulSoup(html, 'lxml') headline = parsed_html.body.find('h1') paragraph = None # Parse Paragraph content_container = parsed_html.body.find( 'div', attrs={'id': 'bodyContent'} ) for p in content_container.findAll('p'): if not p.findParents('table'): paragraph = p break # Parse Image infobox = parsed_html.body.find('table', attrs={'class': 'infobox'}) image = infobox.find('img') if infobox else None return { 'headline': headline.text.strip() if headline else '', 'paragraph': paragraph.text.strip() if paragraph else '', 'image_url': image.attrs.get('src') if image else '', } def crawl(url, **kwargs): response = requests.get(url) response.raise_for_status() data = parse_html(response.content, **kwargs) # TODOs: Persist data return data
import re import requests from bs4 import BeautifulSoup def parse_html(html, **kwargs): is_wikipedia_page = kwargs.get('is_wikipedia_page') parsed_html = BeautifulSoup(html, 'html.parser') headline = parsed_html.body.find('h1') paragraph = None if is_wikipedia_page: # Parse Paragraph content_container = parsed_html.body.find( 'div', attrs={'id': 'bodyContent'} ) for p in content_container.findAll('p'): if not p.findParents('table'): paragraph = p break # Parse Image infobox = parsed_html.body.find('table', attrs={'class': 'infobox'}) image = infobox.find('img') if infobox else None else: content_container_class = kwargs.get('content_container_class') image_container_class = kwargs.get('image_container_class') if not all([ content_container_class, image_container_class ]): return content_container = parsed_html.body.find('div', attrs={'class': content_container_class}) paragraph = content_container.find('p') image_container = parsed_html.body.find('div', attrs={'class': image_container_class}) image = image_container.find('img') return { 'headline': headline.text.strip() if headline else '', 'paragraph': paragraph.text.strip() if paragraph else '', 'image_url': image.attrs.get('src') if image else '', } def crawl(url, **kwargs): response = requests.get(url) response.raise_for_status() is_wikipedia_page = re.compile(r'.*(wikipedia.org)').match(url) is not None if is_wikipedia_page: kwargs.update({ 'is_wikipedia_page': is_wikipedia_page }) data = parse_html(response.content, **kwargs) # TODOs: Persist data return data
Add support for generic pages
Add support for generic pages
Python
mit
alamasfu10/webcrawler
41e0ea623baaff22ed5f436ad563edf52b762bcc
Main.py
Main.py
"""Main Module of PDF Splitter""" import argparse import os from PyPDF2 import PdfFileWriter from Util import all_pdf_files_in_directory, split_on_condition, concat_pdf_pages parser = \ argparse.ArgumentParser( description='Split all the pages of multiple PDF files in a directory by document number' ) parser.add_argument( 'directory', metavar='PATH', type=str, help='path to a directory' ) def width_greater_than_height(page): box = page.mediaBox return box.getWidth() > box.getHeight() if __name__ == '__main__': args = parser.parse_args() all_pdf_files = all_pdf_files_in_directory(args.directory) opened_files = map(lambda path: open(path, 'rb'), all_pdf_files) all_pages = concat_pdf_pages(opened_files) for idx, pages in enumerate(split_on_condition(all_pages, predicate=width_greater_than_height), start=1): pdf_writer = PdfFileWriter() map(pdf_writer.addPage, pages) output_filename = '{0:05}.pdf'.format(idx) with open(output_filename, 'wb') as output_file: pdf_writer.write(output_file) output_file.flush() os.fsync(output_file.fileno()) map(lambda f: f.close, opened_files)
"""Main Module of PDF Splitter""" import argparse import os from PyPDF2 import PdfFileWriter from Util import all_pdf_files_in_directory, split_on_condition, concat_pdf_pages parser = \ argparse.ArgumentParser( description='Split all the pages of multiple PDF files in a directory by document number' ) parser.add_argument( 'directory', metavar='PATH', type=str, help='path to a directory' ) def width_greater_than_height(page): box = page.mediaBox return box.getWidth() > box.getHeight() if __name__ == '__main__': args = parser.parse_args() directory = args.directory all_pdf_files = [os.path.join(directory, filename) for filename in all_pdf_files_in_directory(directory)] opened_files = map(lambda path: open(path, 'rb'), all_pdf_files) all_pages = concat_pdf_pages(opened_files) for idx, pages in enumerate(split_on_condition(all_pages, predicate=width_greater_than_height), start=1): pdf_writer = PdfFileWriter() map(pdf_writer.addPage, pages) output_filename = '{0:05}.pdf'.format(idx) with open(output_filename, 'wb') as output_file: pdf_writer.write(output_file) output_file.flush() os.fsync(output_file.fileno()) map(lambda f: f.close, opened_files)
Fix bug where only PDF files in current directory can be found
Fix bug where only PDF files in current directory can be found
Python
mit
shunghsiyu/pdf-processor
2ee763ae1e4564a57692cb7161f99daab4ae77b7
cookiecutter/main.py
cookiecutter/main.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ cookiecutter.main ----------------- Main entry point for the `cookiecutter` command. The code in this module is also a good example of how to use Cookiecutter as a library rather than a script. """ import argparse import os from .find import find_template from .generate import generate_context, generate_files from .vcs import git_clone def main(): """ Entry point for the package, as defined in setup.py. """ # Get command line input/output arguments parser = argparse.ArgumentParser( description='Create a project from a Cookiecutter project template.' ) parser.add_argument( 'input_dir', help='Cookiecutter project template dir, e.g. {{project.repo_name}}/' ) args = parser.parse_args() # If it's a git repo, clone and prompt if args.input_dir.endswith('.git'): repo_dir = git_clone(args.input_dir) project_template = find_template(repo_dir) os.chdir(repo_dir) else: project_template = args.input_dir # Create project from local context and project template. context = generate_context() generate_files( input_dir=project_template, context=context ) if __name__ == '__main__': main()
#!/usr/bin/env python # -*- coding: utf-8 -*- """ cookiecutter.main ----------------- Main entry point for the `cookiecutter` command. The code in this module is also a good example of how to use Cookiecutter as a library rather than a script. """ import argparse import os from .cleanup import remove_repo from .find import find_template from .generate import generate_context, generate_files from .vcs import git_clone def main(): """ Entry point for the package, as defined in setup.py. """ # Get command line input/output arguments parser = argparse.ArgumentParser( description='Create a project from a Cookiecutter project template.' ) parser.add_argument( 'input_dir', help='Cookiecutter project template dir, e.g. {{project.repo_name}}/' ) args = parser.parse_args() # If it's a git repo, clone and prompt if args.input_dir.endswith('.git'): got_repo_arg = True repo_dir = git_clone(args.input_dir) project_template = find_template(repo_dir) os.chdir(repo_dir) else: project_template = args.input_dir # Create project from local context and project template. context = generate_context() generate_files( input_dir=project_template, context=context ) # Remove repo if Cookiecutter cloned it in the first place. # Here the user just wants a project, not a project template. if got_repo_arg: generated_project = context['project']['repo_name'] remove_repo(repo_dir, generated_project) if __name__ == '__main__': main()
Clean up after cloned repo if needed. (partial checkin)
Clean up after cloned repo if needed. (partial checkin)
Python
bsd-3-clause
atlassian/cookiecutter,dajose/cookiecutter,luzfcb/cookiecutter,foodszhang/cookiecutter,Springerle/cookiecutter,willingc/cookiecutter,utek/cookiecutter,takeflight/cookiecutter,takeflight/cookiecutter,willingc/cookiecutter,stevepiercy/cookiecutter,vincentbernat/cookiecutter,cguardia/cookiecutter,0k/cookiecutter,nhomar/cookiecutter,janusnic/cookiecutter,benthomasson/cookiecutter,0k/cookiecutter,foodszhang/cookiecutter,kkujawinski/cookiecutter,ionelmc/cookiecutter,michaeljoseph/cookiecutter,sp1rs/cookiecutter,jhermann/cookiecutter,lucius-feng/cookiecutter,agconti/cookiecutter,michaeljoseph/cookiecutter,ramiroluz/cookiecutter,letolab/cookiecutter,hackebrot/cookiecutter,audreyr/cookiecutter,sp1rs/cookiecutter,kkujawinski/cookiecutter,vincentbernat/cookiecutter,agconti/cookiecutter,lucius-feng/cookiecutter,drgarcia1986/cookiecutter,drgarcia1986/cookiecutter,atlassian/cookiecutter,ionelmc/cookiecutter,lgp171188/cookiecutter,audreyr/cookiecutter,nhomar/cookiecutter,alex/cookiecutter,cichm/cookiecutter,moi65/cookiecutter,lgp171188/cookiecutter,hackebrot/cookiecutter,tylerdave/cookiecutter,terryjbates/cookiecutter,stevepiercy/cookiecutter,janusnic/cookiecutter,utek/cookiecutter,moi65/cookiecutter,christabor/cookiecutter,vintasoftware/cookiecutter,alex/cookiecutter,cichm/cookiecutter,tylerdave/cookiecutter,letolab/cookiecutter,Vauxoo/cookiecutter,Springerle/cookiecutter,jhermann/cookiecutter,vintasoftware/cookiecutter,cguardia/cookiecutter,Vauxoo/cookiecutter,benthomasson/cookiecutter,dajose/cookiecutter,pjbull/cookiecutter,christabor/cookiecutter,venumech/cookiecutter,luzfcb/cookiecutter,terryjbates/cookiecutter,pjbull/cookiecutter,ramiroluz/cookiecutter,venumech/cookiecutter
a5ddab3208992ca6ab655ddef9a4155d5fc6bc55
tests/grammar_test.py
tests/grammar_test.py
import nose from parser_tool import get_parser, parse sentences = ( # N[s] V[i] "Brad drives", # N[s] V[t] N[p] "Angela drives cars", # N[s] V[t] Det N[s] "Brad buys the house", # Det[s] N[s] V[i] "a dog walks" ) grammar = get_parser("grammars/feat1.fcfg", trace=0) def test_grammar(): global sentences, parser for sent in sentences: print "Testing: %s" % sent trees = parse(grammar, sent) assert len(trees) > 0 if __name__=="__main__": nose.main()
import nose from parser_tool import get_parser, parse sentences = ( # PN V[i] "Brad drives", # PN V[t] N[p] "Angela drives cars", # PN V[t] Det N[s] "Brad buys the house", # Det[s] N[s] V[i] "a dog walks", # Det[p] N[p] V[i] "these dogs walk", # Det[p] N[p] V[t] Det N[s] "the cars enter the house", # A N[p] V[t] Det N[s] "red cars enter the house", # Det A N[s] V[t] Det N[s] "a red car enters the house", # PN V[t] Det A N[s] "Brad buys a red car", ) grammar = get_parser("grammars/feat1.fcfg", trace=0) def test_grammar(): global sentences, parser for sent in sentences: print "Testing: %s" % sent trees = parse(grammar, sent) assert len(trees) > 0 if __name__=="__main__": nose.main()
Increase testing coverage of grammar
Increase testing coverage of grammar * added some sample sentences with ajectives
Python
mit
caninemwenja/marker,kmwenja/marker
2759119ea8a2afe6c47575825aef9ae59c1ce921
python/misc/oo/CSStudent.py
python/misc/oo/CSStudent.py
null
# Code from https://www.geeksforgeeks.org/g-fact-34-class-or-static-variables-in-python/ # Copy - Paste here to test it directly # Python program to show that the variables with a value # assigned in class declaration, are class variables # Class for Computer Science Student class CSStudent: stream = 'cse' # Class Variable def __init__(self,name,roll): self.name = name # Instance Variable self.roll = roll # Instance Variable # Objects of CSStudent class a = CSStudent('Geek', 1) b = CSStudent('Nerd', 2) print(a.stream) # prints "cse" print(b.stream) # prints "cse" print(a.name) # prints "Geek" print(b.name) # prints "Nerd" print(a.roll) # prints "1" print(b.roll) # prints "2" # Class variables can be accessed using class # name also print(CSStudent.stream) # prints "cse" # Now if we change the stream for just a it won't be changed for b a.stream = 'ece' print(a.stream) # prints 'ece' print(b.stream) # prints 'cse' # To change the stream for all instances of the class we can change it # directly from the class CSStudent.stream = 'mech' print(a.stream) # prints 'mech' <-- error here, it prints ece print(b.stream) # prints 'mech'
Test d'un bout de code de GeeksForGeeks
Test d'un bout de code de GeeksForGeeks
Python
mit
TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts,TGITS/programming-workouts
9437b7fa2ef7f581968d6628561940dcb1e3f4ad
test_tws/__init__.py
test_tws/__init__.py
'''Unit test package for package "tws".''' __copyright__ = "Copyright (c) 2008 Kevin J Bluck" __version__ = "$Id$" import socket from tws import EWrapper def test_import(): '''Verify successful import of top-level "tws" package''' import tws assert tws class mock_wrapper(EWrapper): def __init__(self): self.errors = [] def error(self, id, code, text): self.errors.append((id, code, text)) class mock_socket(object): def __init__(self): self._peer = () def connect(self, peer, error=False): if error: raise socket.error() self._peer = peer def getpeername(self): if not self._peer: raise socket.error() return self._peer def makefile(self, mode): return StringIO()
'''Unit test package for package "tws".''' __copyright__ = "Copyright (c) 2008 Kevin J Bluck" __version__ = "$Id$" import socket from tws import EWrapper def test_import(): '''Verify successful import of top-level "tws" package''' import tws assert tws class mock_wrapper(EWrapper): def __init__(self): self.calldata = [] self.errors = [] def error(self, id, code, text): self.errors.append((id, code, text)) def __getattr__(self, name): # Any arbitrary unknown attribute is mapped to a function call which is # recorded into self.calldata. return lambda *args, **kwds: self.calldata.append((name, args, kwds)) class mock_socket(object): def __init__(self): self._peer = () def connect(self, peer, error=False): if error: raise socket.error() self._peer = peer def getpeername(self): if not self._peer: raise socket.error() return self._peer def makefile(self, mode): return StringIO()
Implement a __getattr__() for mock_wrapper that just returns a lambda that records whatever call was attempted along with the call params.
Implement a __getattr__() for mock_wrapper that just returns a lambda that records whatever call was attempted along with the call params.
Python
bsd-3-clause
kbluck/pytws,kbluck/pytws
README.md exists but content is empty. Use the Edit dataset card button to edit it.
Downloads last month
38
Edit dataset card